Next Article in Journal / Special Issue
Damage Detection Based on Power Dissipation Measured with PZT Sensors through the Combination of Electro-Mechanical Impedances and Guided Waves
Previous Article in Journal / Special Issue
3D Visual Data-Driven Spatiotemporal Deformations for Non-Rigid Object Grasping Using Robot Hands
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning and Computer Vision System for Phenotype Data Acquisition and Analysis in Plants

by
Pedro J. Navarro
1,*,†,
Fernando Pérez
2,†,
Julia Weiss
2,† and
Marcos Egea-Cortines
2,†
1
DSIE, Universidad Politécnica de Cartagena, Campus Muralla del Mar, s/n. Cartagena 30202, Spain
2
Genética, Instituto de Biotecnología Vegetal, Universidad Politécnica de Cartagena, Cartagena 30202, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Submission received: 3 March 2016 / Revised: 25 April 2016 / Accepted: 26 April 2016 / Published: 5 May 2016
(This article belongs to the Special Issue State-of-the-Art Sensors Technology in Spain 2015)

Abstract

:
Phenomics is a technology-driven approach with promising future to obtain unbiased data of biological systems. Image acquisition is relatively simple. However data handling and analysis are not as developed compared to the sampling capacities. We present a system based on machine learning (ML) algorithms and computer vision intended to solve the automatic phenotype data analysis in plant material. We developed a growth-chamber able to accommodate species of various sizes. Night image acquisition requires near infrared lightning. For the ML process, we tested three different algorithms: k-nearest neighbour (kNN), Naive Bayes Classifier (NBC), and Support Vector Machine. Each ML algorithm was executed with different kernel functions and they were trained with raw data and two types of data normalisation. Different metrics were computed to determine the optimal configuration of the machine learning algorithms. We obtained a performance of 99.31% in kNN for RGB images and a 99.34% in SVM for NIR. Our results show that ML techniques can speed up phenomic data analysis. Furthermore, both RGB and NIR images can be segmented successfully but may require different ML algorithms for segmentation.

1. Introduction

The advent of the so-called omics technologies has been a major change in the way experiments are designed and has driven new ways to approach biology. One common aspect to these technology-driven approaches is the continuous decrease in price in order to achieve high throughput. As a result biology has become a field where big data accumulates, and which requires analytical tools [1]. The latest newcomer in the field of automatic sampling is the so-called phenomics. It comprises any tool that will help acquire quantitative data of phenotypes. Plant growth and development can be considered as a combination of a default program that interacts with biotic and abiotic stresses, light and temperature to give external phenotypes. And measuring, not only the outcome or end point, but also kinetics and their changes is becoming increasingly important to understand plants as a whole and become more precise at experimental designs. One of the newest developments is automatic image acquisition [2].
One of the fields where automatic image acquisition has defined its development is circadian clock analysis as promoters driving reporter genes such as luciferase or Green Fluorescent Protein allowed the identification of mutants and further characterization of the gene network at the transcriptional level [3,4]. Artificial vision systems have been used to study different aspects of plant growth and development such as root development [5], leaf growth [6], flowers and shoots [7] or seedling [8]. An important challenge of image acquisition in plant biology is the signalling effect of different light wavelengths including blue light, red and far red. As a result image acquisition in the dark requires infrared lightning [7,9].
Phenotyping of small plants such as Arabidopsis thaliana can be performed with a vertical camera taking pictures of rosettes at time intervals [10]. Larger plants or the parallel phenotyping of several traits require image acquisition from lateral positions [11]. Thus obtaining lateral images or the reconstruction of 3-dimensional images is performed by combination of cameras or moving them to acquire images [12].
Although hardware development requires a multidisciplinary approach, the bottleneck lies in image analysis. Ideally images should be analysed in an automatic fashion. The number of images to be processed when screening populations or studying kinetics can easily go into the thousands. The partition of digital images into segments, known as segmentation is a basic process allowing the acquisition of quantitative data that may be a number of pixels of a bidimensional field, determining the boundaries of interest in an object [13]. Segmentation discriminates between background and defines the region under study and is the basis for further data acquision.
The development of artificial intelligence processes based on machine learning (ML) has been an important step in the development of software for omic analysis and modelling [14]. Examples include support vector machines (SVM) for Illumina base calling [15], k-nearest neighbour (kNN) classification for protein localization [16] or Naïve Bayes Classifiers for phylogenetic reconstructions [17]. Furthermore, ML approaches have been used extensively in image analysis applied to plant biology and agriculture [18,19].
Plant growth occurs in a gated manner i.e., it has a major peak during the late night in hypocotyls, stems or large leaves [11,20,21]. This is the result of circadian clock regulation of genes involved in auxin and gibberellin signalling and cell expansion [22]. One of the inputs to the circadian clock is blue light transmitted through proteins that act as receptors such as ZEITLUPE/FLAVIN-BINDING, KELCH REPEAT, F-BOX and LOV KELCH PROTEIN2 [23,24]. Phytochromes absorb red and far red light such as PHYTOCHROME A [25,26]. As a result night image acquisition has to be done with near infrared (NIR) light giving the so-called extended night signal [27]. The aim of this work was to develop the corresponding algorithms to obtain data from day and night imaging. We used machine learning to analyse a set of images taken from different species during day and night. We used the aforementioned SVM, NBC and kNN to obtain image segmentations. Our results demonstrate that ML has great potential to tackle complex problems of image segmentation.

2. Materials and Methods

Figure 1a shows a schematic of the system. The data acquisition system is composed of four modules which we describe below.

2.1. Ilumination Subsystem

We pursued two goals with the illumination subsystem. First we wanted to grow plants under conditions close to their natural environments and second we wanted to acquire pictures during the night-time without interfering with the behaviour of the plant. For this purpose, we have established two illumination periods: daytime and night-time. The illumination subsystem is composed of two LED (light-emitting diode) panels which, allows to carry-out the capture image process and the same time it allows to supply the precise combination of the wavelengths for growing up correctly.
The daytime LED panel is formed by a combination of five types of LEDs emitting wavelengths with peaks in UV light (290 nm), blue light (450 and 460 nm) and red light (630 and 660 nm). The LED panel has a power of fifty watts. It is usually used for indoor growing of crop plants. The merging of wavelengths produces an illumination with a pink-red appearance (Figure 2a).
The night-time LED panel is composed by a bar of 132 NIR LEDs (three rows of forty four LEDs) with a wavelength of 850 nm (Figure 2b).
We programmed a system that would give a day/night timing whereby day light was created by turning on the daytime LED. In order to capture night images, the night-time LED panel was turned on for a period between 3 and 5 s coupled to an image capture trigger. The system can be programmed by the user for different periods of day and night lengths and time course of picture acquisition. The minimal period is one picture every 6 s and the maximal is one picture in 24 h.

2.2. Capture Subsystem

The capture module is in charge of image capture during day and night and the control of the illumination subsystem. The main capture subsystem element is a multispectral 2-channel Charge-Coupled Device (CCD) camera. A prism placed in the same optical path between the lens and CCDs allows a simultaneous capture the visible (or RGB) and NIR image (see Figure 3a). This feature has reduced the amount of cameras being used by the system and has avoided the construction of a mechanical system to move the lenses or the cameras in front of the plants. The camera has a resolution of 1024 (h) × 768 (v) active pixels per channel. During day and night a resolution of 8 bit per pixel was used in all the channels (R-G-B-NIR). Figure 3b,c shows the response of the NIR-CCD and RGB-CCD of the multispectral camera.
Capture and illumination subsystems are controlled via a GUI developed in C/C++ (Figure 4a,b). It comprises eight digital input/output channels and six analog ones in an USB-GPIO module (Figure 5a). The system had 10 bit resolution. It was configured using the termios Linux library in C/C++.
The second component was the optocoupler relay module. It had four optocoupled outputs, optocoupled to a relay triggering at voltages between 3 and 24 V. Both day light and night light LEDs were connected to two relays (Figure 5b), in such a way that the configuration via the control software dictates the beginning of image acquisition, triggers light turning on or off coordinating the light pulses with the camera during the day and night.

2.3. Image Processing Module

Each experiment generates two types of images: one NIR image during the night-time and another RGB image during the daytime. In order to obtain an automatic image segmentation, we designed an algorithm to classify the objects from the images of the experiment in two groups: organs and background. The algorithm developed is divided in three stages.

2.3.1. Extraction of Samples of Images Representative from the Different Classes

During the first stage of the algorithm we have selected a set of representative samples formed by n matrix with size of k × k pixels of each class. The size of the regions can be of 1 × 1, 16 × 16, 32 × 32, 64 × 64 or 128 × 128 pixels. This will depend of size and morphology of the organ to be classified and of the period of the daytime or night-time involved. During the day we used a single pixel per channel while we used the larger pixel regions to increase the information obtained in the IR channel.

2.3.2. Features Vector

The feature vector is usually composed by a wide variety of different types of features. The most utilized features are related to: intensity of image pixels [28], geometries [29] and textures (first and second-order statistical features) [30,31]. In addition the feature vector is computed over image transformations such as Fourier, Gabor, and Wavelet [32]. Colour images comprise three channels for R, G and B. As a result the amount of information is multiplied by three and the number of possible combinations and image transformations are incremented.
We have applied two types of features vector techniques depending whether the image was captured during daytime (RGB image) or night-time (NIR images). We tested several colour spaces to construct the features vector of daytime: RGB primary space, HSV perceptual space and CIE L*a*b* luminance-chrominance space [33]. The use of a single colour space produced poor results. In order to improve the performance we increased the number of feature vectors. We used cross combinations of the RGB, CIE L*a*b*, HSV and found that the best performance was with RGB and CIE L*a*b*. Thus we constructed the features vector formed by the pixel the corresponding pixel values. In the NIR images, we used a features vector computed over two decomposition levels of the Haar wavelet transform.

Colour Images

The features vector of the colour images is composed of six elements extracted from the pixel values of two colour spaces: RGB and CIE L*a*b*. We selected a large set of random pixels of each class to construct the features vector Figure 6a (organs-class1-green and background-class2-white). It was necessary to convert RGB colour space of the original image to CIE L*a*b* colour space. Figure 6b shows the twenty values of the features vector of class 1.

NIR Images

Discrete Wavelet Transformation (DWT) generates a set of values formed by the “wavelet coefficients”. Being f(x, y) an image of M × N size, each level of wavelet decomposition is formed by the convolution of the image f(x, y) with two filters: a low-pass filter (LPF) and a high-pass filter (HPF). The different combinations of these filters result in four images here described as LL, LH, HL and HH. In the first decomposition level four subimages or bands are produced: one smooth image, also called approximation, f LL ( 1 ) ( x , y ) , that represents an approximation of the original image f(x, y) and three detail subimages f LH ( 1 ) ( x , y ) , f HL ( 1 ) ( x , y ) and f HH ( 1 ) ( x , y ) , which represent the horizontal, vertical and diagonal details respectively. There are several wavelet mother functions that can be employed, like Haar, Daubechies, Coiflet, Meyer, Morlet, and Bior, depending on the specific problem to be identified [34,35]. Figure 7 shows the pyramid algorithm of wavelet transform in the first decomposition level.
In this work we have computed a features vector based on the wavelet transform with basis Haar [36]. The features vector is formed of four elements: maximum, minimum, mean and Shannon entropy of coefficients wavelets calculated in the horizontal, vertical and diagonal subimages in two decomposition levels (see Equations (1)–(5)): We have eliminated the approximation subimage due to it contains a representation decimated of the original image:
f 1 ,    .. ,    6 = max { f LH ( l ) ( x , y ) ,   f LH ( l ) ( x , y ) , f LH ( l ) ( x , y ) }   , l = 1 , 2
f 7 , .. , 12 = min { f LH ( l ) ( x , y ) ,   f LH ( l ) ( x , y ) , f LH ( l ) ( x , y ) } ,   l = 1 , 2
f 13 , .. , 18 = mean { f LH ( l ) ( x , y ) ,   f LH ( l ) ( x , y ) , f LH ( l ) ( x , y ) } ,   l = 1 , 2
f 19 ,    .. , 24 = shannon _ entropy { f LH ( l ) ( x , y ) ,   f LH ( l ) ( x , y ) , f LH ( l ) ( x , y ) }   , l = 1 , 2
Shannon Entropy is calculated as the Equation (5):
shannon _ entropy ( f s ( l ) ) = i = 1 M / 2 l j = 1 N / 2 l p ( w i j )   l o g 2 ( p ( w i j ) )
The letter l represents the value of the wavelet decomposition level, s the subimages (LL, HL, LH, HH) created in the wavelet decomposition, and w i j represents the wavelet coefficient (i, j), located in the s-subimage, at l-decomposition level. p represents the occurrence probability of the wavelet coefficient w i j .
Feature vector has been obtained applying the Equations (1)–(5) to each region in two wavelet decomposition levels with Haar basis. The result was a feature vector of twenty-four elements ( f 1 , .. , 24 ) per region of size k × k.

2.3.3. Classification Process

We have tested three machine-learning algorithms: (1) k-nearest neighbour (kNN); (2) naive Bayes classifier (NBC), and Support Vector Machine (SVM). The algorithms selected belong to the type of supervised classification. These type of algorithms require of a training stage before performing the classification process.
kNN classifier is a non-parametric method for classifying objects in a multi-dimensional space. After being trained, kNN assigns a specific class to a new object depending on the majority of votes from its neighbours. This measure is based in metrics such as Euclidean, Hamming or Mahalanobis distances. In the implementation of kNN algorithm it is necessary to assign an integer value to k. This parameter represents the k-neighbors used to carry-out the voting classification. A k optimal determination will allow that the good model adjusts to future data [37]. It is recommendable to use data normalisation coupled to kNN classifiers in order to avoid the predominance of big values over small values in the features vector.
NBC uses a probabilistic learning classification. Classifiers based on Bayesian methods utilize training data to calculate an observed probability of each class based on feature values. When the classifier is used later on unlabeled data, it uses the observed probabilities to predict the most likely class for the new features. As NBC works with probabilities it does not need data normalization.
SVM is a supervised learning algorithm where given labeled training data, it outputs a boundary which divides data by categories and categorizes new examples. The goal of a SVM is to create a boundary, called hyperplane, which leads to homogeneous partitions of data on either side. SVMs can also be extended to problems were the data are not linearly separable. SVMs can be adapted for use with nearly any type of learning task, including both classification and numeric prediction. SVM classifier tend to perform better after data normalisation.
In this work we have used raw data and two types of normalisation procedures which have been computed over features space: dn0: without normalisation; dn1: mean and standard-deviation normalization and dn2: mode and standard-deviation normalisation. Features space of each class is composed by a m × n matrix (see Equation (6)):
f i j n C = [ f 11 1 f 1 n 1 f 1 m 1 f m n 1 · · · f 11 n C f 1 n n C f 1 m n C f m n n C ]   i = 1 , , m   ;   j = 1 , , n   ; n C = 1 ,   2
Being i-th row, the vector of features i-th of features space formed by n features. m represents the number of vectors in the features space and nC represents the number of classes in th space.
The normalised features space, F i j n c , depending on the normalisation types (dn0, dn1, dn2) is computed as is shown in the Equation (7):
F i j n c = [ f 11 1 s t 1 s t 2 f 1 n 1 s t 1 s t 2 f 1 m 1 s t 1 s t 2 f m n 1 s t 1 s t 2 f 11 n C s t 1 s t 2 f 1 n n C s t 1 s t 2 f 1 m n C s t 1 s t 2 f m n n C s t 1 s t 2 ] { d n 0 { s t 1 = 0 s t 2 = 1 r a w d a t a d n 1 { s t 1 = m e a n ( f i j n C ) s t 2 = S t d D e s ( f i j n C ) d n 2 { s t 1 = m o d e ( f i j n C ) s t 2 = S t d D e s ( f i j n C )
To obtain the best result in classification process, the ML algorithms were tested with different configuration parameters. kNN was tested with Euclidean and Minkowski distances with three type of data normalisation, NBC was tested with Gauss and Kernel Smoothing Functions (KSF) without data normalisation, and SVM was tested with linear and quadratic functions, on three types of data normalisation. The ML algorithms used two classes of objects. One for the plant organs and a second one for the background. In all of them we applied the leave-out cross validation (LOOCV) method to measure of the error of the classifier. Basically LOOCV method extracts a sample of the training set and it constructs the classifier with the remaining of the training samples. Then it evaluates the classification error and the process is repeated for all the training samples. At end the LOOCV method computes the mean of the errors and it obtains a measure of how model is adjusted to data. This method allows comparing the results of the different ML algorithms, provided that they will be applied to same sample data. Table 1 shows a summary of parameters used for the classification process.

2.4. Experimental Validation

In order to test and validate the functioning of the system and ML methods, acquired pictures of Antirrhinum majus and Antirrhinum linkianum were used to analyse growth kinetics. The camera was positioned above the plants under study. The daylight LED panel was above the plants while the night-time LED was at a 45°. Data acquisition was performed for a total of six days. We obtained one image every 10 min during day and night. Day night cycles were set to 12:12 h and triggering of the NIR LED for image acquisition during the night was done for a total of 6 s.
In the experiment we obtained 864 colour images from the RGB sensor which were transformed to CIE L*a*b* colour space and 864 gray scale images from the NIR sensor. From each group (RGB and NIR images) we obtained fifty ground-truth images which were segmented manually by human experts. From the fifty ground-truth colour images, we selected 1200 samples of 1 pixel, which we used to train the RGB image processing ML algorithms. From the second fifty ground-truth NIR images we took 1200 regions of 32 × 32 pixels which were to train the NIR image processing ML algorithms. In both cases we selected 600 samples belonging to organs class and 600 samples belonging to background class.
Figure 8 shows two images from each day of the experiment for different capture periods. We can distinguish easily the growth stages of the two species during the daytime and night-time.

3. Results and Discussion

We evaluated the results of training stage of the ML algorithms with a leave-one-out cross-validation method (LOOCV) and with the Receiver Operating Characteristic (ROC) curve over data training sets obtained from RGB and NIR images. LOOCV and ROC curves have been applied under the different ML configurations shown in the Table 1. This allowed to select the optimal ML algorithm to be applied to each type of image depending on when it was captured: during daytime or night-time. Once we determined the optimal ML algorithm for each image set, we used the metric miss-classification [37] to evaluate the final performance of the implemented ML algorithms.

3.1. LOOCV

Table 2 shown the errors obtained after to apply LOOCV to the two images groups respectively. In both images groups the minimum error in data model adjust is produced with kNN classifier. Data normalisation based on in the mean (dn1) produced the best result in both cases, too. The maximum error is produced by the NBC classifier, with KSF and Gauss kernel, respectively.

3.2. ROC Curves

We evaluated the performance of the ML algorithms using Receiver Operating Characteristic (ROC) curve. The ROC curve is created by comparing the sensitivity (the rate of true positives TP, see Equation (8)), versus 1-specificity (the rate of false positives FP see Equation (9)), at various threshold levels [38]. The ROC curves, shown in Figure 8 and Figure 9, allow comparing the results between of ML algorithms per each groups of images:
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
The Area Under the Curve (AUC) is usually used by ML to compare statistical models. AUC can be interpreted as the probability that the classifier will assign a higher score to a randomly chosen positive example than to a randomly chosen negative example [39,40].
Figure 9 shows ROC curves computed over the set of colour images training. The higher values of AUC were obtained by kNN classifier with Euclidean distance. Concerning the data normalisation, we achieved similar results with raw data and normalisation based on the mode and standard deviation (dn0 and dn1).
Figure 10 shows ROC curves computed over the set of NIR images training. We can observe that the best results were obtained with the SVM classifier with quadratic functions and using a normalised data based on the mean and standard deviation (dn1). Table 3 shows AUC values obtained from ROC curves of the Figure 9 and Figure 10.
In both cases after classification stages, we performed a postprocessing stage composed of morphological operations and an area filter to eliminate noise and small particles. The segmented images were merged with the original images (Figure 11 and Figure 12).

3.3. Error Segmentation

The misclassification error (ME) represents the percentage of the background pixels that are incorrectly allocated to the object (i.e., to the foreground) or vice versa. The error can be calculated by means of Equation (10), where BGT (Background Ground-Truth image) and OGT (Object Ground-Truth image) represent the ground-truth image of the background and of the object taken as reference, and BT (Background Test) and OT (Object Test) represent the image to be assessed. In the event that the test image coincides with the pattern image, the classification error will be zero and therefore the performance of the segmentation will be the maximum [41].
M E = | B G T B T | + | O G T O T | | B G T | + | O G T |
The performance of the implemented algorithms is assessed according to the Equation (11):
η = 100 · ( 1 M E )
Table 4 shows mean values computed after to segment the fifty ground-truth images of each group with optimal ML algorithm (kNN and SVM) selected in the previous subsection.
The performance of the image segmentation calculated shows excellent results in both groups (Table 4). It has been necessary to increase the complexity of the features vector in the NIR images (using Wavelet transform) to obtain a similar results of performance. This is probably an expected result as RGB images had three times more of information than a NIR image.

4. Conclusions and Future Work

In this work we have developed a system based on ML algorithms and computer vision intended to solve the automatic phenotype data analysis. The system is composed by a growth-chamber with capacities to perform experiments with numerous species. The design of the growth-chamber has allowed easy positioning of different cameras and illuminations. The system can take thousands of images through the capture subsystem, capture spectral images and it creates time-lapse series of the specie during the experiment. One of the main goals of this work has been to capture images during the night-time without affecting plant growth.
We have used three different ML algorithms for image segmentation: k-nearest neighbour (kNN), Naive Bayes classifier (NBC), and Support Vector Machine. Each ML algorithm was executed with different kernel functions: kNN with Euclidean and Minkowski distances, NBC with Gauss and KSF functions and SVM with linear and quadratic functions. Furthermore ML algorithms have been trained with two types of data normalisation: dn0 (raw data), dn1 (mean and standard deviation) and dn2 (mode & standard deviation). Our results show that RGB images are better classified with the kNN classifier, Euclidean distance and without data normalisation. In contrast, NIR images performed better with SVM classifier with quadratic function and with data normalisation dn1.
In the last stage we have applied ME metrics to measure the image segmentation performance. We have achieved a performance of 99.3% in both ground-truth colour images and ground-truth NIR images. Currently the algorithms are being used in an automatic image segmentation processing to study circadian rhythm in wild type lines, transposon-tagged mutants and transgenic lines with modifications in genes involved in the control of growth and the circadian clock. Regarding future work, we consider important to identify a new feature vector which produces better performance rates, improve the illumination subsystem, reduce the computation time of the windowing segmentation in NIR images.

Acknowledgments

The work has been partially supported and funded by the Spanish Ministerio de Economía y Competitividad (MINECO) under the projects ViSelTR (TIN2012-39279), BFU2013-45148-R and cDrone (TIN2013-45920-R).

Author Contributions

Pedro J. Navarro, Fernándo Pérez, Julia Weiss and Marcos Egea-Cortines conceived and designed the experiments; Pedro J. Navarro, Fernándo Pérez, Julia Weiss, and Marcos Egea-Cortines performed the experiments; Pedro J. Navarro and Fernándo Pérez analyzed the data; Julia Weiss contributed reagents/materials/analysis tools; Pedro J. Navarro, Fernándo Pérez and Marcos Egea-Cortines wrote the paper. Pedro J. Navarro, Fernándo Pérez, Julia Weiss and Marcos Egea-Cortines corrected the draft and approved the final version.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LMMachine learning
KSFKernel smoothing function
kNNk-nearest neighbour
NBCNaïve Bayes classifier
SVMSupport vector machines
LOOCVLeave-one-out cross validation method
AUCArea under curve
MEMisclassification error

References

  1. Fahlgren, N.; Gehan, M.; Baxter, I. Lights, camera, action: High-throughput plant phenotyping is ready for a close-up. Curr. Opin. Plant Biol. 2015, 24, 93–99. [Google Scholar] [CrossRef] [PubMed]
  2. Deligiannidis, L.; Arabnia, H. Emerging Trends in Image Processing, Computer Vision and Pattern Recognition; Elsevier: Boston, MA, USA, 2014. [Google Scholar]
  3. Dee, H.; French, A. From image processing to computer vision: Plant imaging grows up. Funct. Plant Biol. 2015, 42, iii–v. [Google Scholar] [CrossRef]
  4. Furbank, R.T.; Tester, M. Phenomics—Technologies to relieve the phenotyping bottleneck. Trends Plant Sci. 2011, 16, 635–644. [Google Scholar] [CrossRef] [PubMed]
  5. Dhondt, S.; Wuyts, N.; Inzé, D. Cell to whole-plant phenotyping: The best is yet to come. Trends Plant Sci. 2013, 18, 428–439. [Google Scholar] [CrossRef] [PubMed]
  6. Tisné, S.; Serrand, Y.; Bach, L.; Gilbault, E.; Ben Ameur, R.; Balasse, H.; Voisin, R.; Bouchez, D.; Durand-Tardif, M.; Guerche, P.; et al. Phenoscope: An automated large-scale phenotyping platform offering high spatial homogeneity. Plant J. 2013, 74, 534–544. [Google Scholar] [CrossRef] [PubMed]
  7. Li, L.; Zhang, Q.; Huang, D. A review of imaging techniques for plant phenotyping. Sensors 2014, 14, 20078–20111. [Google Scholar] [CrossRef] [PubMed]
  8. Honsdorf, N.; March, T.J.; Berger, B.; Tester, M.; Pillen, K. High-throughput phenotyping to detect drought tolerance QTL in wild barley introgression lines. PLoS ONE 2014, 9, 1–13. [Google Scholar] [CrossRef] [PubMed]
  9. Barron, J.; Liptay, A. Measuring 3-D plant growth using optical flow. Bioimaging 1997, 5, 82–86. [Google Scholar] [CrossRef]
  10. Aboelela, A.; Liptay, A.; Barron, J.L. Plant growth measurement techniques using near-infrared imagery. Int. J. Robot. Autom. 2005, 20, 42–49. [Google Scholar] [CrossRef]
  11. Navarro, P.J.; Fernández, C.; Weiss, J.; Egea-Cortines, M. Development of a configurable growth chamber with a computer vision system to study circadian rhythm in plants. Sensors 2012, 12, 15356–15375. [Google Scholar] [CrossRef] [PubMed]
  12. Nguyen, T.; Slaughter, D.; Max, N.; Maloof, J.; Sinha, N. Structured light-based 3d reconstruction system for plants. Sensors 2015, 15, 18587–18612. [Google Scholar] [CrossRef] [PubMed]
  13. Spalding, E.P.; Miller, N.D. Image analysis is driving a renaissance in growth measurement. Curr. Opin. Plant Biol. 2013, 16, 100–104. [Google Scholar] [CrossRef] [PubMed]
  14. Navlakha, S.; Bar-joseph, Z. Algorithms in nature: The convergence of systems biology and computational thinking. Mol. Syst. Biol. 2011, 7. [Google Scholar] [CrossRef] [PubMed]
  15. Kircher, M.; Stenzel, U.; Kelso, J. Improved base calling for the Illumina Genome Analyzer using machine learning strategies. Genome Biol. 2009, 10, 1–9. [Google Scholar] [CrossRef] [PubMed]
  16. Horton, P.; Nakai, K. Better prediction of protein cellular localization sites with the k nearest neighbors classifier. Proc. Int. Conf. Intell. Syst. Mol. Biol. 1997, 5, 147–152. [Google Scholar] [PubMed]
  17. Yousef, M.; Nebozhyn, M.; Shatkay, H.; Kanterakis, S.; Showe, L.C.; Showe, M.K. Combining multi-species genomic data for microRNA identification using a Naive Bayes classifier. Bioinformatics 2006, 22, 1325–1334. [Google Scholar] [CrossRef] [PubMed]
  18. Tellaeche, A.; Pajares, G.; Burgos-Artizzu, X.P.; Ribeiro, A. A computer vision approach for weeds identification through Support Vector Machines. Appl. Soft Comput. J. 2011, 11, 908–915. [Google Scholar] [CrossRef] [Green Version]
  19. Guerrero, J.M.; Pajares, G.; Montalvo, M.; Romeo, J.; Guijarro, M. Support Vector Machines for crop/weeds identification in maize fields. Expert Syst. Appl. 2012, 39, 11149–11155. [Google Scholar] [CrossRef]
  20. Covington, M.F.; Harmer, S.L. The circadian clock regulates auxin signaling and responses in Arabidopsis. PLoS Biol. 2007, 5, 1773–1784. [Google Scholar] [CrossRef] [PubMed]
  21. Nusinow, D.A.; Helfer, A.; Hamilton, E.E.; King, J.J.; Imaizumi, T.; Schultz, T.F.; Farré, E.M.; Kay, S.A.; Farre, E.M. The ELF4-ELF3-LUX complex links the circadian clock to diurnal control of hypocotyl growth. Nature 2011, 475, 398–402. [Google Scholar] [CrossRef] [PubMed]
  22. De Montaigu, A.; Toth, R.; Coupland, G. Plant development goes like clockwork. Trends Genet. 2010, 26, 296–306. [Google Scholar] [CrossRef] [PubMed]
  23. Baudry, A.; Ito, S.; Song, Y.H.; Strait, A.A.; Kiba, T.; Lu, S.; Henriques, R.; Pruneda-Paz, J.L.; Chua, N.H.; Tobin, E.M.; et al. F-box proteins FKF1 and LKP2 Act in concert with ZEITLUPE to control arabidopsis clock progression. Plant Cell 2010, 22, 606–622. [Google Scholar] [CrossRef] [PubMed]
  24. Kim, W.-Y.Y.; Fujiwara, S.; Suh, S.-S.S.; Kim, J.; Kim, Y.; Han, L.Q.; David, K.; Putterill, J.; Nam, H.G.; Somers, D.E. ZEITLUPE is a circadian photoreceptor stabilized by GIGANTEA in blue light. Nature 2007, 449, 356–360. [Google Scholar] [CrossRef] [PubMed]
  25. Khanna, R.; Kikis, E.A.; Quail, P.H. EARLY FLOWERING 4 functions in phytochrome B-regulated seedling de-etiolation. Plant Physiol. 2003, 133, 1530–1538. [Google Scholar] [CrossRef] [PubMed]
  26. Wenden, B.; Kozma-Bognar, L.; Edwards, K.D.; Hall, A.J.W.; Locke, J.C.W.; Millar, A.J. Light inputs shape the Arabidopsis circadian system. Plant J. 2011, 66, 480–491. [Google Scholar] [CrossRef] [PubMed]
  27. Nozue, K.; Covington, M.F.; Duek, P.D.; Lorrain, S.; Fankhauser, C.; Harmer, S.L.; Maloof, J.N. Rhythmic growth explained by coincidence between internal and external cues. Nature 2007, 448, 358–363. [Google Scholar] [CrossRef] [PubMed]
  28. Fernandez, C.; Suardiaz, J.; Jimenez, C.; Navarro, P.J.; Toledo, A.; Iborra, A. Automated visual inspection system for the classification of preserved vegetables. In Proceedings of the 2002 IEEE International Symposium on Industrial Electronics, ISIE 2002, Roma, Italy, 8–11 May 2002; pp. 265–269.
  29. Chen, Y.Q.; Nixon, M.S.; Thomas, D.W. Statistical geometrical features for texture classification. Pattern Recognit. 1995, 28, 537–552. [Google Scholar] [CrossRef]
  30. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, 610–621. [Google Scholar] [CrossRef]
  31. Zucker, S.W.; Terzopoulos, D. Finding structure in co-occurrence matrices for texture analysis. Comput. Graph. Image Process. 1980, 12, 286–308. [Google Scholar] [CrossRef]
  32. Bharati, M.H.; Liu, J.J.; MacGregor, J.F. Image texture analysis: Methods and comparisons. Chemom. Intell. Lab. Syst. 2004, 72, 57–71. [Google Scholar] [CrossRef]
  33. Navarro, P.J.; Alonso, D.; Stathis, K. Automatic detection of microaneurysms in diabetic retinopathy fundus images using the L*a*b color space. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2016, 33, 74–83. [Google Scholar] [CrossRef] [PubMed]
  34. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  35. Ghazali, K.H.; Mansor, M.F.; Mustafa, M.M.; Hussain, A. Feature extraction technique using discrete wavelet transform for image classification. In Proceedings of the 2007 5th Student Conference on Research and Development, Selangor, Malaysia, 11–12 December 2007; pp. 1–4.
  36. Arivazhagan, S.; Ganesan, L. Texture classification using wavelet transform. Pattern Recognit. Lett. 2003, 24, 1513–1521. [Google Scholar] [CrossRef]
  37. Lantz, B. Machine Learning with R; Packt Publishing Ltd.: Birmingham, UK, 2013. [Google Scholar]
  38. Hastie, T.; Tibshirani, R.; Friedman, J. The elements of statistical learning. Elements 2009, 1, 337–387. [Google Scholar]
  39. Bradley, A. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef] [Green Version]
  40. Hand, D.J. Measuring classifier performance: A coherent alternative to the area under the ROC curve. Mach. Learn. 2009, 77, 103–123. [Google Scholar] [CrossRef]
  41. Sezgin, M.; Sankur, B. Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging 2004, 13, 146–168. [Google Scholar]
Figure 1. Growth chamber: (a) Functional schematic of the system; (b) Experiment with Petunia x hybrida during daytime in the real system.
Figure 1. Growth chamber: (a) Functional schematic of the system; (b) Experiment with Petunia x hybrida during daytime in the real system.
Sensors 16 00641 g001
Figure 2. Illumination subsystem (a) Daytime LED panel; (b) Nightime LED panel.
Figure 2. Illumination subsystem (a) Daytime LED panel; (b) Nightime LED panel.
Sensors 16 00641 g002
Figure 3. Capture subsystem (a) Prism between lens and CCDs; (b) Camera NIR-IR response; (c) Camera RGB response.
Figure 3. Capture subsystem (a) Prism between lens and CCDs; (b) Camera NIR-IR response; (c) Camera RGB response.
Sensors 16 00641 g003
Figure 4. Graphical User Interface for capture subsystem: (a) Image control tap; (b) Time control tab.
Figure 4. Graphical User Interface for capture subsystem: (a) Image control tap; (b) Time control tab.
Sensors 16 00641 g004
Figure 5. Hardware of the capture subsystem. (a) USB-GPIO module (red-board) and opto-coupler relay module (green-board); (b) Electric connections between all hardware modules in the growth-chamber.
Figure 5. Hardware of the capture subsystem. (a) USB-GPIO module (red-board) and opto-coupler relay module (green-board); (b) Electric connections between all hardware modules in the growth-chamber.
Sensors 16 00641 g005
Figure 6. Colour images features vector construction. A matrix of different pixel values corresponding to R, G, B, and L*a*b* colour spaces.
Figure 6. Colour images features vector construction. A matrix of different pixel values corresponding to R, G, B, and L*a*b* colour spaces.
Sensors 16 00641 g006
Figure 7. First level of direct 2D-DWT decomposition.
Figure 7. First level of direct 2D-DWT decomposition.
Sensors 16 00641 g007
Figure 8. Images of the experiment captured every 12 h.
Figure 8. Images of the experiment captured every 12 h.
Sensors 16 00641 g008
Figure 9. ROC results for training colour images. (a) kNN classifier with distance Euclidean and data normalisation: dn0, dn1 and dn2; (b) kNN classifier with distance Minkowski and data normalisation: dn0, dn1 and dn2; (c) BN classifier with Gauss and KSF kernels and data normalisation dn0; (d) SVM classifier with lineal and quadratic polynomial functions and data normalisation: dn1 and dn2.
Figure 9. ROC results for training colour images. (a) kNN classifier with distance Euclidean and data normalisation: dn0, dn1 and dn2; (b) kNN classifier with distance Minkowski and data normalisation: dn0, dn1 and dn2; (c) BN classifier with Gauss and KSF kernels and data normalisation dn0; (d) SVM classifier with lineal and quadratic polynomial functions and data normalisation: dn1 and dn2.
Sensors 16 00641 g009aSensors 16 00641 g009b
Figure 10. ROC results for training NIR images. (a) kNN classifier with distance Euclidean and data normalisation: dn0, dn1 and dn2; (b) kNN classifier with distance Minkowski and data normalisation: dn0, dn1 and dn2; (c) BN classifier with Gauss and KSF kernels and data normalisation dn0; (d) SVM classifier with lineal and quadratic polynomial functions and data normalisation: dn1 and dn2.
Figure 10. ROC results for training NIR images. (a) kNN classifier with distance Euclidean and data normalisation: dn0, dn1 and dn2; (b) kNN classifier with distance Minkowski and data normalisation: dn0, dn1 and dn2; (c) BN classifier with Gauss and KSF kernels and data normalisation dn0; (d) SVM classifier with lineal and quadratic polynomial functions and data normalisation: dn1 and dn2.
Sensors 16 00641 g010
Figure 11. Results of the colour image segmentation based on kNN classifier in different growing stages. Left shows four colour images selected at different growth stages. The second column presents the result of segmentation using kNN classifier with Euclidean distance and without normalisation data (dn0). The third column shows the results after post-processing stage.
Figure 11. Results of the colour image segmentation based on kNN classifier in different growing stages. Left shows four colour images selected at different growth stages. The second column presents the result of segmentation using kNN classifier with Euclidean distance and without normalisation data (dn0). The third column shows the results after post-processing stage.
Sensors 16 00641 g011
Figure 12. Results of the NIR image segmentation based on SVM. Left shows four NIR images chosen at different growth stages growing during night-time. The second column shows the segmentation results after application of the SVM classifier with quadratic function and with data normalisation dn1. The third column shows the results after post-processing stage.
Figure 12. Results of the NIR image segmentation based on SVM. Left shows four NIR images chosen at different growth stages growing during night-time. The second column shows the segmentation results after application of the SVM classifier with quadratic function and with data normalisation dn1. The third column shows the results after post-processing stage.
Sensors 16 00641 g012
Table 1. kNN, NBC and SVM configuration parameters.
Table 1. kNN, NBC and SVM configuration parameters.
ConfigurationkNNNBCSVM
methodEuclidean, MinkowskiGauss, KSFLinear, quadratic
data normalisationdn0, dn1, dn2dn0dn1, dn2
metricsLOOCV, ROCLOOCV, ROCLOOCV, ROC
classes222
Table 2. Colour images and NIR images. LOOCV error for kNN, NBC, SVM.
Table 2. Colour images and NIR images. LOOCV error for kNN, NBC, SVM.
ClassifierkNNNBCSVM
ConfigurationEuclideanMinkowskiGaussKSFLinearQuadratic
Colourdn00.02830.04330.07500.0758--
dn10.02420.0467--0.05330.0383
dn20.02830.0433--0.06670.0450
NIRdn00.02880.0394,0.03560.0319--
dn10.01690.0281--0.03260.0319
dn20.02880.0394--0.03440.0325
Table 3. Colour images and NIR images. AUC for kNN, NBC, SVM.
Table 3. Colour images and NIR images. AUC for kNN, NBC, SVM.
ClassifierkNNNBCSVM
ConfigurationEuclideanMinkowskiGaussKSFLinearQuadratic
Colourdn00.99840.99740.95420.9778--
dn10.99790.9976--0.96220.9875
dn20.99840.9974--0.94960.9886
NIRdn00.99870.99790.98770.9963--
dn10.99750.9993--0.98671.000
dn20.99870.9979--0.98680.9932
Table 4. Performance in the image segmentation.
Table 4. Performance in the image segmentation.
ClassifierkNNSVM
Performance99.311%99.342%

Share and Cite

MDPI and ACS Style

Navarro, P.J.; Pérez, F.; Weiss, J.; Egea-Cortines, M. Machine Learning and Computer Vision System for Phenotype Data Acquisition and Analysis in Plants. Sensors 2016, 16, 641. https://0-doi-org.brum.beds.ac.uk/10.3390/s16050641

AMA Style

Navarro PJ, Pérez F, Weiss J, Egea-Cortines M. Machine Learning and Computer Vision System for Phenotype Data Acquisition and Analysis in Plants. Sensors. 2016; 16(5):641. https://0-doi-org.brum.beds.ac.uk/10.3390/s16050641

Chicago/Turabian Style

Navarro, Pedro J., Fernando Pérez, Julia Weiss, and Marcos Egea-Cortines. 2016. "Machine Learning and Computer Vision System for Phenotype Data Acquisition and Analysis in Plants" Sensors 16, no. 5: 641. https://0-doi-org.brum.beds.ac.uk/10.3390/s16050641

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop