Next Article in Journal
Mechanization Status, Promotional Activities and Government Strategies of Thailand and Vietnam in Comparison to Bangladesh
Next Article in Special Issue
Detection of Black Spot of Rose Based on Hyperspectral Imaging and Convolutional Neural Network
Previous Article in Journal
Essential Oil Content of Baccharis crispa Spreng. Regulated by Water Stress and Seasonal Variation
Previous Article in Special Issue
Deep Learning Application in Plant Stress Imaging: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images

1
Faculty of Engineering, Universidad Nacional de Colombia, Bogotá 111321, Colombia
2
External Consultant, Bogotá 111221, Colombia
3
Faculty of Engineering, Universidad de Cundinamarca, Fusagasuga 252212, Colombia
*
Author to whom correspondence should be addressed.
Submission received: 30 June 2020 / Revised: 19 August 2020 / Accepted: 20 August 2020 / Published: 28 August 2020
(This article belongs to the Special Issue Precision Agriculture Technologies for Management of Plant Diseases)

Abstract

:
Weed management is one of the most important aspects of crop productivity; knowing the amount and the locations of weeds has been a problem that experts have faced for several decades. This paper presents three methods for weed estimation based on deep learning image processing in lettuce crops, and we compared them to visual estimations by experts. One method is based on support vector machines (SVM) using histograms of oriented gradients (HOG) as feature descriptor. The second method was based in YOLOV3 (you only look once V3), taking advantage of its robust architecture for object detection, and the third one was based on Mask R-CNN (region based convolutional neural network) in order to get an instance segmentation for each individual. These methods were complemented with a NDVI index (normalized difference vegetation index) as a background subtractor for removing non photosynthetic objects. According to chosen metrics, the machine and deep learning methods had F1-scores of 88%, 94%, and 94% respectively, regarding to crop detection. Subsequently, detected crops were turned into a binary mask and mixed with the NDVI background subtractor in order to detect weed in an indirect way. Once the weed image was obtained, the coverage percentage of weed was calculated by classical image processing methods. Finally, these performances were compared with the estimations of a set from weed experts through a Bland–Altman plot, intraclass correlation coefficients (ICCs) and Dunn’s test to obtain statistical measurements between every estimation (machine-human); we found that these methods improve accuracy on weed coverage estimation and minimize subjectivity in human-estimated data.

1. Introduction

As the world population increases, so does the demand for food. Taking into account that land, water, and labor are limited resources, it is estimated that the efficiency of agricultural productivity will increase by 25% by the year 2050 [1]. Therefore, it is crucial to focus on the problems faced by the agricultural industry. Liakos et al. [2] propose different categories to classify the challenges faced by machine learning in precision agriculture, such as livestock management, water management, soil management, detection of plant diseases, crop quality, species recognition, and weed detection. New developments in this last category would help to face the most important biological threat in crop productivity. According to Wang [3], an average of 34% of production is lost because weeds directly compete for nutrients, water, and sunlight. Furthermore, weeds are harder to detect due to their non-uniform presence and their overlap with other crops.
The oldest technique used to control weeds in crops is manual weeding. However, it is labor- and time-consuming, which makes it inefficient for larger scale crops. Today, agricultural industry has chemical weeding systems, and to a lesser extent, mechanic weeding systems, but in our context (Andean Highlands), 75% of the vegetables produced, such as lettuces, have manual weeding, which makes production even more inefficient and expensive [4]. Moreover, there is a very high margin of error regarding weeding and these systems may end up damaging the plants [2].
Authors should discuss the results and how they can be interpreted from the perspectives of previous studies and the working hypotheses. The findings and their implications should be discussed in the broadest context possible. Future research directions may also be highlighted.
Due to the herbicide usage policies in some countries and the high cost in areas with underdeveloped agricultural processes, there is a need to carry out weed detection and calculation processes in order to reduce the environmental impact of herbicides and optimize their use [5].
However, regardless of the control method, field information will always be needed to make decisions. This has been done through conventional sampling, which consists of visually estimating the percentage of weeds with respect to the soil (coverage), the number of individuals by area (density), the number of times that each species appears in a certain number of samplings (frequency), or the estimation of weed weight by area (biomass) [6].
The methodologies of conventional sampling widely vary depending on the objective of the information required. There are variations regarding the number of sampling sites, their distribution in the field, and the variables used, among other things. Ambrosio et al. [7] affirms that conventional sampling by square grids is one of the simplest and most effective methods, since weeds have an aggregate distribution and weed density can be determined as a variable that can follow a Poisson distribution or a negative binomial probability distribution [8].
According to López Granados et al. [9] the main purpose of SSWM (site-specific weed management) systems is to spray herbicides on groups of weeds, taking into account their density and the type of weed. In early growth stages, crops are highly vulnerable so it is important to have the information to support management decisions and efficiently control weeds. Additionally, in our context, in a plot there could be 20 or more different species of weeds in a one-hectare field, reaching 100% coverage in 1–2 months, with some of the weeds seeding in 1–2 months and having four to five cohorts of weeds in a 4–6 month crop production cycle [10]. This could happen in any month of the year, since we do not have temperate climate seasons (only wet and not so wet seasons).
However, their detection might be difficult because the crop plants are small and may have some characteristics similar to those of weeds. Therefore, satellite images do not provide the necessary information. Unmanned aircraft systems (UASs) offer a better solution, since they can be automatically or remotely controlled at short and long distances, capturing digital images at the height and frequency required by the user, which varies depending on the type of UAV (unmanned aerial vehicle) that operates the system [11,12].
Additionally, these UAVs can be equipped with multispectral cameras that offer more information than an RGB digital image, since they capture spectral bands not recognized by the human eye—such as near infrared (NIR)—providing information on aspects such as the reflectance of visible light and vegetation indices. These components allow one to find important correlations that help with making different estimates.
Multispectral sensors also offer other benefits such as facilitating the differentiation of some types of plants from their environment, thanks to the different vegetation indices such as the chlorophyll absorption index and the cellulose absorption index, among others. Moreover, they provide more information during the feature extraction processes, such as those used in the OBIA methods (object-based image analysis) [11,13,14,15,16], and SVM-based (vector support machines) classification techniques [8,17,18].
Multispectral images depend on the climatic conditions of the days in which they are captured, since climate changes a plant’s reflectance due to the difference in the amount of light absorbed. Additionally, in the early stages of some crops, certain vegetation indices are similar to those of weeds that surround them; that fact and cropping systems with mulch soil cover pose some challenges for the classification algorithms, since they do the identification on images that are very complex. These and other challenges were faced using artificial neural networks (ANN), which in the last four years have been the most widely used method of weed detection, as Behmann et al. mentioned in their research on weed detection using images [19,20,21,22,23,24,25,26].
The convolutional neural networks (CNN) are setting a strong trend for the classification and identification of image-based patterns. Among the articles reviewed are some of the convolutional neural network architectures that are well-accepted for weed identification, such as AlexNet, ResNet, RCNN, VGG, GoogleNet, and FCN (fully convolutional network) [27,28,29,30,31,32,33,34,35,36,37,38]. The main advantage of such methods is that they speed up the manual labeling of data sets, while maintaining a good classification performance. This approach stands out among the other conventional index-based approaches (such as NDVI) because it is 98.7% accurate. Additionally, since images are the input itself, the problems of SVM and ANN regarding the selection and extraction of features are reduced. Therefore, it represents a major trend in terms of machine learning and pattern recognition, in addition to its great performance with all types of images; so CNNs are increasingly used for weed detection [38,39].
Previous works in lettuce fields such as those of Raja et al. [40] and Elstone et al. Raja et al. [41] obtained good results for weed and crop identification using RGB and multiespectral images, but weeds in highland tropical conditions have different forms and grow in big patches, making detection difficult; therefore, we developed methods to firstly perform vegetation detection, secondly perform crop detection, and lastly identify and quantify weeds in the field.
This study compared the weed quantification results of three methods which used machine learning and multispectral images to the results of three experts in agronomy-weed management with different weed estimation levels of training in the same set of images. The first one was based on histograms of oriented gradients and support vector machines (method 1: HOG-SVM); the second one used a CNN (convolutional neural network) that specializes in object detection (method 2: YOLO); and the third one used masks for edge detection and the "regions with CNN features" algorithm for crop recognition. The main objective in this study was to implement and establish a comparison between one machine learning method and two deep learning approaches for weed detection. Additionally, we wanted to compare those results with three different estimations of weed science experts. Notice that each expert had a different level of academic background and the subjectivity component in every estimation could establish significant differences in statistical terms.

2. Materials and Methods

The images and the methods used are described below, along with their most important elements. Then, the performances of these weed quantification methods are described, and finally they are compared against the estimates of three weed experts.

2.1. Dataset

The drone used was a Mavic Pro with the Parrot Sequoia multispectral camera (Figure 1a), which followed an automatic flight plan at 1 m/s and 2 m high. Regarding the weather conditions, the flight was made around 15 h, partly cloudy but with good sun exposure. The total route of the 100 images used corresponds to approximately 600 m 2 .
The photographs were captured in a commercial lettuce crop, 60 days after seeding and 20 days after the first manual weeding. An unmanned aerial vehicle was used to capture images at 2 m height. Later, 100 images were selected, each with a pixel size of 1280 × 960 (0.22 cm/px) and four spectral bands: green 550 nm, red 660 nm, red edge 735 nm, and near infrared 790 nm. Finally, a false green image was generated, which is the union of the red, green, and near infrared bands, in order to highlight the vegetation (Figure 2).

2.2. Experts’ Evaluations

Three experts in agronomy-weed management made visual estimations pf weed coverage for 100 images with 2219 lettuces and high, medium, and low levels of weeds. It is noteworthy that the experts had different training levels in weed estimation: expert 1 had a BSc in agronomy with some weed coverage estimation training, expert 2 had an MSc in agronomy with proper training, and expert 3 had a PhD in weed science and extensive experience in estimation of coverage of weeds.

2.3. Method 1: HOG-SVM

Support vector machines are mathematical models whose main objective is to find the optimal separating hyperplane between two different classes. This paradigm is based on a type of vectors known as support vectors that can be in a space of infinite dimensions mapped by a kernel and a cost parameter [42]. In this case, the histograms of oriented gradients (HOG) were used as the technique to extract the features in each image object. Subsequently, a training process was carried out using the polynomial kernel with the best performance in [8], using 1400 images. The main objective of this method is to identify the class (weed or lettuce) of a specific object using a mask created through the NDVI and the OTSU method; the Figure 3 shows the flow diagram.
A histogram of oriented gradients represents an object based on the magnitude and orientation of the gradient from a specific set of pixel blocks [8]. This type of feature allows obtaining the shape of a plant with a well-defined geometry. Figure 3 shows the detection process, including the preprocessing phase using the NDVI index as a background estimator. This index creates a mask that removes some objects such as the ground and elements that are not related to vegetation. On the other hand, the edges of this mask are calculated and their respective coordinates are transferred to the original image, thereby providing objects with photosynthetic activity. The features are extracted from these objects using histograms of oriented gradients, which in turn serve as input to a pre-trained support vector machine. The SVM determines whether the detected objects belong to the lettuce class. The objects belonging to that class are stored in a new mask called crop mask (see Figure 3).
Then, this mask is multiplied by the mask obtained through the NDVI index, eliminating all the lettuce objects from the image and leaving only the objects with photosynthetic activity that generally belong to the weed class. Finally, the percentage of white pixels with respect to the total area of the image is calculated in this new weed mask, in order to determine weed coverage as shown in Figure 4b.

2.4. Method 2: CNN-YOLOv3

YOLO (you look only once), as its name implies, is a neural network capable of detecting the bounding boxes of objects in an image and the probability that they belong to a class in a single step. YOLO uses convolutional networks and it was selected for its good performance in object and pattern recognition, which has given it a recent good reputation in fields such as the recognition of means of transportation and animals, and the tracking of moving objects. The first version of YOLO came out in 2016 [43]; its architecture consisted of 24 convolutional layers working as feature extractors and two dense or fully connected layers that performed predictions. YOLOv3 was used for its significant enhancements and feature extraction layers which were replaced by the Darknet-53 architecture [44].
As seen in Figure 5, once the model is trained to identify the crop (Figure 4c), an algorithm uses bounding box coordinates from the model to remove crop samples from the image. Later, a green filter binarizes the image, so pixels without vegetation become black, while the pixels accepted by the green filter become white. Finally, vegetation that does not correspond to the crop is highlighted, thereby simplifying the percentage calculation of weeds per image.
In order to get the most out of YOLO in terms of effective detection of objects and speed, it was decided not to use edge detection and to consider the entire bounding box generated by the model as crops, although this might affect weed calculation since the closest weed to the crop could be lost during estimation.

2.5. Method 3: Mask R-CNN

RCNN uses the “selective search for object recognition” algorithm [45] to extract 2000 regions from the image. They feed a convolutional neural network—in this case Inception V2—which extracts features that will be introduced to an SVM in order to classify the object in corresponding class. Additionally, when using masks in the training process of this method, an instance segmentation [46] is obtained, which provides detailed information about pixels that belong to the class. The training method used in this phase involves labeling crops as a “bounding box” and labeling masks related to each class. This mask comes from the closed region corresponding to crop that is inside corresponding “bounding box”. The architecture used [46] allows one to associate pre-trained masks with detected instances. These masks allow for obtaining the edges of an object without having to calculate them, as in case of the SVM-HOG method.
Unlike SVM-HOG, weeds are estimated by initially processing the complete image with a convolutional network. This network delivers corresponding crop masks in a binary image (Figure 4d). Furthermore, the NDVI index allows one to eliminate background image, leaving only objects corresponding to weeds and lettuce (Figure 6). Finally, this image is mixed with masks generated by neural network to obtain the image of weeds; the percentage of weeds in the image is calculated with respect to the total area.

2.6. Experimental Setup

For this study, three models of each method were trained with 913 samples labeled manually by experts. In the training process, a machine with an 8-core Xeon processor, 16 GB of RAM, and a GTX 1060 6 GB graphic card was used. Training metrics included accuracy, specificity, and precision, among others.
In the performance evaluation and comparison, the best model of each method was selected by taking into account their confusion matrices, coverage values, and evaluation times. Later, they were implemented in separate environments with the same characteristics: 4-core Xeon processor and 8 GB of RAM, without the graphics card. Regarding programming languages, method 1 used C++, method 2 used the Darknet framework for training and Python for evaluation, and method 3 used Python with TensorFlow library.

2.7. Performance of Methods

Several models were trained; we reduced number of training images to leave the maximum possible number of images for validation and comparison with experts (58 images with 1306 lettuces), with a limit of 85% of F1-score for the three methods. In total for training of the final models, 42 images were used, which contained 913 lettuces, approximately 41% of total samples.
Table 1 shows metrics obtained by each model for crop detection calculated using confusion matrices of each method; the three models performed well. However, in this case, accuracy does not provide enough information, since this model was evaluated using a single class and without negative samples, as evidenced by the results of specificity. In terms of sensitivity and precision, models that used YOLO and R-CNN stand out. Their high scores indicate that crop estimation was very accurate for these models without detracting from performance of HOG-SVM model. This is well summarized in the F1-scores of the three models.
To compare the time spent by each technique, several of the non-parametric methods used were reviewed (Nemenyi, Wilcox, Kruskal, Dunn). Due to the lack of normality, Wilcox was chosen, since data corresponded to dependent samples; in fact, they were the same population. There were statistical differences in the time used by different methods. In the boxplot,Figure 7, it can be seen that HOG-SVM and YOLO are more consistent in terms of time, while time is more variable in RCNN.
According to Ambrosio [7], the samples with square grids suggest that the variable of weed density follows a negative binomial probability distribution, a distribution similar to those presented in Figure 8a, which indicates that both estimates (models and experts) behaved as expected. However, curves can be classified into three groups: First, models that used HOG-SVM and RCNN, which had similar and homogeneous distributions with respect to others. The second group includes the third expert and model trained with YOLO; their behavior suggests their estimates have a similar trend. Finally, the third group includes the first and second experts, whose distributions were similar, but very different in comparison to the others. This suggests that their data tended to overestimate the high coverage of weeds compared to the others, which proves that subjectivity is one of the main problems in human estimation of weeds. Opposite results where found in Andujar et al., as they found that visual estimations bias is to underestimate high-coverage weeds [47].

3. Results and Discussion

Regarding the coverage value, a boxplot was developed (Figure 8b), in order to explore possible differences. Then, the Dunn’s test (Table 2) was performed to verify the existence of significant differences. Intraclass correlation coefficients (ICC Table 3) were obtained to verify agreement, and finally [47], exploratory analysis was done to compare the correlations and agreement of methods using a correlation matrix and Bland–Altman plots.
Figure 8b shows that RCNN and HOG-SVM are almost identical; YOLO differs a little and is more similar to experts 1 and 2. Conversely, expert 3 is similar to RCNN and HOG-SVM but has more scattered data. Experts 1 and 2 have a higher dispersion since they overestimated the largest coverage values, as illustrated in the Bland–Altman plots.
Table 2 shows that the RCNN and HOG-SVM methods do not have significant differences with respect to YOLO method, nor with respect to expert 3. As for the YOLO method, it only differs from expert 1; expert 1 is different from all methods and experts. On the other hand, expert 2 is different from all except the YOLO method. Finally, expert 3 is statistically different from the other two experts, but not from any computational method.
Those variations were due to the differences in education and training, since the expert 1 was an agronomy professional, but with little training in the quantification of weeds. Expert 2 was an agronomy professional with training in the quantification of weeds. Expert 3 was a highly trained agronomy doctor. This was corroborated later with the ICC (intraclass correlation coefficient).
The correlation matrix (Figure 8c) shows that method 2 (YOLO)—among the computational techniques—and expert 1—among the experts—were the most inconsistent. However, comparing different evaluation methods is a complex task, since usually there is no standard or true value. Thus, one of the most used, but also the most criticized methods is the Pearson correlation. It evaluates the linear association between measurements (variables), but does not evaluate their agreement. Therefore, two methods can have a good correlation but not agreement, or vice versa [48].

3.1. Analysis and Results Using the Bland–Altman Method

The Bland–Altman graphical method allows one to compare two measurement techniques while taking into account the same quantitative variable, in order to determine agreement limits, biases, and variability. To perform the analysis based on the Blant-Alman method, the R-CNN method and SVM (Figure 9a) were compared. The y axis represents differences in the coverage values for each method, while the x axis represents their mean values.
There are three dotted lines in the Bland–Altman plots. The one in the middle quantifies the mean difference between both methods. The other two represent upper and lower limits of agreement which establish range of differences between one method and the other. The smaller this range is between these limits, the better agreement between methods; a difference more than 10–15 could be unacceptable, since the most common scales for weed management decision support systems are ordinal and its ranks are higher than these values [47]. In Figure 9a, the limits of agreement are between −4 and +5 (dashed lines). As for the blue line, it indicates the regression calculated for the differences, which shows a non-constant systematic bias. In the case of Figure 9a, it has a small positive trend but tended to remain parallel to the x axis. This shows that there is no variability bias, i.e., the data do not scatter more as values of coverage increase. In conclusion, either method can replace the other [49].
When comparing method 3 (RCNN) with method 2 (YOLO based), as coverage increases, YOLO overestimates weed coverage (Figure 9b). The limits of agreement are wider (−15 to +10, dashed lines), and as coverage increases, so does the differences’ variability.
By comparing the most consistent computational method (RCNN) against values obtained by the experts (Figure 9c), it is possible to corroborate the existence of human bias; i.e., high values of coverage are overestimated, in this case up to 32 coverage units. Additionally, limits of agreement increased from −32 to 9 (dashed lines). The average of the expert 1’s data is greater than the average of method 3 by 12 coverage units. Furthermore, it can be seen that experts rounded to integers or multiples of 5 or 10, while the computational methods gave continuous variables.
Expert 1 was not accurate, since for several images, with the same coverage value, the estimate increased or decreased the coverage value by ±10 (Figure 9c). Expert 2 was more accurate because the data for the same coverage value had a difference of ±5 compared to RCNN method. Limits of agreement were slightly lower, from −28 to +8 (dashed lines). However, they were also biased in overestimating the high coverage values (Figure 9d).
For expert 3, limits of agreement were lower, from −16 to +12 (dashed lines, Figure 9c). There was also an overestimation bias but to a lesser extent. However, the low coverage values were also underestimated. Finally, there was no variability bias and average value of the coverage was very close to the average of method 3.

3.2. A Comparison with Previously Published Results

Regarding other authors, there are a few works related to weed detection in lettuce crops; however, most of them are focused on detection with herbicide application purposes via robotics. For example, Raja et al. [40] proposed a weed detection system in lettuce crops using image processing with illumination control incorporated with an RGB camera. Additionally, this system uses a simple thresholding technique in order to detect weeds with an accuracy of 98.1%. This technique is often limited because it depends on light control and as consequence increases the cost of the system. On the other hand, different machine approaches exposed in this paper are able to be invariant to abrupt changes in illumination due to the transfer learning feature of deep learning systems. Thus, in this case, collecting data from environments with changes in their illumination and by re-training the model could be a solution to this problem.
Otherwise, Elstone et al. [41], has her own approach of using a robotic system for detecting lettuces trough a multispectral camera and illumination control system. This approach focuses on NIR reflectivity pixels values and a size classificator assuming that crops will be always larger than weeds. In this case, this assumption could get to many false positives in each detection. This approached achieved 88% accuracy.
Table 4 presents a comparison between the methods that were mentioned and the approaches presented here.
Another important remark about the comparison established in Table 4 is related with the use of deep learning models for weed detection instead of classical image processing techniques reflected in [40,41]. This is because deep learning models have great generalization capabilities and can be used with different kinds of lettuce crops, no matter the geographical location. In contrast, classical image processing methods are involved in local calibrations for the illumination control system. Additionally, reflectivity values depend drastically on sun conditions and geolocation. Moreover, those techniques definitely will not work with crops with many weeds in an advanced growth stage. Meanwhile, those methods that include convolutional neural networks such as YOLO and Mask R-CNN can get the main features of a lettuce by themselves, even features that are outside of human visual perception, no matter how many weeds are in the picture. This advantage is very useful when the weed is the same size as the lettuce—solving in this way the limitation presented in [41]. Finally, the fact that this approach does not involve the construction of a complex system that needs lighting control and that it works on a drone allows greater flexibility with quite acceptable results regarding its performance as a weed detector. This set of deep learning models can be easily adapted to aerial fumigation systems to perform herbicide application tasks in a faster way and with similar results to those found in previous literature.

4. Conclusions

Weed detection using multispectral imaging continues to be an important task, since weed control is of vital importance for agricultural productivity. Therefore, the development and improvement of current computational methods is essential. Although the process can be difficult due to the lack of consistency and distinctive morphological features to locate weed, there have been significant improvements in terms of accuracy and speed of detection for a specific crop, thereby enabling one to discriminate between weeds and non-weeds.
Comparing the estimation made by the experts and the trained models reveals one of the main problems that must be addressed using machine learning in precision agriculture: the high error related to subjectivity of the expert when estimating weeds. This makes it more difficult to obtain a correct consensus from the experts and extends the error to the calculations of working hours, and quantity and cost of materials, among others. Therefore, this study demonstrates that pre-trained models for weed estimation are a reliable source with less uncertainty that can be adopted by professionals dedicated to weed control.
The good performances of the methods were mainly due to different approaches to address the problem of weed estimation: vegetation identification, crop detection, and weed quantification. In principle, two weed and crop classes were selected, taking positive and negative samples of both classes. The problem was the identification of the weed because it has just a few fixed characteristics, which makes identification difficult. This generated a high error rate, so the vegetation was identified through multispectral bands. In this way, the model focuses on identifying only the crop and thus facilitates calculation of the remaining vegetation—in this case weeds.
The YOLO model is attractive, not only because of its demonstrated speed regarding other deep neural network architectures, but because although the crops were analyzed as rectangles while ignoring the closest weed to the crop, it did not significantly affect estimation regarding evaluation from experts. However, the R-CNN model stood out for its accuracy in locating the crop and showing the edges very precisely, a method that can be recommended to address other problems in the sector, such as fruit identification.
It was demonstrated that the HOG-SVM method performed very well, and considering that it needs less processing capacity, it is a very good option for IoT solutions.
The Bland–Altman method showed that methods based on HOG-SVM and RCNN are the most similar in terms of evaluation; therefore, they are considered the most consistent. In contrast, the YOLO method overestimates the high values of weed coverage when compared to the other two. Now, although the human bias of rounding values made the experts less precise, when comparing them with methods 1 and 3, it was clear that there was overestimation and underestimation in all three cases, but to different extents. This indicates that the variability in evaluation is linked to the level of experience and is perhaps the most serious problem in human visual estimation. This happens because over-estimating can lead to wrong management decisions. For example, an expert may find the action threshold at experimental level, but a field evaluator will overestimate coverage and this will lead to unnecessary management, increased production costs, and increased contamination.

Author Contributions

Conceptualization, C.P.; Data curation, A.P. and D.J.; Funding acquisition, L.R.; Investigation, K.O. and A.P.; Methodology, C.P.; Project administration, K.O. and C.P.; Resources, A.P. and D.J.; Software, K.O. and A.P.; Supervision, C.P., D.J. and L.R.; Validation, D.J. and L.R.; Visualization, K.O. and D.J.; Writing original draft, K.O.; Writing review and editing, A.P. All authors have read and agreed to the published version of the manuscript.

Funding

Project “Cuantificación de maleza en cultivos de hortalizas por medio de procesamiento de imágenes digitales multiespectrales”. Convocatoria de proyectos de investigación conjuntos entre la Universidad de Cundinamarca y la Universidad Nacional de Colombia sede Bogotá. Código: 39815.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NDVINormalized difference vegetation index
SSWMSite-Specific Weed Management
UASsUnmanned Aircraft Systems
UAVUnmanned Aerial Vehicle
HOGHistograms of Oriented Gradients
SVMSupport Vector Machines
ANNArtificial Neural Network
CNNConvolutional Neural Network
FCNFully Convolutional Network
R-CNNRegion based Convolutional Neural Network
YOLOV3You only look once (Deep learning method based in CCN)
ICCIntraclass Correlation Coefficient
Exp1Expert 1 (B.Sc in agronomy)
Exp2Expert 2 (M.sc in agronomy)
Exp3Expert 3 (Ph.D in weed science)

References

  1. Cheng, B.; Matson, E.T. A Feature-Based Machine Learning Agent for Automatic Rice and Weed Discrimination. In International Conference on Artificial Intelligence and Soft Computing, Proceedings of the ICAISC 2015: Artificial Intelligence and Soft Computing, Zakopane, Poland, 14–28 June 2015; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; pp. 517–527. [Google Scholar]
  2. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine learning in agriculture: A review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef] [Green Version]
  3. Wang, A.; Zhang, W.; Wei, X. A review on weed detection using ground-based machine vision and image processing techniques. Comput. Electron. Agric. 2019, 158, 226–240. [Google Scholar] [CrossRef]
  4. Rodríguez, M.; Plaza, G.; Gil, R.; Chaves, B.; Jiménez, J. Reconocimiento y fluctuación poblacional arvense en el cultivo de espinaca (Spinacea oleracea L.) para el municipio de Cota, Cundinamarca. Agron. Colomb. 2018, 26, 87–96. [Google Scholar]
  5. Hamuda, E.; Mc Ginley, B.; Glavin, M.; Jones, E. Improved image processing-based crop detection using Kalman filtering and the Hungarian algorithm. Comput. Electron. Agric. 2018, 148, 37–44. [Google Scholar] [CrossRef]
  6. Jamaica, D.; Plaza, G. Evaluation of various conventional methods for sampling weeds in potato and spinach crops. Agron. Colomb. 2014, 32, 36–43. [Google Scholar] [CrossRef]
  7. Ambrosio, L.; Iglesias, L.; Marin, C.; Del Monte, J. Evaluation of sampling methods and assessment of the sample size to estimate the weed seedbank in soil, taking into account spatial variability. Weed Res. 2004, 44, 224–236. [Google Scholar] [CrossRef]
  8. Lara, A.E.P.; Pedraza, C.; Jamaica-Tenjo, D.A. Weed Estimation on Lettuce Crops Using Histograms of Oriented Gradients and Multispectral Images. In Pattern Recognition Applications in Engineering; IGI Global: Hershey, PA, USA, 2020; pp. 204–228. [Google Scholar]
  9. López-Granados, F. Weed detection for site-specific weed management: Mapping and real-time approaches. Weed Res. 2011, 51, 1–11. [Google Scholar] [CrossRef] [Green Version]
  10. Jamaica Tenjo, D.A. Dinámica Espacial y Temporal de Poblaciones de Malezas en Cultivos de papa, Espinaca y caña de Azúcar y su Relación con Propiedades del Suelo en dos Localidades de Colombia. Master’s Thesis, Universidad Nacional de Colombia, Bogotá, Columbia, 2013. [Google Scholar]
  11. Peña, J.M.; Torres-Sánchez, J.; de Castro, A.I.; Kelly, M.; López-Granados, F. Weed Mapping in Early-Season Maize Fields Using Object-Based Analysis of Unmanned Aerial Vehicle (UAV) Images. PLoS ONE 2013, 8, e77151. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Kharuf-Gutierrez, S.; Hernández-Santana, L.; Orozco-Morales, R.; Aday, Díaz, O.D.L.C.; Delgado, Mora, I. Análisis de imágenes multiespectrales adquiridas con vehículos aéreos no tripulados. Ingeniería Electrónica Automática y Comunicaciones 2018, 39, 79–91. [Google Scholar]
  13. López-Granados, F.; Torres-Sánchez, J.; Serrano-Pérez, A.; de Castro, A.I.; Mesas-Carrascosa, F.J.; Peña, J.M. Early season weed mapping in sunflower using UAV technology: Variability of herbicide treatment maps against weed thresholds. Precis. Agric. 2016, 17, 183–199. [Google Scholar] [CrossRef]
  14. López-Granados, F.; Torres-Sánchez, J.; De Castro, A.-I.; Serrano-Pérez, A.; Mesas-Carrascosa, F.-J.; Peña, J.-M. Object-based early monitoring of a grass weed in a grass crop using high resolution UAV imagery. Agron. Sustain. Dev. 2016, 36, 67. [Google Scholar] [CrossRef]
  15. Ahmed, O.S.; Shemrock, A.; Chabot, D.; Dillon, C.; Williams, G.; Wasson, R.; Franklin, S.E. Hierarchical land cover and vegetation classification using multispectral data acquired from an unmanned aerial vehicle. Int. J. Remote Sens. 2017, 38, 2037–2052. [Google Scholar] [CrossRef]
  16. Tao, T.; Wu, S.; Li, L.; Li, J.; Bao, S.; Wei, X. Design and experiments of weeding teleoperated robot spectral sensor for winter rape and weed identification. Adv. Mech. Eng. 2018, 10. [Google Scholar] [CrossRef] [Green Version]
  17. Binch, A.; Fox, C.W. Controlled comparison of machine vision algorithms for Rumex and Urtica detection in grassland. Comput. Electron. Agric. 2017, 140, 123–138. [Google Scholar] [CrossRef] [Green Version]
  18. Rumpf, T.; Römer, C.; Weis, M.; Sökefeld, M.; Gerhards, R.; Plümer, L. Sequential support vector machine classification for small-grain weed species discrimination with special regard to Cirsium arvense and Galium aparine. Comput. Electron. Agric. 2012, 80, 89–96. [Google Scholar] [CrossRef]
  19. Dyrmann, M.; Mortensen, A.K.; Midtiby, H.S.; Jørgensen, R.N. Pixel-wise classification of weeds and crops in images by using a fully convolutional neural network. In Proceedings of the International Conference on Agricultural Engineering, Aarhus, Denmark, 26–29 June 2016. [Google Scholar]
  20. Dyrmann, M.; Karstoft, H.; Midtiby, H.S. Plant species classification using deep convolutional neural network. Biosyst. Eng. 2016, 151, 72–80. [Google Scholar] [CrossRef]
  21. McCool, C.; Pérez, T.; Upcroft, B. Mixtures of lightweight deep convolutional neural networks: Applied to agricultural robotics. IEEE Rob. Autom. Lett. 2017, 2, 1344–1351. [Google Scholar] [CrossRef]
  22. Milioto, A.; Lottes, P.; Stachniss, C. Real-time blob-wise sugar beets vs weeds classification for monitoring fields using convolutional neural networks. In Proceedings of the International Conference on Unmanned Aerial Vehicles in Geomatics, Bonn, Germany, 4–7 September 2017. [Google Scholar]
  23. Potena, C.; Nardi, D.; Pretto, A. Fast and accurate crop and weed identification with summarized train sets for precision agriculture. In Proceedings of the International Conference on Intelligent Autonomous System, Shanghai, China, 3–7 July 2016; Springer: Cham, Switzerland, 2016; pp. 105–121. [Google Scholar]
  24. Partel, V.; Charan, Kakarla, S.; Ampatzidis, Y. Development and evaluation of a low-cost and smart technology for precision weed management utilizing artificial intelligence. Comput. Electron. Agric. 2019, 157, 339–350. [Google Scholar] [CrossRef]
  25. Pantazi, X.-E.; Moshou, D.; Bravo, C. Active learning system for weed species recognition based on hyperspectral sensing. Biosyst. Eng. 2016, 146, 193–202. [Google Scholar] [CrossRef]
  26. Sun, J.; He, X.; Ge, X.; Wu, X.; Shen, J.; Song, Y. Detection of Key Organs in Tomato Based on Deep Migration Learning in a Complex Background. Agriculture 2018, 8, 196. [Google Scholar] [CrossRef] [Green Version]
  27. Xinshao, W.; Cheng, C. Weed seeds classification based on PCANet deep learning baseline. In Proceedings of the IEEE Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, China, 16–19 December 2015; pp. 408–415. [Google Scholar]
  28. Huang, H.; Deng, J.; Lan, Y.; Yang, A.; Deng, X.; Zhang, L. A fully convolutional network for weed mapping of unmanned aerial vehicle (UAV) imagery. PLoS ONE 2018, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Chavan, T.R.; Nandedkar, A.V. AgroAVNET for crops and weeds classification: A step forward in automatic farming. Comput. Electron. Agric. 2018, 154, 361–372. [Google Scholar] [CrossRef]
  30. Dyrmann, M.; Jørgensen, R.N.; Midtiby, H.S. RoboWeedSupport—Detection of weed locations in leaf occluded cereal crops using a fully convolutional neural network. In Proceedings of the 11th European Conference on Precision Agriculture (ECPA), Edinburgh, UK, 16–20 July 2017. [Google Scholar]
  31. Sa, I.; Popovic, M.; Khanna, R.; Chen, Z.; Lottes, P.; Liebisch, F.; Nieto, J.; Stachniss, C.; Walter, A.; Siegwart, R. WeedMap: A large-scale semantic weed mapping framework using aerial multispectral imaging and deep neural network for precision farming. Remote Sens. 2018, 10, 1423. [Google Scholar] [CrossRef] [Green Version]
  32. Sa, I.; Chen, Z.; Popovic, M.; Khanna, R.; Liebisch, F.; Nieto, J.; Siegwart, R. WeedNet: Dense Semantic Weed Classification Using Multispectral Images and MAV for Smart Farming. IEEE Robot. Autom. Lett. 2018, 3, 588–595. [Google Scholar] [CrossRef] [Green Version]
  33. Huang, H.; Lan, Y.; Deng, J.; Yang, A.; Deng, X.; Zhang, L.; Wen, S. A semantic labeling approach for accurate weed mapping of high resolution UAV imagery. Sensors 2018, 18, 2113. [Google Scholar] [CrossRef] [Green Version]
  34. Huang, H.; Deng, J.; Lan, Y.; Yang, A.; Deng, X.; Wen, S.; Zhang, H.; Zhang, Y. Accurate weed mapping and prescription map generation based on fully convolutional networks using UAV imagery. Sensors 2018, 18, 3299. [Google Scholar] [CrossRef] [Green Version]
  35. Suh, H.K.; IJsselmuiden, J.; Hofstee, J.W.; van Henten, E.J. Transfer learning for the classification of sugar beet and volunteer potato under field conditions. Biosyst. Eng. 2018, 174, 50–65. [Google Scholar] [CrossRef]
  36. Behmann, J.; Mahlein, A.K.; Rumpf, T.; Römer, C.; Plümer, L. A review of advanced machine learning methods for the detection of biotic stress in precision crop protection. Precis. Agric. 2015, 16, 239–260. [Google Scholar] [CrossRef]
  37. Pantazi, X.E.; Tamouridou, A.A.; Alex, ridis, T.K.; Lagopodi, A.L.; Kashefi, J.; Moshou, D. Evaluation of hierarchical self-organising maps for weed mapping using UAS multispectral imagery. Comput. Electron. Agric. 2017, 139, 224–230. [Google Scholar] [CrossRef]
  38. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef] [Green Version]
  39. Lameski, P.; Zdravevski, E.; Kulakov, A. Review of Automated Weed Control Approaches: An Environmental Impact Perspective. In Proceedings of the International Conference on Telecommunications, ICT 2018: ICT Innovations 2018, Engineering and Life Sciences, Vienna, Austria, 4–6 December 2018; Springer: Cham, Switzerland, 2018; pp. 132–147. [Google Scholar]
  40. Raja, R.; Nguyen, T.T.; Slaughter, D.C.; Fennimore, S.A. Real-time weed-crop classification and localisation technique for robotic weed control in lettuce. Biosyst. Eng. 2020, 192, 257–274. [Google Scholar] [CrossRef]
  41. Elstone, L.; How, K.Y.; Brodie, S.; Ghazali, M.Z.; Heath, W.P.; Grieve, B. High Speed Crop and Weed Identification in Lettuce Fields for Precision Weeding. Sensors 2020, 20, 455. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  43. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  44. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. CoRR 2018, abs/1804.02767. [Google Scholar]
  45. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective Search for Object Recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar]
  46. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  47. Andújar, D.; Ribeiro, A.; Carmona, R.; Fernández-Quintanilla, C.; Dorado, J. An assessment of the accuracy and consistency of human perception of weed cover. Weed Res. 2010, 50, 638–647. [Google Scholar] [CrossRef]
  48. Altman, D.G.; Bland, J.M. Measurement in medicine: The analysis of method comparison studies. Statistician 1983, 32, 307–317. [Google Scholar] [CrossRef]
  49. Cardemil, F. Comparison analysis and applications of the Bland-Altman method: Correlation or agreement? Medwave 2017, 17, e6852. [Google Scholar] [CrossRef]
Figure 1. (a) Mavic Pro with the Parrot Sequoia multispectral camera. (b) Image taken by the RGB sensor of the Parrot Sequoia camera during the flight.
Figure 1. (a) Mavic Pro with the Parrot Sequoia multispectral camera. (b) Image taken by the RGB sensor of the Parrot Sequoia camera during the flight.
Agriengineering 02 00032 g001
Figure 2. Dataset: (a) Fake Green. (b) Red band 660 nm. (c) Green band 550 nm. (d) Near infrared band 790 nm.
Figure 2. Dataset: (a) Fake Green. (b) Red band 660 nm. (c) Green band 550 nm. (d) Near infrared band 790 nm.
Agriengineering 02 00032 g002
Figure 3. A diagram of method 1 based on HOG (histograms of oriented gradients) and SVM (support vector machines): it shows the sequence of preparation, processing, and weed coverage analysis.
Figure 3. A diagram of method 1 based on HOG (histograms of oriented gradients) and SVM (support vector machines): it shows the sequence of preparation, processing, and weed coverage analysis.
Agriengineering 02 00032 g003
Figure 4. Resulting images: (a) Preprocessed image. (b) Method 1. (c) Method 2. (d) Method 3.
Figure 4. Resulting images: (a) Preprocessed image. (b) Method 1. (c) Method 2. (d) Method 3.
Agriengineering 02 00032 g004
Figure 5. A diagram of method 2 based on convolutional neural network YOLOv3: it shows the sequence for preparation, crop detection, weed identification using an NDVI (normalized difference vegetation index) mask, and weed coverage analysis.
Figure 5. A diagram of method 2 based on convolutional neural network YOLOv3: it shows the sequence for preparation, crop detection, weed identification using an NDVI (normalized difference vegetation index) mask, and weed coverage analysis.
Agriengineering 02 00032 g005
Figure 6. A diagram of method 3 based on RCNN (region convolutional neural network), using masks for contour detection: it shows the sequence for preparation, crop-weed detection, and segmentation using an NDVI mask for weed coverage analysis.
Figure 6. A diagram of method 3 based on RCNN (region convolutional neural network), using masks for contour detection: it shows the sequence for preparation, crop-weed detection, and segmentation using an NDVI mask for weed coverage analysis.
Agriengineering 02 00032 g006
Figure 7. Evaluation time per method. The overall processing time is more variable in the R-CNN method because process to obtain binary masks that fit each detection is often computationally expensive, so this causes a longer detection time of number of lettuces per image. SVM and Yolo obtained less variable processing time because their architectures are less complex.
Figure 7. Evaluation time per method. The overall processing time is more variable in the R-CNN method because process to obtain binary masks that fit each detection is often computationally expensive, so this causes a longer detection time of number of lettuces per image. SVM and Yolo obtained less variable processing time because their architectures are less complex.
Agriengineering 02 00032 g007
Figure 8. (a) Weed coverage probability distribution according to each established method (machine and expert methods). (b) Box plot for each method. Note that experts have greater variability in their methods than machine methods. Meanwhile, machine methods have lesser variability among themselves. (c) Correlation matrix with expert and machine learning methods. Values between machine learning methods are higher than the rest of the matrix. However, expert 3 achieved the closest values to each artificial approach. Notice that YOLO architecture obtained smaller correlation values with R-CNN and SVM. This was because YOLO did not get contours in each detection like SVM and R-CNN methods do.
Figure 8. (a) Weed coverage probability distribution according to each established method (machine and expert methods). (b) Box plot for each method. Note that experts have greater variability in their methods than machine methods. Meanwhile, machine methods have lesser variability among themselves. (c) Correlation matrix with expert and machine learning methods. Values between machine learning methods are higher than the rest of the matrix. However, expert 3 achieved the closest values to each artificial approach. Notice that YOLO architecture obtained smaller correlation values with R-CNN and SVM. This was because YOLO did not get contours in each detection like SVM and R-CNN methods do.
Agriengineering 02 00032 g008aAgriengineering 02 00032 g008b
Figure 9. Bland–Altman Plots for machine–expert (human) and machine–machine methods in order to analyze the agreement between two weed estimations. (a) Method 1 (R-CNN) vs. method 3 (SVM); (b) method 1 (R-CNN) vs. method 2 (YOLO); (c) method 1 (R-CNN) vs. expert 1; (d) method 1 (R-CNN) vs. expert 2; (e) method 1 (R-CNN) vs. expert 3.
Figure 9. Bland–Altman Plots for machine–expert (human) and machine–machine methods in order to analyze the agreement between two weed estimations. (a) Method 1 (R-CNN) vs. method 3 (SVM); (b) method 1 (R-CNN) vs. method 2 (YOLO); (c) method 1 (R-CNN) vs. expert 1; (d) method 1 (R-CNN) vs. expert 2; (e) method 1 (R-CNN) vs. expert 3.
Agriengineering 02 00032 g009
Table 1. The performance evaluation of ML models in testing phase.
Table 1. The performance evaluation of ML models in testing phase.
MetricsModel
HOG-SVMYOLOR-CNN
Accuracy79%89%89%
Sensitivity83%98%91%
Specificity0%0%0%
Precision95%91%98%
F1-Score88%94%94%
Table 2. Dunn test matrix with the different p-values between methods and experts.
Table 2. Dunn test matrix with the different p-values between methods and experts.
RCNNExp1Exp2Exp3HOG-SVM
Exp18.7 × 10 7 ----
Exp20.004300.45020---
Exp30.993701.7 × 10 5 0.02827--
HOG-SVM0.994053.5 × 10 8 0.000480.87911-
YOLO0.259000.013790.699510.599220.07420
Table 3. Comparison using ICC (intraclass correlation coefficient) methods and experts.
Table 3. Comparison using ICC (intraclass correlation coefficient) methods and experts.
MethodsICCp-Value
RCNN vs. YOLO0.5727.76 × 10 7
RCNN vs. HOG-SVM0.9199.72 × 10 26
RCNN vs. Exp10.1120.197
RCNN vs. Exp20.3730.00158
RCNN vs. Exp30.6942.99 × 10 10
Table 4. Comparison between lettuce detection methods.
Table 4. Comparison between lettuce detection methods.
ApproachAccuracyF1-ScoreLight SensitiveUAVRoverCameraSize Sensitive
Thresholding [40]98.1%N/AYESNOYESRGBNO
Size and reflectivity [41]88%N/AYESNOYESMULTIYES
SVM79%88%NOYESNOMULTINO
YOLO89%94%NOYESNOMULTINO
R-CNN89%94%NOYESNOMULTINO

Share and Cite

MDPI and ACS Style

Osorio, K.; Puerto, A.; Pedraza, C.; Jamaica, D.; Rodríguez, L. A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images. AgriEngineering 2020, 2, 471-488. https://0-doi-org.brum.beds.ac.uk/10.3390/agriengineering2030032

AMA Style

Osorio K, Puerto A, Pedraza C, Jamaica D, Rodríguez L. A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images. AgriEngineering. 2020; 2(3):471-488. https://0-doi-org.brum.beds.ac.uk/10.3390/agriengineering2030032

Chicago/Turabian Style

Osorio, Kavir, Andrés Puerto, Cesar Pedraza, David Jamaica, and Leonardo Rodríguez. 2020. "A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images" AgriEngineering 2, no. 3: 471-488. https://0-doi-org.brum.beds.ac.uk/10.3390/agriengineering2030032

Article Metrics

Back to TopTop