Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 12 May 2022
Sec. Technical Advances in Plant Science

Detection and Localization of Tip-Burn on Large Lettuce Canopies

\nBenjamin FranchettiBenjamin Franchetti1Fiora Pirri,
Fiora Pirri2,3*
  • 1Agricola Moderna, Milan, Italy
  • 2Alcor Lab, DIAG, Sapienza University of Rome, Rome, Italy
  • 3Deep Plants, Rome, Italy

Recent years have seen an increased effort in the detection of plant stresses and diseases using non-invasive sensors and deep learning methods. Nonetheless, no studies have been made on dense plant canopies, due to the difficulty in automatically zooming into each plant, especially in outdoor conditions. Zooming in and zooming out is necessary to focus on the plant stress and to precisely localize the stress within the canopy, for further analysis and intervention. This work concentrates on tip-burn, which is a plant stress affecting lettuce grown in controlled environmental conditions, such as in plant factories. We present a new method for tip-burn stress detection and localization, combining both classification and self-supervised segmentation to detect, localize, and closely segment the stressed regions. Starting with images of a dense canopy collecting about 1,000 plants, the proposed method is able to zoom into the tip-burn region of a single plant, covering less than 1/10th of the plant itself. The method is crucial for solving the manual phenotyping that is required in plant factories. The precise localization of the stress within the plant, of the plant within the tray, and of the tray within the table canopy allows to automatically deliver statistics and causal annotations. We have tested our method on different data sets, which do not provide any ground truth segmentation mask, neither for the leaves nor for the stresses; therefore, the results on the self-supervised segmentation is even more impressive. Results show that the accuracy for both classification and self supervised segmentation is new and efficacious. Finally, the data set used for training test and validation is currently available on demand.

1. Introduction

Plant stress detection is a long-standing research field and, among the stresses, tip-burn affecting, particularly, lettuce has been intensively studied, refer for example Termohlen and Hoeven (1965), Lutman (1919), Cox et al. (1976), and Gozzovelli et al. (2021).

Nowadays, the combination of new methods arising from computer vision and deep learning, the availability of new low-cost sensors together with increased attention on the transparency, quality, and healthiness of the farm to fork process is making plant stress analysis a challenging research topic.

Classification of plant diseases is becoming a relevant topic thanks to a number of new data sets, such as PlantLeaves (Chouhan et al., 2019), PlantsDoc (Singh et al., 2020), PlantsVillage (Hughes and Salathe, 2016), Plantae-K (Vippon Preet Kour, 2019), Cassava (Mwebaze et al., 2019), Citrus leaves (Rauf et al., 2019), etc. made available as tensorflow datasets at tensorflow.org. Examples from these datasets are shown in Figure 1. These new datasets and their ease of accessibility have thrived the research improving deep learning models for stress detection applications.

FIGURE 1
www.frontiersin.org

Figure 1. Images from the plant disease classification datasets: Cassava (Mwebaze et al., 2019), Citrus leaves (Rauf et al., 2019), PlantLeaves (Chouhan et al., 2019), and PlantVillage (Hughes and Salathe, 2016). The images clearly illustrate the difference with the proposed task of stress detection on large canopies. (A) Cassava, (B) Citrus leaves, (C) PlantLeaves, and (D) PlantVillage.

A limit of the currently available datasets is their inadequateness for stress analysis in Controlled Environment Agriculture (CEA) and specifically in plant factories, where plants are grown indoors under artificial lights, densely packed together, and stacked on multiple layers. In such a highly densely growing conditions, the plants are compacted on tables of trays, and stress problems need to be studied from this specific perspective, as shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Large canopies of plants grown in Plant Factories (A). In (B) we see the operators controlling the canopy to visually detect tip-burn, on the rolling tables.

The detection and localization of stress in plant factories have to deal with complex surfaces agglomerating several plants, where the single leaf shape is not specifically relevant, and at the same time, stresses, such as tip-burn, occur on the leaf tip. Moreover, typically plants affected by tip-burn are few, sparse, and hidden in the canopy of other healthy leaves. The underlying cause of tip-burn is a lack of calcium intake by the plants. This, however, is a result of multiple factors, such as lack of airflow, high humidity, excessive lighting, inadequate watering, and nutrient supply. A key advantage of growing plants indoors is the possibility to control all aspects of the plant growth including the light recipe and climate, thereby providing the optimal mix of conditions to optimize plant development and quality. However, high-density crop production, limited dimensions, lack of natural ventilation, and the need for artificial lighting for photosynthesis makes plants grown in plant factories, especially, vulnerable to tip-burn. Consequently, tip-burn has become a metric for the healthiness of the plants, and being able to monitor its advent is extremely relevant in indoor growing conditions. By automatically detecting tip-burn, the vertical farm control software can adjust the growing recipes in real time to provide the plants with the optimal growing conditions.

In this work, we propose a novel model for tip-burn detection in lettuce that fills the gap between already explored techniques of deep learning applied to plant stress detection and their practical implementation in plant factories. Our work includes the realization of an adequate dataset made of real and generated images. Yet, to emphasize the generality of our contribution we have also tested our model on PlantLeaves (Chouhan et al., 2019), PlantsVillage (Hughes and Salathe, 2016), and Citrus leaves (Rauf et al., 2019) and compared with other works, whose results have a state of the art.

2. Related Works on Disease Detection

Plants disease detection is nowadays a quite impressive research field collecting methods and studies on a good diversity of diseases, crops, plant species, conditions, and contexts. In particular, most of the recent studies are based on deep learning methods, yet consider different cameras and datasets.

Disease detection. A number of approaches are based on dedicated sensors, such as hyperspectral cameras, or generate their own datasets. For example, Nagasubramanian et al. (2018) studied charcoal rot disease identification in soybean leaves by implementing a 3D Deep-CNN on data collected by a hyperspectral camera. Zhang et al. (2019) carried out a similar study using high-resolution hyperspectral images to detect the presence of yellow rust in winter wheat. Refer to Terentev et al. (2022) for a recent overview of hyperspectral approaches.

On the other hand, the publicly available datasets designed for disease classification, such as those introduced above, have played a crucial role in most of the deep learning methods.

Approaches exploiting the publicly available datasets have obtained very high accuracy for classification. For example, Agarwal et al. (2020) trained a CNN on tomato leaves images taken from the PlantVillage dataset obtaining 91.20% accuracy on 10 classes of diseases. On the other hand, on the same set of tomato classes, Abbas et al. (2021) obtained 97.11% accuracy with DenseNet121 + Synthetic images.

Patidar et al. (2020) obtained 95.38% accuracy in diseases classification on the Rice Leaf Disease Dataset (Prajapati et al., 2017) from the UCI Machine Learning Repository. Mishra et al. (2020) achieved 88.46% accuracy on corn plant disease detection, at the same time, obtaining real-time performance of a deep model capable of running on smart devices. Saleem et al. (2020) experimented a number of deep networks on the Plant Village dataset, proposing a comparative evaluation study between multiple CNNs and optimizers for the task of plant disease classification, in order to find the combination with the best performances, obtaining quite challenging results. Sharma et al. (2020) obtained 98.6% accuracy on PlantVillage by manually segmenting a subset of the images. Hassan and Maji (2022) obtain significant results on three datasets: 99.39% on PlantVillage, 99.66% on Rice, and 76.59% on imbalance cassava. Syed-Ab-Rahman et al. (2022) obtained 94.37% accuracy in detection and an average precision of 95.8% on the Citrus leaves dataset, distinguishing between three different citrus diseases, namely citrus black spot, citrus bacterial canker, and Huanglongbing.

Overall, results on the publicly available datasets are saturating toward super human performance, showing that new steps for diseases detection need to be taken.

Other digital images based deep learning approaches have experimented with their own datasets. Examples are DeChant et al. (2017) and Shrivastava et al. (2019). DeChant et al. (2017) consider the classification of the Northern Leaf Blight in maize plants, taking images of leaves in the field. While (Shrivastava et al., 2019) studied the strength of transfer learning for the identification of three different rice plant diseases. A recent review on computer vision and machine learning methods for disease detection is done in Barbedo (2019a), Abade et al. (2020), and Lu et al. (2021).

Large canopies and tip-burn studies. Tip-burn studies date back long ago (Lutman, 1919; Termohlen and Hoeven, 1965; Cox and McKee, 1976), essentially exploring causes induced by lack of nutrients absorption, such as in Son and Takakura (1989) and Watchareeruetai et al. (2018). As far as we know, only (Shimamura et al., 2019) conducted tip-burn identification in plant factories using GoogLeNet, for binary classification of single lettuce images. They check from manually collected images of a single plant whether it has tip-burn or not.

Similarly, in Gozzovelli et al. (2021), a dataset for tip-burn detection on large dense canopies of indoor grown plants is generated with specific attention to cover the data imbalance. To cope with the imbalance, a huge amount of data were generated with Wasserstein Generative Adversarial Network (GANs) and verified using the realism score of Kynkäänniemi et al. (2019). Classification was performed with two class-classifier architecture highly inspired from DarkNet-19, YOLOv2 backbone (Redmon and Farhadi, 2016), while the tip-burn region was identified preparing a ground-truth with a conditional random field, further generalized with a U-Net (Noh et al., 2015).

GANs were already used in Giuffrida et al. (2017) to generate Arabidopsis leaf using the number of leaves as the label. Similar to Gozzovelli et al. (2021) in Douarre et al. (2019), the authors explore segmentation at the canopy level, of apple scab. They augment the segmentation training set with conditional GANS.

Plants stress and disease segmentation. Segmentation for enhancing plant stress and disease detection has been explored by the works of Tian and LI (2004) and Zhang and Wang (2007). Most of the methods, even recently, tend to use image processing methods, such as filtering, thresholding, Gaussian mixtures, and color transforms to segment the disease or part of the leaf. Barbedo (2017) noted that when the disease symptoms show a difference in color with respect to surrounding areas, then ROI segmentation can be easily exploited. This observation has led to the study of the improvements in disease classification led by segmentation. This indeed was the choice in Gozzovelli et al. (2021) and Sharma et al. (2020), despite in the latter, segmentation is done manually. A leaf segmented version of Plant Village is used by Abdu et al. (2018) to introduce an automatic extended region of interest (EROI) algorithm for simplified detection. The segmentation of the disease is obtained by thresholding while the leaf segmentation is not treated and segmented leaf images are provided as a dataset. Following the work of Abdu et al. (2018) in Abdu et al. (2019), an extended EROI version is provided to study individual diseased segments, still based on a segmented version of PlantVillage, provided as a dataset.

In Douarre et al. (2019), the authors segment a canopy apple leaves extending the manual training set with cGAN (Mirza and Osindero, 2014) generated images. Sodjinou et al. (2021) propose a segmentation method to separate plants and weeds, based on initial semi-manual preprocessing, using cropping and thresholding, further U-Net semantic segmentation refines the segmentation and, finally, the results are post processed with a subtractive clustering algorithm.

As a matter of fact, despite the observation of Barbedo (2017), better and more generalized results can be obtained using deep learning methods that do not rely on specific image processing practices to come out with a segmentation result, as shown in other application fields.

Weakly Self Supervised segmentation. As far as we know, no method has so far explored self-supervised segmentation of plants disease, based on the class annotation only. Our work is the first one providing both the leaf segmentation (for PlantLeaves and PlantVillage, and Citrus Leaves) and the tip-burn stress segmentation without any manual annotation of pixel labels for segmentation.

We recall that weakly self-supervised segmentation (WSSS) is self-supervised segmentation using only image-level annotation. This means that only the information of the category in the image (e.g., “diseased” or “healthy”) is used to segment the object(s) of interest. Namely, the method consists of predicting a pseudo-label mask of the objects belonging to the class of interest, only relying on the image class label. Recent research has dedicated significant attention to the problem, introducing new methods based on weakly supervised learning, such as self training (Zou et al., 2018; Gu et al., 2020; Wang et al., 2020), domain adaptation (Pan et al., 2020; Yang and Soatto, 2020), noisy label learning (Xie and Huang, 2021) and class activation maps (CAM). CAM, introduced by Zeiler and Fergus (2014) and Zhou et al. (2016) localize the object of interest only relying on the image classes and backpropagating the probability to layers before the logits. The CAM-based methods have motivated a huge amount of works, such as Sun et al. (2020), Chan et al. (2021), Araslanov and Roth (2020), Yao and Gong (2020), and Wang et al. (2020). The method we propose in this work is WSSS using only the image class label, to segment the plants' lesions. The only available knowledge is whether the image represents a stressed or not-stressed region. Our method works on domains where the task is to generate pseudo label masks of quite small high-deformable shapes. Despite our elective application domain being large canopies of plants grown in plant factories, it can be used for other applications, as we show applying our method to publicly available datasets.

3. Materials and Methods

3.1. Data Collection

Since tip-burn manifests on the leaves tip, it is mandatory to acquire images with a top view of the whole table. We do so by taking images with an HR digital camera fixed above the rolling table shown in Figure 2. A table is a base on which plants are grown. Each table assembles into multiple trays, which in turn are further divided into multiple cells where plant seeds are placed.

We collected images of size 4.64E + 3×6.96E + 3×3 of the tables, using a camera Canon32.5 APS-C of 32.5 megapixels located above the rolling tables (shown in Figure 2). The whole set is made of 43 images, 30 for training and 13 for validation and testing. Images were collected in a period of tip-burn spread. As a tip burn is about 5×5 pixels in the camera image, we have devised a splitting process that allows to zoom into the table image. We split the 43 images of size 4, 640×6, 960×3 into smaller images of size 64×64×3, with an interface we have prepared for the task, and collected 2,127 images of tip-burn. We have automatically selected the same number of images of healthy plants, ensuring to be healthy by correlated with the stressed images from the same table. The images collected by splitting the original table images have been then augmented to finally obtain a training set of 16,323 images of tip-burn stressed and healthy plants, a validation set of 5,596 images and a test set of 1,399 images.

For the purpose of illustrating our method on other datasets, we used the PlantVillage, PlantLeaves, and CitrusLeaves datasets available on Tensorflow.org.

3.2. Method

Preliminaries. The main practicality of weakly-supervised semantic segmentation methods (WSSS) is to avoid the resource-demanding manual labeling of each pixel of the categories of interest in an image, which is an impossible task for large canopies. Indeed, WSSS transforms the semantic segmentation task into the much less demanding effort of image-level class annotations. The problem is ill-conditioned and difficult, and a large literature is dedicated to the solution of it starting from Zeiler and Fergus (2014) and Zhou et al. (2016), up to most recent contributions (Chang et al., 2020; Sun et al., 2020, 2022; Wang et al., 2020; Wu et al., 2021; Zhang et al., 2021). Semantic segmentation is critical for detecting tip-burn on large canopies due to the difficulty of both identifying it on a dense set of plants and to individually localizing each tip-burned plant within the canopy, as shown in Figure 3. To ensure both identification and localization, we develop a new method for weakly-supervised semantic segmentation for the tip-burn stress (and for the visible disease in plants disease datasets) by defining a network pipeline using attention-based splitting, classification, and graph convolution.

FIGURE 3
www.frontiersin.org

Figure 3. The problem: given the image of a large canopy find all regions with tip-burn. Because tip-burn regions are very small and maybe each other close, segmentation is better than simple localization with bounding boxes. We propose a novel method for weakly supervised semantic segmentation, with only image class-labels annotations (classification accuracy 97.3%).

A crucial aspect of our model is that we adopt the same classifier for both the image and the patches, suitably resized. For this idea to work, it is required that feature properties are shared between the image of the object as a whole and the image of sub-parts of the object. For example, any subset of the image of a canopy shares similar features with the image as a whole. See Figure 4, last image of the upper strip captioned as ‘input image'. Similarly, a leaf and part of a leaf have the same feature properties. This often occurs in natural images, though it is not true, for example, for a tree, which has different features for the trunk and the crown. We define this characteristic of an object feature property as the principle of decomposition. In this work, we show that this principle is valid for both stress detection and segmentation on large canopies and for disease detection and segmentation for leaves (from the cited datasets), which is the domain of interest in this work.

FIGURE 4
www.frontiersin.org

Figure 4. Main idea of the tip-burn semantic segmentation requiring only image-level class annotation: decompose table canopy images up to an image X of size 352×576×3. Split X into overlapping patches of size 64×64×3 and use classification trained on these patches to obtain an attention map. Use the attention map as supervision for training a convolutional graph transferring probabilities on similar patches. Finally, results are automatically merged together forming the segmentation map of the canopy.

We consider a classification model f(C|X, Y, θ), where C indicates the class a sample image X belongs to, Y = {1, …, c} is the vector of training labels, ℓ indicates the size of the images accepted by the network, and θ are the network parameters. The classification model maps each sample X to pX[C], which is the probability vector for the class C given X, as estimated by the softmax activation.

Let X be an N×M×d tensor specifying an image, and X be any connected sub-tensor of it of size n×m×d, with mM and nN, where connected means that chosen row and column elements n and m from X are consecutive. We say that X enjoys the principle of decomposition if, given a deep classifier fM×N(C|X, Y, θ) with probability p of correctly classifying X, with respect to classes C, we expect that it correctly classifies S(X) with approximately the same p. Here, S is a suitable scaling transformation, including appropriate filtering, transforming X* to X′ having the same size as X.

Pre-processing and classification model for tip-burn on large canopies. Tip-burn pre-processing exposes three components. The first component is splitting the canopy image into two images of size (4.64E+ 3×6.96E + 3×3) representing half-table, then into all the sub-trays, and further each tray into 16 input images of size 352×576×3. The second component is the augmentation of the training images of size 64×64×3, by random rotation between 0 up to 90 degrees, flipping up and down and left to right, color quantization to 8 colors, zooming in by scaling and cropping, zooming out by padding, and finally by Gaussian blurring with random variance σ ∈ (0.5, 2). For the classification model, we have used as backbone Resnet50V2, trained on ImageNet 1000 and fine tuned with Global Average Pooling, drop-out (to introduce stochasticity in the training) and dense layers.

Pre-processing and classification model for the single leaf image datasets. For classifying the single leaf images of datasets, like PlantVillage, CitrusLeaves, and PlantLeaves, we used the same backbone as for the tip-burn. On the other hand, for weakly supervised semantic segmentation, we have also used a multi layer perceptron (MLP) to separate the background from the foreground. Profiting from the simple arrangement of a single leaf on a background of these datasets, we have automatically sampled from each image a patch of size 8×8×3 from each corner, labeling them background, and 6 patches of the same size from the image center as foreground and gave these data to the MLP to learn to separate the background from the foreground.

Local attention by splitting with hard strides. The main interest of splitting an image into patches with hard strides is to obtain the attention map, in a way similar to how the human gaze glimpses a scene focusing on interesting regions. Here, by hard strides, we intend strides that allow for a significant overlapping of the patches or, more specifically, strides that have a dimension much lower than the patch size.

Most of the work for attention estimation is done by the overlapping induced by the strides like when the gaze goes back to an interesting region of the scene several times. Yet, this kind of attention is local, as it does not capture the whole context. To obtain the context, we shall refine this spitting-based attention with spectral graph convolution, described in the next paragraph.

The splitting process and patch classification. Breaking an image into patches is a well-known technique (see Nowak et al., 2006; Zhou et al., 2009; Dong et al., 2011), requiring only algebraic manipulations of tensors. Consider the image X of size M×N×d. The splitting operation, along the spatial dimension, extracts from X patches of dimension (px, py, d). Here, the splitting combined with strides allows for overlapping the neighbors' patches according to the stride values (sx, sy). In some sense, it is like taking the inner product of X with the lower and upper shift matrices A, A′, and their transposed A and A with suitable shifts, and then cropping the non-zero values. Or, similarly, convolving X with a shifting kernel and cropping the non zero elements. The number of obtained patches and their configuration depends on the sizes M, N of the image, the number of channels d, the patch sizes (px, py, d), and the spatial stride (sx, sy). k1 and k2 are obtained like in the convolution output, though here we do not consider padding:

k1=(M-px)/sx+1 and k2=(N-py)/sy+1    (1)

We denote CO as a configuration of k1×k2 patches, each of size px×py×d. Namely, it is the shaping of k1 patches on a row and k2 patches on a column, shown in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. The figure illustrates the splitting process and patch classification taking as an example an image X∈ℝM×N×3, with M=160 and N=195. The first plate above shows the configuration CO obtained by splitting X into patches of size px×py×d and a stride (sx, sy) with px=py=64, d=3 and sx=sy=8. In the example, CO has a configuration of k1×k2 patches, with k1=13 and k2=17. The plate on the upper-right shows the configuration of probabilities CoP obtained via the softmax by classifying each patch in CO. CoP has the same shape as CO, by fine-tuning Resnet50V2. The plates below, on the right of the image X, show the matrix AoverlapM×N, collecting the number of times patches overlap on a pixel when collapsing the configuration into the reconstructed image, according to the stride. The Reconstructed Attention Map (RAM) is obtained by collapsing CoP, and it has the same size as X.

Given X∈ℝM×N×d, the configuration CO, and a patch Xj in CO, with 0≤jk1·k2, the patch Xj is resized as S(Xj) (including required filtering modes), where ℓ indicates the size of input images accepted by the network f(C|X, Y, θ). The value of ℓ changes according to the considered dataset. For the plants, disease datasets, the classification entry corresponds to the size of the image in the dataset, may be reduced as for PlantLeaves, while a patch is proportional to the image size. This shows the extreme flexibility of the splitting process followed by classification which is adaptable to several kinds of backbones.

For each patch in the configuration CO, obtained by splitting the original image, the probability that it belongs to the class of interest (e.g., tip-burn) is estimated by the network f resizing the patch to the input size ℓ accepted by the classification network f.

The estimation amounts to the softmax applied to the logits of the classifier f(C|X, Y, θ), here we used Resnet50V2 as a backbone. After each patch, probability to belong to C is estimated, a configuration of probabilities (CoP) is obtained, as shown in Figure 5. CoP has the same configuration as CO, though each patch π is defined by repeating at each pixel, the probability p computed by the softmax on classifying the patch. When we indicate the probability pr,c of the patch located at indexes (r, c), we mean the probability p.

A mapping h from CoP to the reconstructed attention map (RAM) is defined by collapsing the patches in CoP into the corresponding pixels of the matrix RAM. Note that while the whole size of CoP is k1·px×k2·py, namely (M·⌈px/sx⌉)×(N·⌈py/sy⌉), with sxpx and sypy, RAM has the same spatial dimension as the original image, namely M×N. Given a patch πr,c in row r and column c in CoP, and a pixel at location (i, j) in πr,c, the tuple ((r, i), (c, j)) is mapped by h to the pixel (x, y) in RAM, as follows:

(x,y)RAM=(h(r,i),h(c,j))=((i+r·sx),(j+c·sy))for πr,cCoP and (i,j)πr,c.    (2)

Given Equation (2), we also obtain a matrix Aoverlap by counting all times a pixel from CoP hits the corresponding pixel of RAM. Indeed, this matrix specifies how many overlapping patches contribute to a pixel in RAM:

Aoverlap((h(r,i),h(c,j)))(t)=Aoverlap((h(r,i),h(c,j)))(t-1)+1.    (3)

The matrix Aoverlap is used to count the accuracy of the classification at the pixel level, and it allows to suitably average RAM. The averaged RAM is obtained as follows:

RAM=RAM/Aoverlap    (4)

Figure 5 shows an example of CO representation, of CoP, of the matrix Aoverlap, of RAM given a random image X with tip-burn highlighted. Where, here and from now on, we denote RAM RAM.

We can see that the accuracy of RAM is determined by the strides. For example, if the stride is sx = sy = 10, we have an accuracy at the level of a region of size 10×10, and if the stride is sx = sy = 1, the accuracy is at the pixel level allowing to effectively label each pixel. Clearly reducing the stride increases the number of patches of the same image. The average increase of the number of pixel is by a factor of 9.

Refining by graph convolution. The RAM results in pseudo segmentation masks for the tip-burn stressed leaves found in the RGB image, following the pipeline splitting-classification-reconstruction. Differently from CAM (Zeiler and Fergus, 2014; Zhou et al., 2016), RAM highlights in the same map all objects of interest quite accurately. Moreover, while in CAM the result is obtained by the gradient of the softmax outcome, with respect to the last feature map, which has very low resolution, thus requiring significant resizing inducing blurring, here we do not need any resizing, as we can obtain the original image by a single step merging, according to Equation 2. Despite classification accuracy for tip-burn is 98.3%, and for the other datasets is no less than 96%, there is still noise on the attention map because classification is done on S(X), namely on the resized patch, given the decomposition principle.

Comparing the size of a patch in CoP and the size of the probabilities highlighted in RAM in Figure 5, we can note they have different sizes. This is due to overlapping and projection, which augment the resolution of the probability from uniform in a patch of size 64×64 to uniform on a patch of size 8×8. Indeed, the RAM probability resolution is 8×8. Having in RAM a higher probability resolution, we re-propose the splitting into sub-patches with size (sx, sy, d), namely of size 8×8×d, d ∈ {1, 3}, for both the RGB image of size 352×576×3 (see the paragraph above, on Preprocessing and Classification, and Figure 6) and the RAM of size 352×576×1. A schema of this further splitting follows:

COnewsize=42×72×3,sub-patch size=8×8×3,num of sub-patches=3168CoPnewsize=42×72×1,sub-patch size=8×8×1,num of sub-patches=3168    (5)

The goal is to use CoPnew as supervision for training a convolutional graph network (GCN) improving the semantic segmentation accuracy obtained by classifying the patches. This further splitting step obtains a COnew and a CoPnew, as specified in Equation (5), from which we obtain the features and the labels for the GCN. The softmax of the GCN classifies the nodes of the GCN, inducing an effective semantic segmentation for tip-burn for each image in the dataset, and similarly for the other datasets.

FIGURE 6
www.frontiersin.org

Figure 6. Overview of the weakly supervised semantic segmentation process of tip-burn leaves stress supervised with image-level class annotation only. Above, the splitting process starts at images X ∈ ℝM×N×3, M=4, 640, N=6, 960. Note that starting from patches of size 352×576×3 the splitting process plays with overlapping patches of size 64×64×3 with a hard stride of 8 pixels. The classification is trained to detect tip-burn stresses, and via the softmax to predict a probability for each patch to belong to the class tip-burn. After classification, a reconstruction step obtains the Reconstructed Attention Map (RAM) for patches of size 352×576×3. These are again split into patches 8×8×3 and used to supervise the Graph Convolutional network (GCN). Namely, the GCN features nodes are the flattened 8×8×3 patches and the labels are one hot encoded vectors obtained from the classification predictions, see the plate with the GCN, The GCN estimates a refined semantic segmentation of pixels. A final reconstruction does the inverse splitting process reconstructing the table canopies from patches.

Following Kipf and Welling (2016), a number of approaches have experimented with graph convolution (GCN), especially on non grid structures. Though, recently, an increasing interest is devolved to apply graph convolution on images, for segmentation and attention purposes, such as Li and Gupta (2018) and Hu et al. (2020). Here, we apply an unsupervised node classification, conditioning the graph model both on the data and on the adjacency matrix via graph convolution. As a matter of fact, we are going to generate the adjacency matrix for the graph G = (V, E), fully unsupervised. We construct a graph for each RGB input image X of size 352×576×3, fixing the size of the nodes, so as to put the graphs in a batch.

We take the patches as node features, labeling them with a one hot encoding vector obtained by thresholding the score pc,r in CoPnew of patch πc,rCOnew. More precisely, we flatten each mini-patch of size 8×8×3 into a vector x ∈ ℝk, k = 192 and stack all the flattened patches into a matrix Xφn×k, n = 3168. To obtain a corresponding ordering, we use an index function idx:(r, c) → i, idx(r, c)=w(c−1)+r=i with w the number of rows, r and c the row and column indexes in COnew and in CoPnew, respectively, and i the corresponding index in Xφ. Xφ is the input matrix to the network. At each layer of the network, a feature matrix is generated, starting with Xφ.

To connect subsets of nodes, based on their feature similarity, we generate the adjacency matrix Adj, which is symmetric and of size n×n, as follows. We keep the indexing idx to maintain the correspondence between COnew and Xφ and between CoPnew and the labels. For each sub-patch, we estimate a non-parametric probability by computing the histogram using both the RGB and the HSV color transformation of the sub-patch and collapsing the 64·3·2 vector into a histogram with 64 bin-edges. For each pair of histograms qi, qj, we compute the Shannon-Jensen divergence:

JSD(qiqj)=0.5KL(qim)+0.5KL(qjm)withm=0.5(qi+qj) and KL(qim)=xXqi(x)log(qi(x)/m(x))    (6)

The choice of JSD is required by the need of the adjacency matrix Adj to be symmetric. Then, two nodes vi, vj feature vectors xi and xj are similar, hence connected by an edge ei,jE if JSD<β, we have chosen β=0.005 for the tip-burn dataset. Given the n nodes in V, the diagonal degree matrix D adds for each node the number of its connected ones. The normalized graph Laplacian matrix is Lnorm=In-D-12Adj D-12=UΛU. Here, Λ is the matrix of the eigenvalues and U is the orthogonal eigenvectors. The graph convolution gθ(Lnorm) ⋆ Xj using Lnorm is the spectral convolution, based on obtaining parameter filters from the eigenvectors of Lnorm in the Fourier domain. Several simplifications have been proposed, we refer the reader to Defferrard et al. (2016) for spectral convolution in the Fourier domain and the approximation of the L eigenvectors by Chebyshev polynomial up to the K-th order. Kipf and Welling (2016) obtain a GCN by a first-order approximation spectral graph convolution. They define K = 1 and reduce the Chebyshev coefficients to a matrix of filter parameters.

The feed forward propagation of the GCN is recursively defined as:

H(t+1)=σ(ÂH(t)W(t)), with H(0)=Xφ    (7)

Here σ is an activation function, H(t)={h1(t),,hn(t)} are the hidden vectors of the t-th layer, with hi(t) the hidden feature vector of the node vi.  is defined as follows. A = Adj + In, to include self loops, since nodes are self similar, D~ii=jAij and Â=D~-12AD~-12, so as to be normalized, In is the identity matrix of size n×n. The role of  is to aggregate information from connected nodes. W(t) is the weight matrix to be learned. The dimensions are as follows:

Ân×n,H(t)n×ut,W(t)ut×ut+1    (8)

Here, ut and ut+1 are the sizes of the hidden layers. A 3 layer GCN has the form:

Z=g(Xφ|Â,W)=softmax(G)    (9)

Where the softmax is applied row-wise and

G= ReLU(Â(ReLU(ÂH(0)W(0))W(1))W(2))    (10)

The optimization of the GCN uses cross-entropy loss on all labeled nodes (Kipf and Welling, 2016), where here the labels are the one hot encoded values obtained from RAMnew. Let us denote by I the indexes of the nodes and by Yi,l an indicator which has value 1 if node vi has label l:

L=-iIl=1nYi,llogZi,l    (11)

According to the number K of layers, a GCN convolves the K-hop neighbors of a node, essentially clustering similar nodes, according to their probability labels and features. We use simple 3 layers GCN, since in the end tip-burn stresses on leaves are very small and rare. An overview of the whole learning process is given in Figure 6.

The GCN adjusts the RAM by looking at the context, going beyond the localized estimation of splitting plus classification. GCN estimates the probability that a node, corresponding to features of a patch 8×8×3, belongs to tip-burn or not, by updating the belief that two patches are similar. At the end of the training, CoPnew is updated with the new distribution. In Table 2, in Section 5, we show the advantages of the GCN by ablation.

Reconstruction. Given the initial image of the dense canopy, the question to be explored is “which plant suffers tip-burn stress and where it is?” including counting would not be so useful. Consider that when tables are unrolled, from the position of the plant on the table, it is possible to go back to the cell the plant comes from, and possibly revise its growing conditions, or make useful statistics. It is therefore pivotal to localize the stress segment on the table image. It turns out that by the proposed model it is extremely easy.

In fact, as noted in the paragraph on splitting, reconstruction is automatically done by projecting back a pixel in CoP into a pixel in RAM by Equation (2). Obviously, it can be done for any image, not only for the maps but also for RGB images.

Reconstruction is done both when the stride sx>1 and sy>1, that is when the splitting generates sub-images that overlap and, obviously, when they do not overlap. This, in fact, can be done for all the steps of splitting, from the table canopy up to Ramnew and for its dual RGB image, and again back to the large table canopy.

The back process requires preserving just the patches size for the maps and the scores, at each layer of the splitting. Then, the process is simply recursively applied to go from the patch up to the image of the whole canopy. Note that for the semantic segmentation, we need only to preserve the score vectors estimated by GCN. An image of a partial reconstruction of the half table is given in Figures 4, 6.

4. Application of the Model to Other Datasets

As gathered in the introduction, we have collected three datasets, namely PlantLeaves (Chouhan et al., 2019), PlantsVillage (Hughes and Salathe, 2016), and Citrus leaves (Rauf et al., 2019), to evaluate our approach. Usually these datasets are tested for classification, which has nowadays obtained striking results. Here, instead, we consider the semantic segmentation of the leaves lesions using only images class-labels, which is actually the only information available for these datasets.

Our goal here is to discuss mildly classification and most of all the whole pipeline we used to segment both the leaves and the disease spots and lesions. Clearly, segmenting the disease is more difficult when the leaf is almost completely covered by the disease spots, which are discolored regions or dark necrotic spots. As we shall see, the best results are actually obtained for CitrusLeaves and PlantLeaves, where the disease spots are localized.

From each dataset, we have chosen a class of diseases and the corresponding healthy images, for segmentation. For PlantVillage, we used the whole dataset, but we have chosen only Pepper Bell bacterial and Pepper Bell healthy for segmentation. Consider that the only burden for classification once the model is defined is to load the data. In turn, the model is just a fine-tuning of an already existing model, such as Resnet50V2. All parameters and accuracies for each network model available for fine-tuning classification are provided on the Keras Application page.

PlantLeaves consists of 4,502 images of healthy and diseased leaves divided into 22 categories including species and disease. From this dataset, we have chosen Pomegranate (P9) both diseased and healthy. There are 272 images of diseased Pomegranate (P9) and 287 healthy ones.

PlantVillage consists of 54,303 images of healthy and diseased leaves divided into 38 categories including species and disease. It is possible to download either the augmented or the non augmented set of images. As gathered above we have considered the whole dataset for classification, PepperBell healthy and PepperBell Bacterial spot for segmentation. PepperBell Bacterial spot are 998 images, and PepperBell healthy are 1,478 images.

CitrusLeaves consists of 594 images of healthy and diseased leaves with 4 diseased categories and one healthy. We have chosen healthy and Canker. Canker contains 163 images, while healthy contains 58 images.

The model for the above indicated datasets, from splitting up to segmentation, is similar to the tip-burn stress segmentation, starting from splitting. Yet, the preprocessing is quite different. Preprocessing, for the three datasets consists of removing the background and tightening the image within its bounding box. This last step is crucial for complying with the principle of decomposition, discussed at the beginning of the method, and also for avoiding overfitting due to background. We have automatically sampled from each image, the four corners with 8×8×3 pixels and 6 patches of the same size from the image center. We have then defined a MLP that could accept sub-patches of the size 8×8×3 to separate background and foreground. Some results, compared with the original image are shown in Figure 7. Note that the background has a value (0, 0, 0), thus not influencing the CNN classification.

FIGURE 7
www.frontiersin.org

Figure 7. Preliminary leaves segmentation for PlantLeaves, PlantVillage, and CitrusLeaves, from left to right in the order.

Being the image of PlantLeaves of size 4E + 3×6E + 3, we reduce them to 264×400 after automatic cropping with the MLP. On the other hand, we resize both CitrusLeaves and PlantLeaves to their original size 256×256, after automatic cropping with MLP. We do augmentation by flipping up and down, left and right, and blurring with a Gaussian filter with random variance σ ∈ (0.5, 2), for both CitrusLeaves and PlantLeaves. We have not augmented PlantVillage, since it comes already augmented.

Also, differently from the model for tip-burn stress detection, for these datasets, we do the first splitting to a size of 70×70, with the same stride of sx = sy = 8, for tip-burn, and then we resize each patch to the original image size for classification. For classification, we have fine-tuned Resnet50V2, as for tip-burn data. The remaining of the model, from further splitting up to the GCN and the reconstruction, here just for the leaves, follows the same steps, which are the relevant novelties.

5. Experiments and Results

5.1. Setup

The whole model is implemented in Tensorflow 2.5, on a GeForce RTX 3080, 300 HZ. For the ResNet50V2, we use the Keras API in Tensorflow. We used the Keras functional API for fine tuning the model, with all the provided advances, such as early stopping, and learning rate decay. For early stopping we used patience 4, with delta 0.001. For reducing the initial learning rate when on a plateau, we used a factor of 0.2 and a minimum learning rate of 0.001 starting with an initial learning rate of 0.1. For the loss, we used categorical cross entropy with Adam as the optimizer (Kingma and Ba, 2014).

For the splitting and reconstruction, we use Tensorflow GradientTape, as the gradient computes both Aoverlap and the mapping between CoP and RAM for both Equations (2) and (3).

We have implemented a good part of the GCN including the adjacency matrix, the features vectors and the joining step, to transfer probabilities, in Tensorflow. We used much intuition from DGL, an open-source graph library introduced by Zheng et al. (2021), though DGL is implemented in PyTorch. We also get inspired by Spektral of Grattarola and Alippi (2021), an open-source Python library, for building graph neural networks with TensorFlow and Keras interface. As specified in the Method Section, we have been using cross entropy loss and Adam optimizer like in the classifier together with early stopping.

5.2. Comparison With State of the Art

The main contribution of our work is weakly-supervised semantic segmentation with the only supervision being the image class labels, whether there is tip-burn or not. For classification, we have been using Resnet50V2, because it is quite flexible, and fine-tuned it. As we have already mentioned, we expect that if f is a classifier that classifies correctly X with probability p, if the classifier generalizes well, it would classify the resized X, namely S(X), approximately with the same probability p. This is shown to be correct for tip-burn and plant disease datasets CitrusLeaves, PlantLeaves, and PlantVillage, according to the principle of decomposition.

Stress and disease detection. For training tip-burn CNN classification, we used 30 out of 43 images and compared our work with Gozzovelli et al. (2021), where DarkNet was used. For classification of the plant disease datasets, we considered the following recent works: Sujatha et al. (2021), Khattak et al. (2021), and Syed-Ab-Rahman et al. (2022) for CitrusLeaves; Mohameth et al. (2020), Mohanty et al. (2016), Chen et al. (2020), Agarwal et al. (2020), and Abbas et al. (2021) for PlantVillage; for PlantLeaves, we expose only our approach as there are no recent contributions.

Results are shown in Table 1, considering validation accuracy, as usual. The best accuracy in class is highlighted in bold. We note that for a number of species in PlantVillage, we obtained a validation accuracy of 1.0, in few epochs. Since we use early dropping, this was not caused by overfitting, as shown in Figure 8, where it can be observed that for all datasets we have a small number of epochs. Note that we have also removed the background because, according to Barbedo (2019b), accuracy drops without the background. It seems possible that the background induces overfitting. In Figure 8, we show some results motivated by changing the patience value in early stopping.

TABLE 1
www.frontiersin.org

Table 1. Tip-burn stress and plant disease detection.

FIGURE 8
www.frontiersin.org

Figure 8. The first graph on the left shows the validation accuracy and the loss for Tip-burn, PlantsVillage, CitrusLeaves, and PlantLeaves, with patience 4 for both early stopping and for updating the learning rule. We can observe that the maximum number of epochs is 13 for PlantsVillage. In the central graph, we see a paradox, validation hits 1% before training, for the grape class in PlantsVillage. On the right, we observe the convergence for tip-burn at 6 epochs.

We split the leaves datasets 70/30% between train and validation plus test as in Mohanty et al. (2016). As shown by Mohanty et al. (2016), GoogleLeNet is the best backbone for plant disease detection, and we recall that the assessed performance of Resnet50V2 on the Imagenet validation set is 0.760 on Top-1 accuracy and 0.930 on Top-5 accuracy.

5.3. Tip-Burn Segmentation and Plant Datasets Segmentation

Testing accuracy by manually labeling ground-truth. Typically, accuracy metrics for segmentation are F1, in the context of segmentation referred to as Dice similarity coefficient, and Intersection over Union (IoU), both required to compute the corresponding pixels (true positive), the exceeding pixels (false positives) or lacking pixels (false negative) between the ground truth mask and the estimated masks. Because, in none of the available datasets, we have ground truth masks available, we introduce the patch-based method that is not too demanding to obtain an approximate Dice coefficient. Here, approximate means that instead of computing the pixels we compute the super-pixels and also it means that we use a reduced number of test samples.

We consider 1 test image from the table canopy images and 20 test images for each of the plant disease datasets. Note that 1 test image is the half image of a table canopy, and it amounts to 346,464 images of size 64×64×3, that is 2, 406·16·9.

Now, assuming that we have segmented the test images, by the automatic decomposition and recomposition process, by definition of the model, we have made available all the patches that contribute to the final estimated segmentation. These patches are actually vectors Z holding the probability that the corresponding RGB vector is of class tip-burn or not, as estimated by the GCN and similarly, for the other datasets. At the same time, according to the described model, there is a one to one correspondence between the patches in CoPnew and COnew and there is a correspondence, by Equation (2), between the patches and the attention map RAM, hence the image, and the final segmentation map estimated. So, it is enough to choose the patches in COnew. The manually chosen patches are immediately aligned with CoPnew and the final segmentation. That is, suppose we have chosen a patch Xj which will be of size 8×8×3, by definition of the model, then we have automatically selected 64 pixels, and the process is significantly sped up. Once the patches are selected, we know the corresponding value in the segmentation map and can compute both the Dice similarity coefficient at sub-patch or super-pixel level, instead of pixel level, and the IoU. Let X = {Xj|Xj ∈ selected}, where each selected patch has value 1, and Y the corresponding patches in COPnew with value Z computed by the GCN, the DSCpatch= 2(XY)/(|X| + |Y|), with | · | the cardinality.

Table 2 gives the results for the approximate meanIoU and F1 (Dice similarity coefficient), computed according to sub-patches (super-pixels) in place of pixels, and according to a limited subset of the test images. We give some ablation too.

TABLE 2
www.frontiersin.org

Table 2. Segmentation and ablations.

Ablation. Consider first the segmentation using just thresholding of the attention map RAM. Introducing the CGN with two layers, we observe an improvement for all models. Extending the GCN to three layers, we observe that accuracy improves for all models but for PlantVillage. It is interesting to note also that doubling the number of nodes in the GCN lowers the accuracy for all models, which is reasonable because we have to choose patches with lower probability of being tip-burn.

In Figure 9, we provide some qualitative results facilitating an understanding of the extremely good results of the model.

FIGURE 9
www.frontiersin.org

Figure 9. Qualitative results of the weakly supervised semantic segmentation of tip-burn stress and of disease spot and lesions on PlantVillage, PlantLeaves, and CitrusLeaves.

6. Conclusion

In this paper, we have introduced a new method for detection and localization of tip-burn stress in large plant canopies grown in plant factories. The idea is very simple to implement, and the only supervised step is a classification of the image, namely just knowing the class in the image. We have shown that the method obtains quite nice refined weakly self-supervised segmentation for tip-burn stress.

We have tested our method both on publicly available datasets, such as PlantVillage, PlantLeaves, and CitrusLeaves, and in operating conditions inside a plant factory showing the flexibility of our model. The results show that plant stress detection and localization can be done automatically in Controlled Environment Agriculture conditions.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Funding

The research has been partly funded by Agricola Moderna.

Conflict of Interest

BF was employed by Agricola Moderna.

The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors would like to acknowledge Malik Bekmurat for having collected the data set, and the plant scientists at Agricola Moderna for their insight into plant physiology.

References

Abade, A. S., Ferreira, P. A., and Vidal, F., d. B. (2020). Plant diseases recognition on images using convolutional neural networks: a systematic review. arXiv preprint arXiv:2009.04365. doi: 10.1016/j.compag.2021.106125

CrossRef Full Text | Google Scholar

Abbas, A., Jain, S., Gour, M., and Vankudothu, S. (2021). Tomato plant disease detection using transfer learning with c-gan synthetic images. Comput. Electron. Agric. 187, 106279. doi: 10.1016/j.compag.2021.106279

CrossRef Full Text | Google Scholar

Abdu, A. M., Mokji, M., and Sheikh, U. (2018). An investigation into the effect of disease symptoms segmentation boundary limit on classifier performance in application of machine learning for plant disease detection. Int. J. Agric. Forestry Plantation 7, 33–40.

Google Scholar

Abdu, A. M., Mokji, M. M., and Sheikh, U. U. (2019). “An automatic plant disease symptom segmentation concept based on pathological analogy,” in 2019 IEEE 10th Control and System Graduate Research Colloquium (ICSGRC) (Shah Alam: IEEE), 94–99.

Google Scholar

Agarwal, M., Singh, A., Arjaria, S., Sinha, A., and Gupta, S. (2020). ToLeD: tomato leaf disease detection using convolution neural network. Procedia Comput. Sci. 167, 293–301. doi: 10.1016/j.procs.2020.03.225

CrossRef Full Text | Google Scholar

Araslanov, N., and Roth, S. (2020). “Single-stage semantic segmentation from image labels,” in CVPR, 4253–4262. doi: 10.1109/CVPR42600.2020.00431

CrossRef Full Text | Google Scholar

Barbedo, J. G. A. (2017). A new automatic method for disease symptom segmentation in digital photographs of plant leaves. Eur. J. Plant Pathol. 147, 349–364. doi: 10.1007/s10658-016-1007-6

CrossRef Full Text | Google Scholar

Barbedo, J. G. A. (2019a). Detection of nutrition deficiencies in plants using proximal images and machine learning: a review. Comput. Electron. Agric. 162, 482–492. doi: 10.1016/j.compag.2019.04.035

CrossRef Full Text | Google Scholar

Barbedo, J. G. A. (2019b). Plant disease identification from individual lesions and spots using deep learning. Biosyst. Eng. 180, 96–107. doi: 10.1016/j.biosystemseng.2019.02.002

CrossRef Full Text | Google Scholar

Chan, L., Hosseini, M. S., and Plataniotis, K. N. (2021). A comprehensive analysis of weakly-supervised semantic segmentation in different image domains. Int. J. Comput. Vision 129, 361–384. doi: 10.1007/s11263-020-01373-4

CrossRef Full Text | Google Scholar

Chang, Y.-T., Wang, Q., Hung, W.-C., Piramuthu, R., Tsai, Y.-H., and Yang, M.-H. (2020). “Weakly-supervised semantic segmentation via sub-category exploration,” in CVPR, 8991–9000. doi: 10.1109/CVPR42600.2020.00901

CrossRef Full Text | Google Scholar

Chen, J., Chen, J., Zhang, D., Sun, Y., and Nanehkaran, Y. A. (2020). Using deep transfer learning for image-based plant disease identification. Comput. Electron. Agric. 173, 105393. doi: 10.1016/j.compag.2020.105393

PubMed Abstract | CrossRef Full Text | Google Scholar

Chouhan, S., Koul, A., Singh, D. U., and Jain, S. (2019). “A data repository of leaf images: Practice towards plant conservation with plant Pathology,” in 2019 4th International Conference on Information Systems and Computer Networks (ISCON) (Mathura: IEEE).

Google Scholar

Cox, E., and McKee, J. (1976). A comparison of tipburn susceptibility in lettuce under field and glasshouse conditions. J. Hortic. Sci. 51, 117–122. doi: 10.1080/00221589.1976.11514671

CrossRef Full Text | Google Scholar

Cox, E., McKee, J., and Dearman, A. (1976). The effect of growth rate on tipburn occurrence in lettuce. J. Hortic. Sci. 51, 297–309. doi: 10.1080/00221589.1976.11514693

CrossRef Full Text | Google Scholar

DeChant, C., Wiesner-Hanks, T., Chen, S., Stewart, E., Yosinski, J., Gore, M., et al. (2017). Automated identification of northern leaf blight-infected maize plants from field imagery using deep learning. Phytopathology 107, 1426–1432. doi: 10.1094/PHYTO-11-16-0417-R

PubMed Abstract | CrossRef Full Text | Google Scholar

Defferrard, M., Bresson, X., and Vandergheynst, P. (2016). Convolutional neural networks on graphs with fast localized spectral filtering. Adv. Neural Inf. Process. Syst. 29, 3844–3852.

Google Scholar

Dong, W., Li, X., Zhang, L., and Shi, G. (2011). “Sparsity-based image denoising via dictionary learning and structural clustering,” in CVPR, 457–464. doi: 10.1109/CVPR.2011.5995478

CrossRef Full Text | Google Scholar

Douarre, C., Crispim-Junior, C. F., Gelibert, A., Tougne, L., and Rousseau, D. (2019). Novel data augmentation strategies to boost supervised segmentation of plant disease. Comput. Electron. Agric. 165, 104967. doi: 10.1016/j.compag.2019.104967

CrossRef Full Text | Google Scholar

Giuffrida, V., Scharr, H., and Tsaftaris, S. A. (2017). “Arigan: Synthetic arabidopsis plants using generative adversarial network,” in Proceedings of the IEEE International Conference on Computer Vision Workshops (Venice: IEEE), 2064–2071.

Google Scholar

Gozzovelli, R., Franchetti, B., Bekmurat, M., and Pirri, F. (2021). “Tip-burn stress detection of lettuce canopy grown in plant factories,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (Montreal, BC: IEEE), 1259–1268.

Google Scholar

Grattarola, D., and Alippi, C. (2021). Graph neural networks in tensorflow and keras with spektral [application notes]. IEEE Comput. Intell. Mag. 16, 99–106. doi: 10.1109/MCI.2020.3039072

CrossRef Full Text | Google Scholar

Gu, J., Kuen, J., Joty, S., Cai, J., Morariu, V., Zhao, H., et al. (2020). Self-supervised relationship probing. Adv. Neural Inf. Process. Syst. 33, 1841–1853.

Google Scholar

Hassan, S. M., and Maji, A. K. (2022). Plant disease identification using a novel convolutional neural network. IEEE Access 10, 5390–5401. doi: 10.1109/ACCESS.2022.3141371

CrossRef Full Text | Google Scholar

Hu, H., Ji, D., Gan, W., Bai, S., Wu, W., and Yan, J. (2020). “Class-wise dynamic graph convolution for semantic segmentation,” in European Conference on Computer Vision (Cham: Springer), 1–17.

Google Scholar

Hughes, D. P., and Salathe, M. (2016). An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv preprint.

Google Scholar

Khattak, A., Asghar, M. U., Batool, U., Asghar, M. Z., Ullah, H., Al-Rakhami, M., et al. (2021). Automatic detection of citrus fruit and leaves diseases using deep neural network model. IEEE Access 9, 112942–112954. doi: 10.1109/ACCESS.2021.3096895

CrossRef Full Text | Google Scholar

Kingma, D. P., and Ba, J. (2014). Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980.

Google Scholar

Kipf, T. N., and Welling, M. (2016). Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907.

Google Scholar

Kynkäänniemi, T., Karras, T., Laine, S., Lehtinen, J., and Aila, T. (2019). Improved precision and recall metric for assessing generative models. arXiv preprint arXiv:1904.06991.

Google Scholar

Li, Y., and Gupta, A. (2018). Beyond grids: learning graph representations for visual recognition. Adv. Neural Inf. Process. Syst. 31, 9225–9235.

Google Scholar

Lu, J., Tan, L., and Jiang, H. (2021). Review on convolutional neural network (cnn) applied to plant leaf disease classification. Agriculture 11, 707. doi: 10.3390/agriculture11080707

CrossRef Full Text | Google Scholar

Lutman, B. F. (1919). Tip burn of the potato and other plants. Vermont Agric. Exp. Station.

Mirza, M., and Osindero, S. (2014). Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784.

Google Scholar

Mishra, S., Sachan, R., and Rajpal, D. (2020). Deep convolutional neural network based detection system for real-time corn plant disease recognition. Procedia Comput. Sci. 167, 2003–2010. doi: 10.1016/j.procs.2020.03.236

CrossRef Full Text | Google Scholar

Mohameth, F., Bingcai, C., and Sada, K. A. (2020). Plant disease detection with deep learning and feature extraction using plant village. J. Comput. Commun. 8, 10–22. doi: 10.4236/jcc.2020.86002

PubMed Abstract | CrossRef Full Text | Google Scholar

Mohanty, S., Hughes, D., and Salathe, M. (2016). Using deep learning for image-based plant disease detection. Front Plant Sci. 7, 1419. doi: 10.3389/fpls.2016.01419

PubMed Abstract | CrossRef Full Text | Google Scholar

Mwebaze, E., Gebru, T., Frome, A., Nsumba, S., and Tusubira, J. (2019). icassava 2019fine-grained visual categorization challenge. arXiv preprint.

Google Scholar

Nagasubramanian, K., Jones, S., Singh, A., Singh, A., Ganapathysubramanian, B., and Sarkar, S. (2018). Explaining hyperspectral imaging based plant disease identification: 3d cnn and saliency maps. arXiv preprint.

Google Scholar

Noh, H., Hong, S., and Han, B. (2015). “Learning deconvolution network for semantic segmentation,” in Proceedings of the IEEE International Conference on Computer Vision (Santiago: IEEE), 1520–1528.

Google Scholar

Nowak, E., Jurie, F., and Triggs, B. (2006). “Sampling strategies for bag-of-features image classification,” in ECCV (Berlin: Springer), 490–503.

Google Scholar

Pan, F., Shin, I., Rameau, F., Lee, S., and Kweon, I. S. (2020). “Unsupervised intra-domain adaptation for semantic segmentation through self-supervision,” in CVPR, 3764–3773. doi: 10.1109/CVPR42600.2020.00382

CrossRef Full Text | Google Scholar

Patidar, S., Pandey, A., Shirish, B. A., and Sriram, A. (2020). “Rice plant disease detection and classification using deep residual learning,” in International Conference on Machine Learning, Image Processing, Network Security and Data Sciences (Springer), 278–293.

Google Scholar

Prajapati, H. B., Shah, J. P., and Dabhi, V. K. (2017). Detection and classification of rice plant diseases. Intell. Decis. Technol. 11, 357–373. doi: 10.3233/IDT-170301

CrossRef Full Text | Google Scholar

Rauf, H. T., Saleem, B. A., Lali, M. I. U., Khan, M. A., Sharif, M., and Bukhari, S. A. C. (2019). A citrus fruits and leaves dataset for detection and classification of citrus diseases through machine learning. Data Brief 26, 104340. doi: 10.1016/j.dib.2019.104340

PubMed Abstract | CrossRef Full Text | Google Scholar

Redmon, J., and Farhadi, A. (2016). “Yolo9000: Better, faster, Stronger,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Honolulu, HI: IEEE).

Google Scholar

Saleem, M. H., Khanchi, S., Potgieter, J., and Arif, K. M. (2020). Image-based plant disease identification by deep learning meta-architectures. Plants 9, 1451. doi: 10.3390/plants9111451

PubMed Abstract | CrossRef Full Text | Google Scholar

Sharma, P., Berwal, Y. P. S., and Ghai, W. (2020). Performance analysis of deep learning cnn models for disease detection in plants using image segmentation. Inf. Process. Agric. 7, 566–574. doi: 10.1016/j.inpa.2019.11.001

CrossRef Full Text | Google Scholar

Shimamura, S., Uehara, K., and Koakutsu, S. (2019). Automatic identification of plant physiological disorders in plant factories with artificial light using convolutional neural networks. Int. J. New Comput. Archi. Appl. 9, 25–31. doi: 10.17781/P002611

CrossRef Full Text | Google Scholar

Shrivastava, V., Pradhan, M., Minz, S., and Thakur, M. (2019). “Rice plant disease classification using transfer learning of deep convolution neural network,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-3/W6, 631–635. doi: 10.5194/isprs-archives-XLII-3-W6-631-2019

CrossRef Full Text | Google Scholar

Singh, D., Jain, N., Jain, P., Kayal, P., Kumawat, S., and Batra, N. (2020). “Plantdoc,” in Proceedings of the 7th ACM IKDD CoDS and 25th COMAD, 249–253. doi: 10.1145/3371158.3371196

CrossRef Full Text | Google Scholar

Sodjinou, S. G., Mohammadi, V., Mahama, A. T. S., and Gouton, P. (2021). A deep semantic segmentation-based algorithm to segment crops and weeds in agronomic color images. Inf. Process. Agric. doi: 10.1016/j.inpa.2021.08.003

CrossRef Full Text | Google Scholar

Son, J. E., and Takakura, T. (1989). Effect of ec of nutrient solution and light condition on transpiration and tipburn injury of lettuce in a plant factory. J. Agric. Meteorol. 44, 253–258. doi: 10.2480/agrmet.44.253

CrossRef Full Text | Google Scholar

Sujatha, R., Chatterjee, J. M., Jhanjhi, N., and Brohi, S. N. (2021). Performance of deep learning vs machine learning in plant leaf disease detection. Microprocess Microsyst. 80, 103615. doi: 10.1016/j.micpro.2020.103615

CrossRef Full Text | Google Scholar

Sun, G., Wang, W., Dai, J., and Van Gool, L. (2020). “Mining cross-image semantics for weakly supervised semantic segmentation,” in ECCV, 347–365.

PubMed Abstract | Google Scholar

Sun, W., Zhang, J., and Barnes, N. (2022). “Inferring the class conditional response map for weakly supervised semantic segmentation,” in WACV, 2878–2887.

Google Scholar

Syed-Ab-Rahman, S. F., Hesamian, M. H., and Prasad, M. (2022). Citrus disease detection and classification using end-to-end anchor-based deep learning model. Appl. Intell. 52, 927–938. doi: 10.1007/s10489-021-02452-w

CrossRef Full Text | Google Scholar

Terentev, A., Dolzhenko, V., Fedotov, A., and Eremenko, D. (2022). Current state of hyperspectral remote sensing for early plant disease detection: a review. Sensors 22, 757. doi: 10.3390/s22030757

PubMed Abstract | CrossRef Full Text | Google Scholar

Termohlen, G., and Hoeven, A. V. (1965). Tipburn symptoms in lettuce. Sympos. Veget. Growing Glass 4, 105–110. doi: 10.17660/ActaHortic.1966.4.21

CrossRef Full Text | Google Scholar

Tian, Y.-W., and Li, C.-H. (2004). Color image segmentation method based on statistical pattern recognition for plant disease diagnose. J Jilin Univer. Technol. 2, 28.

Vippon Preet Kour, S. A. (2019). Plantaek: A Leaf Database of Native Plants of Jammu and Kashmir (Singapore: Springer), 359–368.

Google Scholar

Wang, Y., Zhang, J., Kan, M., Shan, S., and Chen, X. (2020). “Self-supervised equivariant attention mechanism for weakly supervised semantic segmentation,” in CVPR, 12275–12284. doi: 10.1109/CVPR42600.2020.01229

CrossRef Full Text | Google Scholar

Watchareeruetai, U., Noinongyao, P., Wattanapaiboonsuk, C., Khantiviriya, P., and Duangsrisai, S. (2018). “Identification of plant nutrient deficiencies using convolutional neural networks,” in 2018 International Electrical Engineering Congress (iEECON) (Krabi,: IEEE), 1–4.

Google Scholar

Wu, T., Huang, J., Gao, G., Wei, X., Wei, X., Luo, X., et al. (2021). “Embedded discriminative attention mechanism for weakly supervised semantic segmentation,” in CVPR, 16765–16774.

Google Scholar

Xie, M.-K., and Huang, S.-J. (2021). “Partial multi-label learning with noisy label identification,” in IEEE Transactions on Pattern Analysis and Machine Intelligence.

PubMed Abstract | Google Scholar

Yang, Y., and Soatto, S. (2020). “Fda: fourier domain adaptation for semantic segmentation,” in CVPR, 4085–4095.

Google Scholar

Yao, Q., and Gong, X. (2020). Saliency guided self-attention network for weakly and semi-supervised semantic segmentation. IEEE Access 8, 14413–14423. doi: 10.1109/ACCESS.2020.2966647

CrossRef Full Text | Google Scholar

Zeiler, M. D., and Fergus, R. (2014). “Visualizing and understanding convolutional networks,” in ECCV (Cham: Springer), 818–833.

Google Scholar

Zhang, F., Gu, C., Zhang, C., and Dai, Y. (2021). “Complementary patch for weakly supervised semantic segmentation,” in ICCV, 7242–7251.

Google Scholar

Zhang, J., and Wang, S.-X. (2007). A study on the segmentation method in image processing for plant disease of gree nhouse. J. Inner Mongolia Agric. Univer. 3, 1009–3575.

Zhang, X., Han, L., Dong, Y., Shi, Y., Huang, W., HAN, L., et al. (2019). A deep learning-based approach for automated yellow rust disease detection from high-resolution hyperspectral uav images. Remote Sens. 11, 1554. doi: 10.3390/rs11131554

CrossRef Full Text | Google Scholar

Zheng, D., Wang, M., Gan, Q., Song, X., Zhang, Z., and Karypis, G. (2021). “Scalable graph neural networks with deep graph library,” in Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 1141–1142.

PubMed Abstract | Google Scholar

Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., and Torralba, A. (2016). “Learning deep features for discriminative localization,” in CVPR, 2921–2929.

Google Scholar

Zhou, M., Chen, H., Paisley, J. W., Ren, L., Sapiro, G., and Carin, L. (2009). “Non-parametric bayesian dictionary learning for sparse image representations,” in NIPS, Vol. 9, 2295–2303.

Google Scholar

Zou, Y., Yu, Z., Kumar, B., and Wang, J. (2018). “Unsupervised domain adaptation for semantic segmentation via class-balanced self-training,” in ECCV, 289–305.

Google Scholar

Keywords: tip-burn detection and localization, self supervised segmentation, plant disease classification, segmentation of large canopies, indoor farming

Citation: Franchetti B and Pirri F (2022) Detection and Localization of Tip-Burn on Large Lettuce Canopies. Front. Plant Sci. 13:874035. doi: 10.3389/fpls.2022.874035

Received: 11 February 2022; Accepted: 29 March 2022;
Published: 12 May 2022.

Edited by:

Shawn Carlisle Kefauver, University of Barcelona, Spain

Reviewed by:

Jianjun Du, Beijing Research Center for Information Technology in Agriculture, China
Ana María Mendez-Espinoza, Instituto de Investigaciones Agropecuarias, Chile
Kun Li, Chinese Academy of Agricultural Sciences (CAAS), China

Copyright © 2022 Franchetti and Pirri. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Fiora Pirri, pirri@diag.uniroma1.it; fiora@deepplants.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.