Next Article in Journal
Wavelet Energy Evolution Characteristics of Acoustic Emission Signal under True-Triaxial Loading during the Rockburst Test
Next Article in Special Issue
Intelligent Data Analytics Framework for Precision Farming Using IoT and Regressor Machine Learning Algorithms
Previous Article in Journal
A Rotor Winding Internal Short-Circuit Fault Protection Method for Variable-Speed Pumped Storage Units
Previous Article in Special Issue
Fuzzy Neural Network PID Strategy Based on PSO Optimization for pH Control of Water and Fertilizer Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks

1
Faculty of Software and Information Science, Iwate Prefectural University, Takizawa 020-0693, Japan
2
Faculty of Systems Science and Technology, Akita Prefectural University, Yurihonjo 015-0055, Japan
3
Faculty of Bioresource Sciences, Akita Prefectural University, Akita 010-0195, Japan
4
Graduate School of International Resource Sciences, Akita University, Akita 010-8502, Japan
5
Agri-Innovation Education and Research Center, Akita Prefectural University, Ogata 010-0444, Japan
*
Author to whom correspondence should be addressed.
Submission received: 14 July 2022 / Revised: 25 July 2022 / Accepted: 27 July 2022 / Published: 2 August 2022
(This article belongs to the Collection Agriculture 4.0: From Precision Agriculture to Smart Farming)

Abstract

:
Classification, segmentation, and recognition techniques based on deep-learning algorithms are used for smart farming. It is an important and challenging task to reduce the time, burden, and cost of annotation procedures for collected datasets from fields and crops that are changing in a wide variety of ways according to growing, weather patterns, and seasons. This study was conducted to generate crop image datasets for semantic segmentation based on an image style transfer using generative adversarial networks (GANs). To assess data-augmented performance and calculation burdens, our proposed framework comprises contrastive unpaired translation (CUT) for a conditional GAN, pix2pixHD for an unconditional GAN, and DeepLabV3+ for semantic segmentation. Using these networks, the proposed framework provides not only image generation for data augmentation, but also automatic labeling based on distinctive feature learning among domains. The Fréchet inception distance (FID) and mean intersection over union (mIoU) were used, respectively, as evaluation metrics for GANs and semantic segmentation. We used a public benchmark dataset and two original benchmark datasets to evaluate our framework of four image-augmentation types compared with the baseline without using GANs. The experimentally obtained results showed the efficacy of using augmented images, which we evaluated using FID and mIoU. The mIoU scores for the public benchmark dataset improved by 0.03 for the training subset, while remaining similar on the test subset. For the first original benchmark dataset, the mIoU scores improved by 0.01 for the test subset, while they dropped by 0.03 for the training subset. Finally, the mIoU scores for the second original benchmark dataset improved by 0.18 for the training subset and 0.03 for the test subset.

1. Introduction

Smart farming provides labor conservation, automation, and improved efficiency of farming operations, thereby leading to advancement of value-added agricultural activities [1]. Some representative case studies of successful smart farming are automatic operations of farm machinery in terms of tractors and combine harvesters using multiple sensors and global navigation satellite systems (GNSS), automatic water control and management systems based on internet of things (IoT) sensors and devices [2], crop monitoring, and pesticide and fertilizer spreaders using drones [3]. Although smart farming has already progressed to the implementation, demonstration, and dissemination phases, the major tasks are limited to automated repetition of simplified operations and locomotion based on operational procedures and routes that must be set in advance.
Practical applications of deep learning (DL) [4] technologies for smart farming are progressing rapidly [5]. Compared with general DL applications in terms of automated driving [6], biometric identification [7], and medical image processing [8], technical challenges remain, especially for classification, recognition, and decisions for agricultural applications [9,10,11]. Moreover, the collection of high-quality and large agricultural datasets [12] entails great burdens because of the wide variations in crop growing periods, weather patterns, air temperature, seasonal changes, and various external factors [13]. Although extensive amounts of elaborately annotated data improve DL performance and avoid overfitting [14], the burdens and time of annotation [15], especially for pixel-wise segmentation labels, make acquiring and using them labor-intensive and expensive. Therefore, numerous research tasks remain for DL applications in smart farming [16].
This study specifically examines semantic segmentation, which is a major task in computer vision studies, to explore DL-based solutions as challenges for smart farming [17]. Particularly, this study examines semantic segmentation of crop images [18] related to disease detection, weeding, and cultivation management. Image segmentation can be of three types: semantic segmentation, instance segmentation, and panoptic segmentation [19]. Segmentation targets comprise things as countable objects and stuff as matter, material, and textures. Semantic segmentation can handle both targets. Things within the same object category are assigned the same label. Instance segmentation handles only things. Things in each object are assigned with unique labels for the respective instances. Panoptic segmentation includes both semantic and instance segmentation mechanisms.
Champ et al. [20] proposed an instant segmentation method for crop and weed plants using a region-based convolutional neural network (R-CNN) [21]. Instance segmentation provides flexible detection of each crop and weed for sparse vegetation. Their research project specifically examined sparse vegetation images because they were intended to be mounted on a precision agricultural robot. By contrast, instance segmentation will be challenging if the vegetation distribution is dense. This paper presents a semantic segmentation framework to exploit agricultural images with dense vegetation.
Numerous benchmark image datasets are available for semantic segmentation in agriculture [22]. These datasets were annotated using ground truth (GT) labels through an enormous effort. Annotations for agricultural datasets are conducted for the respective purposes and tasks. Special training is also necessary for annotators in advance, depending on objects such as crop damage and leaf types. To reduce burdens for annotating and producing synthesized data [23], this study is intended to generate crop datasets based on image style transfers [24] using generative adversarial networks (GANs) [25]. Results obtained from experimentation using the proposed framework demonstrate that the quantity and quality expansion of images and labels using GANs improve accuracy and convergence compared to existing semantic segmentation models using source images and labels.
This paper is structured as follows. Section 2 briefly reviews state-of-the-art data augmentation methods based on image style transfer and GANs. Subsequently, Section 3 and Section 4 present our framework for semantic segmentation combined with data augmentation using two GAN models with different mechanisms and their setup properties, including benchmark datasets of three types and evaluation metrics. The preliminary and evaluation experiment results are presented in Section 5 and Section 6. Finally, Section 7 presents the conclusions and explains some intentions for future work.

2. Related Studies

Existing data augmentation studies are divisible into two approaches. The first approach is based on model-based algorithms. Rozantsev et al. [26] proposed an algorithm that synthesizes virtual 3D object images superimposed on specific background images. Rematas et al. [27] proposed an image generation method from unknown viewpoints using structural features extracted from 3D objects. They applied their method to data augmentation for creating super-high-resolution images combined with 2D to 3D transformations. Pishchulin et al. [28] proposed a data augmentation method for pedestrian detection and pose estimation. They specifically controlled the poses and appearances of the existing data. However, these data augmentation methods were applied to unsupervised classification tasks because the augmentation targets were merely source data without labels. As one example of research into automatic label generation, Vázquez et al. [29] detected pedestrians using labels generated from outdoor scene images that were created using a simulator. Their proposed domain-adapted pedestrian classifier was trained by combining a small number of pedestrian samples from a real-world target domain with numerous ones from a virtual-world source domain. Ros et al. [30] proposed a method to synthesize images and labels in a virtual urban environment. They demonstrated application experiments for semantic segmentation using synthesized multi-camera images from a virtual car after annotation, including different labels for each season.
The second approach is based on learning algorithms. This approach includes numerous methods based on conventional machine-learning (CMT) algorithms and DL algorithms using state-of-the-art backbones such as ConvNets [31,32,33,34,35,36,37,38,39] and Transformers [40,41,42,43,44]. Recently, numerous studies of GANs have been conducted for data generation and augmentation [45,46,47,48,49]. Fundamentally, GANs comprise a generator and a discriminator. A generator supplies virtual images generated from noise signals to a discriminator. The learning objective for a discriminator is the classification of real images and generated pseudo-images. The learning objective for a generator is the opposite of that of a discriminator. This relation is consistent with the mini-max principle [50]. A discriminator learns from both images while avoiding false segmentation of augmented images provided by a generator. The respective networks modify weights simultaneously and in parallel to opposite objectives as adversarial learning.
Radford et al. [51] proposed a deep convolutional GAN (DCGAN) for unsupervised representation learning. They introduced convolutional layers as an architectural topology inside of the generator. Based on the energy-based GAN (EBGAN) [52], Berthelot et al. [53] proposed a boundary equilibrium GAN (BEGAN) that uses a convolutional auto-encoder (CAE) [54] for a discriminator. Although DCGAN contained no mechanism to judge the convergence of learning for a generator and a discriminator, BEGAN solved this difficulty after introducing loss functions to the respective networks. Karras et al. [55] proposed a progressive growing GAN (PGGAN) that adds new layers according to learning progress. They demonstrated not only the creation of high-resolution images, but also a reduction in computational cost for training because their proposed method gradually learned the overall image structure.
As an alternative generator architecture based on the arbitrary style transfer mechanism [56], Karras et al. [57] proposed StyleGAN for learning unsupervised separation of high-level attributes and stochastic variation in the augmented images. They introduced not only a learning mechanism to increase resolutions resembling PGGAN progressively, but also local vectors to add probabilistic variation and global vectors to decide the object orientations and respective parts in the generative process. StyleGAN has been used as a baseline in numerous studies because of its high accuracy and suitability for various applications of image generation. Moreover, the process of image generation based on GANs includes practical applications [58]. Wang et al. [59] proposed an enhanced super-resolution GAN (ESRGAN) for applications that transform low-resolution images to high-resolution images. Zhang et al. [60] proposed a new model called StackGAN to convert text to images. These studies demonstrated the wide range of GAN applications, with research and development becoming more active in recent years.
GAN frameworks are also applied to image style transfer [61]. Zhao et al. [62] proposed a crop detection method using StyleGAN [57] combined with deep visual transfer learning. We consider that data augmentation with image style transfer based on GANs contributes to the improvement of segmentation accuracy in agricultural applications. GANs can be of two models: conditional GANs and unconditional GANs [63]. This difference derives from the correspondence between source images and labels. On the one hand, unconditional GANs learn generative models from source images. On the other hand, conditional GANs learn generative models from source images and labels similarly to supervised learning. As a representative conditional GAN, Isola et al. [64] proposed pix2pix, which uses two conditional GANs to enable a conversion task between images. Wang et al. [65] proposed an improved model pix2pixHD that added a learning mechanism for high-resolution images in two generators. Moreover, this model learns distinct boundaries for similar objects using instance maps.
Lee et al. [66] proposed MaskGAN, which specializes in the generation of face images from face labels. Park et al. [67] proposed an improved model, spatially adaptive denormalization (SPADE), which tackled shortcomings of class information deficits in the input process of label images for batch normalization. SPADE provides various image transformations for simple input labels with flexibility in scaling and bias rates in each layer. Sushko et al. [68] proposed an improved model, with only adversarial supervision for semantic image segmentation (OASIS), which applied segmentation networks to SPADE discriminators. Although SPADE used VGGNet [69], which contains numerous parameters as a backbone for the discriminator of SPADE [67], OASIS provides a lightweight network through the introduction of a simplified regionally segmented network. Moreover, OASIS provides multimodal images with the direct input of 3D noise tensors and labels images to a discriminator, with transformed images based on multidimensional information [68].
As a representative unconditional GAN, Zhu et al. [70] proposed CycleGAN for actualizing cycle consistency. To learn the relational features between input features and output values, cycle consistency loss (CCL) was introduced to CycleGAN for reconstructing input images from output images as inverse mapping in low-dimensional feature spaces. Park et al. [71] proposed contrastive unpaired translation (CUT) based on contrastive learning (CL). They extracted simulated features such as background and object structure in input images to learn patch regions without using comprehensive learning for entire images. Based on the U-Net [72] architecture, Eskandar et al. [73] proposed an unsupervised paradigm for semantic image synthesis (USIS) consisting of a generator and a discriminator derived from SPADE [67]. This is a novel framework combined with an unconditional GAN with contrastive self-supervised learning (SSL) [74] based on image segmentation. Moreover, entered images were recognized based on wavelet transformations for matching the color and texture distributions in source images. The experimentally obtained results obtained using three public datasets demonstrated bridging of the performance gap separating paired and unpaired translation models [75].
Numerous data augmentation methods using GANs have been proposed in conjunction with academic research and industrial DL applications [76,77,78]. Although research on data augmentation is advancing rapidly, no meta approach has yet been proposed. Recognition accuracies are competitive and interchangeable depending on benchmark datasets. Moreover, popular benchmark datasets comprise images of general objects taken as scenes in various environments. Because sensing targets such as crops and weeds grow and change diversely over time [79], research projects and case studies using agricultural datasets are more limited than those presented for other computer vision tasks [80].

3. Proposed Framework

3.1. Overall Network Architecture

Figure 1 presents the overall network architecture of our proposed framework. This framework globally comprises three modules: GAN modules of two types and a semantic segmentation module.
For conventional semantic segmentation methods [81,82,83] based on supervised learning, source images and corresponding label images were prepared in advance. Training images including validation images and test images were extracted from these image sets. Our proposed framework provides image augmentation using GANs of two types. Subsequently, source and augmented images and labels were provided to the semantic segmentation module for pixel-wise segmentation. The utilization of both GAN modules provides not only data augmentation based on image generation, but also automatic labeling for the limited number of source images. The conditional GAN module is applied if a source image and its corresponding label are present. By contrast, the unconditional GAN module is applied if source images are solely present without label images.

3.2. GAN

Enormous computational resources must be used for GANs [84]. To assess both the extended capacity and computational cost, we used CUT [71] for an unconditional GAN and pix2pixHD [65] for a conditional GAN as the baseline for this study. Using both outstanding GANs, our framework provides not only distinctive feature learning among domains, but also automatic labeling and data augmentation based on simulated image creation. Moreover, both models are useful to achieving semi-conditional translation with multiple GANs as a future application. Figure 2 depicts structural differences between a conditional GAN and an unconditional GAN. The flows of label images given to a generator are distinctly different.

3.2.1. Unconditional GAN

Based on the framework of CycleGAN [70], CUT [71] introduced contrastive representation learning mechanism [85]. Let x, y ^ , and z respectively denote an input image, an output image, and a patch region and letting N denote the number of image sampling; then, a positive x sample, z n + in N and y ^ in Z, is trained as a similar pair. By contrast, for a negative x sample, z n in N and y ^ in Z are trained as a non-similar pair.
For a classification problem of N + 1 patterns, softmax cross-entropy L s [86] is calculated as
L s ( z , z + , z ) = log exp ( z · z + / τ ) exp ( z · z + / τ ) + n = 1 N exp ( z · z n / τ ) ,
where τ represents a metric correction coefficient for z + and z n in z.
To calculate the obtained patch values, L p was introduced as a loss function for a particular patch combined with the conventional loss function L g for a GAN. L g and L p are defined as
L g ( G , D , X , Y ) = E y log D ( y ) + E x log ( 1 D ( G ( x ) ) ) ,
L p ( G , X ) = E x s = 1 L t = 1 T ( z s , t , z s , t + , z s , t ) ,
where X, Y, s S , and t T respectively denote an input image domain, an output image domain, an index value for sampling layers, and the sampling number in a patch region at the s-th layers.
After adding the Formulas (2) and (3), the final L for CUT is obtained as the following.
min G max D L = L g ( G , D , X , Y ) + L p ( G , X ) + L p ( G , Y ) .
After learning patch regions, structural object features and the background of input images are extracted simultaneously.

3.2.2. Conditional GAN

To translate images with higher resolutions and quality, pix2pixHD [65] was proposed as an improved version of pix2pix [64]. The pix2pixHD network includes two subnetworks: a global generator ( G 1 ) and a local enhanced network ( G 2 ). First, G 1 is trained using low-resolution images. After training, G 2 is connected to G 1 . Subsequently, both networks are trained using high-resolution images. The G 1 weights are fixed partially while the training proceeds. Finally, both networks are retained after releasing the G 1 weights. Stability and convergence are improved based on a mechanism that allows the system to learn incrementally for images of different resolutions.
To enhance the network architecture, pix2pixHD copes with feature unevenness caused by scale changes using three discriminators. Each discriminator extracts salient features from each layer. To assess the calculation of the numerical feature matching degrees obtained from source images and augmented images, feature matching loss L m based on the Manhattan metric in the taxicab geometry [87] is defined as
L m ( G , D k ) = E ( x , y ) t = 1 T 1 N t D k t ( x , y ) D k t ( x , G ( x ) ) ,
where D k and T respectively represent the k-th discriminator and the number of layers in each discriminator.
After adding Equations (2) and (5), the final loss L of pix2pixHD is presented as
min G max D L = k = 1 K L g ( G , D k , X , Y ) + λ k = 1 K L m ( G , D k ) ,
where λ and K respectively stand for a weight coefficient of L m and the number of discriminators.

3.3. Semantic Segmentation

Semantic segmentation provides object and stuff class labels for every pixel in a presented image. This process is also known as a high-density labeling task [88]. Images include widely diverse explicit and implicit information and meanings. Moreover, semantic segmentation provides a comprehensive scene description that includes information related to object category, location, shape, and context [89]. Semantic segmentation has numerous applications such as medical image processing [90], self-driving cars [91], structural damage mapping [92], vehicle detection [93], remote sensing [94], monitoring photovoltaic solar plants [95], power line cable inspection [96], building extraction [97], drone detection [98], classifying cooking activities [99], socially assistive robots [100], and industrial automation [101]. Representative semantic segmentation models include SegNet [102], U-Net [72], DeepLabV3+ [103], HRNet [104], SegFormer [41], Segmenter [42], and ConvNeXt [105].
For visual tasks, vision transformer (ViT) [106] models perform better than ConvNets [4] if segmentation targets are objects [107]. By contrast, for stuff classes, especially in textures, the convolution mechanism provided by ConvNets demonstrated superiority, including in feature representations based on multilayered relations, compared to attention mechanisms provided by ViT [108]. Steiner et al. [109] indicated that the weaker ViT inductive bias is generally found to cause an increased reliance on model regularization or data augmentation when training on smaller training subsets. Therefore, for extracting sufficient ViT performance, very large datasets such as ImageNet-21K [110] or JFT-300M [111] are necessary for pre-training [112].
Considering a balance between classification accuracy, implementation platform, and processing efficiency, DeepLabV3+ [103] was used for this study. This network flexibly extracts features with parallel processing of multiple filters that contain arbitrary resolutions by spatial pyramid pooling. The automatic formation of a decoding structure that propagates spatial information held by an encoder provides extraction of clear boundaries between objects. DeepLabV3+ [103] was pre-trained using the COCO-Stuff [113] open benchmark dataset, which provided comprehensive annotations of 172 classes. Based on a comparison of results obtained from our earlier studies [114,115], Xception-65 [116] was used for the network backbone in this study. Here, the semantic segmentation module can be switched flexibly to state-of-the-art or practical backbones according to technological progress, target problems, and specifications, such as ConvNeXt [105], Segmenter [42], SegFormer [41], TransFG [117], MetaFormer [118], and K-Net [119].

4. Requirements

4.1. Benchmark Datasets

For this study, we used three benchmark datasets. The first benchmark dataset is Cityscapes [120], which is a public dataset for semantic urban scene understanding. The second and third datasets are our originally developed datasets: Rice blast, which is a rice disease detection dataset developed for our earlier study [80], and Soybeans, which is a leaf and weed classification dataset developed for this study.

4.1.1. Cityscapes

Cityscapes [120] is a benchmark suite for training and validating pixel-wise and instance-level semantic labeling models. Based on this pioneering work, derivative datasets have been proposed such as Citypersons [121], CityWalks [122], and Cityscapes 3D [123]. The Cityscapes suite consists of a large and diverse set of stereo video sequences recorded on the streets of 50 German cities. All images converted from the video sequences were annotated as 20 classes for semantic labels and eight classes for instance labels. For this study, we used images with semantic labels because of the similar specifications in our original datasets.
Figure 3 depicts sample images and respective annotations as GT. The semantic labels and corresponding colors are presented in Table 1. The original image resolution is 2048 × 1024 pixels. The region of interest (RoI) sizes for random sampling are 512 × 256 pixels for the semantic segmentation module and 256 × 256 pixels for the GAN modules. We set these RoI sizes considering the computation time and memory capacity in our implementation environment. We split 3000 randomly extracted images in a 4:1:1 ratio for training, validation, and test subsets.

4.1.2. Rice Blast

For the second benchmark dataset, we used this benchmark dataset from our earlier work [80]. Figure 4 depicts sample images and corresponding labels after annotation. This dataset comprises three classes: background (BG), rice leaves (RL), and rice blast (RB). The black, green, and red color pixels respectively correspond to BG, RL, and RB classes. This dataset comprises 119 images with a resolution of 6240 × 4160 pixels. Random sampling RoI sizes are 480 × 320 pixels for the semantic segmentation module and 256 × 256 pixels for the GAN modules. Images were split according to a 7:2:3 ratio for training, validation, and test subsets.

4.1.3. Soybeans

For this study, we created a new benchmark dataset designated as Soybeans. Figure 5 depicts sample images and corresponding labels after annotation. This dataset comprises four classes: BG, soybean leaves (SL), Poaceae weeds (PW), and other weeds (OW). The black, green, red, and blue color pixels respectively correspond to BG, SL, PW, and OW classes. We obtained an HD movie using an onboard camera mounted on a drone (Phantom 4; SZ DJI Technology Co., Ltd., Shenzhen, China). We extracted 300 images with resolution of 1280 × 720 pixels from the movie. The RoI sizes for random sampling were set as 512 × 256 pixels for the semantic segmentation module and 480 × 360 pixels for the GAN modules. The images were split according to a 4:1:1 ratio for training, validation, and test subsets.

4.2. Evaluation Metrics

We used the Fréchet Inception distance (FID) [124] to assess the quality of images generated using GANs. FID is an evaluation criterion based on the Inception [33] network backbone, which was pre-trained using ImageNet [125]. Similarities are evaluated quantitatively from feature vectors extracted from source images and augmented images. Letting m w and m respectively denote norms of feature vectors obtained from augmented images and source images, and letting C w and C respectively denote the respective feature vector and covariance matrixes, the definition of FID can be presented as
FID = m m w 2 + T r ( C + C w 2 C C w ) ,
where T r stands for a trace of matrixes.
Subsequently, we used intersection over union (IoU), which is also known as the Jaccard index [126], as a semantic segmentation evaluation metric. IoU is the de facto evaluation metric used for object challenges such as PASCAL VOC [127] and MS COCO [128]. Using IoU, the degree of overlapping in the GT regions and prediction regions is defined as
IoU = T P T P + F P + F N ,
where T P , F P , and F N respectively denote true positive, false positive, and false negative rates.

5. Preliminary Experiment

Our proposed method generates labels from images and images from labels using CUT [71] as an unconditional GAN and pix2pixHD [65] as a conditional GAN. This preliminary experiment provides accuracy and characteristics of labels generated from images and images generated from labels. For this study, the processes to be generated from images to labels and labels to images are represented as I2L and L2I.

5.1. Cityscapes

Figure 6 depicts time-series training loss trends for the respective GANs in Cityscapes. For the trends of CUT, the I2L loss from the discriminator was lower than that from the generator after 150 epochs. By contrast, the L2I losses from the generators and the discriminators of both GAN modules are competitive, respectively, at around 0.4 and 0.6.
Figure 7 depicts FID trends at 25 epoch intervals in respective GANs. The FID scores, except for I2L for CUT, dropped according to the progress of epochs. As the minimum, the L2I FID score for CUT showed 27.6 at the final epoch. By contrast, the I2L FID for CUT is unchanged after increasing the score from 75 epoch to 100 epoch. This result demonstrates the difficulty of the generation process from images to labels, with CUT being harder than the other three generators.
Figure 8 depicts sample images of style transfer results. These results were obtained from the final generation of 200 epochs except for the I2L result for CUT. For the FID presented in Figure 7, the result shown in Figure 8a was obtained at the generation of 175 epochs. The first and third columns respectively present the source images and GT images. The second and fourth columns respectively present I2L results and L2I results. Therefore, label generation results can be compared subjectively in the images of the second and third columns. Moreover, augmented images can be compared subjectively to the images in the first and fourth columns. Two sets of results are shown for each transfer and each GAN. The first and third rows present transfer results of high FID scores. The second and fourth rows present transfer results of low FID scores.
Although pixels of the source images and the label image are widely matched in stuff in terms of roads and vegetation as stuff for I2L of CUT in Figure 8a, the objects in terms of cars and riders are unmatched. By contrast, the I2L results of pix2pixHD in Figure 8c ensure high reproducibility for both stuff and objects. The L2I result for CUT in Figure 8b demonstrates that distortion occurred throughout the images. Although cars and riders were generated as objects, discriminating them is difficult without consideration of the traffic scene context. The L2I result for pix2pixHD in Figure 8d demonstrates that an image resembling GT was generated. Although small differences are apparent, such as color patterns, road signs painted on the road surface, traffic dashes and solid lines, and overall shapes of the windows and doors of the buildings, this result can be evaluated as having sufficient quality, similar to an actual image. As depicted in Figure 6, the pix2pixHD loss is lower than the CUT loss. As depicted in Figure 7, the FID of pix2pixHD is lower than that of CUT. Visual observation-based evaluation results verify that the incongruity of pix2pixHD is lower than that of CUT. From these results, we infer that the quality of generated labels and images for pix2pixHD is higher than those for CUT. We consider that this trend corresponds to the property of the objective evaluation metrics of loss and FID.

5.2. Rice Blast

Figure 9 depicts time-series training loss trends for the respective GANs in Rice blast. Although the L2I losses for CUT and the I2L losses for pix2pixHD are mutually close, the I2L losses for CUT are separated. For I2L of CUT and L2I of pix2pixHD, the discriminator loss is lower than the generator loss. However, the I2L losses for CUT are mutually close.
Figure 10 depicts time-series FID trends at 25 epoch intervals in the respective GANs. The L2I FID scores for pix2pixHD dropped smoothly according to the forwarded generations. The FID at 100 epoch is 205.1. By contrast, FID scores of the other three models showed no change according to the increased generations. The I2L FID for pix2pixHD was 170.9 at 100 epoch. This score is low compared to that of other models. The L2I FID for CUT shows two-times higher results than those obtained with other models. This tendency indicates difficulties in ensuring accuracy compared with other models, especially in the L2I process for CUT.
Figure 11 depicts sample images of style transfer results. These results were obtained from the final generation of 100 epochs. Unmatched labels appeared in label images of I2L for CUT, especially in RB regions. The L2I result for CUT demonstrates that object boundaries were created according to the boundaries of labels; the augmented images are unnatural. For I2L results for pix2pixHD, the RB positions are correct, but the shapes and sizes are unmatched. The L2I results for pix2pixHD show the occurrence of shadow regions that dropped the contrast of RL regions partially. Although RB regions lack clarity, virtual blasts similar to those in the source images were generated.

5.3. Soybeans

Figure 12 presents time-series training loss trends for respective GANs in Soybeans. The pix2pixHD result demonstrates similar I2L loss trends in the generator and the discriminator. Results for the other three models indicate that the discriminator losses are lower than the generator losses.
Figure 13 portrays time-series FID trends at 25 epoch intervals in the respective GANs. Overall, the L2I FID scores tended to decrease as the generations progressed. The I2L FID scores for pix2pixHD showed no change through generations. The I2L FID scores for CUT increased after 50 generations.
Figure 14 depicts sample images of style transfer results. These results were obtained from the final generation of 100 epochs. The I2L results for CUT include labels that were converted incorrectly from BG to SL. The L2I results for CUT include numerous pixels that were converted incorrectly from OW to BG. The I2L results for pix2pixHD show labels generated according to GT with slight shape differences. The L2I results for pix2pixHD show no false transformation pixels that differ from those for CUT. Compared to the source images, transferred images were generated with high reality with only slight differences in leaf shape and sharpness.

6. Evaluation Experiment

6.1. Setup

For this study, we evaluated the proposed framework quantitatively and qualitatively using five experiments, including baseline evaluation. Table 2 presents a summary of the data source combinations for each experiment. First, semantic segmentation was conducted using only the source images. We considered this result as the baseline. Subsequently, Experiment A was conducted using images generated from CUT [71] as an unconditional GAN. Experiment B was conducted using images generated from pix2pixHD [65] as a conditional GAN. Experiment C was conducted using images generated from both CUT and pix2pixHD. Finally, Experiment D was conducted using all the images.
Table 3 presents hyperparameters of each model for the experiments. The CUT and pix2pixHD parameters were set through the literature [65,71] and our preliminary experiments. The DeepLabV3+ parameters were set empirically through our earlier studies [114,115] and several references [129,130,131]. These settings were not necessarily optimal values because we used no automated parameter setting tools [132,133,134].
We used two platforms for implementing and executing our framework. First, we used Colaboratory (Colab) (Google LLC; Mountain View, CA, USA), which is a cloud platform for accelerating DL applications [135]. The major Colab specifications are the following: 12 GB Tesla K80 (NVIDIA Corp., Santa Clara, CA, USA), GPU; XEON 2.20 GHz (Intel Corp., Santa Clara, CA, USA), CPU; Ubuntu 18.04.5 LTS, OS; 12.69 GB, RAM; 107.72 GB, HDD for CPU and TPU; and 78.19 GB, HDD for GPU. The second platform was a Windows desktop computer. The major specifications of this platform are the following: 48 GB RTX A6000 (NVIDIA Corp., Santa Clara, CA, USA), GPU; Ryzen Threadripper 3970X 3.7 GHz (Advanced Micro Devices, Inc., Santa Clara, CA, USA), CPU; Ubuntu 18.04.5 LTS on Windows Subsystem for Linux 2 (Microsoft Corp., Redmond, WA, USA), OS; 32 GB, RAM; and 512 GB, HDD.
Although numerous GAN and semantic segmentation models have been proposed, a tradeoff exists between the performance and computational cost. For example, StyleGAN2 [136] and StyleGAN3 [137] were calculated using eight GPUs (Tesla V100, NVIDIA Corp., Santa Clara, CA, USA). For our implementation and execution environment, we used the models and parameters presented in Table 3. These models can be changed arbitrarily according to the computational platform because the proposed method is the framework depicted in Figure 1.

6.2. Baseline

As a baseline for this study, we trained DeepLabV3+ using only the source images without augmented images generated from GANs.

6.2.1. Cityscapes

Figure 15 depicts the confusion matrix for the Cityscapes test subset images. After normalization in each class, each matrix element is displayed in heatmap color distributions. The diagonal color pattern from the upper left to the bottom right represents high temperature. However, the color patterns for IDs 7, 8, 9, 14, 18, and 19, which correspond to pole, traffic light, traffic sign, rider, train, and motorcycle, show high temperatures in the other classes of the non-diagonal positions.
Table 4 presents IoU scores of the semantic segmentation result in each class of Cityscapes. The corresponding class IDs are apparent in Table 1. For the training subsets, the IoU scores of 10 classes, which respectively correspond to unlabeled, road, sidewalk, building, vegetation, terrain, sky, car, bus, and train, are greater than the mean score. For the test subsets, the IoU scores of eight classes, which respectively correspond to unlabeled, ground, road, sidewalk, building, vegetation, sky, and car, are greater than the mean score.
Figure 16 and Figure 17 respectively denote representative successful and failed test subset images. The mIoU threshold to classify these cases was set as 0.266. Figure 17 depicts false segmentation samples in the c a r , b u s , and t r u c k classes. We infer that these similar classes have difficulty learning differences among object features.

6.2.2. Rice Blast

Table 5 represents the confusion matrix for the Rice blast test subset images. Numbers in the parentheses correspond to recognition rates on the diagonal cells and the false recognition rates on the other cells. Although high accuracy was obtained in RL, the accuracy was low in RB. The false recognition between RB and RL is greater than in other combinations of classes.
Table 6 presents the semantic segmentation results evaluated using IoU scores for the respective Rice blast classes. The respective mIoU scores for the training and test subset images are 0.87 and 0.45. The respective highest IoU scores for the training and test RL images are 0.94 and 0.72. By contrast, the lowest IoU scores for the training and test RB images are, respectively, 0.77 and 0.10. The IoU scores of RB for the training and test subset images are, respectively, 0.10 and 0.35 lower than the mean score. We consider that the low IoU score for RB is attributable to its lower frequency of occurrence than other classes.
Figure 18 and Figure 19 respectively present representative successful and failed test subset images. The mIoU threshold to classify these cases was set as 0.10. As the depicted images in Figure 18, RB pixels were roughly segmented from the RL regions. This result indicates the relation that RB pixels are apparent in RL regions. By contrast, as depicted images in Figure 19, false segmentation pixels exist in small RB regions with low segmentation accuracy. False segmentation pixels are apparent in dead leaf tips and RB pixels that are segmented as BG pixels. This finding from experimentation demonstrates the difficulty of learning object boundaries between RB and RL pixels.

6.2.3. Soybeans

Table 7 represents a confusion matrix for the Soybeans test subset images. The numbers of false segmentation pixels are greater between PW and SL, and between PW and BG. However, the false segmentation rate is as high as 3% because the number of PW pixels is approximately 60% in this dataset.
Table 8 presents the semantic segmentation results evaluated using IoU scores for the respective Soybeans classes. The mIoU scores for the training and test subset images are, respectively, 0.59 and 0.56. The gap score between them is only 0.03, which is approximately a tenth of that for Rice blast. The highest IoU scores are for the training and test PW images at 0.78 and 0.83, respectively. By contrast, the lowest IoU scores are for the training and test OW images at 0.12 and 0.32, respectively. The IoU scores for the training and test OW images are, respectively, 0.47 and 0.37 lower than the mean score. We infer that this tendency is related to the OW data amount, which is merely 2% of the total data amount.
Figure 20 and Figure 21 respectively portray representative successful and failed test subset images. Resultant images in Figure 20 depict noise caused by unsteady segmentation at the boundaries between BG and OW pixels. Resultant images in Figure 21 show false segmentation pixels between SL and OW. Although the PW pixels are segmented correctly, several false segmentation pixels exist at the leaf tips.

6.3. Experiment A

Experiment A provides semantic segmentation obtained using augmented images generated from the CUT as an unconditional GAN. Table 9, Table 10 and Table 11 respectively show IoU scores for Cityscapes, Rice blast, and Soybeans. The mIoU scores for Soybeans were 0.49 for the training subset and 0.20 for the test subset. The highest and lowest IoU scores were achieved, respectively, for the road and traffic light classes. The Rice blast mIoU scores were 0.76 for the training subset and 0.25 for the test subset. The highest and lowest IoU scores were achieved, respectively, for the RL and RB classes. The Soybeans mIoU scores were 0.73 for the training subset and 0.55 for the test subset. The highest and lowest IoU scores were achieved, respectively, for the PW and OW classes.

6.4. Experiment B

Experiment B provides semantic segmentation obtained using augmented images generated from the pix2pixHD as a conditional GAN. Table 12, Table 13 and Table 14 respectively present IoU scores for Cityscapes, Rice blast, and Soybeans. The mIoU scores for Cityscapes, Rice blast, and Soybeans were, respectively, 0.51, 0.69, and 0.76 for the training subset and 0.27, 0.42, and 0.59 for the test subset. Trends of the highest and lowest IoU scores were similar to those for Experiment A.

6.5. Experiment C

Experiment C provides semantic segmentation obtained using augmented images generated from CUT and pix2pixHD. Table 15, Table 16 and Table 17 respectively denote IoU scores for Cityscapes, Rice blast, and Soybeans. The mIoU scores for Cityscapes, Rice blast, and Soybeans were, respectively, 0.47, 0.63, and 0.77 for the training subset and 0.25, 0.38, and 0.58 for the test subset. Trends of the highest and lowest IoU scores are equivalent to Experiments A and B.

6.6. Experiment D

Experiment D provides semantic segmentation obtained using the source images and augmented images generated from CUT and pix2pixHD. Table 18, Table 19 and Table 20 respectively present IoU scores for Cityscapes, Rice blast, and Soybeans. The mIoU scores for Cityscapes, Rice blast, and Soybeans were, respectively, 0.50, 0.69, and 0.76 for the training subset and 0.27, 0.42, and 0.59 for the test subset. Trends of the highest and lowest IoU scores are equivalent to Experiments A, B, and C.

6.7. Overall Comparison and Discussion

6.7.1. Cityscapes

Table 21 presents the result of a comparison of mIoU scores in each Cityscapes experiment. Experiment B produced the highest mIoU score. This result demonstrates the effectiveness of using augmented images obtained using the conditional GAN. By contrast, the mIoU score obtained from Experiment C was lower than that of the baseline. We infer that the augmented images obtained using the conditional and unconditional GANs provided no contribution to improvement of the accuracy. Experiment D results demonstrated that, using all images, including the original, improved the mIoU score.
Figure 22 depicts representative images of comparative semantic segmentation results in each Cityscapes experiment. The road, car, vegetation, and building classes were segmented correctly. False segmentation pixels occurred at the sidewalks on the left and right sides of the images. Based on these results, we infer that segmentation between fence and vegetation is a challenging task for this dataset.
Figure 23 depicts results obtained from comparison of the respective experiments for the lower and upper five classes ordered by the IoU scores in Table 4. The lower five classes are traffic light, pole, rider, ground, and traffic sign in the training subset images and are traffic light, train, motorcycle, terrain, and rider in the test subset images. Two classes of traffic light and rider are common in both datasets. Results obtained from Experiment B improve accuracy in the respective classes. Although the test subset showed remarkable improvement in accuracy for the g r o u n d class, it dropped considerably for the rider class. The upper five classes are building, vegetation, car, sky, and road. They are car, unlabeled, building, vegetation, and road for the test subset. Three classes of building, vegetation, and road are common to both. The accuracy for the training subset exhibits no remarkable difference. For the test subset, the argument data combinations in Experiments B and D were slightly predominant. We consider that Experiments A and C contributed no accuracy improvement. However, the improved accuracy results in Experiments B and D verify a tendency that is similar to the overall results, as shown in Table 21.

6.7.2. Rice Blast

Table 22 presents a comparison of the respective results obtained for mIoU scores in experiments for Rice blast. For the training subset, mIoU scores of the four experiments were found to be lower than those of the baseline. For the test subset, mIoU scores for Experiments A, C, and D were found to be lower than the baseline score. The mIoU score for Experiment B was found to be 0.01 higher than the baseline score.
Figure 24 depicts representative images of comparative semantic segmentation results in each Rice blast experiment. Compared to other experiments, the segmentation result in Experiment B is similar to the GT images, especially in RB pixels annotated in red. By contrast, the segmentation result in Experiment A shows a trend toward incorrect segmentation of pixels in the respective classes.
Figure 25 presents compared results for the respective classes. For the training subset, no class achieved accuracy improvement compared to the baseline. For the test subset, IoU scores for RB were improved by 0.05 from 0.10 of the baseline to 0.15 in Experiment B.

6.7.3. Soybeans

Table 23 shows compared results of mIoU scores from the respective Soybeans experiments. For the training subset, the mIoU scores of all four experiments were found to be higher than those of the baseline. For the test subset, the mIoU scores obtained from Experiments A and B were found to be lower than those of the baseline. However, the mIoU scores in Experiments C and D were found to be 0.02 and 0.03 higher than those of the baseline for the test subset, respectively.
Figure 26 depicts representative images of comparative semantic segmentation results for each experiment conducted with Soybeans. The overall segmentation trends are similar among the four experiments. The results of the PW pixels in the bottom center panels of the images are different parts. These pixels were segmented as SL, except for the Experiment D result. However, the PW pixels were expanded in the bottom right panels of the images.
Figure 27 depicts comparison results in each class. For the training subset, all experiment results showed improved accuracy compared to the baseline. Particularly for OW, although the baseline result was 0.12, Experiments C and D results were 0.63 and 0.62. The BG and PW classes achieved comparable accuracy for the test subset. Although the IoU score for SL is lower than the mean score, the IoU scores for OW in Experiments C and D were 0.49 and 0.48, which were, respectively, 0.17 and 0.16 higher than those of the baseline score.

7. Conclusions

This paper presented a data augmentation framework to improve semantic segmentation accuracy based on style transfer using GANs of two types. In particular, this study specifically addressed data generation of crop images for remote and smart farming to reduce the burden associated with image annotation. The proposed framework comprised pix2pixHD, CUT, and DeepLabV3+ as a conditional GAN, an unconditional GAN, and a semantic segmentation network. We used the Cityscapes public benchmark dataset and two originally collected benchmark datasets termed Rice blast and Soybeans. In addition, we conducted four experiments using the source images and augmented images from the combination of two GANs. The mIoU scores for the public benchmark dataset improved by 0.03 for the training subset, while remaining similar on the test subset. For the first original benchmark dataset, the mIoU scores improved by 0.01 for the test subset, while they dropped by 0.03 for the training subset. The mIoU scores for the second original benchmark dataset improved by 0.18 for the training subset and 0.03 for the test subset. These results demonstrated the effectiveness of our proposed framework using augmented images generated from two GANs in response to dataset combinations.
As future work, we must evaluate the effectiveness of our framework for use with various agricultural benchmark datasets. We will analyze the replacement of GANs and semantic segmentation backbones by considering a tradeoff between the accuracy and computational cost. We will consider the introduction of an automatic parameter optimization method for improving accuracy. Finally, we would like to collaborate in the application of our framework with a weed-removing robot or a harvesting robot to achieve remote and smart farming.

Author Contributions

Conceptualization, S.Y.; methodology, H.M.; software, K.T.; validation, K.T. and T.K.S.; formal analysis, K.S. (Kazuki Saruta); investigation, S.C.; resources, Y.N.; data curation, Y.N.; writing—original draft preparation, H.M.; writing—review and editing, K.S. (Kazuhito Sato); visualization, S.N.; supervision, K.S. (Kazuhito Sato); project administration, Y.N.; funding acquisition, S.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Japan Society for the Promotion of Science (JSPS) KAKENHI Grant Number 21H02321.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets described as a result of this study are available on request to the corresponding author.

Acknowledgments

We would like to show our appreciation to the students at our lab for their great cooperation in the experiments.

Conflicts of Interest

The authors declare that they have no conflict of interest. The funders had no role in the design of the study, in the collection, analyses, or in interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
BGBackground
BEGANBoundary equilibrium generative adversarial network
CAEConvolutional auto encoder
CCLCycle consistency loss
CLContrastive learning
CMTConventional machine-learning
CUTContrastive unpaired translation
DCGANDeep convolutional generative adversarial network
DLDeep learning
EBGANEnergy-based generative adversarial network
ESRGANEnhanced super-resolution generative adversarial network
FIDFréchet Inception distance
GANGenerative adversarial network
GNSSGlobal navigation satellite systems
GTGround truth
I2LImage to label
IoTInternet of things
IoUIntersection over union
L2ILabel to image
OASISOnly adversarial supervision for semantic image segmentation
OWOther weeds
PGGANProgressive growing generative adversarial network
PWPoaceae weeds
RBRice blast
RoIRegion of interest
RLRice leaves
SLSoybean leaves
SPADESpatially adaptive denormalization
SSLSelf-supervised learning
USISUnsupervised paradigm for semantic image synthesis
ViTVision transformer

References

  1. Saiz-Rubio, V.; Rovira-Más, F. From Smart Farming towards Agriculture 5.0: A Review on Crop Data Management. Agronomy 2020, 10, 207. [Google Scholar] [CrossRef] [Green Version]
  2. Farooq, M.S.; Riaz, S.; Abid, A.; Abid, K.; Naeem, M.A. A Survey on the Role of IoT in Agriculture for the Implementation of Smart Farming. IEEE Access 2019, 7, 156237–156271. [Google Scholar] [CrossRef]
  3. Murugan, D.; Garg, A.; Singh, D. Development of an Adaptive Approach for Precision Agriculture Monitoring with Drone and Satellite Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 12, 5322–5328. [Google Scholar] [CrossRef]
  4. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  5. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef] [Green Version]
  6. Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Sallab, A.A.A.; Yogamani, S.; Pérez, P. Deep Reinforcement Learning for Autonomous Driving: A Survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4909–4926. [Google Scholar] [CrossRef]
  7. Sundararajan, K.; Woodard, D.L. Deep Learning for Biometrics: A Survey. ACM Comput. Surv. 2019, 51, 65. [Google Scholar] [CrossRef]
  8. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Laak, J.A.W.M.; Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  9. Sharma, A.; Jain, A.; Gupta, P.; Chowdary, V. Machine Learning Applications for Precision Agriculture: A Comprehensive Review. IEEE Access 2021, 9, 4843–4873. [Google Scholar] [CrossRef]
  10. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine Learning in Agriculture: A Review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Benos, L.; Tagarakis, A.C.; Dolias, G.; Berruto, R.; Kateris, D.; Bochtis, D. Machine Learning in Agriculture: A Comprehensive Updated Review. Sensors 2021, 21, 3758. [Google Scholar] [CrossRef] [PubMed]
  12. Balducci, F.; Impedovo, D.; Pirlo, G. Machine Learning Applications on Agricultural Datasets for Smart Farm Enhancement. Machines 2018, 6, 38. [Google Scholar] [CrossRef] [Green Version]
  13. Wolfert, S.; Ge, L.; Verdouw, C.; Bogaardt, M.-J. Big Data in Smart Farming: A review. Agric. Syst. 2017, 153, 69–80. [Google Scholar] [CrossRef]
  14. Lu, X.; Wang, W.; Shen, J.; Tai, Y.-W.; Crandall, D.J.; Hoi, S.C.H. Learning Video Object Segmentation From Unlabeled Videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 8960–8970. [Google Scholar]
  15. Du, X.; Jiang, C.; Xu, H.; Zhang, G.; Li, Z. How to Save your Annotation Cost for Panoptic Segmentation? In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 1282–1290. [Google Scholar]
  16. Unal, Z. Smart Farming Becomes Even Smart with Deep Learning—A Bibliographical Analysis. IEEE Access 2020, 8, 105587–105609. [Google Scholar] [CrossRef]
  17. Thoma, M. A survey of semantic segmentation. arXiv 2016, arXiv:1602.06541. [Google Scholar]
  18. Milioto, A.; Lottes, P.; Stachniss, C. Real-Time Semantic Segmentation of Crop and Weed for Precision Agriculture Robots Leveraging Background Knowledge in CNNs. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 2229–2235. [Google Scholar]
  19. Kirillov, A.; He, K.; Girshick, R.; Rother, C.; Dollar, P. Panoptic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 9404–9413. [Google Scholar]
  20. Champ, J.; Mora-Fallas, A.; Goëau, H.; Mata-Montero, E.; Bonnet, P.; Joly, A. Instance segmentation for the fine detection of crop and weed plants by precision agricultural robots. Appl. Plant Sci. 2020, 8, e11373. [Google Scholar] [CrossRef] [PubMed]
  21. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. MaskR-CNN. In Proceedings of the IEEE International Conference on Computer Vision, (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  22. Zheng, Y.-Y.; Kong, J.-L.; Jin, X.-B.; Wang, X.-Y.; Su, T.-L.; Zuo, M. CropDeep: The Crop Vision Dataset for Deep-Learning-Based Classification and Detection in Precision Agriculture. Sensors 2019, 19, 1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Singh, A.; Jaiswal, V.; Joshi, G.; Sanjeeve, A.; Gite, S.; Kotecha, K. Neural Style Transfer: A Critical Review. IEEE Access 2021, 9, 131583–131613. [Google Scholar] [CrossRef]
  24. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image Style Transfer Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2414–2423. [Google Scholar]
  25. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  26. Rozantsev, A.; Lepetit, V.; Fua, P. On rendering synthetic images for training an object detector. arXiv 2014, arXiv:1411.7911. [Google Scholar] [CrossRef] [Green Version]
  27. Rematas, K.; Ritschel, T.; Fritz, M.; Tuytelaars, T. Image-based Synthesis and Re-Synthesis of Viewpoints Guided by 3D Models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 24–27 June 2014; pp. 3898–3905. [Google Scholar]
  28. Pishchulin, L.; Jain, A.; Andriluka, M. Articulated People Detection and Pose Estimation: Reshaping the Future. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 3178–3185. [Google Scholar]
  29. Vázquez, D.; López, A.M.; Marin, J.; Ponsa, D.; Gerónimo, D. Virtual and real world adaptation for pedestrian detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 4, 797–809. [Google Scholar] [CrossRef]
  30. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3234–3243. [Google Scholar]
  31. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  32. Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition in videos. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  33. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  35. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  36. Huang, G.; Liu, Z.; Maaten, L.; Weinberger, K. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  37. Howard, A.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  38. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  39. Radosavovic, I.; Kosaraju, R.P.; Girshick, R.; He, K.; Dollár, P. Designing network design spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 5549–5558. [Google Scholar]
  40. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  41. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Virtual, 6–14 December 2021. [Google Scholar]
  42. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. In Proceedings of the International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  43. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, Z.P.H.S.; et al. Rethinking Semantic Segmentation From a Sequence-to-Sequence Perspective With Transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 6881–6890. [Google Scholar]
  44. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  45. Tran, N.-T.; Tran, V.-H.; Nguyen, N.-B.; Nguyen, T.-K.; Cheung, N.-M. On Data Augmentation for GAN Training. IEEE Trans. Image Process. 2021, 30, 1882–1897. [Google Scholar] [CrossRef]
  46. Tanaka, F.H.K.S.; Aranha, C. Data Augmentation Using GANs. arXiv 2019, arXiv:1904.09135. [Google Scholar]
  47. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  48. Huang, S.-W.; Lin, C.-T.; Chen, S.-P.; Wu, Y.-Y.; Hsu, P.-H.; Lai, S.-H. AugGAN: Cross Domain Adaptation with GAN-based Data Augmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 718–731. [Google Scholar]
  49. Choi, J.; Kim, T.; Kim, C. Self-Ensembling with GAN-Based Data Augmentation for Domain Adaptation in Semantic Segmentation. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2017; pp. 6830–6840. [Google Scholar]
  50. Vilkas, E. Axiomatic definition of the value of a matrix game. Theory Probab. Appl. 1963, 8, 304–307. [Google Scholar] [CrossRef]
  51. Radford, A.; Metz, L. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  52. Zhao, J.; Mathieu, M.; LeCun, Y. Energy-based Generative Adversarial Network. arXiv 2017, arXiv:1609.03126. [Google Scholar]
  53. Berthelot, D.; Schumm, T.; Metz, L. BEGAN: Boundary Equilibrium Generative Adversarial Networks. arXiv 2017, arXiv:1703.10717. [Google Scholar]
  54. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive Growing of GAN for Improved Quality, Stability, and Variation. arXiv 2018, arXiv:1710.10196. [Google Scholar]
  56. Huang, X.; Belongie, S.J. Arbitrary style transfer in real-time with adaptive instance normalization. arXiv 2017, arXiv:1703.06868. [Google Scholar]
  57. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4401–4410. [Google Scholar]
  58. Bermano, A.H.; Gal, R.; Alaluf, Y.; Mokady, R.; Nitzan, Y.; Tov, O.; Patashnik, O.; Cohen-Or, D. State-of-the-Art in the Architecture, Methods and Applications of StyleGAN. arXiv 2022, arXiv:2202.14020. [Google Scholar] [CrossRef]
  59. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Loy, C.C.; Qiao, Y.; Tang, X. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  60. Zhang, H.; Xu, T.; Li, H.; Zhang, S.; Wang, X.; Huang, X.; Metaxas, D. StackGAN: Text to Photo-realistic Image Synthesis with Stacked Generative Adversarial Networks. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5907–5915. [Google Scholar]
  61. Jing, Y.; Yang, Y.; Feng, Z.; Ye, J.; Yu, Y.; Song, M. Neural Style Transfer: A Review. IEEE Trans. Vis. Comput. Graph. 2020, 26, 3365–3385. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Zhao, W.; Yamada, W.; Li, T.; Digman, M.; Runge, T. Augmenting Crop Detection for Precision Agriculture with Deep Visual Transfer Learning—Case Study of Bale Detection. Remote Sens. 2021, 13, 23. [Google Scholar] [CrossRef]
  63. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  64. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  65. Wang, T.C.; Liu, M.Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-Resolution Image Synthesis and Semantic Manipulation with Conditional GAN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 19–21 June 2018; pp. 8798–8807. [Google Scholar]
  66. Lee, C.H.; Liu, Z.; Wu, L.; Luo, P. Maskgan: Towards diverse and interactive facial image manipulation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 5549–5558. [Google Scholar]
  67. Park, T.; Liu, M.Y.; Wang, T.C.; Zhu, J.Y. Semantic image synthesis with spatially adaptive normalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4401–4410. [Google Scholar]
  68. Sushko, V.; Schonfeld, E.; Zhang, D.; Gall, J.; Schiele, B.; Khoreva, A. You Only Need Adversarial Supervision for Semantic Image Synthesis. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021. [Google Scholar]
  69. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  70. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  71. Park, T.; Efros, A.A.; Zhang, R.; Zhu, J.Y. Contrastive Learning for Unpaired Image-to-Image Translation. In Proceedings of the European Conference on Computer Vision (ECCV), Virtual, 23–28 August 2020; pp. 319–345. [Google Scholar]
  72. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  73. Eskandar, G.; Abdelsamad, M.; Armanious, K.; Yang, B. USIS: Unsupervised Semantic Image Synthesis. arXiv 2021, arXiv:2109.14715. [Google Scholar]
  74. Jaiswal, A.; Babu, A.R.; Zadeh, M.Z.; Banerjee, D.; Makedon, F.A. Survey on Contrastive Self-Supervised Learning. Technologies 2021, 9, 2. [Google Scholar] [CrossRef]
  75. Eskandar, G.; Abdelsamad, M.; Armanious, K.; Zhang, S.; Yang, B. Wavelet-Based Unsupervised Label-to-Image Translation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 1760–1764. [Google Scholar]
  76. Cao, Y.-J.; Jia, L.-L.; Chen, Y.-X.; Lin, N.; Yang, C.; Zhang, B.; Liu, Z.; Li, X.-X.; Dai, H.-H. Recent Advances of Generative Adversarial Networks in Computer Vision. IEEE Access 2018, 7, 14985–15006. [Google Scholar] [CrossRef]
  77. Pan, Z.; Yu, W.; Yi, X.; Khan, A.; Yuan, F.; Zheng, Y. Recent Progress on Generative Adversarial Networks (GANs): A Survey. IEEE Access 2019, 7, 36322–36333. [Google Scholar] [CrossRef]
  78. Hajarolasvadi, N.; Ramírez, M.A.; Beccaro, W.; Demirel, H. Generative Adversarial Networks in Human Emotion Synthesis: A Review. IEEE Access 2020, 8, 218499–218529. [Google Scholar] [CrossRef]
  79. Dekker, J. Weed diversity and weed management. Weed Sci. 1997, 45, 357–363. [Google Scholar] [CrossRef]
  80. Madokoro, H.; Yamamoto, S.; Nishimura, Y.; Nix, S.; Woo, H.; Sato, K. Prototype Development of Small Mobile Robots for Mallard Navigation in Paddy Fields: Toward Realizing Remote Farming. Robotics 2021, 10, 63. [Google Scholar] [CrossRef]
  81. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef]
  82. Liu, X.; Deng, Z.; Yang, Y. Recent progress in semantic image segmentation. Artif. Intell. Rev. 2019, 52, 1089–1106. [Google Scholar] [CrossRef] [Green Version]
  83. Yu, H.; Yang, Z.; Tan, L.; Wang, Y.; Sun, W.; Sun, M.; Tang, Y. Methods and datasets on semantic segmentation: A review. Neurocomputing 2018, 304, 82–103. [Google Scholar] [CrossRef]
  84. Fahim, M.A.-N.I.; Jung, H.Y. A Lightweight GAN Network for Large Scale Fingerprint Generation. IEEE Access 2020, 8, 92918–92928. [Google Scholar] [CrossRef]
  85. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the 36th International Conference on Machine Learning (ICML), Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  86. Pang, T.; Xu, K.; Dong, Y.; Du, C.; Chen, N.; Zhu, J. Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness. arXiv 2020, arXiv:1905.10626. [Google Scholar]
  87. Krause, E.F. Taxicab Geometry. Math. Teach. 1973, 66, 8, 695–706. [Google Scholar] [CrossRef]
  88. Toldo, M.; Maracani, A.; Michieli, U.; Zanuttigh, P. Unsupervised Domain Adaptation in Semantic Segmentation: A Review. Technologies 2020, 8, 35. [Google Scholar] [CrossRef]
  89. Zhang, S.; Benenson, R.; Schiele, B. CityPersons: A Diverse Dataset for Pedestrian Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3213–3221. [Google Scholar]
  90. Huang, M.L.; Wu, Y.Z. Semantic segmentation of pancreatic medical images by using convolutional neural network. Biomed. Signal Process. Control 2022, 73, 1746–8094. [Google Scholar] [CrossRef]
  91. Ivanovs, M.; Ozols, K.; Dobrajs, A.; Kadikis, R. Improving Semantic Segmentation of Urban Scenes for Self-Driving Cars with Synthetic Images. Sensors 2022, 22, 2252. [Google Scholar] [CrossRef] [PubMed]
  92. Kerle, N.; Nex, F.; Gerke, M.; Duarte, D.; Vetrivel, A. UAV-Based Structural Damage Mapping: A Review. ISPRS Int. J. Geo-Inf. 2020, 9, 14. [Google Scholar] [CrossRef] [Green Version]
  93. Yu, R.; Li, H.; Jiang, Y.; Zhang, B.; Wang, Y. Tiny Vehicle Detection for Mid-to-High Altitude UAV Images Based on Visual Attention and Spatial–Temporal Information. Sensors 2022, 22, 2354. [Google Scholar] [CrossRef]
  94. Xu, Z.; Zhang, W.; Zhang, T.; Li, J. HRCNet: High-Resolution Context Extraction Network for Semantic Segmentation of Remote Sensing Images. Remote Sens. 2021, 13, 71. [Google Scholar] [CrossRef]
  95. Costa, M.V.C.V.d.; Carvalho, O.L.F.d.; Orlandi, A.G.; Hirata, I.; Albuquerque, A.O.d.; Silva, F.V.e.; Guimarães, R.F.; Gomes, R.A.T.; Júnior, O.A.d.C. Remote Sensing for Monitoring Photovoltaic Solar Plants in Brazil Using Deep Semantic Segmentation. Energies 2021, 14, 2960. [Google Scholar] [CrossRef]
  96. Yang, L.; Fan, J.; Liu, Y.; Li, E.; Peng, J.; Liang, Z. A Review on State-of-the-Art Power Line Inspection Techniques. IEEE Trans. Instrum. Meas. 2020, 69, 9350–9365. [Google Scholar] [CrossRef]
  97. Shin, Y.-H.; Son, K.-W.; Lee, D.-C. Semantic Segmentation and Building Extraction from Airborne LiDAR Data with Multiple Return Using PointNet++. Appl. Sci. 2022, 12, 1975. [Google Scholar] [CrossRef]
  98. Craye, C.; Ardjoune, S. Spatiotemporal Semantic Segmentation for Drone Detection. In Proceedings of the 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Taipei, Taiwan, 18–21 September 2019; pp. 1–5. [Google Scholar]
  99. Ramirez-Amaro, K.; Minhas, H.N.; Zehetleitner, M.; Beetz, M.; Cheng, G. Added value of gaze-exploiting semantic representation to allow robots inferring human behaviors. ACM Trans. Interact. Intell. Syst. 2017, 7, 1–30. [Google Scholar] [CrossRef]
  100. Lundgren, A.V.A.; Santos, M.A.O.d.; Bezerra, B.L.D.; Bastos-Filho, C.J.A. Systematic Review of Computer Vision Semantic Analysis in Socially Assistive Robotics. AI 2022, 3, 229–249. [Google Scholar] [CrossRef]
  101. Xu, H.; Chen, G.; Wang, Z.; Sun, L.; Su, F. RGB-D-Based Pose Estimation of Workpieces with Semantic Segmentation and Point Cloud Registration. Sensors 2019, 19, 1873. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  102. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  103. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder–Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  104. Sun, K.; Zhao, Y.; Jiang, B.; Cheng, T.; Xiao, B.; Liu, D.; Mu, Y.; Wang, X.; Liu, W.; Wang, J. High-Resolution Representations for Labeling Pixels and Regions. arXiv 2019, arXiv:1904.04514. [Google Scholar]
  105. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 11976–11986. [Google Scholar]
  106. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  107. Tuli, S.; Dasgupta, I.; Grant, E.; Griffiths, T.L. Are Convolutional Neural Networks or Transformers more like human vision? arXiv 2021, arXiv:2105.07197. [Google Scholar]
  108. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do Vision Transformers See Like Convolutional Neural Networks? In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Virtual, 6–14 December 2021. [Google Scholar]
  109. Steiner, A.; Kolesnikov, A.; Zhai, X.; Wightman, R.; Uszkoreit, J.; Beyer, L. How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers. arXiv 2021, arXiv:2106.10270. [Google Scholar]
  110. Ridnik, T.; Ben-Baruch, E.; Noy, A.; Lihi Zelnik-Manor, L. ImageNet-21K Pretraining for the Masses. arXiv 2021, arXiv:2104.10972. [Google Scholar]
  111. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  112. Kolesnikov, A.; Beyer, L.; Zhai, X.; Puigcerver, J.; Yung, J.; Gelly, S.; Houlsby, N. Big Transfer (BiT): General Visual Representation Learning. In Proceedings of the European Conference on Computer Vision (ECCV), Virtual, 23–28 August 2020; pp. 491–507. [Google Scholar]
  113. Caesar, H.; Uijlings, J.; Ferrari, V. COCO-Stuff: Thing and stuff Classes in Context. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 19–21 June 2018; pp. 1209–1218. [Google Scholar]
  114. Sato, K.H.; Madokoro, H.; Nagayoshi, T.; Chiyonobu, S.; Martizzi, P.; Nix, S.; Woo, H.; Saito, T.K.; Sato, K. Semantic Segmentation of Outcrop Images using Deep Learning Networks Toward Realization of Carbon Capture and Storage. In Proceedings of the 17th International Conference on Control, Automation and Systems (ICCAS), Jeju, Korea, 12–15 October 2021; pp. 436–441. [Google Scholar]
  115. Takahashi, K.H.; Madokoro, H.; Yamamoto, S.; Nishimura, Y.; Nix, S.; Woo, H.; Saito, T.K.; Sato, K. Domain Adaptation for Agricultural Image Recognition and Segmentation Using Category Maps. In Proceedings of the 17th International Conference on Control, Automation and Systems (ICCAS), Jeju, Korea, 12–15 October 2021; pp. 1680–1685. [Google Scholar]
  116. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  117. He, J.; Chen, J.-N.; Liu, S.; Kortylewski, A.; Yang, C.; Bai, Y.; Wang, C. TransFG: A Transformer Architecture for Fine-grained Recognition. arXiv 2021, arXiv:2103.07976. [Google Scholar] [CrossRef]
  118. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. MetaFormer is Actually What You Need for Vision. arXiv 2021, arXiv:2111.11418. [Google Scholar]
  119. Zhang, W.; Pang, J.; Chen, K.; Loy, C.C. K-Net: Towards Unified Image Segmentation. arXiv 2021, arXiv:2106.14855. [Google Scholar]
  120. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3213–3223. [Google Scholar]
  121. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context Encoding for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 19–21 June 2018; pp. 7151–7160. [Google Scholar]
  122. Sulistiyo, M.D.; Kawanishi, Y.; Deguchi, D.; Ide, I.; Hirayama, T.; Murase, H. CityWalks: An Extended Dataset for Attribute-aware Semantic Segmentation. In Proceedings of the Tokai-Section Joint Conference on Electrical, Electronics, Information, and Related Engineering, Nagoya, Japan, 9–10 September 2019. [Google Scholar]
  123. Gählert, N.; Jourdan, N.; Cordts, M.; Franke, U.; Denzler, J. Cityscapes 3D: Dataset and Benchmark for 9 DoF Vehicle Detection. arXiv 2020, arXiv:2006.07864. [Google Scholar]
  124. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GAN Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 6626–6637. [Google Scholar]
  125. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; FeiFei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami Beach, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  126. Jaccard, P. The distribution of the flora in the alpine zone. New Phytol. 1912, 11, 37–50. [Google Scholar] [CrossRef]
  127. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 2, 303–338. [Google Scholar] [CrossRef] [Green Version]
  128. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  129. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 19–21 June 2018; pp. 4510–4520. [Google Scholar]
  130. Hussein, B.R.; Malik, O.A.; Ong, W.; Slik, J.W.F. Semantic Segmentation of Herbarium Specimens Using Deep Learning Techniques. Comput. Sci. Technol. Lect. Notes Electr. Eng. 2020, 603, 321–330. [Google Scholar]
  131. Yu, T.; Zhu, H. Hyper-Parameter Optimization: A Review of Algorithms and Applications. arXiv 2020, arXiv:2003.05689. [Google Scholar]
  132. He, X.; Zhao, K.; Chu, X. AutoML: A survey of the state-of-the-art. Knowl.-Based Syst. 2021, 212, 106622. [Google Scholar] [CrossRef]
  133. Wever, M.; Tornede, A.; Mohr, F.; Hullermeier, E. AutoML for Multi-Label Classification: Overview and Empirical Evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3037–3054. [Google Scholar] [CrossRef]
  134. Truong, A.; Walters, A.; Goodsitt, J.; Hines, K.; Bruss, C.B.; Farivar, R. Towards Automated Machine Learning: Evaluation and Comparison of AutoML Approaches and Tools. In Proceedings of the IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI), Portland, OR, USA, 4–6 November 2019; pp. 1471–1479. [Google Scholar]
  135. Carneiro, T.; Medeiros Da NóBrega, R.V.; Nepomuceno, T.T.; Bian, G.-B.; De Albuquerque, V.H.C.; Filho, P.P.R. Performance Analysis of Google Colaboratory as a Tool for Accelerating Deep Learning Applications. IEEE Access 2018, 6, 61677–61685. [Google Scholar] [CrossRef]
  136. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of StyleGAN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 14–19 June 2020; pp. 4401–4410. [Google Scholar]
  137. Karras, T.; Aittala, M.; Laine, S.; Härkönen, E.; Hellsten, J.; Lehtinen, J.; Aila, T. Alias-Free Generative Adversarial Networks. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Virtual, 6–14 December 2021. [Google Scholar]
Figure 1. Overall network architecture of the proposed framework consisting of three modules.
Figure 1. Overall network architecture of the proposed framework consisting of three modules.
Applsci 12 07785 g001
Figure 2. Structural differences between conditional GAN and unconditional GAN.
Figure 2. Structural differences between conditional GAN and unconditional GAN.
Applsci 12 07785 g002
Figure 3. Sample images of Cityscapes: source images (upper) and GT images (lower).
Figure 3. Sample images of Cityscapes: source images (upper) and GT images (lower).
Applsci 12 07785 g003
Figure 4. Sample images of Rice blast: source images (upper) and GT images (lower).
Figure 4. Sample images of Rice blast: source images (upper) and GT images (lower).
Applsci 12 07785 g004
Figure 5. Sample images of Soybeans: source images (upper) and GT images (lower).
Figure 5. Sample images of Soybeans: source images (upper) and GT images (lower).
Applsci 12 07785 g005
Figure 6. Time-series training loss trends for the respective GANs in Cityscapes.
Figure 6. Time-series training loss trends for the respective GANs in Cityscapes.
Applsci 12 07785 g006
Figure 7. Time-series FID trends for the respective GANs in Cityscapes.
Figure 7. Time-series FID trends for the respective GANs in Cityscapes.
Applsci 12 07785 g007
Figure 8. I2L and L2I style transfer results for the respective GANs in Cityscapes.
Figure 8. I2L and L2I style transfer results for the respective GANs in Cityscapes.
Applsci 12 07785 g008
Figure 9. Time-series training loss trends for respective GANs in Rice blast.
Figure 9. Time-series training loss trends for respective GANs in Rice blast.
Applsci 12 07785 g009
Figure 10. Time-series FID trends for the respective GANs in Rice blast.
Figure 10. Time-series FID trends for the respective GANs in Rice blast.
Applsci 12 07785 g010
Figure 11. I2L and L2I style transfer results for the respective GANs in Rice blast.
Figure 11. I2L and L2I style transfer results for the respective GANs in Rice blast.
Applsci 12 07785 g011
Figure 12. Time-series training loss trends for the respective GANs in Soybeans.
Figure 12. Time-series training loss trends for the respective GANs in Soybeans.
Applsci 12 07785 g012
Figure 13. Time-series FID trends for the respective GANs in Soybeans.
Figure 13. Time-series FID trends for the respective GANs in Soybeans.
Applsci 12 07785 g013
Figure 14. I2L and L2I style transfer results for the respective GANs in Soybeans.
Figure 14. I2L and L2I style transfer results for the respective GANs in Soybeans.
Applsci 12 07785 g014
Figure 15. Confusion matrix for Cityscapes test subset images.
Figure 15. Confusion matrix for Cityscapes test subset images.
Applsci 12 07785 g015
Figure 16. Representative successful Cityscapes images: source images (left), GT (center), and results (right).
Figure 16. Representative successful Cityscapes images: source images (left), GT (center), and results (right).
Applsci 12 07785 g016
Figure 17. Representative failed Cityscapes images: source images (left), GT (center), and results (right).
Figure 17. Representative failed Cityscapes images: source images (left), GT (center), and results (right).
Applsci 12 07785 g017
Figure 18. Representative successful Rice blast images: source images (left), GT (center), and results (right).
Figure 18. Representative successful Rice blast images: source images (left), GT (center), and results (right).
Applsci 12 07785 g018
Figure 19. Representative failed Rice blast images: source images (left), GT (center), and results (right).
Figure 19. Representative failed Rice blast images: source images (left), GT (center), and results (right).
Applsci 12 07785 g019
Figure 20. Representative successful Soybeans images: source images (left), GT (center), and results (right).
Figure 20. Representative successful Soybeans images: source images (left), GT (center), and results (right).
Applsci 12 07785 g020
Figure 21. Representative failed Soybeans images: source images (left), GT (center), and results (right).
Figure 21. Representative failed Soybeans images: source images (left), GT (center), and results (right).
Applsci 12 07785 g021
Figure 22. Comparison results of semantic segmentation in each Cityscapes experiment.
Figure 22. Comparison results of semantic segmentation in each Cityscapes experiment.
Applsci 12 07785 g022
Figure 23. Comparison of IoU scores between the respective classes: upper five class for training subset (upper left), upper five class for test subset (upper right), lower five class for training subset (bottom left), and lower five class for test subset (bottom right).
Figure 23. Comparison of IoU scores between the respective classes: upper five class for training subset (upper left), upper five class for test subset (upper right), lower five class for training subset (bottom left), and lower five class for test subset (bottom right).
Applsci 12 07785 g023
Figure 24. Compared results of semantic segmentation for Rice blast experiments.
Figure 24. Compared results of semantic segmentation for Rice blast experiments.
Applsci 12 07785 g024
Figure 25. Compared results of IoU scores in the respective classes: Training dataset (left) and Test dataset (right).
Figure 25. Compared results of IoU scores in the respective classes: Training dataset (left) and Test dataset (right).
Applsci 12 07785 g025
Figure 26. Results obtained from comparing semantic segmentation in the respective experiments with Soybeans.
Figure 26. Results obtained from comparing semantic segmentation in the respective experiments with Soybeans.
Applsci 12 07785 g026
Figure 27. Comparison results of IoU scores in each class: training dataset (left) and test dataset (right).
Figure 27. Comparison results of IoU scores in each class: training dataset (left) and test dataset (right).
Applsci 12 07785 g027
Table 1. Semantic labels and corresponding colors.
Table 1. Semantic labels and corresponding colors.
IDClassColorIDClassColorIDClassColor
0unlabeled 1ground 2road
3sidewalk 4building 5wall
6fence 7pole 8traffic light
9traffic sign 10vegetation 11terrain
12sky 13person 14rider
15car 16truck 17bus
18train 19motorcycle 20bicycle
Table 2. Data combination for each experiment.
Table 2. Data combination for each experiment.
Data SourceBaselineExp. AExp. BExp. CExp. D
Source images
CUT [71]
pix2pixHD [65]
Table 3. Hyperparameters and setting values.
Table 3. Hyperparameters and setting values.
ModelCUT [71]pix2pixHD [65]DeepLabV3+ [103]
BackboneResNet-50 [34]VGG-19 [69]Xeception-65 [116]
Training iterations [epoch]200300
Training coefficient0.00020.0010
Mini-batch size42
Sampling size [pixel] 256 × 256 480 × 360
Number of sampling25002000
Table 4. IoU scores for the respective Cityscapes classes.
Table 4. IoU scores for the respective Cityscapes classes.
ID012345678910
Training0.660.200.930.660.800.290.290.130.010.250.82
Test0.690.030.880.480.700.040.040.090.000.110.74
ID11121314151617181920Mean
Training0.510.870.500.170.840.340.480.710.270.400.48
Test0.010.640.160.020.660.030.060.010.010.160.27
Table 5. Confusion matrix for Rice blast test subset images [pixels].
Table 5. Confusion matrix for Rice blast test subset images [pixels].
BGRLRB
BG43,544,28832,753,68114,875
(57.06%)(42.92%)(0.02%)
RL2,926,68389,268,83238,594
(3.17%)(96.79%)(0.04%)
RB26,151127,39228,898
(14.33%)(69.83%)(15.84%)
Table 6. IoU scores for the respective Rice blast classes.
Table 6. IoU scores for the respective Rice blast classes.
BGRLRBMean
Training0.910.940.770.87
Test0.520.720.100.45
Table 7. Confusion matrix for Soybeans test subset images [pixels].
Table 7. Confusion matrix for Soybeans test subset images [pixels].
BGSLPWOW
BG383,495905223,4633159
(91.49%)(2.16%)(5.60%)(0.75%)
SL9767162,27739,9037350
(4.45%)(74.00%)(18.20%)(3.35%)
PW50,33337,1941,069,90417,227
(4.28%)(3.17%)(91.08%)(1.47%)
OW11096206965211,482
(3.90%)(21.81%)(33.93%)(40.36%)
Table 8. IoU scores for the respective Soybeans classes.
Table 8. IoU scores for the respective Soybeans classes.
BGSLPWOWMean
Training0.700.750.780.120.59
Test0.520.560.830.320.56
Table 9. IoU scores for Cityscapes in Experiment A.
Table 9. IoU scores for Cityscapes in Experiment A.
ID012345678910
Training0.690.650.950.830.780.500.490.100.0000.140.81
Test0.540.010.820.330.530.010.010.050.0020.030.62
ID11121314151617181920Mean
Training0.590.850.420.010.830.420.400.490.020.230.49
Test0.080.500.060.010.530.010.020.010.010.020.20
Table 10. IoU scores for Rice blast in Experiment A.
Table 10. IoU scores for Rice blast in Experiment A.
BGRLRBMean
Training0.830.860.590.76
Test0.260.490.040.25
Table 11. IoU scores for Soybeans in Experiment A.
Table 11. IoU scores for Soybeans in Experiment A.
BGSLPWOWMean
Training0.770.810.800.550.73
Test0.520.530.810.320.55
Table 12. IoU scores for Cityscapes in Experiment B.
Table 12. IoU scores for Cityscapes in Experiment B.
ID012345678910
Training0.690.270.940.680.810.360.340.130.030.260.83
Test0.680.030.880.480.690.050.040.090.010.130.73
ID11121314151617181920Mean
Training0.570.880.500.190.850.400.610.700.350.430.51
Test0.100.630.170.030.650.030.070.010.010.150.27
Table 13. IoU scores for Rice blast in Experiment B.
Table 13. IoU scores for Rice blast in Experiment B.
BGRLRBMean
Training0.880.930.720.84
Test0.520.720.150.46
Table 14. IoU scores for Soybeans in Experiment B.
Table 14. IoU scores for Soybeans in Experiment B.
BGSLPWOWMean
Training0.780.800.800.320.68
Test0.530.530.810.330.55
Table 15. IoU scores for Cityscapes in Experiment C.
Table 15. IoU scores for Cityscapes in Experiment C.
ID012345678910
Training0.690.430.940.720.810.320.350.110.010.210.82
Test0.660.020.880.450.680.020.040.070.010.080.71
ID11121314151617181920Mean
Training0.570.860.450.090.830.410.360.520.140.330.47
Test0.090.620.140.020.630.020.060.010.010.130.25
Table 16. IoU scores for Rice blast in Experiment C.
Table 16. IoU scores for Rice blast in Experiment C.
BGRLRBMean
Training0.620.780.470.63
Test0.420.680.050.38
Table 17. IoU scores for Soybeans in Experiment C.
Table 17. IoU scores for Soybeans in Experiment C.
BGSLPWOWMean
Training0.830.790.810.630.77
Test0.520.530.800.490.58
Table 18. IoU scores for Cityscapes in Experiment D.
Table 18. IoU scores for Cityscapes in Experiment D.
ID012345678910
Training0.690.370.940.710.810.310.330.120.010.220.82
Test0.690.040.880.470.700.030.050.080.010.100.73
ID11121314151617181920Mean
Training0.580.870.460.200.830.510.540.580.240.360.50
Test0.100.630.170.030.650.020.070.010.010.140.27
Table 19. IoU scores for Rice blast in Experiment D.
Table 19. IoU scores for Rice blast in Experiment D.
BGRLRBMean
Training0.730.820.520.69
Test0.480.700.090.42
Table 20. IoU scores for Soybeans in Experiment D.
Table 20. IoU scores for Soybeans in Experiment D.
BGSLPWOWMean
Training0.820.790.810.620.76
Test0.530.520.820.480.59
Table 21. Comparison results of mIoU scores for the respective Cityscapes experiments (Bold values show maximum scores).
Table 21. Comparison results of mIoU scores for the respective Cityscapes experiments (Bold values show maximum scores).
ModeBaselineABCD
Training0.480.490.510.470.50
Test0.270.200.270.250.27
Table 22. Compared results of mIoU scores from the Rice blast experiments (Bold values show maximum scores).
Table 22. Compared results of mIoU scores from the Rice blast experiments (Bold values show maximum scores).
ModeBaselineABCD
Training0.870.760.840.630.69
Test0.450.250.460.380.42
Table 23. Compared results of mIoU scores in Soybeans experiments (Bold values show maximum scores).
Table 23. Compared results of mIoU scores in Soybeans experiments (Bold values show maximum scores).
ModeBaselineABCD
Training0.590.730.680.770.76
Test0.560.550.550.580.59
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Madokoro, H.; Takahashi, K.; Yamamoto, S.; Nix, S.; Chiyonobu, S.; Saruta, K.; Saito, T.K.; Nishimura, Y.; Sato, K. Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks. Appl. Sci. 2022, 12, 7785. https://0-doi-org.brum.beds.ac.uk/10.3390/app12157785

AMA Style

Madokoro H, Takahashi K, Yamamoto S, Nix S, Chiyonobu S, Saruta K, Saito TK, Nishimura Y, Sato K. Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks. Applied Sciences. 2022; 12(15):7785. https://0-doi-org.brum.beds.ac.uk/10.3390/app12157785

Chicago/Turabian Style

Madokoro, Hirokazu, Kota Takahashi, Satoshi Yamamoto, Stephanie Nix, Shun Chiyonobu, Kazuki Saruta, Takashi K. Saito, Yo Nishimura, and Kazuhito Sato. 2022. "Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks" Applied Sciences 12, no. 15: 7785. https://0-doi-org.brum.beds.ac.uk/10.3390/app12157785

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop