Next Article in Journal
Creating a Lowland and Peatland Landscape Digital Terrain Model (DTM) from Interpolated Partial Coverage LiDAR Data for Central Kalimantan and East Sumatra, Indonesia
Previous Article in Journal
Fitted PROSAIL Parameterization of Leaf Inclinations, Water Content and Brown Pigment Content for Winter Wheat and Maize Canopies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Effective Classification Scheme for Hyperspectral Image Based on Superpixel and Discontinuity Preserving Relaxation

College of Urban and Environment, Liaoning Normal University, Dalian 116029, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(10), 1149; https://0-doi-org.brum.beds.ac.uk/10.3390/rs11101149
Submission received: 9 April 2019 / Revised: 6 May 2019 / Accepted: 6 May 2019 / Published: 14 May 2019
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Hyperspectral image (HSI) classification is one of the most active topics in remote sensing. However, it is still a nontrivial task to classify the hyperspectral data accurately, since HSI always suffers from a large number of noise pixels, the complexity of the spatial structure of objects and the spectral similarity between different objects. In this study, an effective classification scheme for hyperspectral image based on superpixel and discontinuity preserving relaxation (DPR) is proposed to discriminate land covers of interest. A novel technique for measuring the similarity of a pair of pixels in HSI is suggested to improve the simple linear iterative clustering (SLIC) algorithm. Unlike the existing application of SLIC technique to HSI, the improved SLIC algorithm can be directly used to segment HSI into superpixels without using principal component analysis in advance, and is free of parameters. Furthermore, the proposed three-step classification scheme explores how to effectively use the global spectral information and local spatial structure of hyperspectral data for HSI classification. Compared with the existing two-step classification framework, the use of DPR technology in preprocessing significantly improves the classification accuracy. The effectiveness of the proposed method is verified on three public real hyperspectral datasets. The comparison results of several competitive methods show the superiority of this scheme.

1. Introduction

A hyperspectral image (HSI) is acquired by hyperspectral remote sensors and composed of hundreds of bands over the same spatial area. It can provide high spectral resolution and rich spatial information. These features of HSIs contain a wealth of information to identify land covers of interest effectively. As a fundamental problem, HSI classification has been paid more and more attention recently, partially due to its widely successful applications in various fields, such as precision agriculture [1], urban planning [2], environment monitoring [3], target detection [4], and anomaly detection [5].
HSI classification is to assign each pixel to a meaningful and physical class based on their spectral features and the land coverage surface. During the last decades, plenty of HSI classification methods have been proposed. Some typical approaches; for example, support a vector machine (SVM)-based method [6,7,8,9], multinomial logistic regression-based techniques [10,11], various feature selection-based methods [12,13], and sparse representation-based methods [14,15], have been demonstrated to be very successful techniques. The aforementioned methods make full use of the spectral features of hyperspectral data in the process of classification. Generally, the classification results obtained by those above-mentioned methods that do not consider the spatial information of HSI are often not ideal. To further improve the classification performance, a large number of methods combining spectral information with spatial information have been investigated in recent years [16,17,18,19,20,21,22].
Among of the spectral–spatial based approaches, some spatial filtering techniques are developed [23,24,25,26,27,28]. The purpose of applying these techniques to HSI classification is to denoise HSI in the pre-processing task or improve the classification accuracy in post-processing. Several popular methods are discontinuity preserving relaxation (DPR) [28], Markov random fields (MRF) [29,30], probabilistic relaxation (PR) [31], the local binary pattern filter [32], the edge-preserving filter [33], Gabor filter [34], the extended morphological profiles filter [35] and so on.
Although the application of spatial filtering techniques in HSI classification can effectively reduce some noise pixels and provide a better classification result, the adopted spatial region is fixed in size and shape. In other words, the spatial information of HSI may be exploited insufficiently. Recent research shows that superpixel segmentation provides a powerful way to address this problem, as a superpixel is with adaptive shape and size. According to the techniques used to generate a superpixel, the known superpixel segmentation methods can be categorized into graph-based methods [36,37,38], gradient-ascent-based methods [39,40] and cluster-based methods [41]. Among of these segmentation algorithms, simple linear iterative clustering (SLIC) algorithm seems to be more popular because it is fast to compute, simple to use and good to adhere to boundaries [41].
Over the past decade, superpixel segmentation methods have also been extended to HSI classification [42,43,44,45,46,47,48,49,50], aiming at making full use of spectral information and spatial structure in hyperspectral data. By the combination of different segmentation techniques with various classification methods, a number of approaches for HSI classification have been developed, such as ER with sparse representation [42,51,52,53], SVM [54] or extreme learning machines [55], SLIC with multi-morphological method [56], SVM [57] or convolutional neural network [58] and so on. These HSI classification methods based on superpixel segmentation display good performance in experiments. Generally, when using SLIC algorithms to segment HSI, the principal component analysis (PCA) method must be adopted previously and the first three components are taken as the Lab color space. However, it will still inevitably encounter the difficulties of selecting the optimal weight in SLIC algorithms. To address the above-mentioned problems, therefore, it is necessary to improve SLIC used in color image segmentation so that it can be directly applied to segment HSI.
In this study, a novel technique to measure the similarity of a pair of pixels in HSI is suggested. The proposed similarity is designed especially for the SLIC algorithm so that SLIC can be used to segment HSI directly without adopting PCA in advance. The advantages of the proposal lie in: on the one hand, we can better measure the similarity between two pixels since the spectral distance, spectral correlation and spatial distance are considered simultaneously; on the other hand, it cleverly avoids the problem of optimal parameter setting in SLIC.
In addition, the acquired HSIs contain a large number of noise data for a variety of reasons. There is no doubt that the presence of a multiple of noise will seriously affect the accurate classification and interpretation of HSIs. Furthermore, the high-dimensional and big data features of HSI also bring great difficulties to the accurate classification of land cover. Therefore, it is very urgent and important to put forward effective methods to tackle the above-mentioned problems in the field of remote sensing. Based on the improved SLIC algorithm and DPR strategy, this study develops an effective semi-supervised HSI classification scheme. The DPR method is adopted to preserve the class boundary information well in the process of denoising. In post-processing, the use of the superpixels with adaptive shape and size can effectively improve the classification accuracy.
The main contributions of the proposal are summarized as follows.
  • An effective classification scheme for HSI is developed based on DPR strategy, SVM and the improved SLIC method.
  • A novel technique to measure the similarity of a pair of pixels in HSI is proposed to improve the SLIC algorithm.
  • The improved SLIC algorithm can be directly applied to superpixel segmentation of hyperspectral data and is free of parameters.

2. The Proposed Classification Scheme for HSI

We first introduce some notations used in this study.
Let HSI = ( x 1   ,   x 2 , , x n ) be a hyperspectral image with n pixels; x i = ( x i , 1 ,   x i , 2 , , x i , B ) T indicate the spectral vector with regard to the pixel xi; B is the number of bands.
N ( x i ) denotes a set of spatial neighbors of the pixel xi. We herein adopt the Moore neighbor, which is defined as
N ( x i ) = { x j : | p i p j | 1   o r | q i q j | 1   }
where ( p i , q i ) is the spatial coordinate of the pixel xi.

2.1. Discontinuity Preserving Relaxation

In hyperspectral data, it is possible that the same class ground objects are with different spectra and the same spectrum may be corresponding to different ground objects. This feature of HSI leads to a lot of noise pixels. In addition, the acquired spectral reflectance values are also affected by water vapor and atmospheric radiation, also resulting in noisy pixels that remain despite the atmospheric compensation step. Noise pixels in hyperspectral data will obviously affect the final classification accuracy. One of the effective ways to solve this problem is to denoise hyperspectral data by using a fixed size moving window technique. The use of this technique will make smooth areas smoother and smoother, but at the same time, it blurs the boundary of the class. Fortunately, the DPR method provides a good solution to this problem, due to the fact that the DPR method preserves the class boundary information well while de-noising.
In HSI preprocessing, DPR, initially developed by Li et al. [28], adopts the local spatial relation among adjacent pixels to denoise the hyperspectral data, and attempts to preserve the class boundary information as much as possible. It is in fact an iterative relaxation procedure.
Specifically, the DPR method can be depicted as follows.
For a given HSI, let P = [ p 1   , , p n ] R k × n , p i = [ p i ( 1 ) , , p i ( k ) ] T ; p i ( j ) is the probability of pixel xi belonging to the j-th class. The probability matrix U = [ u 1 , , u n ] R k × n , u i = [ u i ( 1 ) , , u i ( k ) ] T can be obtained by solving the following optimization problem
{ min u { ( 1 β ) | | U P | | 2 + β i = 1 n x j N ( x i ) γ j | | u j u i | | 2 } , u i 0 , i = 1 n u i   = 1 ,
where β is a weight to balance the first term and the second term in Equation (2); γ j is a value of the pixel xj of the edge image. γ is calculated by Equation (3)
γ = e x p ( i = 1 B sobel ( b i ) )
where Sobel() represents the Sobel filter that detects the discontinuities in a band image. bi denotes the i-th band of the original hyperspectral data cube.
Our earlier work [59] showed that using Roberts cross operator instead of Sobel filter operator in Equation (3) can provide better relaxation effect. The reason of doing so is that Roberts cross operator is simple, easy to calculate, and more accurate to find the class boundary information. Thus, in this study, we still use the Roberts cross operator in Equation (3), and experimental results confirm the effectiveness of this substitution again.
While applying DPR in preprocessing stage, appropriate changes are to be made since it is impossible to know p i ( j ) previously for a given HSI.
The Sobel filter is replaced by the Roberts cross operator in Equation (3), i.e.,
γ = e x p ( i = 1 B R o b e r t s ( b i ) )
For each band b, we use Equation (5) to update it constantly
x ˜ i , b ( t + 1 ) = ( 1 β ) x i , b + β x j N ( x i ) γ j x ˜ j , b ( t ) ( 1 β ) + β x j N ( x i ) γ j   , b = 1 , 2 , , B
where x ˜ i , b ( t ) is the value of the bth band of the pixel xi in the t-th iteration.
The update process will terminate if Equation (6) is satisfied
| | E b ( t + 1 ) E b ( t ) | | < ε , E b ( t + 1 ) = | | x ˜ b ( t + 1 ) x ˜ b ( t ) | | | | x ˜ b ( t ) | |
where x ˜ b indicates the b-th band image; ε is a predetermined threshold.

2.2. Superpixel Segmentation

The superpixel algorithm partitions an image into small non-overlapping homogeneous regions with adaptive shape and size. Previous works show that superpixel segmentation has achieved great success in computer vision and image analysis. Among of the popular superpixel algorithms, the simple linear iterative cluster (SLIC) [41] is a more powerful method because of its advantages of simple use, fast computing and better preservation of the class boundary. SLIC adopts the k-means cluster technique to segment an image according to the spatial structure and color.
The distance between two pixels is defined as follows,
d = ( d c N c ) 2 + ( d s s ) 2
where dc and ds denote the color distance and spatial distance of a pair of pixels, respectively; Nc is the maximum color distance; s is the segmentation scale.
As pointed out in [41], determining the maximum color distance Nc is nontrivial as color distances can vary significantly from cluster to cluster and image to image. To tackle this problem, the authors fixed Nc to a constant m and rewrote Equation (7) in the following form
d = d c 2 + m 2 ( d s s ) 2
where m is a weight balancing color similarity and spatial proximity.
To achieve better segmentation effect, the CIELAB color space (instead of RGB color space) was adopted in the classical SLIC algorithm. Euclidean distance was taken in the computation of the color distance dc and spatial distance ds. In this case, m can be in the range [1,40].

2.3. The Improved SLIC Algorithm

In SLIC method, the segmentation effect of HSI obviously depends on the choice of the weight m. In general, it is inappropriate to use the SLIC algorithm and the suggested parameter value directly in superpixel segmentation of HSI, as significant differences in reflectance among hundreds of bands would result in a tremendous difference between spectral distance and spatial distance. In addition, it can be seen from Figure 1 that even for two pixels in the same class, there still is a distinctive difference in spectral distance if only Euclidean distance is adopted. This obviously increases the possibility of misclassification in hyperspectral data processing.
To address the aforementioned problem, a novel technique to measure the similarity of a pair of pixels in HSI is suggested, aiming at applying SLIC algorithm handily in superpixel segmentation of HSI.
The spectral distance dspec, the spatial distance dspat and the correlation coefficient ρ are defined as follows.
d s p e c ( x i , x j ) = h = 1 B | x i , h x j , h |
d s p a t ( x i , x j ) = ( p i p j ) 2 + ( q i q j ) 2
ρ ( x i , x j ) = B h = 1 B x i , h x j , h h = 1 B x i , h h = 1 B x j , h B h = 1 B x i , h 2 ( h = 1 B x i , h ) 2 B h = 1 B x j , h 2 ( h = 1 B x j , h ) 2
The Manhatten distance is adopted in the calculation of spectral distance in Equation (9). The value of ρ ( x i , x j ) is obvious in the range [0,1] as all components of xi and xj are positive.
For the convenience of narration, we define r ( x i , x j ) in the following form:
r ( x i , x j ) = 1 ρ ( x i , x j )
The proposed strategy to measure the similarity of a pair of pixels in HSI is as follows.
Suppose that there are k cluster centroids, c1, c2, …, ck, around the tested pixel xi. By using Equations (9), (10) and (12), the spectral distance dspec(xi, cj), the spatial distance dspat(xi, cj) and r ( x i , c j ) from xi to cj (j=1, 2, …, k) is calculated, respectively. In each group, we rank them in ascending order, i.e.,
d s p e c ( x i , c 1 ) d s p e c ( x i , c 2 ) d s p e c ( x i , c k )
d s p a t ( x i , c 1 ) d s p a t ( x i , c 2 ) d s p a t ( x i , c k )
r ( x i , c 1 ) r ( x i , c 2 ) r ( x i , c k )
where ( c 1 ,   c 2 , , c k ) , ( c 1 ,   c 2 , , c k ) and ( c 1 ,   c 2 , , c k ) is a permutation of ( c 1 , c 2 , , c k ) , respectively.
The tested pixel xi will be assigned to j-th cluster, if among of dspec (xi, cj), dspat(xi, cj) and r ( x i , c j ) , at least two of them are the smallest in the corresponding groups. Otherwise, spatial distance is considered only.
This strategy can be thought of as the generalization of Equation (8). When all of them, or spectral distance and spatial distance are minimal, this is exactly the case of Equation (8). When spectral distance and r ( x i , c j ) are minimal, it indicates that the spectral curves of the pixel xi and centroid cj are similar both in shape and reflectance. The assignment of pixel xi to j-th cluster is propitious to enhance the homogeneity of superpixels. As for the case of dspat(xi, cj) and r ( x i , c j ) , it can be regarded as the supplement of the case of spectral distance and spatial distance. The difference is that they are very similar in shape, but there are some differences in their reflection values. If there is only one smallest among of dspec( x i , c j ), dspat( x i , c j ) and r ( x i , c j ) , it is reasonable to assign pixel xi to the closest cluster according to the first law of geography.
The proposed technique not only measures the similarity between two pixels better, but also gets rid of the trouble of choosing optimal weight m in Equation (8). It is worth noting that you can choose the distance you like to calculate the spectral distance between two pixels because the comparison takes place within each group.
The improved SLIC algorithm can be summarized as follows.
  • Initialize the cluster centers by sampling pixels at scale s. Move cluster center to the pixel with the lowest gradient in a neighborhood.
  • Assign each pixel to the nearest cluster in a region by using Equations (9), (10) and (12) and the proposed strategy.
  • Update each cluster center.
  • Repeat clustering until a given threshold is met.

2.4. An Effective Classification Scheme for HSI

In what follows, an effective classification scheme for HSI is presented based on the DPR method, SVM and improved SLIC algorithm. The proposed scheme can be divided into three steps.
Step 1. Data preprocessing.
There is no doubt that a large number of noise pixels in hyperspectral data will affect the HSI classification result. In order to get a better classification result, it is very necessary to preprocess hyperspectral data before classification. In this work, we prefer to take DPR as the data preprocessing method to denoise a given HSI. Specifically, we use Equations (4)–(6) to do this work.
Step 2. Classification at pixel wise.
Previous studies have demonstrated that SVM is a more powerful classifier in machine learning and hyperspectral data analysis. In this step, SVM with five-fold cross-validation (SVM-5) is used to classify the preprocessed HSI in a pixel-wise fashion. To enhance the performance of SVM-5, the Gaussian radial basis function (RBF) kernel is taken as its kernel function. The optimization process of the parameters used in SVM-5 is implemented by means of five-fold cross-validation.
Step 3. Post-proprecessing
After data preprocessing, we have also used the improved SLIC algorithm to carry out the superpixel segmentation of HSI. The obtained superpixels are herein used to improve the classification result provided in step 2.
The superpixel segmentation of HSI is based on the assumption that the spatial adjacent pixels in the remote sensing image have similar spectral properties and, thus, should fall into the same class. This assumption allows us to apply superpiexls to the post-processing of HSI classification. Unlike the MRF method with fixed-size window, superpixels have adaptive shapes and sizes. This feature of a superpixel makes it favorable for its application in the post-processing of HSI classification. Experimental results of this study confirm this idea.
To better understand the proposed classification scheme, the overall workflow of the integration of DPR, SVM and superpixel segmentation is elaborated in Figure 2. In this scheme, the DPR technique is responsible for denoising, while retaining the class boundary information better. The SVM method has good classification performance for high-dimensional data with small labeled samples. Similar to the other two-step classification methods, the use of a superpixel in post-processing can significantly improve the classification accuracy. The proposed approach effectively integrates their strengths, and the experimental results also confirm the effectiveness of the proposal.

3. Experimental Results and Analysis

3.1. Datasets

The effectiveness of the proposed DPR-SVM-SP method is tested on three public real hyperspectral datasets, i.e., Indian Pines dataset, Pavia University dataset and Salinas dataset. The reasons why these three datasets are widely used to test the performance of HSI classification algorithms are: (1) The Indian Pines dataset is a severely unbalanced dataset with 16 classes. The maximum class and the minimum class contain 2455 pixels and only 20 pixels, respectively. (2) Although there are only nine classes in Pavia University dataset, objects of different classes, even objects in the same class, show great differences in spatial structure. (3) In the Salinas dataset, there are two spatial adjacent classes that have a little difference in spectral features. All of these factors pose a challenge to the HSI classification methods. These three hyperspectral datasets are available online at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.
The main features of these three hyperspectral datasets can be found in Table 1. Their false color images and ground-truths are shown in Figure 3, Figure 4 and Figure 5.

3.2. Experimental Design

In the experiments, 5% of pixels from each class for the Indian Pines dataset and 1% of pixels per class for the Pavia University and Salinas dataset are randomly labeled as training sets, respectively. The rest make up the test set. To overcome the deviation caused by random sampling on classification results, the classification accuracy provided in this study is the statistical results of ten independent trials with randomly selected training sets, i.e., the mean and standard deviation of ten classification results. The classification results are evaluated by three popular indices, overall accuracy (OA), average accuracy (AA), and kappa coefficient (κ).

3.3. Classification Result

Table 2 reports the statistical results of ten independent classifications provided by SVM only, SVM and superpixel obtained by improved SLIC algorithm (SVM-SP), DPR+SVM+superpixel obtained by PCA+original SLIC algorithm (DPR-SVM-POS), and the proposed scheme DPR-SVM-SP on the Indian Pines dataset. The SVM-SP method means that the classification is done pixel-wise by using SVM, and then the superpixel is used to improve the classification result in post-processing. The classification results obtained by SVM using only spectral information are unsatisfied on this hyperspectral dataset. The use of a superpixel in the SVM-SP approach has led to a significant improvement in the classification results, that is, the classification accuracy is increased by about 10%. According to the classification accuracy, it seems that all samples in class Alfalfa (46 samples) and class Oats (20 samples) are misclassified, except the labeled sample itself. This is probably because the volumes of the two categories are so small, and their spectral features are so similar to those of their spatial neighbor classes that they are divided into other classes in superpixel segmentation. Satisfactory results have been acquired by the proposed classification scheme, and the classification accuracies of almost all classes are more than 90%, except for Corn. The proposal shows a good classification performance on the three minimum classes. It indicates that using DPR and superpixel together in HSI classification is good for discriminating land covers of interest, as shown in Figure 6.
In Table 3, compared with the result provided by SVM on the Pavia University dataset, the classification accuracy (OA) obtained by SVM-SP and DPR-SVM-SP is improved by about 6% and 10%, respectively. Statistical results on class Soil-vineyard-develop explain that the spectral characteristics of its pixels are very similar to each other. The use of superpixels in post-processing has a slight effect on its classification result. This shows that the improved SLIC algorithm can still better segment the class in which the spatial distribution of pixels is very scattered, as shown by the light green class “Fallowroughplow” in Figure 4b. However, the breakdown boundary and the very small fragments resulted in a decline in classification accuracy of this class, due to the application of the DPR technique. Maybe the DPR method does not deal with this case very well. The visual classification results of the Pavia University dataset are shown in Figure 7.
Table 4 lists an accurate classification result achieved by the proposed method for the Salinas dataset. For this hyperspectral dataset, the correct classification between class “Grapes-untrained” and class “Grapes-untrained” is a challenging problem for existing methods, because these two categories are actually the same [60]. That is to say that their spectral features are highly similar. We can easily know from Figure 8a that most of misclassified pixels belong to these two classes. As can be seen from Table 4 and Figure 8d, the proposed method is a good solution to the problem. In addition, we have also observed that the standard deviation of classification accuracy is very small, even for each category. It means that there is good stability of the proposed classification scheme.
As can be seen form Table 2, Table 3 and Table 4, compared with the SVM-SP method, the proposed method improves significantly the classification accuracy because of the use of DPR. Compared to DPR-SVM-POS approach, the proposed scheme still has good classification performance. This comparison also confirms the effectiveness of the improved SLIC. It should be noted that the improved SLIC algorithm can directly segment HSI into superpixels without using PCA in advance, and is free of parameters.

3.4. Comparative Test

To compare the proposed classification scheme with other state-of-the-art methods, it is important to select the same dataset, the same number of labeled samples and those classification approaches related to superpixels or spatial structure. Based on this consideration, we compare the proposal with different HSI classification methods, including SVM with the Extended Morphological Profile (EMP) and superpixels (EMP-SP-SVM) [56], multi-scale superpixel (MSP) and subspace-based SVM (MSP-SVMsub) [57], superpixel-based discriminative sparse model (SBDSM) [52], superpixel and extreme learning machines (SP-ELM) [55], superpixel-based spatial pyramid representation (SP-SPR) [61], multiple kernel learning-based low rank representation at superpixel level (SP-MKL-LRR) [48] segmented stacked autoencoder (S-SAE) [18] and spectral–spatial correlation segmentation-based classifier (SoCRATE) [21], SuperPCA [54]. All of these approaches try to use superpixels or spatial structure to improve the accuracy of classification results. Table 5 reports the comparison results of nine methods on the same three hyperspectral datasets. The classification accuracy of these nine methods comes from their papers.
There is no significant difference between the different classification methods because of the higher classification accuracy. For example, there is only a slight difference between our proposal and SP-SPR, SBDSM, S-SAE, and SP-ELM methods on the corresponding datasets. However, our method outperforms MSP-SVMsub (87.62% vs. 84.01%) and EMP-SP-SVM (93.74% vs. 91.64%) on the Indian Pines dataset with small training samples. What needs to be explained is that, on the Pavia University dataset, the classification accuracy (93.97%) of SP-MKL-LRR with labeled ratio 15% is still lower than our result (97.7%) with a labeled ratio of 1%. It indicates that our method makes better use of spatial information in the classification process. In case of the segmentation-aided sampling, SoCRATE is superior to our method (98.18% vs. 96%) on the Indian Pines dataset, but classification results (97.7% vs. 93.44%) on the Pavia University dataset and that (99.16% vs. 98.64%) on the Salinas dataset indicate that the proposed scheme is better than SoCRATE. Although the proposal is slightly superior to the SuperPCA method on the Indian Pines and Pavia University datasets, the classification result (98.97% vs. 98.12%) on the Salinas dataset is better than ours. When the SuperPCA method and multi-scale superpixel were combined, the MSuperPCA method presented a good classification result [54]. The results of Table 5 show that the classification accuracy can be significantly improved by using the spatial information and the improved SLIC algorithm effectively.

4. Parameter Analysis

In this section, we discuss the effects of the parameters adopted in the proposed scheme on classification results on the three hyperspectral datasets. Parameter β controls the contribution of discontinuity calculated by edge detection operator to the smooth image in the DPR method. Segmentation scale s dominates the size of superpixels in SLIC algorithm. Furthermore, the effects of different edge detection operators on the classification results are also analyzed.
Figure 9 shows the classification results on three hyperspectral datasets for four different edge detection operators. For Indian Pines dataset, with the increase of superpixel volume, the classification accuracy of the four operators shows an obvious downward trend. This may be due to the fact that the number of object classes contained in large superpixels exceeds the number of clusters, thus undermining the homogeneity of superpixels. All the four operators show better classification performance on Pavia University dataset, and almost all classification accuracies are more than 90%, especially Robert operator and Sobel operator. The reason should be that Pavia University dataset has only nine classes and that they have significant spatial separation. The satisfactory classification results can be obtained on Salinas dataset by using Robert, Sobel and Prewitt operators, most of which are above 95%. The best results achieved by these four operators on different datasets are recorded in Table 6.
It is easy to see from table 6 that Robert operator provides two of the three best classification accuracy. In particular, compared with the other three operators, Robert operator has an obvious advantage on Indian Pines dataset. Almost perfect classification result is obtained by using Robert operator on Salinas dataset. Canny operator outperforms the other three operators on Pavia University dataset. This is because the Canny operator is a second-order differential operator and has a good ability to recognize irregular boundaries. Unlike Indian Pines and Salinas datasets, the difference of classification accuracy among the four operators on Pavia University dataset is less than 0.3%. In other words, it is acceptable to adopt Robert operator in the proposed method.
Figure 10 exhibits the variation of the average classification accuracy with the increase of β. Each point in Figure 10 is the average of the classification results obtained by the proposed scheme on ten different superpixel segmentation scales s and with the same β. Furthermore, the classification accuracy for the fixed segmentation scale s is the average of the classification results of the training set generated randomly for ten times. It is known from Equation (5) that β = 0 means to classify HSI on the original dataset without data smoothing; β = 1 represents to denoise HSI completely using the boundary information contained in the neighbors of the tested pixel, while neglecting the spectral band information of the tested pixel itself.
For two AVIRIS datasets, the Indian Pines and Salinas, the average classification accuracy becomes better and better as β value increase. This indicates that all the other three operators, except canny operators, can detect effectively class boundary, and this information plays a dominant role in the DPR method. The reason may be that, as shown in Figure 3b and Figure 5b, the boundary of each category in these two datasets is essentially regular, but the acquirement of the best β value (0.8 or 0.9) depends on the operator used in the DPR strategy. Although the classification results obtained by using the three operators are not much different, the Robert operator still has some advantage.
The variation of average classification accuracy does not show obvious regularity on the Pavia University dataset. The great increasing of the average classification accuracy with the increase of β value from 0 to 0.1 or 0.2 explains that it is necessary to adopt edge detection operators in DPR method. While β value varies from 0.7 to 1, the classification accuracy obtained by Canny operator is still superior, because of its good ability to extract an irregular class boundary. On this dataset, the acquirement of the best β is seriously depended on the operator applied in DPR strategy.
Each datum in Figure 11 is the average of the classification results obtained by the proposal for 11 different β values and with the same segmentation scale s. For the Indian pines dataset, the average classification accuracy shows a significant downward trend as the segmentation scale s changes from 9 to 21. Since the Indian Pines dataset with size 145 × 145 has 16 classes, large-scale segmentation will inevitably reduce the homogeneity of superpixels. In post-processing, the heterogeneity of superpixels obviously leads to a distinctive decrease in classification accuracy. Contrary to the case of the Indian Pines dataset, satisfactory classification results have been obtained on the Pavia University dataset when s was from 9 to 21. One can see from Figure 8b that all the average classification accuracies are over 95.5%. As shown in Figure 4b, a great difference in spatial structure of the objects in the same class implies that using large scales in the SLIC algorithm will achieve a better segmentation effect on this dataset. In superpixel addition, the advantage of the Canny operator in detecting an irregular boundary is proved again. On the Salinas dataset, the Sobel operator shows good stability with the increase of the volume of. There is no significant difference in average classification accuracy among Sobel operator, Robert operator and Prewitt operator when the segmentation scale s varies from 7 to 19. The reason behind this lies in that in this hyperspectral dataset, the regular class shape and the absence of the class with a very small volume make the proposed method insensitive to the segmentation scale.
Table 7 shows how the average classification accuracy on three datasets varies with the increasing of the number of training samples. The visualization of result in Table 7 is displayed in Figure 12. As seen from Table 7 and Figure 12, the classification accuracy has been greatly improved by using spatial information in pre-processing or post-processing. For the Pavia University dataset, the classification accuracy of SVM-SP and DPR-SVM-SP increases slightly as the number of training samples is changed from 2 to 10%. This is because the classification accuracy is so high that there is no room for improvement. The classification accuracy of DPR-SVM-SP on AVIRIS dataset Salinas is more than 95% when the labeled ratio is great than or equal to 0.3%. It indicates that the proposed method can achieve the satisfactory classification result with small training samples. The classification results on the Indian Pines dataset have not shown the growth we expect, when the labeled ratio is greater than 6%. Maybe, most of the pixels that are correctly newly classified are in the superpixels that have already been correctly classified previously.

5. Conclusions

In this study, a technique for measuring the similarity of two pixels in HSI is proposed to address the problem of dividing HSI into superpixels directly by using the SLIC algorithm without using a PCA method. The adoption of this new similarity in the SLIC algorithm will also make it a non-parametric algorithm and easy to use. The experimental results of this work confirm the success of this attempt. Based on the improved SLIC algorithm, SVM and DPR methods, an effective classification scheme for hyperspectral data is developed. The classification accuracy can be significantly improved by utilizing DPR method in preprocessing and superpixel in post-processing. Experimental results on three real hyperspectral datasets widely used to test the performance of HSI classification approaches demonstrate the effectiveness of the proposed method.
In the proposed classification scheme, the DPR strategy, SVM and superpixels obtained by the improved SLIC can be replaced with other denoising techniques, classification algorithms or superpixels detected by other methods. That is to say that among the many combinations, only one of them has been considered in this paper. To an extent, the proposal is a general semi-supervised classification scheme. The DPR method with Robert operator has a good denoising effect on the hyperspectral data with relatively regular class boundary. However, for the HSI in which some classes are composed of many very small fragments, DPR with Canny operator seems to have a good de-noising performance. In summary, one can see from Table 6 and Table 7 that the proposal can achieve good classification results.

Author Contributions

C.L. and F.X. conceived and designed the experiments; C.L. performed the experiments; J.Y. and C.J. analyzed the data and developed the graphs and tables; F.X. and C.J. wrote the paper.

Funding

This research was funded by the National Natural Science Foundation of China, grants number 41801340 and 41771178; and Natural Science Foundation of Liaoning Province, grant number 20180550238.

Acknowledgments

We would like to sincerely thank anonymous reviewers for their very valuable suggestions and professional comments to improve the manuscript. We would also like to express our sincere gratitude to Editor Xia for her kind help.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kanning, M.; Siegmann, B.; Jarmer, T. Regionalization of uncovered agricultural soils based on organic carbon and soil texture estimations. Remote Sens. 2016, 8, 927. [Google Scholar] [CrossRef]
  2. Heldens, W.; Heiden, U.; Esch, T.; Stein, E.; Müller, A. Can the future Enmap mission contribute to urban applications? A literature survey. Remote Sens. 2011, 3, 1817–1846. [Google Scholar] [CrossRef]
  3. Clark, M.L.; Roberts, D.A. Species-level differences in hyperspectral metrics among tropical rainforest trees as determined by a tree-based classifier. Remote Sens. 2012, 4, 1820–1855. [Google Scholar] [CrossRef]
  4. Zhang, L.; Tao, D.; Xin, H. Sparse transfer manifold embedding for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 1030–1043. [Google Scholar] [CrossRef]
  5. Du, B.; Zhang, L. Random-selection-based anomaly detector for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2011, 49, 1578–1589. [Google Scholar] [CrossRef]
  6. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  7. Chi, M.; Bruzzone, L. Semisupervised classification of hyperspectral images by svms optimized in the primal. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1870–1880. [Google Scholar] [CrossRef]
  8. Bazi, Y.; Melgani, F. Toward an optimal svm classification system for hyperspectral remote sensing images. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3374–3385. [Google Scholar] [CrossRef]
  9. Tuia, D.; Volpi, M.; Mura, M.D.; Rakotomamonjy, A.; Flamary, R. Automatic feature learning for spatio-spectral image classification with sparse svm. IEEE Trans. Geosci. Remote Sens. 2014, 52, 6062–6074. [Google Scholar] [CrossRef]
  10. Böhning, D. Multinomial logistic regression algorithm. Ann. Inst. Statist. Math. 1992, 44, 197–200. [Google Scholar] [CrossRef]
  11. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semi-supervised hyperspectral image segmentation using multinomial logistic regression with active learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar]
  12. Feng, J.; Jiao, L.; Liu, F.; Sun, T.; Zhang, X. Unsupervised feature selection based on maximum information and minimum redundancy for hyperspectral images. Patt. Recog. 2016, 51, 295–309. [Google Scholar] [CrossRef]
  13. Xie, F.; Li, F.; Lei, C.; Yang, J.; Zhang, Y. Unsupervised band selection based on artificial bee colony algorithm for hyperspectral image classification. Applied Soft Computing. 2019, 75, 428–440. [Google Scholar] [CrossRef]
  14. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  15. Ni, D.; Ma, H. Hyperspectral image classification via sparse code histogram. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1843–1847. [Google Scholar]
  16. Tan, K.; Hu, J.; Li, J.; Du, P. A novel semi-supervised hyperspectral image classification approach based on spatial neighborhood information and classifier combination. ISPRS J. Photo. 2015, 105, 19–29. [Google Scholar] [CrossRef]
  17. Ghamisi, P.; Mura, M.D.; Benediktsson, J.A. A survey on spectral–spatial classification techniques based on attribute profiles. IEEE Trans. Geos. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  18. Paul, S.; Kumar, D.N. Spectral–spatial classification of hyperspectral data with mutual information based segmented stacked auto-encoder approach. ISPRS J. Photo. Remot. Sens. 2018, 138, 265–280. [Google Scholar] [CrossRef]
  19. Ghamisi, P.; Benediktsson, J.A.; Ulfarsson, M.O. Spectral–spatial classification of hyperspectral images based on hidden Markov random fields. IEEE Trans. Geos. Remote Sens. 2014, 52, 2565–2574. [Google Scholar] [CrossRef]
  20. Pan, B.; Shi, Z.; Xu, X. Mugnet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogram. Rem. Sens. 2018, 145, 108–119. [Google Scholar] [CrossRef]
  21. Appice, A.; Malerba, D. Segmentation-aided classification of hyperspectral data using spatial dependency of spectral bands. ISPRS J. Photogram. Rem. Sens. 2019, 147, 215–231. [Google Scholar] [CrossRef]
  22. Jia, S.; Zhang, X.; Li, Q. Spectral–spatial hyperspectral image classification using regularized low-rank representation and sparse representation-based graph cuts. IEEE J. Sel. Top. Appl. Earth Obs. 2017, 8, 2473–2484. [Google Scholar] [CrossRef]
  23. Yu, H.; Gao, L.; Li, J.; Li, S.; Zhang, B.; Benediktsson, J. Spectral–spatial hyperspectral image classification using subspace-based support vector machines and adaptive Markov random fields. Remote Sens. 2016, 8, 355. [Google Scholar] [CrossRef]
  24. Li, H.; Zheng, H.; Han, C.; Wang, H.; Miao, M. Onboard spectral and spatial cloud detection for hyperspectral remote sensing images. Remote Sens. 2018, 10, 152. [Google Scholar] [CrossRef]
  25. Shuai, L.; Licheng, J.; Shuyuan, Y. Hierarchical sparse learning with spectral–spatial information for hyperspectral imagery denoising. Sensors 2016, 16, 1718. [Google Scholar] [CrossRef]
  26. Li, W.; Prasad, S.; Fowler, J.E. Hyperspectral image classification using gaussian mixture models and markov random fields. IEEE Geosci. Remote Sens. Lett. 2013, 11, 153–157. [Google Scholar] [CrossRef]
  27. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  28. Li, J.; Khodadadzadeh, M.; Plaza, A.; Jia, X.; Bioucas-Dias, J.M. A discontinuity preserving relaxation scheme for spectral–spatial hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. 2017, 9, 625–639. [Google Scholar] [CrossRef]
  29. He, L.; Li, J.; Plaza, A.; Li, Y. Discriminative Low-Rank Gabor Filtering for Spectral–Spatial Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1381–1395. [Google Scholar] [CrossRef]
  30. Li, A.; Qin, A.; Shang, Z.; Tang, Y.Y. Spectral–spatial Sparse Subspace Clustering Based on Three-Dimensional Edge-Preserving Filtering for Hyperspectral Image. Int. J. Pattern Recognit. Artif. Intell. 2018. [Google Scholar] [CrossRef]
  31. Charles, A.S.; Rozell, C.J. Spectral Super resolution of Hyperspectral Imagery Using Reweighted l1 Spatial Filtering. IEEE Geosci. Remote Sens. Lett. 2014, 11, 602–606. [Google Scholar] [CrossRef]
  32. Chen, C.; Zhou, L.; Guo, J.; Li, W.; Su, H.; Guo, F. Gabor-Filtering-Based Completed Local Binary Patterns for Land-Use Scene Classification. In Proceedings of the IEEE International Conference on Multimedia Big Data, Beijing, China, 20–22 April 2015. [Google Scholar]
  33. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification with edge-preserving filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  34. Jia, S.; Shen, L.; Li, Q. Gabor feature-based collaborative representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1118–1129. [Google Scholar]
  35. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  36. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient graph-based image segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
  37. Shi, J.; Malik, J. Normalized cuts and image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar] [Green Version]
  38. Lu, T.; Wang, J.; Zhou, H.; Jiang, J.; Ma, J.; Wang, Z. Rectangular-Normalized Superpixel Entropy Index for Image Quality Assessment. Entropy 2018, 20, 947. [Google Scholar] [CrossRef]
  39. Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef]
  40. Levinshtein, A.; Stere, A.; Kutulakos, K.N.; Fleet, D.J.; Dickinson, S.J.; Siddiqi, K. Turbopixels: Fast superpixels using geometric flows. IEEE Trans. Patt. Anal. Mach. Intell. 2009, 31, 2290–2297. [Google Scholar] [CrossRef]
  41. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Sabine, S. SLIC superpixels compared to state-of- the-art superpixel methods. IEEE Trans. Patt. Analy. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  42. Li, J.; Zhang, H.; Zhang, L. Efficient superpixel-level multitask joint sparse representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5338–5351. [Google Scholar]
  43. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral–spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef]
  44. Saranathan, A.M.; Parente, M. Uniformity-based superpixel segmentation of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1419–1430. [Google Scholar] [CrossRef]
  45. Tan, N.; Xu, Y.; Goh, W.B.; Liu, J. Robust multi-scale superpixel classification for optic cup localization. Comput. Med. Imag. Grap. 2015, 40, 182–193. [Google Scholar] [CrossRef]
  46. Wang, Q.; Lin, J.; Yuan, Y. Salient band selection for hyperspectral image classification via manifold ranking. IEEE Trans. Neural Netw. 2016, 27, 1279–1289. [Google Scholar] [CrossRef]
  47. Wang, Y.; Zhang, Y.; Song, H. A spectral-texture kernel-based classification method for hyperspectral images. Remote Sens. 2016, 8, 919. [Google Scholar] [CrossRef]
  48. Zhan, T.; Sun, L.; Xu, Y.; Yang, G.; Zhang, Y.; Wu, Z. Hyperspectral classification via superpixel kernel learning-based low rank representation. Remote Sens. 2018, 10, 1639. [Google Scholar] [CrossRef]
  49. Fan, J.; Chen, T.; Lu, S. Superpixel guided deep-sparse-representation learning for hyperspectral image classification. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 3163–3173. [Google Scholar] [CrossRef]
  50. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Superpixel-based multitask learning framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 99, 2575–2588. [Google Scholar] [CrossRef]
  51. Jia, S.; Deng, B.; Huang, Q. An efficient superpixel-based sparse representation framework for hyperspectral image classification. Int. J. Wavelets Multiresolution Inf. Process. 2017, 15. [Google Scholar] [CrossRef]
  52. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral–spatial classification of hyperspectral images with a superpixel-based discriminative sparse model. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4186–4201. [Google Scholar] [CrossRef]
  53. Zhang, S.; Li, S.; Fu, W.; Fang, L. Multi-scale superpixel-based sparse representation for hyperspectral image classification. Remote Sens. 2017, 9, 139. [Google Scholar] [CrossRef]
  54. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A Superpixel wise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef]
  55. Duan, W.; Li, S.; Fang, L. Spectral–spatial hyperspectral image classification using superpixel and extreme learning machines. Patt. Recog. 2014, 483, 159–167. [Google Scholar]
  56. Liu, T.; Gu, Y.; Chanussot, J.; Mura, M.D. Multimorphological superpixel model for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6950–6963. [Google Scholar] [CrossRef]
  57. Yu, H.; Gao, L.; Liao, W.; Zhang, B.; Pižurica, A.; Philips, W. Multiscale superpixel-level subspace-based support vector machines for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2142–2146. [Google Scholar] [CrossRef]
  58. Cao, J.; Zhao, C.; Wang, B. Deep convolutional networks with superpixel segmentation for hyperspectral image classification. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Beijing, China, 10–15 July 2016. [Google Scholar]
  59. Xie, F.; Hu, D.; Li, F.; Yang, J.; Liu, D. Semi-supervised classification for hyperspectral images based on multiple classifiers and relaxation strategy. ISPRS Int. J. Geo-Inf. 2018, 7, 284. [Google Scholar] [CrossRef]
  60. Gualtieri, J.; Chettri, S.; Crompb, R.; Johnson, L. Support Vector Machine Classifiers as Applied to Aviris Data. 1999. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.30.2656 (accessed on 9 March 2019).
  61. Fan, J.; Hui, L.T.; Toomik, M.; Lu, S. Spectral–Spatial Hyperspectral Image Classification Using Superpixel-Based Spatial Pyramid Representation. 2016. Available online: https://0-doi-org.brum.beds.ac.uk/10.1117/12.2241033 (accessed on 9 March 2019).
Figure 1. The spectral curves of the fifth class in Pavia University dataset.
Figure 1. The spectral curves of the fifth class in Pavia University dataset.
Remotesensing 11 01149 g001
Figure 2. The framework of the proposed classification scheme.
Figure 2. The framework of the proposed classification scheme.
Remotesensing 11 01149 g002
Figure 3. Indian Pines dataset. (a) False color image. (b) Ground-truth.
Figure 3. Indian Pines dataset. (a) False color image. (b) Ground-truth.
Remotesensing 11 01149 g003
Figure 4. Pavia University dataset. (a) False color image. (b) Ground-truth.
Figure 4. Pavia University dataset. (a) False color image. (b) Ground-truth.
Remotesensing 11 01149 g004
Figure 5. Salinas dataset. (a) False color image. (b) Ground-truth.
Figure 5. Salinas dataset. (a) False color image. (b) Ground-truth.
Remotesensing 11 01149 g005
Figure 6. The classification results of Indian Pines dataset. (a) SVM. (b) SVM-SP (s = 5). (c) DPR-SVM-POS (m = 15, s = 5 and β = 0.9). (d) DPR-SVM-SP (s = 5 and β = 0.9).
Figure 6. The classification results of Indian Pines dataset. (a) SVM. (b) SVM-SP (s = 5). (c) DPR-SVM-POS (m = 15, s = 5 and β = 0.9). (d) DPR-SVM-SP (s = 5 and β = 0.9).
Remotesensing 11 01149 g006
Figure 7. The classification results of Pavia University dataset. (a) SVM. (b) SVM-SP (s = 9). (c) DPR-SVM-POS (m = 15, s = 9 and β = 0.2). (d) DPR-SVM-SP (s = 9 and β = 0.2).
Figure 7. The classification results of Pavia University dataset. (a) SVM. (b) SVM-SP (s = 9). (c) DPR-SVM-POS (m = 15, s = 9 and β = 0.2). (d) DPR-SVM-SP (s = 9 and β = 0.2).
Remotesensing 11 01149 g007
Figure 8. The classification results of Salinas dataset. (a) SVM. (b) SVM-SP (s= 15). (c) DPR-SVM-POS (m = 15, s = 15 and β = 0.9). (d) DPR-SVM-SP (s = 15 and β = 0.9).
Figure 8. The classification results of Salinas dataset. (a) SVM. (b) SVM-SP (s= 15). (c) DPR-SVM-POS (m = 15, s = 15 and β = 0.9). (d) DPR-SVM-SP (s = 15 and β = 0.9).
Remotesensing 11 01149 g008
Figure 9. The classification results of four edge detection operator Robert, Sobel, Canny and Prewitt for different parameter combinations on three hyperspectral datasets. (Note: The color of the line represents the size of the OA accuracy. From dark blue to deep red, it means that the OA accuracy changes from small to greater. It can be clearly seen from (a)).
Figure 9. The classification results of four edge detection operator Robert, Sobel, Canny and Prewitt for different parameter combinations on three hyperspectral datasets. (Note: The color of the line represents the size of the OA accuracy. From dark blue to deep red, it means that the OA accuracy changes from small to greater. It can be clearly seen from (a)).
Remotesensing 11 01149 g009
Figure 10. The variation of the average classification accuracy with the increase of β. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Figure 10. The variation of the average classification accuracy with the increase of β. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Remotesensing 11 01149 g010
Figure 11. The variation of the average classification accuracy with increase of segmentation scale s. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Figure 11. The variation of the average classification accuracy with increase of segmentation scale s. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Remotesensing 11 01149 g011
Figure 12. The visualization of results in Table 7. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Figure 12. The visualization of results in Table 7. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Remotesensing 11 01149 g012
Table 1. The main features of three hyperspectral datasets.
Table 1. The main features of three hyperspectral datasets.
ClassIndian PinesSalinasPavia University
16169
band200204103
Size 145 × 145 512 × 217 610 × 340
sensorAVIRISAVIRISROSIS
resolution20 m3.7 m1.3 m
sample10,24954,12942,776
Table 2. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Indian Pines datasets with 5% labeled samples per class, s = 5 and β = 0.9.
Table 2. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Indian Pines datasets with 5% labeled samples per class, s = 5 and β = 0.9.
ClassesTrain/TestSVMSVM-SPDPR-SVM-POSDPR-SVM-SP
Alfalfa3/438.37 ± 7.535.12 ± 13.8361.17 ± 6.497.67 ± 0.00
Corn-notill72/135661.97 ± 4.3577.07 ± 6.2093.16 ± 0.496.35 ± 0.11
Corn-mintilla42/78846.13 ± 2.7759.49 ± 5.4491.17 ± 0.792.57 ± 0.18
Corn12/22525.47 ± 5.5932.36 ± 14.5484.67 ± 0.675.47 ± 0.86
Grass-pasture25/45879.91 ± 3.7485.33 ± 5.8893.21 ± 0.290.46 ± 0.25
Grass-trees37/69394.14 ± 2.1696.54 ± 2.3199.86 ± 0100.00 ± 0.00
Grass-pasturemowed2/2633.46 ± 14.5160.77 ± 35.4590 ± 2100.00 ± 0.00
Hay-windrowed24/45494.47 ± 2.4796.83 ± 3.3298.94 ± 0.198.75 ± 0.11
Oats1/194.74 ± 8.762.11 ± 6.3260 ± 2.7100.00 ± 0.00
Soybean-notill49/92364.60 ± 4.5475.84 ± 5.4191.51 ± 0.291.75 ± 0.19
Soybean-mintill123/233280.64 ± 2.6793.07 ± 1.8198.81 ± 097.62 ± 0.06
Soybean-clean30/56337.90 ± 6.4954.51 ± 12.2385.61 ± 1.794.23 ± 0.12
Wheat11/19495.15 ± 2.8498.19 ± 0.6699.07 ± 0.399.07 ± 0.22
Woods64/120195.50 ± 1.2099.45 ± 1.0097.94 ± 0.1100.00 ± 0.00
Building-grass-trees20/36631.86 ± 6.1938.36 ± 12.21100 ± 099.76 ± 0.09
Stone-steel-towers5/8858.41 ± 7.5284.77 ± 14.3695.68 ± 0.590.34 ± 0.60
OA 71.04 ± 1.2781.01 ± 2.2795.09 ± 0.196.00 ± 0.04
AA 57.05 ± 5.2166.24 ± 8.8190.05 ± 1.095.25 ± 0.17
κ 66.54 ± 1.5178.02 ± 2.6894.39 ± 0.295.43 ± 0.05
Table 3. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Pavia University datasets with 1% labeled samples per class, s =9 and β = 0.2.
Table 3. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Pavia University datasets with 1% labeled samples per class, s =9 and β = 0.2.
ClassesTrain/TestSVMSVM-SPDPR-SVM-POSDPR-SVM-SP
Broccoligreenweed167/656485.85 ± 2.1897.47 ± 0.9596.86 ± 0.297.60 ± 0.02
Broccoligreenweed2187/1846296.78 ± 1.0799.30 ± 0.7899.25 ± 099.90 ± 0.00
Fallow21/207861.01 ± 4.0572.51 ± 9.5990.74 ± 0.993.43 ± 0.05
Fallowroughplough31/303383.9 ± 4.5787.44 ± 3.1493.83 ± 0.194.95 ± 0.03
Fallowsmooth14/133196.81 ± 3.3499.71 ± 0.0399.77 ± 097.92 ± 0.04
Stubble51/497871.53 ± 4.5578.97 ± 7.6999.9 ± 098.88 ± 0.01
Celery14/131677.31 ± 6.1693.37 ± 4.4699.02 ± 085.91 ± 0.08
Grapesuntrained37/364582.56 ± 3.8995.11 ± 2.6692.15 ± 0.896.72 ± 0.03
Soilvineyarddevelop10/93799.63 ± 0.3998.59 ± 0.0499.26 ± 091.28 ± 0.11
OA 87.67 ± 1.0693.92 ± 1.2397.55 ± 0.197.70 ± 0.01
AA 83.93 ± 3.3591.39 ± 3.2696.75 ± 0.295.18 ± 0.04
κ 83.44 ± 1.4891.82 ± 1.6896.75 ± 0.197.06 ± 0.01
Table 4. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Salinas datasets with 1% labeled samples per class, s = 15 and β = 0.9.
Table 4. Statistical results (mean and standard deviation in percent) of 10 independent classifications provided by SVM, SVM-SP, DPR-SVM-POS and DPR-SVM-SP on Salinas datasets with 1% labeled samples per class, s = 15 and β = 0.9.
ClassesTrain/TestSVMSVM-SPDPR-SVM-POSDPR-SVM-SP
Brocoligreenweed121/198897.81 ± 0.93100.00 ± 0.00100 ± 0100.00 ± 0.00
Brocoligreenweed238/368899.31 ± 0.41100.00 ± 0.00100 ± 0100.00 ± 0.00
Fallow20/195692.32 ± 4.8599.15 ± 2.6898.53 ± 0100.00 ± 0.00
Fallowroughplow14/138098.52 ± 0.8398.01 ± 0.0584.47 ± 194.91 ± 0.07
Fallowsmooth27/265196.68 ± 1.2494.66 ± 0.0482.72 ± 0.799.77 ± 0.01
Stubble40/391999.55 ± 0.2899.59 ± 0.0199.8 ± 0100.00 ± 0.00
Celery36/354399.28 ± 0.2099.86 ± 0.01100 ± 0100.00 ± 0.00
Grapesuntrained113/1115883.85 ± 5.1993.37 ± 1.4199 ± 098.01 ± 0.02
Soilvinyarddevelop63/614098.95 ± 0.4599.88 ± 0.06100 ± 0100.00 ± 0.00
Cornsenescedgreen33/324586.63 ± 1.9193.41 ± 5.7197.9 ± 099.69 ± 0.00
Lettuceromaine 4wk11/105787.71 ± 5.1897.36 ± 4.7594.93 ± 0.199.53 ± 0.00
Lettuceromaine 5wk20/190798.00 ± 1.0198.04 ± 0.6985.82 ± 0.199.33 ± 0.02
Lettuceromaine 6wk10/90698.01 ± 0.8098.14 ± 0.0574.19 ± 0.298.04 ± 0.05
Lettuceromaine 7wk11/105990.01 ± 2.8596.81 ± 0.0468.06 ± 0.196.22 ± 0.08
Vinyarduntrained73/719558.65 ± 4.7365.93 ± 9.0498.49 ± 099.62 ± 0.01
Vinyardvertical19/178893.87 ± 2.9993.37 ± 4.89100 ± 097.67 ± 0.03
OA 88.71 ± 0.8892.81 ± 1.5196.46 ± 0.0199.16 ± 0.01
AA 92.45 ± 2.1295.47 ± 1.8492.74 ± 0.198.92 ± 0.02
κ 87.40 ± 0.9791.97 ± 1.7096.06 ± 0.0199.06 ± 0.01
Table 5. Comparison with other nine competitive methods on Indian Pines (IP), Pavia University (PU) and Salinas (SA) datasets. Train denotes the number of training samples per class (ratio or fixed number).
Table 5. Comparison with other nine competitive methods on Indian Pines (IP), Pavia University (PU) and Salinas (SA) datasets. Train denotes the number of training samples per class (ratio or fixed number).
CompetitorsDPR-SVM-SPOA Improvement
DatasetsTrainOAκOAκ%
SP-SPRIP10%97.2497.0097.14 ± 0.4496.73 ± 0.51−0.1
PU25098.5498.0098.61 ± 0.2198.13 ± 0.280.07
SP-MKL-LRRIP10%96.9096.4097.14 ± 0.4496.73 ± 0.510.24
PU15%93.9791.9297.70 ± 0.0197.06 ± 0.013.73
SBDSMIP10%97.12 ± 0.4197.00 ± 0.5097.14 ± 0.4496.73 ± 0.510.02
PU25097.33 ± 0.3396.00 ± 0.5098.61 ± 0.2198.13 ± 0.281.28
SA1%99.37 ± 0.2399.00 ± 0.2099.16 ± 0.0199.06 ± 0.01−0.21
SP-ELMIP10%97.7897.0097.14 ± 0.4496.73 ± 0.51−0.64
PU30098.1798.0098.95 ± 0.2198.58 ± 0.280.78
MSP-SVMsubIP1584.0182.0086.72 ± 2.5984.94 ± 2.892.71
PU10097.5797.0097.54 ± 0.8996.68 ± 1.17−0.03
EMP-SP-SVMPU5091.64 ± 1.21-93.74 ± 1.4791.81 ± 1.882.1
SA5094.01 ± 0.51-97.23 ± 0.7196.92 ± 0.793.22
SoCRATEIP5%98.18-96.00 ± 0.0495.43 ± 0.05−2.18
PU1%93.44-97.70 ± 0.0197.06 ± 0.014.26
SA1%98.64-99.16 ± 0.0199.06 ± 0.010.52
S-SAEIP10%96.66 ± 0.6696.19 ± 0.8097.14 ± 0.4496.73 ± 0.510.48
PU5%96.66 ± 0.2895.57 ± 0.4098.81 ± 0.2398.42 ± 0.302.15
IP3094.6293.8395.52 ± 0.2194.94 ± 0.120.9
SuperPCAPU3091.388.5692.64 ± 1.2389.91 ± 1.351.34
SA3098.9798.8698.12 ± 0.7698.03 ± 0.01−0.85
Table 6. The best results (mean and standard deviation) obtained by using four operators on Indian Pines (IP), Pavia University (PU) and Salinas (SA) datasets.
Table 6. The best results (mean and standard deviation) obtained by using four operators on Indian Pines (IP), Pavia University (PU) and Salinas (SA) datasets.
CannyPrewittSobelRobert
OA(%)βsOA(%)βsOA(%)βsOA(%)βs
IP92.91 ± 0.090.9793.75 ± 0.080.9594.22 ± 0.090.7596 ± 0.040.95
PU97.86 ± 0.010.51397.75 ± 0.010.41197.59 ± 0.010.6997.7 ± 0.010.29
SA97.57 ± 0.010.91799 ± 0.010.91598.98 ± 0.010.91399.16 ± 0.010.915
Table 7. The variation of average classification accuracy (OA in percent) of three datasets with the increase of the number of training samples.
Table 7. The variation of average classification accuracy (OA in percent) of three datasets with the increase of the number of training samples.
Indian PinesPavia UniversitySalinas
Training samplesSVMSVM-SPDPR-SVM-SPTraining samplesSVMSVM-SPDPR-SVM-SPTraining samplesSVMSVM-SPDPR-SVM-SP
1%44.35 ± 9.2045.58 ± 10.1282.44 ± 2.961%86.88 ± 1.0693.08 ± 1.5397.68 ± 0.220.1%80.07 ± 2.4184.93 ± 2.9591.06 ± 1.93
2%63.31 ± 1.1371.07 ± 2.1589.13 ± 2.052%91.11 ± 0.5097.35 ± 0.5797.92 ± 0.370.2%82.65 ± 1.0987.16 ± 2.2993.96 ± 1.86
3%67.27 ± 1.0975.00 ± 2.7092.72 ± 1.013%92.22 ± 0.4797.88 ± 0.4298.66 ± 0.280.3%84.92 ± 1.5190.16 ± 2.4095.78 ± 0.69
4%69.25 ± 1.0577.27 ± 1.7094.29 ± 0.474%93.01 ± 0.3298.35 ± 0.2398.73 ± 0.200.4%85.31 ± 1.5488.83 ± 3.0996.46 ± 1.06
5%71.55 ± 1.3980.59 ± 2.5696.04 ± 0.895%93.36 ± 0.2098.55 ± 0.1198.81 ± 0.230.5%86.15 ± 1.0291.87 ± 3.1096.95 ± 0.90
6%73.18 ± 1.0682.87 ± 2.0996.26 ± 0.646%93.67 ± 0.2898.56 ± 0.1498.97 ± 0.180.6%86.70 ± 1.3091.10 ± 2.2597.05 ± 0.63
7%75.03 ± 0.5183.51 ± 1.2997.01 ± 0.307%93.78 ± 0.1998.43 ± 0.2199.19 ± 0.180.7%87.83 ± 0.7191.84 ± 1.5897.53 ± 0.77
8%75.89 ± 0.3885.95 ± 1.9296.87 ± 0.348%94.15 ± 0.2498.79 ± 0.1299.25 ± 0.150.8%87.87 ± 0.8292.26 ± 1.5397.41 ± 0.59
9%77.05 ± 0.6186.77 ± 1.2597.24 ± 0.539%94.26 ± 0.1498.68 ± 0.1799.22 ± 0.150.9%88.66 ± 0.7992.52 ± 1.5098.20 ± 0.50
10%77.79 ± 0.6487.95 ± 1.6997.14 ± 0.4410%94.44 ± 0.0898.88 ± 0.1499.25 ± 0.151%89.23 ± 0.3593.79 ± 1.1999.12 ± 0.42

Share and Cite

MDPI and ACS Style

Xie, F.; Lei, C.; Yang, J.; Jin, C. An Effective Classification Scheme for Hyperspectral Image Based on Superpixel and Discontinuity Preserving Relaxation. Remote Sens. 2019, 11, 1149. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11101149

AMA Style

Xie F, Lei C, Yang J, Jin C. An Effective Classification Scheme for Hyperspectral Image Based on Superpixel and Discontinuity Preserving Relaxation. Remote Sensing. 2019; 11(10):1149. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11101149

Chicago/Turabian Style

Xie, Fuding, Cunkuan Lei, Jun Yang, and Cui Jin. 2019. "An Effective Classification Scheme for Hyperspectral Image Based on Superpixel and Discontinuity Preserving Relaxation" Remote Sensing 11, no. 10: 1149. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11101149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop