Next Article in Journal
On the Frequency Sweep Rate Estimation in Airborne FMCW SAR Systems
Next Article in Special Issue
Airborne Radar Super-Resolution Imaging Based on Fast Total Variation Method
Previous Article in Journal
Understanding Ku-Band Ocean Radar Backscatter at Low Incidence Angles under Weak to Severe Wind Conditions by Comparison of Measurements and Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Thick Cloud Removal of Remote Sensing Images Using Temporal Smoothness and Sparsity Regularized Tensor Optimization

1
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, 129 Luoyu Road, Wuhan 430079, China
2
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(20), 3446; https://0-doi-org.brum.beds.ac.uk/10.3390/rs12203446
Submission received: 31 August 2020 / Revised: 15 October 2020 / Accepted: 19 October 2020 / Published: 20 October 2020
(This article belongs to the Special Issue Remote Sensing Image Denoising, Restoration and Reconstruction)

Abstract

:
In remote sensing images, the presence of thick cloud accompanying shadow can affect the quality of subsequent processing and limit the scenarios of application. Hence, to make good use of such images, it is indispensable to remove the thick cloud and cloud shadow as well as recover the cloud-contaminated pixels. Generally, the thick cloud and cloud shadow element are not only sparse but also smooth along the spatial horizontal and vertical direction, while the clean element is smooth along the temporal direction. Guided by the above insight, a novel thick cloud removal method for remote sensing images based on temporal smoothness and sparsity regularized tensor optimization (TSSTO) is proposed in this paper. Firstly, the sparsity norm is utilized to boost the sparsity of the cloud and cloud shadow element, and unidirectional total variation (UTV) regularizers are applied to ensure the smoothness in different directions. Then, through thresholding, the cloud mask and the cloud shadow mask can be acquired and used to guide the substitution. Finally, the reference image is selected to reconstruct details of the repairing area. A series of experiments are conducted both on simulated and real cloud-contaminated images from different sensors and with different resolutions, and the results demonstrate the potential of the proposed TSSTO method for removing cloud and cloud shadow from both qualitative and quantitative viewpoints.

Graphical Abstract

1. Introduction

Remote sensing images have been widely used in many research and application fields, such as classification [1,2], object detection [3], and urban geographical mapping [4]. However, remote sensing images are often affected by cloud cover. According to the report from the International Satellite Cloud Climatology Project (ISCCP) [5], the annual global average cloud cover is up to 66%. Thick clouds usually cover up the earth’s surface information, so the information on cloud-contaminated pixels are always inaccurate or missing, which notably affects usability and limits further analysis of remote sensing images. Since thick cloud cover is inevitable for optical remote sensing images and influences the subsequent processing, using appropriate methods to remove the cloud and cloud shadow is imperative.
Thick cloud leads to loss of land cover information, so the main task of removing thick cloud is to recover the cloud-contaminated pixels using available information. To this end, many methods for thick cloud removal have been proposed in the last few decades. According to the source of information exploited to reconstruct the cloud-cover area, the existing methods can be classified into spatial information-based methods, spectral information-based methods, and temporal information-based methods.
Spatial information-based methods
The spatial information-based methods are dedicated to recovering the cloud-contaminated pixels by making full use of the information from the cloud-free area in the image to be repaired. Missing pixel interpolation is the elementary method of these methods [6,7], which is only suitable for recovering tiny missing areas but cannot deal with the large area absence. To remedy this limitation, many techniques have been introduced [8,9,10]. Criminisi inpainting [8] fulfilled the missing area within the target image and reconstructed pixels in an appropriate order patch by patch. Bandelet-based inpainting exploited the spectro-geometrical information in the cloud-free parts of the cloud-contaminated image [9]. Maximum a posteriori (MAP)-based method got the cloud-free image using apriori knowledge and gradient descent optimization [10]. Recently, Cheng et al. [11] and Zeng et al. [12] proposed methods using similar pixel locations and global optimization to get a seamless repaired image. Additionally, tensor completion-based methods [13] such as FaLRTC and HaLRTC also can be generalized to cloud removal in remotely sensed images. However, spatial information-based methods can only get a visually plausible cloud-free image when the cloud-contaminated area is small. Once the missing area becomes large, these methods’ results are not practicable for subsequent quantitative analysis and applications [14].
Spectral information-based methods
For many multispectral and hyperspectral images, some areas can be contaminated in certain bands, while other bands’ information can be intact. The main idea of the spectral information-based methods is to use information from unaffected and unspotted bands to reconstruct the missing bands. For instance, many algorithms modelled the relationship between the Moderate Resolution Imaging Spectrora-diometer (MODIS) band 7 and MODIS band 6 to remove the periodic strips in MODIS band 6 [15,16,17]. These methods can be applied to cloud removal, as the main task of cloud removal is to reconstruct the missing data. Compared to the visible wavelengths, short-wavelength infrared (SWIR) wavelengths can penetrate through cloud, which allows detection of ground surface information under cloud cover. On this premise, Li et al. [18] exploited shortwave infrared information to obtain haze-free imagery. For Landsat images, Zhang et al. [19] represented a haze optimized transformation (HOT) method that corrected the visible bands radiometrically and removed the cloud in images. In recent years, Malek et al. [20] harnessed autoencoder networks to reconstruct the cloud area of the multispectral image. Although spectral information-based methods can make better use of auxiliary information than spatial information-based methods, they are invalid when thick clouds contaminate all the bands.
Temporal information-based methods
In temporal information-based methods, there has to be at least one reference image which covers the cloud area in the target image (the image to be repaired). As the auxiliary information contained in the reference image can be utilized to reconstruct the cloud area in the target image, the effectiveness of these methods is ensured. The primary method is managing to build a linear relationship between cloud-contaminated pixels in the target image and corresponding pixels in the reference image [21,22,23]. However, the linear regression model has a finite ability to obtain the seamless reconstructed image. To tackle this issue, many tactics have been figured out to lower the radiometric differences between the original cloud-free area and the area recovered by regression. For example, Zeng et al. [24] and Li et al. [5] proposed different residual correction methods to guarantee seamless reconstruction results. As the land cover change increases the difficulty of cloud removal, Markov random field global function [14] and patch-based information reconstruction [25,26] were proposed to minimize the effect of the land cover changes. Meanwhile, the sparse representation techniques have also been applied to cloud removal. Sparsity assumes that a signal has only few non-zero points, and all other sampling points are zero (or around zero). Regarding when the cloud is sparse and the cloud-free element possesses the characteristic of low-rank, Wen et al. [27] and Zhang et al. [28] introduced the robust principal component analysis (RPCA) into the cloud removal task. By applying discrepant RPCA, they presented two-pass RPCA (TRPCA) and group-sparsity constrained RPCA (GRPCA) to detect and remove the cloud. As a frequently used sparse representation method, dictionary learning was also employed to remove the cloud in remote sensing images [29,30,31,32]. Similar to sparse representation, Li et al. [33] proposed a nonnegative matrix factorization-based method but the coefficients of which were not always sparse. Furthermore, many researchers expanded this series of methods from a two-dimensional matrix to high-dimensional tensor and combined them with total variation (TV). Ng et al. [34] and Cheng et al. [35] designed adaptive weighted tensor completion methods to recover the cloud-contaminated pixels. Ji et al. [36] suggested a nonlocal tensor completion method to restore the missing area. Chen et al. [37] proposed a low-rank sparsity decomposition method regularized by total variation, called TVLRSDC, which could remove the cloud and cloud shadow simultaneously. As deep learning has shown its tremendous potential in image processing, Zhang et al. [38] made use of a deep convolutional neural network to extract multiscale features and recover the absent information. Generally, the temporal information-based methods can obtain satisfactory results, as long as the spectral difference and the land cover change are not too discrepant between the target image and reference image.
Different from the methods as mentioned above, this paper proposes a novel tensor optimization model based on temporal smoothness and sparse representation (TSSTO) for thick cloud and cloud shadow removal in remote sensing images. For the sake of clear illustration, the rest of this paper use shorthand “horizontal” and” vertical” to represent “spatial horizontal” and “spatial vertical”, respectively. Meanwhile, “shadow” signifies “cloud shadow”, and “cloud/shadow” represents the cloud and shadow in images. The proposed TSSTO method is mainly on the basis of the discriminatively intrinsic features that the cloud/shadow is often sparse and smooth along the horizontal and vertical direction in images, while the clean images are smooth along the temporal direction between images.
The main contributions of this paper could be listed as follows:
  • This paper ensures temporal smoothness by unidirectional total variation regularizer. Considering the gradient along the temporal direction between images in the clean element is smaller than that in the cloud/shadow element, unidirectional total variation (UTV) regularizer along the temporal direction is exploited to constrain the temporal smoothness of the clean element.
  • Group sparsity is used to enhance the sparsity of the cloud/shadow, and two unidirectional total variation regularizers along the horizontal and vertical direction are designed to deal with the large cloud/shadow area. This scheme enables TSSTO better to remove both the large and small cloud-contaminated area.
The rest of this paper is organized as follows. In Section 2, the proposed TSSTO method and implementation details are introduced. In Section 3, a series of experiments are conducted and analyzed. The conclusions of this paper and the direction of future work are reported in Section 4.

2. Methodology

Usually, the cloud/shadow element would be sparse and smooth along the horizontal and vertical direction, while the clean images would be smooth along the temporal direction, except when the majority of images are covered by cloud, or there are sudden significant surface changes between images. Based on this prior knowledge, the cloud/shadow element and the clean element can be obtained by the proposed temporal smoothness and sparsity regularized model. Concretely, the cloud/shadow element is constrained by the group sparsity and two UTV regularizers, and the UTV regularizer guarantees the temporal smoothness of the clean element along the temporal direction. As shown in Figure 1, the proposed TSSTO method contains three steps: firstly, the images are arranged to a tensor according to the acquisition time. By executing the proposed model that is introduced in Section 2.1.1, the cloud/shadow element and the clean element can be acquired, and the clean element can be seen as the preliminary recovery results. Secondly, thresholds are applied to the cloud/shadow element to get the cloud mask and shadow mask. Then, according to the cloud/shadow mask, the clean area in the original images is substituted for the corresponding area in the preliminary recovery result. This step together with the tensor optimization step is called TOS. Thirdly, to recover subtle texture information, the details of the cloud/shadow area are recovered by information cloning. It is worth noting that the cloud/shadow will be removed only when the clean complementary information is available in the multitemporal images. The input of the proposed TSSTO method is a sequence of images that cover the same area, and the cloud/shadow masks could be provided or not.

2.1. Tensor Optimization Based on Temporal Smoothness and Sparsity

Generally, the cloud/shadow element is supposed to be sparser than the clean element in the cloud-contaminated images, unless the cloud cover obscures the most parts of images. Therefore, group sparsity is used to promote the sparsity of the cloud/shadow element. To better explain the UTV regularizers, Figure 2 shows the original cloud-contaminated images, clean element, and the cloud/shadow element. Figure 3, Figure 4 and Figure 5 show the histogram of the absolute values of the gradients in different directions.
As shown in Figure 3b and Figure 4b, the derivatives of the clean element are not sparse along the horizontal and vertical directions. Conversely, the derivatives of the cloud/shadow element are sparse along the horizontal and vertical direction, which can be seen from Figure 3a and Figure 4a. Hence, we constrain the horizontal and vertical gradient of the cloud/shadow element to be small in horizontal and vertical directions, thereby extracting the cloud/shadow element. In the temporal direction, as shown in Figure 5, the gradient of the clean images is usually small, while the temporal gradient of the cloud/shadow element tends to be extraordinarily finite or extremely large. This graph indicates essential information of the cloud/shadow is in the part of the large gradient. By constraining the temporal gradient of the clean element to be small, the cloud/shadow element can be obtained, and the general spectrum of the cloud/shadow area can be recovered in the clean element.

2.1.1. Temporal Smoothness and Sparsity Regularized Model

Mathematically, a single remote sensing image can be represented as I R m × n × b where m and n denote the numbers of rows and columns, and b denotes the number of spectral bands, respectively. Given a series of remote sensing images, according to the acquisition time, they can be arranged into a tensor R m × n × t , where t denotes the numbers of images, thereby processing images band by band.
Supposing that D , C , and B R m × n × t ,   respectively, represent the original cloud-contaminated images, the cloud/shadow element, and the clean element, the proposed cloud removal model can be formulated as
a r g   min B , C   λ 1 x C 1 + λ 2 y C 1 + λ 3 z B 1 + λ 4 C 2 , 1 s . t .   D = B + C ,   B 0 .
where x ,   y ,   and   z denote the derivative operators of horizontal dimension, vertical dimension, and temporal dimension, respectively. λ 1 ,   λ 2 ,   λ 3 , and   λ 4 are the parameters that adjust the weights of different information sources.
As the derivatives of the cloud/shadow element along the horizontal and vertical direction are sparser than clean element, which can be seen in Figure 3 and Figure 4, the unidirectional total variation (UTV) regularizers x C 1 and y C 1 are utilized to enhance the smoothness of the cloud/shadow element.
Generally, the cloud/shadow elements usually have large gradients in the temporal domain between images as a result of the randomness of the thick cloud. Conversely, the clean element owns small temporal gradients because the features of land cover just change slightly according to a particular trend over time. Therefore, the term z B 1 to constrain the temporal smoothness is applied in Equation (1).
The term   C 2 , 1 represents the group sparsity of the cloud/shadow. When the cloud is not massive, the cloud/shadow can be seen as a sparse term approximately. Commonly, the cloud/shadow element is sparser than the clean element for a useful image. Hence, C 2 , 1 , a 2 , 1 norm, is used to promote the group sparsity of the cloud/shadow element. Concretely, for a matrix X , X 2 , 1 is defined as
X 2 , 1 = i = 1 s x g i 2
where   g i denotes the ith group of the variable X and s is the number of groups. For a tensor C , C 2 , 1 is defined as
C 2 , 1 = unfold 1 ( C ) 2 , 1

2.1.2. Optimization of the Proposed Model

The proposed model formulated by Equation (1) is a convex optimization problem. This paper utilizes the alternating direction method of multipliers (ADMM) framework [39] to solve the proposed model. ADMM framework can convert the complex multivariable optimization task to a few simple subtasks. In the ADMM framework, the variables are updated iteratively, and the optimization problem is solved by alternately optimizing one variable with other variables fixed. To fit the ADMM framework, the auxiliary variables,   A ,   H ,   V ,   and   T are introduced and the constrained Equation (1) is rewritten to
a r g   min A , C , H , V , T   λ 1 H 1 + λ 2 V 1 + λ 3 T 1 + λ 4 A 2 , 1 s . t .   A = C                H = x C                V = y C                             T = z   ( D C )             D C   ,
Then the Augmented Lagrange Multipliers (ALM) method [40] is employed to optimize the Equation (4). ALM replaces constrained optimization problem by an unconstrained problem using Lagrange multiplier and adds a penalty term to the objective. Let Y = [ Y 1 , Y 2 , Y 3 , Y 4 ] be the Lagrange multiplier and   μ be the positive penalty scalar, thereby defining the augmented Lagrangian function as follows
f μ ( A , C ,   H , V , T , Y ) = λ 1 H 1 + λ 2 V 1 + λ 3 T 1 + λ 4 A 2 , 1 +   Y 1 , A C + μ 2 A C F 2 +   Y 2 ,   H x C + μ 2 H x C F 2 +   Y 3 , V y C + μ 2 V y C F 2 +   Y 4 , T z   ( D C ) + μ 2 T z   ( D C ) F 2
and minimizing the Equation (5) is equivalent to
a r g   min A , C ,   H , V , T   f μ ( A , C ,   H , V , T , Y ) = λ 1 H 1 + λ 2 V 1 + λ 3 T 1 + λ 4 A 2 , 1 + μ 2 A C + Y 1 μ F 2 +   μ 2 H x C + Y 2 μ F 2 +   μ 2 V y C + Y 3 μ F 2 + μ 2 T z   ( D C ) + Y 4 μ F 2
Separating the variables [ A , C ,   H , V , T ] into two groups which are [ A ,   H , V , T ] and [ C ] , the optimization problem of Equation (6) can be solved by alternately updating the variables of two groups. The multivariable optimization task can be figured out by optimizing the following five sub-problems.
A k + 1 = a r g   min A f μ ( A , C k , H k , V k , T k , Y k ) C k + 1 = a r g   min C f μ ( A k , C , H k , V k , T k , Y k ) H k + 1 = a r g   min H f μ ( A k , C k ,   H , V k , T k , Y k ) V k + 1 = a r g   min V f μ ( A k , C k , H k , V , T k , Y k ) T k + 1 = a r g   min T f μ ( A k , C k , H k , V k T , Y k )
and as in [40], the Lagrange multiplier Y can be updated as
Y 1   k + 1 =   Y 1 k + μ ( A k C k ) Y 2   k + 1 = Y 2   k + 1 + μ ( H k x C k ) Y 3   k + 1 = Y 3   k + 1 + μ ( V k y C k ) Y 4   k + 1 = Y 4   k + 1 + μ ( T k z   ( D C k ) )
Following, the approaches of updating variables in these five sub-problems are illustrated.
u p d a t e   A :
A k + 1 = a r g   min A λ 4 A 2 , 1 + μ 2 A C k + Y 1 μ F 2
As demonstrated in [41], the above problem is equivalent to
A g i k + 1 =   a r g   min A g i ( min A λ 4 A g i 2 , 1 + μ 2 A g i R g i k F 2 ) R g i : = C ( Y 1 ) g i μ
which can be solved by the soft thresholding formula:
A g i k + 1 = m a x ( R g i k 2 λ 4 μ , 0 ) R g i k R g i k 2
where A g i k + 1   denotes the ith group of the variable A k + 1 .
u p d a t e   C :
C k + 1 = a r g   min C μ 2 A k + 1 C + Y 1 μ F 2 + μ 2 H k x C + Y 2 2 F 2                + μ 2 V k y C + Y 3 μ F 2 + μ 2 T k z   ( D C ) + Y 4 μ F 2
is a least square problem, and the normal equation can be written as
( μ I + μ x T x + μ y T y + μ z T z ) C k + 1                   = μ A k + 1 + Y 1 + x T ( μ H k + Y 2 )               + y T ( μ V k + Y 3 ) + z T ( μ z D μ T k Y 4 )
which can be calculated by a closed-form solution as follows
C k + 1 = F 1 ( L 1 L 2 )
where   L 1 = μ A k + 1 + Y 1 + x T ( μ H k + Y 2 ) + y T ( μ V k + Y 3 ) + z T ( μ z D μ T k Y 4 )   L 2 = μ I + μ x T x + μ y T y + μ z T z
u p d a t e   H , V , T :
H k + 1 = a r g   min H λ 1 H 1 + μ 2 H x C k + 1 + Y 1 μ F 2
V k + 1 = a r g   min V λ 2 V 1 + μ 2 V y C k + 1 + Y 2 μ F 2
T k + 1 = a r g   min T λ 3 T 1 + μ 2 T z ( D C k + 1 ) + Y 3 μ F 2
As in [40], component-wise soft thresholding is an exact updating solution for the above three problems
H k + 1 = S λ 1 μ ( x C k + 1 Y 1 μ )
V k + 1 = S λ 2 μ ( y C k + 1 Y 2 μ )
T k + 1 = S λ 3 μ ( z ( D C k + 1 ) Y 3 μ )
where S denotes the soft-thresholding operator
S ω ( x ) = s i g n ( x ) m a x ( | x | ω , 0 )
The above optimization process can be concluded in Algorithm 1. As the proposed model is convex as well as the variables can be separated into two groups, the convergence of the proposed algorithm is ensured by the ADMM framework in theory.
Algorithm 1 Optimization of the Proposed Model
Input: Matrix D R m × n × t , scalar μ , λ 1 ,   λ 2 ,   λ 3 , λ 4
1: Initialization: B ( 0 ) = D ,   C ( 0 ) = D B ( 0 ) ;
2: while not converged do
3: Update A via Equation (10);
4: Update C via Equation (14);
5: Update H , V   and   T   via Equations (18)–(20);
6: Update Y via Equation (8);
7: end while
Output: B , C

2.2. Substitution of Clean Area

After executing the model mentioned above, the cloud/shadow element and the clean element are acquired. The proposed model obtains the general spectral information of the images as clean element B , while the differences between images are regarded as the cloud/shadow element C . Since the clean element is obtained from the original cloud-contaminated image and the cloud/shadow has been removed, the clean element can be seen as the preliminary recovery results. Generally, the larger difference appears in the cloud/shadow area rather than the clean area, which means the cloud/shadow area can be obtained through thresholding. To be specific, a pixel in the cloud/shadow element will be considered as a cloud pixel if its value is greater than the cloud threshold. Similarly, a pixel will be considered as a shadow pixel if its value is smaller than the shadow threshold. The cloud threshold and shadow threshold are given empirically. By doing this, the specific cloud/shadow area Ω and the clean area Ω can be acquired.
As the sparse representation-based methods process the whole image at the same time, the pixel in the original clear area can be changed, which affects the subsequent processes and application. To address this problem, the clean area in the original cloud-contaminated images is used to fill the corresponding area of the preliminary results. As shown in Figure 6, to facilitate understanding, I o i   is used to denote the original cloud-contaminated image in i th time node, and I B i denotes the corresponding image derived from tensor B . The pixels in I B i Ω are replaced by   I o i Ω . The substitution can be represented as
I B i Ω =   I o i Ω  
For convenience, the procedure of tensor optimization followed by substitution is called TOS.

2.3. Details Reconstruction

After restoring the cloud/shadow area preliminarily and replacing the original cloud-free area, the detailed information of the cloud/shadow area is reconstructed by information cloning. The image that is closest in the time dimension and that is clean in the corresponding area is taken as the reference image. Inspired by the conception of Poisson editing [42], the problem is modelled as a Poisson equation. To make it easier to explain, M is used to denote the reference image, and M indicates the result of TOS. Let   Ω represent the recovery area in M , and the boundary of Ω is denoted as Ω . The intensity function f to be calculated is defined over the recovery region   Ω , and   f denotes the intensity function defined over M minus Ω . The guidance vector field defined over Ω is denoted as W . W is defined as the mixing gradients of recovery area in preliminary result and reference image. The problem of reconstructing the optimized result can be formulated as an optimization problem with the boundary condition   f |   Ω = f |   Ω :
min f Ω | f W | 2 ,   with   f |   Ω = f |   Ω
where represents the gradient operator, which can be calculated as follows.
f ( x , y ) x = f ( x + 1 , y ) f ( x , y ) f ( x , y ) y = f ( x , y + 1 ) f ( x , y ) =   ( ( / x ) , ( / y ) )
Equation (23) devotes to deriving the f with a gradient which is approach W as much as possible, which can be solved as a Poisson equation with Dirichlet boundary condition,
Δ f = d i v   W   over   Ω ,   with   f |   Ω = f |   Ω
where   d i v   W   represents the divergence of the vector field W , Δ represents the Laplacian operator, which is defined as Δ = ( 2 x 2 ) + ( 2 y 2 )
2 f ( x , y ) x 2   =   f ( x + 1 , y ) + f ( x 1 , y ) 2 f ( x , y ) 2 f ( x , y ) y 2   =   f ( x , y + 1 ) + f ( x , y 1 ) 2 f ( x , y )
where ( x , y ) represents the position of the pixel in the image.
The essential formulary of information cloning is shown in Equations (23) and (25), which is an optimization problem constrained by the boundary condition. The boundary condition enforces the consistency of the boundary by constraint formulation f |   Ω = f |   Ω . Minimization of the 2 norm represents that the target gradient should be as close as the guidance gradient field. By solving this problem, a seamless cloud/shadow-free image can be obtained.
Equations (23) and (25) can be discretized using the pixel grid. The 4-neighborhood of the pixel p in M is represented as N p , and one of the pixels in 4-neighborhood is represented as q . f ( p ) denotes the intensity function value at pixel p . The aim is to calculate the pixel intensities in Ω . Equation (23) can be rewritten as the following discrete form:
min f p , q Ω 0 ( f ( p ) f ( q ) w p q ) 2 ,   with   f ( p ) = f ( p )   for   all   p Ω
where w p q = M ( p ) M ( q ) denotes the directional gradient of M at pixel p .
According to (27), the following equation can be obtained:
| N p | f ( p ) q N p Ω f ( q ) = q N p Ω f ( q ) + q N p w p q   for   all   p Ω
where | N p | is the number of neighbors in N p . The above equation can be solved iteratively until f is converged.
When pixels   p   is in Ω and N p Ω , there are no boundary terms in the right side of Equation (28)
| N p | f ( p ) q N p f ( q ) = q N p w p q
As the general spectral information of the image is recovered in M, it is required to combine the properties of f and the reference image. Therefore, the guidance field should be as follows
w ( s ) = { f ( s )         i f   | f ( s ) | > | r ( s ) | , r ( s )                otherwise     for   all   s Ω
where r denotes the reference image, and s indicates a pixel in Ω .
As information cloning uses pixel gradient instead of pixel intensity to process the image, the recovery area looks more realistic.

3. Experimental Results and Analyses

In this part, five groups of experiments that include three simulated experiments and two real-data experiments are conducted on different satellite images with different resolutions and different land features. To verify the performance of the proposed TSSTO method from various aspects, the results of TSSTO are compared with those of the common tensor completion method, FaLRTC and HaLRTC [13], the RPCA-based method TRPCA [27], and the state-of-the-art TVLRSDC [37].
For simulated experiments, the cloud area is simulated on the original clean images. The results of the methods are compared with the original clear images from both visual effect and quantitative assessment. To evaluate the performance of TSSTO for cloud removal in simulated experiments, peak signal-to-noise ratio (PSNR), structural similarity (SSIM) index [43], multiscale structural similarity (MS-SSIM) [44], information content weighted structural similarity (IW-SSIM) [45], figure definition (FD), cross-entropy (CE), and correlation coefficient (CC) index are employed. The simulated experiments and quantitative metrics are introduced in Section 3.1.
For real-data experiments, two datasets covered with real cloud/shadow from different sensors are used to test the performance of TSSTO on real cloud-contaminated images. Since there is no original cloud-free image in the real-data experiments, only no-reference image quality assessments can be used for analysis. Figure definition (FD), standard deviation (SD), information entropy (IE), blind image quality measure of enhanced images (BIQME) [46], and From Patches to Pictures (PaQ-2-PiQ) model [47] are exploited to evaluate the proposed TSSTO method quantitatively. Finally, cloud/shadow detection results are presented and compared to analyze TSSTO further.

3.1. Simulated Experiments

In this part, the cloud areas in different shapes and sizes are simulated on the original clean images that are from different sensors and with different resolutions. Experiments on three groups of simulated cloud-contaminated images are designed to evaluate the effectiveness of the proposed TSSTO method and its sensitivity to cloud size. PSNR, SSIM, MS-SSIM, IW-SSIM, FD, and CE are calculated in the whole image within all bands used in the experiment. Unlike these indexes, CC index is computed between the constructed area and the corresponding area in the original clear image. Following, the quantitative metrics used in simulated experiments are illustrated briefly.
PSNR indicates the ratio of the maximum pixel intensity to the power of the distortion. FD refers to the definition of the images, which is represented by the average gradient of the image. The average gradient reflects the rate of the intensity change in the image of small details. Let O ( i , j ) and R ( i , j )   denote the ( i , j ) th pixel in the original clean image and the cloud removal result, respectively. m and n are the numbers of rows and columns in image. PSNR and FD can be defined as Equations (31) and (32), where M A X represents the possible maximum value of all the pixels in image R.
The SSIM metric is a rich synthesis score of luminance, local image structure, and contrast. In SSIM, structures are modes of pixel intensities, peculiarly among neighboring pixels, after normalizing for luminance and contrast. As the human eyes are skilled at perceiving structure, the SSIM quality metric agrees more closely with the subjective quality score than PSNR. The MS-SSIM metric expands on the SSIM index by compounding luminance information at the highest resolution level with structure and contrast information at several downsampled resolutions, or scales. IW-SSIM is an extension of the SSIM by incorporating the idea of information content weighted pooling. The more similar two images are, the larger SSIM, MS-SSIM, and IW-SSIM will achieve.
CE is a popular measure to evaluate the degree of difference between the two images, and lower CE means the result is closer to the original clear image. Equation (33) shows the computation of CE, in which the g is the pixel value, R ( g ) and O ( g ) are the probability of g in image R and O . CC indicates the consistency between the groups of data. Let C denote the total number of the cloud-contaminated pixels. R k c and O k c are the kth reconstructed pixel value and kth original pixel value of contaminated cloud pixels.   M E A N R C and M E A N O C are the mean values. Specifically, CC can be defined as Equation (34).
The quantitative metrics used in simulated experiments are summarized in Table 1.
P S N R = 10 ×   l o g 10 ( M A X M S E ) = 20 ×   l o g 10 ( M A X M S E ) M S E = 1 m n i = 1 m j = 1 n [ R ( i , j ) O ( i , j ) ] 2
F D = 1 m n i = 1 m j = 1 n { [ R ( i , j ) R ( i , j + 1 ) ] 2 + [ R ( i , j ) R ( i + 1 , j ) ] 2 } / 2
C E = g = 0 M A X 1 R ( g ) log 2 R ( g ) O ( g )
C C = k = 1 C ( R k c M E A N R C ) ( O k c M E A N O C ) k = 1 C ( R k c M E A N R C ) 2 k = 1 C ( O k c M E A N O C ) 2

3.1.1. Effectiveness of the Proposed TSSTO Method

In this section, experiments are performed on datasets from Gaofen-1 WFV images, and SPOT-5 images, respectively. The results of the proposed TSSTO method are compared with those of FaLRTC, HaLRTC, TRPCA, and TVLRSDC in both qualitative and quantitative ways.
The first simulated experiment is conducted on six Gaofen-1 Wide Field View Multispectral (WFV) images with a size of 512 × 512 pixels, three bands (bands 1, 2, and 3) of which are used to undertake the test. The images mainly contain the urban area and vegetation, and the images are acquired in October 2015 and December 2015, in Hubei. The original and simulated images of true color composition (red: band 3; green: band 2; blue: band 1) are shown in Figure 7a,b. The recovery results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, TOS, and the proposed TSSTO method are presented in Figure 7c–h. The zoomed detailed regions of images in Figure 7 is shown in Figure 8. The quantitative metrics of the above methods are listed in Table 2. The scatterplots with CC for different results are shown in Figure 9.
It can be seen that the ground features are complex, and the details are rich in these images. There is a huge difference between the result of TRPCA (Figure 7e) and the original image (Figure 7a), especially the city area in the bottom left. The reason is that TRPCA changes the values of the pixels in the cloud-free area. For the HaLRTC method, as shown in Figure 7d and Figure 8d, it is evident that the texture information of the result is lost, since no auxiliary information out of the image is used to reconstruct the cloud area. For FaLRTC, some details are lost in the result which does not achieve satisfactory performance, as FaLRTC utilizes no reference images. The result of TVLRSDC (Figure 8f) looks more like the original image than those of the above three contrast methods (Figure 8c–e). The ground features are recovered generally, and the spectral consistency is maintained nicely. In Figure 8, compared to the result of TVLRSDC, the result of TSSTO not only achieves spectral consistency with the original image but also restores more details. The texture in the yellow rectangle in Figure 8h is more refined than the same area in Figure 8f, and the quantitative indexes (Table 2) of the result of TSSTO achieve optimal among all the methods. In addition, the result of TOS restored spectral characteristics generally, which shows the reliability of the proposed model that uses UTV regularizer to constraint the temporal smoothness of the clean element. As shown in Figure 9, the CC values of the HaLRTC and FaLRTC results are no more than 0.6, which are unsatisfactory; CC of the TRPCA and TVLRSDC results are 0.707 and 0.793, respectively, which have a moderate increase; and the result of the proposed TSSTO method gives the highest value of 0.990. These results suggest that the cloud removal results from the proposed TSSTO method can deliver a high-caliber reconstruction result and provide a better support for further applications.
In the second simulated experiment, the proposed TSSTO method is tested on eight SPOT-5 images with four bands (bands 1, 2, 3, and 4) which were acquired in 2015, in Beijing. The size of the experimental images is 512 × 512 pixels. The original images of false-color composition (red: band 3; green: band 2; blue: band 1) are shown in Figure 10a. The results of the different methods are shown in Figure 10c–h. Figure 11a–h shows the zoomed regions cropped from Figure 10a–h. The quantitative assessment indexes of the above methods were calculated within four bands and listed in Table 3. Figure 11 lists the scatterplots with CC for different results.
For the results of FaLRTC (Figure 11c) and HaLRTC (Figure 11d), the reason for the blurry resulting images is that restoring a large amount of texture and structure information is hard just by using little information from the cloud-free part of the image to be repaired. From Table 3, the result of TRPCA gets the unsatisfactory value of metrics, due to the differences in the non-cloud area between the original image and the resulting image (e.g., the area in the yellow rectangle in Figure 10e). In the result of TVLRSDC (Figure 11f), the recovery region looks darker than the rest of the image. This phenomenon may result from that TVLRSDC reshapes the original fourth-order dataset A R m × n × b × t into a matrix R m n × b t , which leads to equal treatment of information from the spectral domain and temporal domain. Moreover, with matrix decomposition denoising image and removing cloud, images can be smooth and vague, but there is no processing in the TVLRSDC method to recover the details. For the result of TOS, the values of PSNR, SSIM, MS-SSIM, IW-SSIM, and FD are greater than the results of FaLRTC, HaLRTC, and TRPCA. The reconstructed result of TSSTO is more persuasive with enough details. As shown in Table 3, the quantitative assessments of TSSTO expound the validity of the proposed method as well. In Figure 12, the result of TSSTO is convincing with highest CC.

3.1.2. Sensitivity to Cloud Size

In this section, the insensitivity of the proposed method to the cloud size is proved. The experiment was carried out on 11 well-aligned Landsat-8 OLI images with a size of 512 ×   512 pixels, three bands (bands 2, 3, and 4) of which were used to conduct the test, with different sizes of cloud area. With true color composition (red: band 4; green: band 3; blue: band 2), the simulated images and results of TSSTO are shown in Figure 13a,b. The proportions of cloud pixels in the simulated images are 1.34% in Figure 13a, 3.92% in Figure 13b, 10.35% in Figure 13c, 19.74% in Figure 13d, 34.84% in Figure 13e, respectively. The simulated images with the results of different cloud sizes are shown in Figure 13. The zoomed details region of Figure 13 is shown in Figure 14. Figure 15 shows the PSNR values of the results with different methods and cloud sizes.
Figure 13, Figure 14 and Figure 15 can be analyzed as follows. First, since HaLRTC and FaLRTC both reconstruct the cloud-contaminated pixel using information from the cloud-free area of the target image, the curves of PSNR values of HaLRTC and FaLRTC are similar. Second, the proposed TSSTO method can achieve excellent results for cloud removal. As shown in Figure 14e, the proposed TSSTO method can obtain detail-rich and seamless results of cloud removal, even the proportion of cloud in the test image reaches about 35%. Third, although the PSNR value declines with the increasing of the cloud sizes, PSNR values of TSSTO are always more prominent than those of the compared methods. Fourthly, as shown in Figure 15, it is distinct that the performance of TRPCA is barely satisfactory, which can be explained as follows: HaLRTC, FaLRTC, TVLRSDC, and TSSTO change the pixel values in the cloud-contaminated area only, but TRPCA manipulates the cloud-free pixels twice and introduces errors into the cloud-free area. Through the above experiments and analyses, the insensitivity of TSSTO to cloud size is proved. With two sparsity norms regularizers along the horizontal direction and vertical direction applied on cloud/shadow, TSSTO can obtain satisfactory results regardless of the size of the cloud area.

3.2. Real-Data Experiments

In this section, two groups of real-data experiments are conducted on different real cloud-contaminated images from different sensors, and the cloud areas are in different shapes and different sizes. To evaluate the cloud removal results quantitatively, the FD, SD, IE, BIQME, and PaQ-2-PiQ were calculated and listed. These assessments were calculated in the whole image. To make a comprehensive evaluation of the proposed TSSTO method, the cloud/shadow detection results are provided and compared with those of TRPCA and TVLRSDC.
The assessments used to evaluate the results are described below. SD is the simplest way to evaluate image quality, which represents the dispersion degree of pixel value relative to the mean value of the image. Let R ( i , j )   denote the ( i , j ) th pixel in the result, and m and n are the numbers of rows and columns in the resulting image. Then Equation (35) defines SD, where the mean value of image R is represented as M E A N R . Larger SD indicates a better image with a more dispersed grey level. IE estimates the amount of information in an image from the information theory perspective. The probability of g in image R is signified by R ( g ) , and Equation (36) shows the computation of IE, where M A X denotes the possible maximum value of all the pixels in image R. BIQME is proposed in [46], which is a comprehensive evaluation index of contrast, sharpness, brightness, colorfulness, and naturalness of images. PAQ-2-PiQ is a deep learning model trained on a sizeable subjective picture quality dataset for blind or no-reference perceptual picture quality prediction. A higher score generally means a higher quality of an image. More details of PAQ-2-PiQ can be seen in [47].
The quantitative metrics used in real-data experiments are summarized in Table 4.
S D = 1 m n i = 1 m j = 1 n [ R ( i , j ) M E A N R ] 2
I E = g M A X R ( g ) log 2 R ( g )

3.2.1. Real-Data Experiments Results

The first experiment was performed on five Gaofen-1 WFV images with four bands (bands 1, 2, 3, and 4) in the size of 1000 × 1000 pixels, which were acquired in October 2015 and December 2015, in Hubei. Figure 16a shows the original image (true color composition red: band 3; green: band 2; blue: band 1). Figure 16f shows the recovery results of the proposed TSSTO method, and the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC are listed in Figure 16b–e. Figure 17 shows the zoomed detailed regions of images in Figure 16. Table 5 lists the values of SD, FD, IE, BIQME, and PAQ-2-PiQ of the results generated by different methods.
As shown in Figure 16, the results of FaLRTC and HaLRTC are not adequate for the demands, due to the difficulty to restore the missing texture only using the information in the clean area. The cloud-free area (e.g., area in the yellow rectangle) is the result of TRPCA () and looks different from the same area in the original image (Figure 16). For the zoomed detailed result of TVLRSDC (Figure 17e), there is some apparent color deviation in the recovery area, as the information in the spectral domain and the temporal domain is processed together and the details are not especially recovered. The result of TSSTO (Figure 16f) is satisfactory visually, and the zoomed area shows that TSSTO achieves optimal results with clear textures and enough details. In Table 5, compared with the results of contrast methods, the majority of the quantitative metrics values of the result of TSSTO is optimal. The experiment results testify the effectiveness of the proposed model that ensures temporal smoothness between images by UTV regularizer along the temporal direction, and band-by-band processing avoids large spectral differences.
In the second real-data experiment, the proposed TSSTO method was tested on 10 well-aligned SPOT-5 images with a size of 2000 × 2000 pixels. The images were acquired in 2015, in Beijing. Four bands (bands 1, 2, 3, and 4) of the images were used to conduct the test. The original cloud-contaminated image and results of different methods are shown in Figure 18 (red: band 3; green: band 2; blue: band 1). Figure 18a is the original cloud-contaminated image. Figure 18b–f shows the results of FaLRTC, HaLRTC, TRPCA, TVLRDC, and the proposed TSSTO method, respectively. The zoomed detailed regions of images in Figure 18 are represented in Figure 19. The values of quantitative assessments of the results obtained by different methods are listed in Table 6.
As shown in Figure 18 and Figure 19, the results of FaLRTC and HaLRTC are not acceptable, as large unrepaired area and complex texture information make it difficult to reconstruct the cloud-contaminated area. For the results of TRPCA, TVLRSDC, TSSTO, which are with decent visual effects, it is noticed that the result of TRPCA (Figure 18d) has large spectral differences from the original image (Figure 18a). The transition between the original clean area and the recovery area is not smooth in the result of TVLRDC, as shown in Figure 19e. This deficiency can be explained as follows. There are large differences in the cloud area between the temporal images, while the differences in the cloud area between different bands within a single image are small. Hence, the strategy taken by TVLRSDC that processes the information from the temporal dimension and spectral dimension together can make the result unsatisfactory. Compared with the result of TVLRSDC, the result of TSSTO obtains the refined textures and rich details, as the proposed model guarantees the general spectral information to accord with the cloud-free area and information cloning guarantees fine-grained details in the cloud recovery area.
In this section, to testify the performance of the proposed method, a series of comprehensive experiments were conducted on different datasets. Due to the abundant and various spatial details, the tensor complement method, HaLRTC and FaLRTC cannot reach stable results, in which the reconstructed area appears lack of texture, especially when the cloud area and shadow area contain sophisticated land cover (e.g., river and artificial facilities). TRPCA changes the values of pixels in the original clean area, which makes the result untrustworthy. TVLRSDC can process the information from spectral and temporal domains together. However, the information from different domains is treated equally without discrimination, which may cause abnormal spectral differences. For the proposed TSSTO method, with the recovery of the details in place, the result is optimal in most cases. The visual effect, as well as quantitative indexes, shows that tensor optimization can get the nearly acceptable result and information cloning restore the details effectively.

3.2.2. Cloud/Shadow Detection Result

To further evaluate the performance of the proposed TSSTO methods, the cloud/shadow detection results of TRPCA, TVLRSDC, and TSSTO are presented in Figure 20c–e. As there are three categories (cloud, shadow, and clear background) in real-data images to be detected, original images were manually delineated as ground truth (Figure 20b), and overall accuracy (OA), average accuracy (AA), and Kappa coefficient (K) were calculated and displayed in Table 7 to demonstrate the feasibility of the proposed TSSTO. Concretely, OA represents the ratio of the correct classifications of total pixels. AA means the average accuracy of all categories, and the K measures the consistency between the ground truth and detection result. The higher these three metric values are, the better the detection result is.
From Figure 20, cloud/shadow detection results of TSSTO look convincing, according to the original cloud-contaminated images. For Gaofen-1 WFV image, it can be noticed that some cloud-free areas are detected as the cloud by TRPCA, while some cloud-free regions are recognized as the shadow by TVLRSDC. For SPOT-5 image, TSSTO also delivers a more accurate mask, which enables cloud removal methods not to change the pixels in the cloud-free area as much as possible and helps to preserve the spectral characteristics of cloud-free regions. The quantity indexes also demonstrate the effectiveness of the proposed TSSTO, which can be seen in Table 7.

4. Conclusions

This paper proposed a novel TSSTO method for remote sensing images cloud/shadow removal using tensor optimization based on temporal smoothness and sparsity. Specifically, the smoothness along different directions is guaranteed by the UTV regularization terms, and the sparsity is enhanced by the group sparse norm. Furthermore, the convergence of the model is ensured by the ADMM framework in theory. The flow of the proposed TSSTO method encompasses three steps: tensor optimization, the substitution of cloud/shadow-free area, and information cloning. In the tensor optimization step, the cloud/shadow element and the clean element is generated by executing the proposed model. In the substitution step, the cloud/shadow area is determined by the threshold segmentation to obtain the cloud/shadow mask, thereby replacing the original clean area in the preliminary result with the same area of the original image guided by the generated mask. Finally, information cloning is applied to restore more details in the recovery area. A series of experiments have been conducted on the different datasets, which verifies that the proposed TSSTO method not only can recover the cloud area and cloud shadow area with fine-grained details but also can keep spectral consistency and the continuity well. What is more, the increasing size of the cloud areas shows little influence on the proposed method.
Meanwhile, TSSTO still has room for improvement. First, the result of TSSTO is related to the result of cloud/shadow detection, which means the performance of TSSTO leverages on credible cloud/shadow detection result. The authors will make a good effort to study and to overcome this limitation. Second, reliance on the quality of input images reduces some application scenarios of TSSTO. In the future, it is worth making appropriate use of the information that in the spectral domain, spatial domain, and temporal domain jointly weaken the dependency on the quality of the input images and better reconstruct the cloud-contaminated area.

Author Contributions

Funding acquisition, J.P.; Methodology, C.D.; Project administration, J.P.; Resources, J.P.; Validation, R.L.; Writing—original draft, C.D.; Writing—review and editing, C.D., J.P., and R.L. All authors have read and agree to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 91738301, 41971422), National Major Project on High Resolution Earth Observation System (No. GFZX0403260306), and Land and Resources Research Plan of Hubei Province [2018] No. 844-11.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef] [Green Version]
  2. Amini, S.; Homayouni, S.; Safari, A.; Darvishsefat, A.A. Object-based classification of hyperspectral data using Random Forest algorithm. Geo-Spat. Inf. Sci. 2018, 21, 127–138. [Google Scholar] [CrossRef] [Green Version]
  3. Yokoya, N.; Iwasaki, A. Object detection based on sparse representation and Hough voting for optical remote sensing imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2053–2062. [Google Scholar] [CrossRef]
  4. Inglada, J.; Vincent, A.; Arias, M.; Tardy, B.; Morin, D.; Rodes, I. Operational high resolution land cover map production at the country scale using satellite image time series. Remote Sens. 2017, 9, 95. [Google Scholar] [CrossRef] [Green Version]
  5. Li, Z.; Shen, H.; Cheng, Q.; Li, W.; Zhang, L. Thick cloud removal in high-resolution satellite images using stepwise radiometric adjustment and residual correction. Remote Sens. 2019, 11, 1925. [Google Scholar] [CrossRef] [Green Version]
  6. Rossi, R.E.; Dungan, J.L.; Beck, L.R. Kriging in the shadows: Geostatistical interpolation for remote sensing. Remote Sens. Environ. 1994, 49, 32–40. [Google Scholar] [CrossRef]
  7. Van der Meer, F. Remote-sensing image analysis and geostatistics. Int. J. Remote Sens. 2012, 33, 5644–5676. [Google Scholar] [CrossRef]
  8. Criminisi, A.; Pérez, P.; Toyama, K. Region filling and object removal by exemplar-based image inpainting. IEEE Trans. Image Process. 2004, 13, 1200–1212. [Google Scholar] [CrossRef]
  9. Maalouf, A.; Carré, P.; Augereau, B.; Fernandez-Maloigne, C. A bandelet-based inpainting technique for clouds removal from remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2363–2371. [Google Scholar] [CrossRef]
  10. Shen, H.; Zhang, L. A MAP-based algorithm for destriping and inpainting of remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1490–1500. [Google Scholar]
  11. Cheng, Q.; Shen, H.; Zhang, L.; Zhang, L.; Peng, Z. Missing information reconstruction for single remote sensing images using structure-preserving global optimization. IEEE Signal Process. Lett. 2017, 24, 1163–1167. [Google Scholar] [CrossRef]
  12. Zeng, C.; Shen, H.; Zhang, L. Recovering missing pixels for Landsat ETM+ SLC-off imagery using multi-temporal regression analysis and a regularization method. Remote Sens. Environ. 2013, 131, 182–194. [Google Scholar] [CrossRef]
  13. Liu, J.; Musialski, P.; Wonka, P.; Ye, J. Tensor completion for estimating missing values in visual data. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 208–220. [Google Scholar] [CrossRef] [PubMed]
  14. Cheng, Q.; Shen, H.; Zhang, L.; Yuan, Q.; Zeng, C. Cloud removal for remotely sensed images by similar pixel replacement guided with a spatio-temporal mrf model. ISPRS J. Photogramm. Remote Sens. 2014, 92, 54–68. [Google Scholar] [CrossRef]
  15. Rakwatin, P.; Takeuchi, W.; Yasuoka, Y. Restoration of aqua MODIS band 6 using histogram matching and local least squares fitting. IEEE Geosci. Remote Sens. Lett. 2009, 47, 613–627. [Google Scholar] [CrossRef]
  16. Shen, H.; Zeng, C.; Zhang, L. Recovering reflectance of AQUA MODIS band 6 based on within-class local fitting. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2011, 4, 185–192. [Google Scholar] [CrossRef]
  17. Gladkova, I.; Grossberg, M.D.; Shahriar, F.; Bonev, G.; Romanov, P. Quantitative restoration for MODIS band 6 on aqua. IEEE Trans. Geosci. Remote Sens. 2012, 50, 2409–2416. [Google Scholar] [CrossRef]
  18. Li, H.; Zhang, L.; Shen, H.; Li, P. A variational gradient-based fusion method for visible and SWIR imagery. Photogramm. Eng. Remote Sens. 2012, 78, 947–958. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Guindon, B.; Cihlar, J. An image transform to characterize and compensate for spatial variations in thin cloud contamination of Landsat images. Remote Sens. Environ. 2002, 82, 173–187. [Google Scholar] [CrossRef]
  20. Malek, S.; Melgani, F.; Bazi, Y.; Alajlan, N. Reconstructing cloud-contaminated multispectral images with contextualized autoencoder neural networks. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2270–2282. [Google Scholar] [CrossRef]
  21. Storey, J.; Scaramuzza, P.; Schmidt, G.; Barsi, J. Landsat 7 scan line corrector-off gap-filled product development. In Proceedings of the Pecora 16 Conference on Global Priorities in Land Remote Sensing, Sioux Falls, SD, USA, 23–27 October 2005. [Google Scholar]
  22. Zhang, X.; Qin, F.; Qin, Y. Study on the thick cloud removal method based on multi-temporal remote sensing images. In Proceedings of the 2010 International Conference on Multimedia Technology, Ningbo, China, 29–31 October 2010; IEEE: Piscataway, NJ, USA; pp. 1–3. [Google Scholar]
  23. Du, W.; Qin, Z.; Fan, J.; Gao, M.; Wang, F.; Abbasi, B. An efficient approach to remove thick cloud in VNIR bands of multi-temporal remote sensing images. Remote Sens. 2019, 11, 1284. [Google Scholar] [CrossRef] [Green Version]
  24. Zeng, C.; Long, D.; Shen, H.; Wu, P.; Cui, Y.; Hong, Y. A two-step framework for reconstructing remotely sensed land surface temperatures contaminated by cloud. ISPRS J. Photogramm. Remote Sens. 2018, 141, 30–45. [Google Scholar] [CrossRef]
  25. Lin, C.H.; Tsai, P.H.; Lai, K.H.; Chen, J.Y. Cloud removal from multitemporal satellite images using information cloning. IEEE Trans. Geosci. Remote Sens. 2012, 51, 232–241. [Google Scholar] [CrossRef]
  26. Lin, C.H.; Lai, K.H.; Chen, Z.B.; Chen, J.Y. Patch-based information reconstruction of cloud-contaminated multitemporal images. IEEE Trans. Geosci. Remote Sens. 2013, 52, 163–174. [Google Scholar] [CrossRef]
  27. Wen, F.; Zhang, Y.; Gao, Z.; Ling, X. Two-pass robust component analysis for cloud removal in satellite image sequence. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1090–1094. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Wen, F.; Gao, Z.; Ling, X. A Coarse-to-Fine Framework for Cloud Removal in Remote Sensing Image Sequence. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5963–5974. [Google Scholar] [CrossRef]
  29. Li, X.; Shen, H.; Zhang, L.; Zhang, H.; Yuan, Q.; Yang, G. Recovering quantitative remote sensing products contaminated by thick clouds and shadows using multitemporal dictionary learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7086–7098. [Google Scholar]
  30. Li, X.; Shen, H.; Zhang, L.; Li, H. Sparse-based reconstruction of missing information in remote sensing images from spectral/temporal complementary information. ISPRS J. Photogramm. Remote Sens. 2015, 106, 1–15. [Google Scholar] [CrossRef]
  31. Li, X.; Shen, H.; Li, H.; Zhang, L. Patch matching-based multitemporal group sparse representation for the missing information reconstruction of remote-sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3629–3641. [Google Scholar] [CrossRef]
  32. Xu, M.; Jia, X.; Pickering, M.; Plaza, A.J. Cloud removal based on sparse representation via multitemporal dictionary learning. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2998–3006. [Google Scholar] [CrossRef]
  33. Li, X.; Wang, L.; Cheng, Q.; Wu, P.; Gan, W.; Fang, L. Cloud removal in remote sensing images using nonnegative matrix factorization and error correction. ISPRS J. Photogramm. Remote Sens. 2019, 148, 103–113. [Google Scholar] [CrossRef]
  34. Ng, M.K.P.; Yuan, Q.; Yan, L.; Sun, J. An adaptive weighted tensor completion method for the recovery of remote sensing images with missing data. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3367–3381. [Google Scholar] [CrossRef]
  35. Cheng, Q.; Yuan, Q.; Ng, M.K.P.; Shen, H.; Zhang, L. Missing Data Reconstruction for Remote Sensing Images with Weighted Low-Rank Tensor Model. IEEE Access 2019, 7, 142339–142352. [Google Scholar] [CrossRef]
  36. Ji, T.Y.; Yokoya, N.; Zhu, X.X.; Huang, T.Z. Nonlocal tensor completion for multitemporal remotely sensed images’ inpainting. IEEE Trans. Geosci. Remote Sens. 2017, 56, 3047–3061. [Google Scholar] [CrossRef]
  37. Chen, Y.; He, W.; Yokoya, N.; Huang, T.Z. Blind cloud and cloud shadow removal of multitemporal images based on total variation regularized low-rank sparsity decomposition. ISPRS J. Photogramm. Remote Sens. 2019, 157, 93–107. [Google Scholar] [CrossRef]
  38. Zhang, Q.; Yuan, Q.; Zeng, C.; Li, X.; Wei, Y. Missing data reconstruction in remote sensing image with a unified spatial-temporal-spectral deep convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4274–4288. [Google Scholar] [CrossRef] [Green Version]
  39. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Foundations and Trends in Machine Learning; Now Publishers Inc.: Breda, The Netherlands, 2010; Volume 3, pp. 1–122. [Google Scholar]
  40. Lin, Z.; Chen, M.; Ma, Y. The augmented Lagrange multiplier method for exact recovery of corrupted low-rank matrices. arXiv 2010, arXiv:1009.5055. [Google Scholar]
  41. Deng, W.; Yin, W.-T.; Zhang, Y. Group sparse optimization by alternating direction method. In Proceedings of the International Society of Optics and Photonics (SPIE), San Diego, CA, USA, 25–29 August 2013; SPIE: Bellingham, WA, USA, 2013. [Google Scholar]
  42. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. ACM Trans. Graph. 2003, 22, 313–318. [Google Scholar] [CrossRef]
  43. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  44. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multi-scale structural similarity for image quality assessment. In Proceedings of the Thirty-Seventh Asilomar Conference on Signals, Systems and Computers 2003, Pacific Grove, CA, USA, 9–12 November 2013; Volume 2, pp. 1398–1402. [Google Scholar]
  45. Wang, Z.; Li, Q. Information Content Weighting for Perceptual Image Quality Assessment. IEEE Trans. Image Process. 2011, 20, 1185–1198. [Google Scholar] [CrossRef]
  46. Gu, K.; Tao, D.; Qiao, J.; Lin, W. Learning a No-Reference Quality Assessment Model of Enhanced Images with Big Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1301–1313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Ying, Z.; Niu, H.; Gupta, P.; Mahajan, D.; Ghadiyaram, D.; Bovik, A. From Patches to Pictures (PaQ-2-PiQ): Mapping the Perceptual Space of Picture Quality. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
Figure 1. Flowchart of the proposed Temporal Smoothness and Sparsity-regularized Tensor Optimization (TSSTO) method.
Figure 1. Flowchart of the proposed Temporal Smoothness and Sparsity-regularized Tensor Optimization (TSSTO) method.
Remotesensing 12 03446 g001
Figure 2. The images used to draw histograms (a) original cloud-contaminated images; (b) clean element; (c) cloud/shadow element.
Figure 2. The images used to draw histograms (a) original cloud-contaminated images; (b) clean element; (c) cloud/shadow element.
Remotesensing 12 03446 g002
Figure 3. The histogram of the absolute values of the horizontal gradient of (a) cloud/shadow element and (b) clean element.
Figure 3. The histogram of the absolute values of the horizontal gradient of (a) cloud/shadow element and (b) clean element.
Remotesensing 12 03446 g003
Figure 4. The histogram of the absolute values of the vertical gradient of (a) cloud/shadow element and (b) clean element.
Figure 4. The histogram of the absolute values of the vertical gradient of (a) cloud/shadow element and (b) clean element.
Remotesensing 12 03446 g004
Figure 5. The histogram of the absolute values of the temporal gradient of (a) cloud/shadow element and (b) clean element.
Figure 5. The histogram of the absolute values of the temporal gradient of (a) cloud/shadow element and (b) clean element.
Remotesensing 12 03446 g005
Figure 6. Explanation of substitution.
Figure 6. Explanation of substitution.
Remotesensing 12 03446 g006
Figure 7. Gaofen-1 WFV images for the first simulated experiment: (a) original clean image; (b) simulated cloud image; (ch) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, TOS, and TSSTO, respectively.
Figure 7. Gaofen-1 WFV images for the first simulated experiment: (a) original clean image; (b) simulated cloud image; (ch) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, TOS, and TSSTO, respectively.
Remotesensing 12 03446 g007
Figure 8. (ah) Detailed regions clipped from Figure 7a–h.
Figure 8. (ah) Detailed regions clipped from Figure 7a–h.
Remotesensing 12 03446 g008
Figure 9. The scatterplots of the original and the reconstructed values in the first simulated experiment: (a) the FaLRTC method; (b) the HaLRTC method; (c) the TRPCA method; (d) the TVLRSDC method; (e) the TOS method; (f) the proposed TSSTO method.
Figure 9. The scatterplots of the original and the reconstructed values in the first simulated experiment: (a) the FaLRTC method; (b) the HaLRTC method; (c) the TRPCA method; (d) the TVLRSDC method; (e) the TOS method; (f) the proposed TSSTO method.
Remotesensing 12 03446 g009
Figure 10. SPOT-5 images for the second simulated experiment: (a) original clean image; (b) simulated cloud image; (ch) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, TOS, and TSSTO, respectively.
Figure 10. SPOT-5 images for the second simulated experiment: (a) original clean image; (b) simulated cloud image; (ch) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, TOS, and TSSTO, respectively.
Remotesensing 12 03446 g010
Figure 11. (ah) Detailed regions clipped from Figure 10a–h.
Figure 11. (ah) Detailed regions clipped from Figure 10a–h.
Remotesensing 12 03446 g011
Figure 12. The scatterplots of the original and the reconstructed values in the second simulated experiment: (a) the FaLRTC method; (b) the HaLRTC method; (c) the TRPCA method; (d) the TVLRSDC method; (e) the TOS method; (f) the proposed TSSTO method.
Figure 12. The scatterplots of the original and the reconstructed values in the second simulated experiment: (a) the FaLRTC method; (b) the HaLRTC method; (c) the TRPCA method; (d) the TVLRSDC method; (e) the TOS method; (f) the proposed TSSTO method.
Remotesensing 12 03446 g012
Figure 13. The simulated images and results with different cloud sizes. The proportions of cloud pixels in the simulated images are 1.34% (a), 3.92% (b), 10.35% (c), 19.74% (d), 34.84% (e), respectively.
Figure 13. The simulated images and results with different cloud sizes. The proportions of cloud pixels in the simulated images are 1.34% (a), 3.92% (b), 10.35% (c), 19.74% (d), 34.84% (e), respectively.
Remotesensing 12 03446 g013
Figure 14. (ae) Detailed regions clipped from Figure 13a–e.
Figure 14. (ae) Detailed regions clipped from Figure 13a–e.
Remotesensing 12 03446 g014
Figure 15. The PSNR values of the results with different methods and cloud sizes.
Figure 15. The PSNR values of the results with different methods and cloud sizes.
Remotesensing 12 03446 g015
Figure 16. Gaofen-1 WFV images for the second real-data experiment: (a) cloud image; (bf) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, and TSSTO, respectively.
Figure 16. Gaofen-1 WFV images for the second real-data experiment: (a) cloud image; (bf) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, and TSSTO, respectively.
Remotesensing 12 03446 g016
Figure 17. (af) Detailed regions clipped from Figure 16a–f.
Figure 17. (af) Detailed regions clipped from Figure 16a–f.
Remotesensing 12 03446 g017
Figure 18. SPOT-5 images for the third real-data experiment: (a) cloud image; (bf) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, and TSSTO, respectively.
Figure 18. SPOT-5 images for the third real-data experiment: (a) cloud image; (bf) the results of FaLRTC, HaLRTC, TRPCA, TVLRSDC, and TSSTO, respectively.
Remotesensing 12 03446 g018
Figure 19. (af) Detailed regions clipped from Figure 18a–f.
Figure 19. (af) Detailed regions clipped from Figure 18a–f.
Remotesensing 12 03446 g019
Figure 20. Cloud/shadow detection results for the real-data experiments. (a) Cloud image; (b) ground truth; (ce) the detection results of TRPCA, TVLRSDC, and TSSTO, respectively.
Figure 20. Cloud/shadow detection results for the real-data experiments. (a) Cloud image; (b) ground truth; (ce) the detection results of TRPCA, TVLRSDC, and TSSTO, respectively.
Remotesensing 12 03446 g020
Table 1. The induction of the quantitative metrics used in simulated experiments.
Table 1. The induction of the quantitative metrics used in simulated experiments.
IndexFull NameFormula or ReferenceNote
PSNRPeak signal-to-noise ratioEquation (31)The higher the better
SSIMStructural similarity index(Wang et al., 2004) [43]The higher the better
MS-SSIMMultiscale structural similarity(Wang et al., 2003) [44]The higher the better
IW-SSIMInformation content weighted structural similarity(Wang et al., 2011) [45]The higher the better
FDFigure definitionEquation (32)The higher the better
CECross-entropyEquation (33)The lower the better
CCCorrelation coefficient indexEquation (34)The higher the better
Table 2. The PSNR, SSIM, MS-SSIM, IW-SSIM, FD, and CE of the first simulated experimental results. The values of the highest accuracies are in bold.
Table 2. The PSNR, SSIM, MS-SSIM, IW-SSIM, FD, and CE of the first simulated experimental results. The values of the highest accuracies are in bold.
IndexFaLRTCHaLRTCTRPCATVLRSDCTOSTSSTO
PSNR35.8035.8324.2746.2947.2555.84
SSIM0.97140.97160.69360.99820.99560.9998
MS-SSIM0.97970.97970.85680.99960.99900.9998
IW-SSIM0.96310.96330.93210.99960.99830.9998
FD1204.271204.36671.691213.51205.141220.96
CE5.04 × 10−42.80 × 10−43.98 × 10−15.55 × 10−59.20 × 10−53.37 × 10−5
Table 3. The PSNR, SSIM, MS-SSIM, IW-SSIM, FD, and CE of the second simulated experimental results. The values of the highest accuracies are in bold.
Table 3. The PSNR, SSIM, MS-SSIM, IW-SSIM, FD, and CE of the second simulated experimental results. The values of the highest accuracies are in bold.
IndexFaLRTCHaLRTCTRPCATVLRSDCTOSTSSTO
PSNR43.2043.4236.3049.9847.7552.25
SSIM0.96440.96340.90100.99760.99430.9989
MS-SSIM0.97580.97500.96350.98660.98290.9890
IW-SSIM0.95710.95830.95910.98360.97880.9871
FD1153.771144.97940.561159.461148.951171.05
CE3.24 × 10−41.55 × 10−36.45 × 10−28.59 × 10−52.84 × 10−48.09 × 10−5
Table 4. The induction of the quantitative metrics used in real-data experiments.
Table 4. The induction of the quantitative metrics used in real-data experiments.
IndexFull NameFormula or ReferenceNote
FDFigure definitionEquation (32)The higher the better
SDStandard deviationEquation (35)The higher the better
IEInformation entropyEquation (36)The higher the better
BIQMEBlind image quality measure of enhanced images(Gu et al., 2018) [46]The higher the better
PAQ-2-PiQFrom Patches to Pictures model(Ying et al., 2020) [47]The higher the better
Table 5. The SD, FD, IE, BIQME, and PAQ-2-PiQ of the first real-data experimental results. The values of the highest accuracies are in bold.
Table 5. The SD, FD, IE, BIQME, and PAQ-2-PiQ of the first real-data experimental results. The values of the highest accuracies are in bold.
IndexFaLRTCHaLRTCTRPCATVLRSDCTSSTO
SD89.972889.311084.869288.358188.9597
FD2044.101992.41959.802104.802087.41
IE5.96395.89645.54845.77035.9279
BIQME0.54310.55310.55650.55480.5691
PAQ-2-PiQ62.3588662.3641164.0616568.4303770.99075
Table 6. The SD, FD, IE, BIQME, and PAQ-2-PiQ of the second real-data experimental results. The values of the highest accuracies are in bold.
Table 6. The SD, FD, IE, BIQME, and PAQ-2-PiQ of the second real-data experimental results. The values of the highest accuracies are in bold.
IndexFaLRTCHaLRTCTRPCATVLRSDCTSSTO
SD83.220782.847982.827382.153083.2207
FD1275.141276.561333.001259.681420.49
IE7.29767.29767.37627.27847.3312
BIQME73.5636473.5616476.8467677.2234777.36649
PAQ-2-PiQ0.53770.52650.54730.55860.5620
Table 7. The AA, OA, and Kappa in the real-data experiments. The bold values denote the highest accuracies in each experiment.
Table 7. The AA, OA, and Kappa in the real-data experiments. The bold values denote the highest accuracies in each experiment.
ImageMethodOAAA K × 100
Gaofen-1 WFV image TRPCA91.2891.9187.30
TVLRSDC91.7291.9488.05
TSSTO93.0893.3090.92
SPOT-5 imageTRPCA89.5678.9977.93
TVLRSDC91.1582.4182.91
TSSTO92.5185.8087.07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Duan, C.; Pan, J.; Li, R. Thick Cloud Removal of Remote Sensing Images Using Temporal Smoothness and Sparsity Regularized Tensor Optimization. Remote Sens. 2020, 12, 3446. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12203446

AMA Style

Duan C, Pan J, Li R. Thick Cloud Removal of Remote Sensing Images Using Temporal Smoothness and Sparsity Regularized Tensor Optimization. Remote Sensing. 2020; 12(20):3446. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12203446

Chicago/Turabian Style

Duan, Chenxi, Jun Pan, and Rui Li. 2020. "Thick Cloud Removal of Remote Sensing Images Using Temporal Smoothness and Sparsity Regularized Tensor Optimization" Remote Sensing 12, no. 20: 3446. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12203446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop