Next Article in Journal
Non-Enzymatic Glucose Sensing Based on Incorporation of Carbon Nanotube into Zn-Co-S Ball-in-Ball Hollow Sphere
Previous Article in Journal
Wristbands Containing Accelerometers for Objective Arm Swing Analysis in Patients with Parkinson’s Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Fine Registration of Multisensor Remote Sensing Images Based on Enhanced Subpixel Phase Correlation

1
College of Surveying and Geo-Informatics, Tongji University, 1239 Siping Road, Shanghai 200092, China
2
Faculty of Electrical Engineering and Computer Science, Technische Universität Berlin, 10587 Berlin, Germany
3
School of Mathematics and Statistics, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Current address: Photogrammetry and Remote Sensing, Technische Universität München, 80333 Munich, Germany.
Submission received: 9 July 2020 / Revised: 30 July 2020 / Accepted: 2 August 2020 / Published: 4 August 2020
(This article belongs to the Section Remote Sensors)

Abstract

:
Automatic fine registration of multisensor images plays an essential role in many remote sensing applications. However, it is always a challenging task due to significant radiometric and textural differences. In this paper, an enhanced subpixel phase correlation method is proposed, which embeds phase congruency-based structural representation, L1-norm-based rank-one matrix approximation with adaptive masking, and stable robust model fitting into the conventional calculation framework in the frequency domain. The aim is to improve the accuracy and robustness of subpixel translation estimation in practical cases. In addition, template matching using the enhanced subpixel phase correlation is integrated to realize reliable fine registration, which is able to extract a sufficient number of well-distributed and high-accuracy tie points and reduce the local misalignment for coarsely coregistered multisensor remote sensing images. Experiments undertaken with images from different satellites and sensors were carried out in two parts: tie point matching and fine registration. The results of qualitative analysis and quantitative comparison with the state-of-the-art area-based and feature-based matching methods demonstrate the effectiveness and reliability of the proposed method for multisensor matching and registration.

1. Introduction

Image registration, which is the process of geometrically aligning two or more images of the same scene taken at different conditions, is essential to image analysis tasks involving information extraction from different overlapping images [1]. With the rapid development of sensor technology, remote sensing images have attracted more and more attention due to their increasing spatial and spectral resolution, convenience, and coverage [2]. Remote sensing images from different sensors are able to provide useful complementary information. Multisensor image registration is a fundamental preprocessing step for utilizing these images in a wide variety of applications, such as image fusion, change detection, and environmental monitoring [3,4,5]. However, due to the temporal difference and the diverse properties of sensors or regions in the scene, the image pairs acquired from different optical sensors exist the issues of non-linear intensity differences, textural changes and local distortions [6]. Therefore, automatic registration of multisensor images is a challenging task.
Image registration can be generally divided into coarse registration and fine registration. The coarse registration stage pre-registers the reference and sensed images to eliminate significant rotation and scale differences and shorten the search range through a global transformation model, while the fine registration stage corrects the misalignment and refines the registration performance commonly through a more local or higher-order transformation model [7,8]. Most current remote sensing images are usually attached with georeferencing information that can be employed to remove the obvious geometric differences between images, such as rotation, scale, and global translation [9,10]. In other words, coarse registration of remote sensing images can be achieved by direct georeferencing using sensor models, and the pre-registered image pairs only exist an offset of several or dozens of pixels that require the fine registration stage to compensate. In this study, we focus exclusively on fine registration of remote sensing images.
A typical image registration method consists of two basic steps, i.e., image matching and image warping [11]. The former step extracts and matches the tie points between reference and sensed images that are the distinctive and representative points of the investigated scenes, while the latter step estimates a transformation model from the set of corresponding tie points and then transforms the sensed image to the reference image using image resampling. In order to realize precise and reliable fine registration of multisensor remote sensing images, the image matching part that determines the correspondence relationship of the tie points plays the most crucial role. In the literature, there are two major types of image matching methods: feature-based methods and area-based methods [1,12,13]. The feature-based methods match the features detected separately from each image based on their spatial structure or distance of invariant descriptor vectors. The most widely used local invariant features applied in image registration are the scale-invariant feature transform (SIFT) feature and its variants [14,15,16,17]. However, one of the main limitations of feature-based methods is that they require a sufficient number of highly repeatable features extracted from both images, which is especially difficult in the multisensor cases with obvious radiometric and textural changes.
In contrast, area-based methods rely on the similarity measure directly calculated from the intensity in the corresponding window pairs or even the entire images, which usually outperform feature-based methods in the aspect of precision, distribution, and robustness [18]. These merits enable the area-based methods more effective in fine registration of multisensor remote sensing images [19]. Phase correlation (PC) is an area-based matching technique according to the image information and operation in the frequency domain. By means of fast Fourier transform (FT) and phase information, PC can achieve outstanding performance in theoretical accuracy, computational efficiency, and robustness against the frequency-dependent noise and illumination changes [20]. These merits make it quite feasible for multisensor image registration. When used in coarse registration, PC can be extended to deal with rotation and scale estimation without the need for initialization and iteration using the Fourier-Mellin transform [21,22,23]. For fine registration, PC can be adopted in local template matching even pointwise dense matching with subpixel estimation. Additional operations that ensure the best approximation of the theoretical phase difference model play an important role in the subpixel PC methods. In this study, an enhanced subpixel PC method calculated in the frequency domain is proposed. Three additional operations are embedded into the conventional line fitting-based PC method to improve the practical performance of tie point matching: (1) phase congruency information is adopted as feature representations to reduce the influence of nonlinear intensity differences in multisensor cases; (2) a L1-norm-based robust low-rank matrix factorization algorithm is used with effective frequency masking to find the best rank-one approximation of the normalized cross-power spectrum matrix in the presence of corrupted components; and (3) a stable robust estimation algorithm is employed to effectively eliminate the residual outliers during line fitting. In addition, a fine registration method on the basis of the enhanced subpixel PC method is introduced, which is able to reduce the local misalignments between multisensor and multisource remote sensing images. The experiments carried out on remote sensing images from different satellites demonstrated the feasibility and reliability of the proposed method. In summary, the main contributions of this paper are: (1) an accurate and robust subpixel PC method for translation estimation is proposed, additionally embedding phase congruency-based structural representation, robust masked rank-one matrix approximation and robust model fitting using higher than minimal subset sampling; and (2) based on the enhanced subpixel PC matching, an automatic and reliable fine registration method for multisensor remote sensing images is presented, combining with the block-based phase congruency feature detector and local warping model.
The remainder of this paper is organized as follows. Related work is briefly reviewed in Section 2. The details of the proposed subpixel PC method and fine registration method are described in Section 3 and Section 4, respectively. Section 5 presents the experimental results and analysis, including the tie point matching experiment and fine registration experiment. Finally, the concluding remarks and considerations for future work are given in Section 6.

2. Related Work

2.1. Fine Registration Using Area-Based Methods

Area-based matching methods directly utilize intensity-based information to match images or regions. This type of matching method is widely used to optimize the coarse registration of remote sensing images due to the superiority in accuracy [24]. The adopted similarity measure is a decisive component of area-based methods. The conventional ones mainly include the sum of squared difference, the sum of absolute difference, the normalized cross correlation (NCC) [25], but are sensitive to nonlinear intensity changes [26]. In order to enhance the illumination robustness, several more sophisticated similarity measures such as mutual information (MI) [27], cross cumulative residual entropy [28], Jeffrey’s divergence [29], and matching by tone matching (MTM) [30] have been developed and broadly applied in remote sensing image registration [31,32]. In [33] and [34], MI-based metrics were utilized in optimization procedure to refine the coarse results of feature-based registration. In [35], normalized gradient field was adopted as a similarity measure to align the georeferenced airborne light detection and ranging (LiDAR), hyperspectral and photographic imagery. Moreover, some structure and shape features have been recently adopted as the replacement of image intensity and combined with the conventional similarity measure to reduce the influence of complicated radiometric difference on image registration [19,36]. The histogram of orientated phase congruency (HOPC) descriptor and the scene shape similarity feature descriptor were proposed in [9] and [37] respectively, and combined with NCC to achieve multimodal remote sensing image registration. In [38], a novel similarity measure was developed for optical-to-synthetic aperture radar (SAR) image matching as the NCC between dense rank-based local self-similarity descriptors. However, these similarity measures are somewhat computationally expensive or merely determine the subpixel measurement through simple polynomial fitting [39].

2.2. Phase Correlation

PC is a special area-based method calculated through frequency-domain operation. The theoretical basis of PC matching is the translation property of FT that links the shift of two relevant images in the spatial domain with the phase difference in the frequency domain. Assuming an image f ( x , y ) and another shifted image g ( x , y ) = f ( x Δ x , y Δ y ) , the normalized cross-power spectrum can be calculated by [40]:
Q ( u , v ) = f ( u , v ) * g ( u , v ) | f ( u , v ) * g ( u , v ) | = exp ( i ( u Δ x + v Δ y ) )
where f ( u , v ) and g ( u , v ) are the corresponding frequency representations of two images after FT, i is the first solution to the equation i 2 = 1 , and * denotes the complex conjugate. The correlation function of PC is derived as the inverse FT of the normalized cross-power spectrum. In the ideal case of integer shifts, this correlation function corresponds to a Dirac delta function centered at ( Δ x , Δ y ) . Accordingly, the pixel-level results of PC can be obtained by locating the peak of the correlation function.
In the case of subpixel shifts, the signal power of PC is not concentrated in a single peak, and leads to a downsampled 2-D Dirichlet kernel [40]. The existing subpixel PC methods can be found in two categories [20]. The first category is implemented in the spatial domain. The objective is to determine the fractional peak location of the correlation function with maximum correlation value, similar to the pixel-level matching. This can be achieved through similarity fitting with a certain set of neighbors using analytical derivations [40] or empirical fitting models [41], as well as upsampling the correlation function to a desired resolution in the frequency domain [42]. These methods have been successfully applied in the fine registration of multisensor remote sensing images [10,43], but they are vulnerable to the actual noise and aliasing.
The second category is realized in the frequency domain, which relies on the phase difference between two images, which is defined as the phase angle of the complex normalized cross-power spectrum. According to Euler’s formula, the phase difference can be expressed by:
φ ( u , v ) = Q ( u , v ) = ( u Δ x + v Δ y ) .
It can be found that the phase difference is a linear function of the shift vector, and the shifts can thus be estimated from the slope of phase difference. In this case, subpixel PC methods in the frequency domain are calculated by plane fitting [44,45], line fitting [46,47], or nonlinear optimization [48] with the linear phase difference between images. Note that the phase difference is 2π wrapped when dealing with discrete image signals, and phase unwrapping is needed in practice when estimating the shifts greater than 0.5 pixels. Due to avoiding the inverse FT process and relying on a theoretical expression, the second category usually has advantages in matching accuracy and robustness over the first category [49].

3. Enhanced Subpixel Phase Correlation

3.1. Workflow of the Enhanced Subpixel Method

The proposed subpixel PC method calculates the translation parameters in the frequency domain by means of the phase difference between input images. The overall workflow of the proposed method, which mainly consists of four steps, is depicted in Figure 1 and introduced in the following.
(1)
Construction of phase congruency-based structural representation. In order to minimize the influence of complicated intensity differences and emphasize the useful structural information for matching, we adopt the phase congruency [50] to generate a complex structural representation. The magnitude and orientation of the phase congruency features are combined to replace the original image intensity for the following image matching.
(2)
Calculation of normalized cross-power spectrum. The structural representations are transferred to the frequency domain using discrete FT. However, the periodicity of discrete FT induces the edge effect that affects the performance of PC. Therefore, we use an image decomposition algorithm [51] to extract the periodic component to eliminate the edge effects. Compared with the conventional windowing operation, this decomposition avoids narrowing the effective matching region and loss of image information [52]. The normalized cross-power spectrum matrix Q is then calculated as Equation (1).
(3)
Frequency masking and rank-one matrix approximation. In uncontrolled conditions, noise, aliasing, and other interference factors will contaminate the spectral components and degrade the following rank-one approximation and line fitting processing. In this case, we apply an adaptive frequency masking operation to filter out the corrupted frequency components [48]. Subsequently, two 1-D column vectors are factorized from the normalized cross-power spectrum matrix by determining the best rank-one approximation using a low-rank matrix approximation algorithm [53] which is robust to missing masked data and outliers.
(4)
Estimation of translation parameters. With the low-rank vectors obtained, the phase difference is separately extracted in each dimension after 1-D phase unwrapping. The correct slopes ( s x , s y ) of the unwrapped phase angles are identified by a robust estimation algorithm using higher than minimal subset sampling [54] in the presence of residual outliers, and finally converted to the results of translation parameters according to Δ x = s x M / 2 π , Δ y = s y N / 2 π , where M and N denote the size of the input images.

3.2. Details of the Enhanced Subpixel Method

To ensure the high accuracy and robustness, the enhanced subpixel PC method additionally integrates phase congruency-based structural representation, robust rank-one matrix approximation with adaptive frequency masking, and stable robust line fitting. All of these operations aim to guarantee that the practical phase difference calculated in tie point matching better agrees with the theoretical model in Equation (2).

3.2.1. Phase Congruency-Based Structural Representation

Although PC is insensitive to image content and intensity changes to some extent since it relates solely to phase information, the complicated radiometric changes can still deteriorate the linear relationship of the phase difference between input images [55]. The illumination robustness can be improved by constructing a structural representation combining the magnitude and orientation of phase congruency [56]. Phase congruency is a feature measure based on local frequency analysis, which perceives the corner and edge features where the Fourier components are maximal in phase. Phase congruency conforms to the human visual perception of image features, and has been widely applied in multimodal registration and matching [9,11,36]. By convolving a 2-D image f ( x , y ) through log-Gabor filters over several scales and orientations, the magnitude A n o and phase ϕ n o of the filter responses at a scale n and orientation o are given by:
A n = e n ( x , y ) 2 + o n ( x , y ) 2 ϕ n = atan 2 ( e n ( x , y ) , o n ( x , y ) ) [ e n ( x , y ) , o n ( x , y ) ] = [ f ( x , y ) M n e , f ( x , y ) M n o ]
where M n e and M n o denote the log Gabor even-symmetric and odd-symmetric wavelets that are the real and imaginary components of log-Gabor filters, respectively, e n o and o n o denote the convolution results of these two wavelets. The magnitude of the phase congruency can be expressed as [50]:
P C ( x , y ) = o n W o ( x , y ) A n o ( x , y ) Δ Φ n o ( x , y ) T o n A n o ( x , y ) + ε Δ Φ n o ( x , y ) = cos ( ϕ n o ( x , y ) ϕ ¯ ( x , y ) ) | sin ( ϕ n o ( x , y ) ϕ ¯ ( x , y ) ) |
where ϕ ¯ is the mean phase, W is a weighting term based on the frequency spread, T is a noise threshold, ε is a small constant and the symbol denotes that the enclosed quantity is equal to itself when its value is positive, or is zero otherwise. The orientation of the phase congruency can be calculated using the log Gabor odd-symmetric wavelets of multiple directions, which is expressed as:
Φ ( x , y ) = atan 2 ( θ ( o n o ( x , y ) sin ( θ ) ) , θ ( o n o ( x , y ) cos ( θ ) ) ) ,
where θ is the orientation angle. Then, the phase congruency-based structural representation is constructed as:
R P C ( x , y ) = P C ( x , y ) cos ( Φ ( x , y ) ) + i P C ( x , y ) sin ( Φ ( x , y ) ) .
The following subpixel PC is performed on the complex structural representations of both images instead of the original intensity. Both phase congruency and PC matching take advantage of the phase information of the image and are independent of magnitude information. Phase congruency relies on the local phase of images to preserve local topological information, while PC matching relies on the global phase difference to estimate the translation and similarity between images. Therefore, PC matching with phase congruency-based representations combines the global and local phase information to underline the frequency response of structural features and improve the robustness to local radiometric differences for translation estimation.

3.2.2. Robust Rank-One Matrix Approximation with Adaptive Frequency Masking

According to the expression in Equation (1), the normalized cross-power spectrum matrix Q is theoretically a rank-one matrix [46], i.e., Q = q x q y T , where qx and qy are complex column vectors. This implies that the 2-D translation estimation can be converted to two separate 1-D problems by finding the dominant rank-one subspace of Q. The most straightforward way is to use the singular value decomposition algorithm. However, the corrupted spectral components caused by noise, aliasing, and other interference factors in practice will potentially bias the ideal rank-one computation and the final estimation results [48,57]. Therefore, an effective frequency masking operation to remove the corrupted components and a robust low-rank matrix approximation algorithm to deal with missing data and outliers are adopted.
Since the high frequencies and the frequencies with small spectral magnitude that are most likely to be corrupted, the masking operation firstly masks out the high-frequency components at each periphery (e.g., 15% as suggested in [44]) of Q. Then, the unreliable frequency components with small magnitude are identified according to the normalized log-spectrum [48]. Therefore, the frequency mask is defined as:
W ( u , v ) = { 0 , u < 0.15 M ; u > 0.85 M ; v < 0.15 N ; v > 0.85 N 0 , NLS ( u , v ) p mean { NLS ( u , v ) } 1 , others LS ( u , v ) = log 10 | f ( u , v ) * g ( u , v ) | NLS ( u , v ) = LS ( u , v ) max { LS ( u , v ) }
where p is a specific parameter, we fix p = 0.9 the same as [48] for all the experiments.
The robust rank-one approximation is formulated as an optimization problem based on L1-norm loss and nuclear-norm regularizer [53], which is able to effectively handle the masked data and residual outliers. The objective function is written as:
min q x , q y W ( Q q x q y T ) 1 + λ q x q y T ,
where λ is a balancing parameter, W is the frequency masking matrix, the operator denotes the element-by-element matrix product, the symbol 1 denotes L1-norm, and * denotes nuclear-norm which is defined as the sum of singular values. The regularized optimization problem can be solved by an augmented Lagrange multiplier method. By introducing a matrix E = q x q y T and some constraints, Equation (8) becomes:
min E , q x , q y W ( Q E ) 1 + λ q y T * s . t . , E = q x q y T , q x T q x = 1
The unconstrained augmented Lagrange function after adding a penalty term and a Lagrange multiplier L is given by:
f ( E , q x , q y , L , μ ) = W ( Q E ) 1 + λ q y T * + μ 2 E q x q y T F 2 + L , E q x q y T ,
where μ is a penalty parameter, the symbol F denotes Frobenius norm, and A , B is equivalent to the trace of A T B . The complex column vectors can be solved by Gauss-Seidel iteration that iteratively solve one set of variables in E , q x , q y while fixing the other two with the Lagrange multiplier L and the penalty parameter μ updated in each iteration. More details of the optimization and implementation settings can be found in [53].

3.2.3. Stable Robust Line Fitting

To automatically exclude the corrupted phase angle values when fitting the slopes of the unwrapped phase angle vectors, a robust estimation algorithm using higher than minimal subset sampling (HMSS) [54] is introduced. Compared with the conventional random sample consensus algorithm [58], HMSS has two refinements: (1) it increases the initial sampling size beyond the minimal size to ensure the closeness of the hypothesis generation to the true model; (2) it is not a pure random sampling strategy, but a greedy strategy that starts from a random hypothesis and is iterated towards an optimized solution using the least k-th order statistic cost function until a stopping criterion is reached. These enable HMSS to achieve advantages on stability, accuracy, computational efficiency, and parameter insensitivity. The routine of our HMSS fitting is presented as follows. (1) For 2-D line fitting, five points (minimal size three plus two) from the unwrapped phase angle vectors are randomly selected to generate an initial model using a least-squares fitting. (2) In each iteration l, the residuals of all points are calculated and sorted, and the least k-th order statistic is calculated as a cost function:
F ( δ l ) = r i k , δ l 2 ( δ l ) ,
where r i 2 ( δ ) denotes the i-th squared residual regarding model δ , i k , δ denotes the index of k-th sorted squared residual regarding model δ , and k is an inlier parameter fixed at k = 0.2 T N , where TN is the total data number. The model of next iteration δ l + 1 is updated using the new five sample points around the k-th sorted square residual. (3) Iterations are continued until reaching a stopping criterion. The criterion is designed to check if the samples selected in consecutive iterations are from similar models, and is given by:
F s t o p = ( r i k , δ l 2 ( δ l ) < 1 5 j = k 5 + 1 k r i j , δ l 1 2 ( δ l ) ) ( r i k , δ l 2 ( δ l ) < 1 5 j = k 5 + 1 k r i j , δ l 2 2 ( δ l ) ) ,
where r i j , δ l 1 2 ( δ l ) and r i j , δ l 2 2 ( δ l ) denote the residuals of the sample points selected in two previous iterations with regard to the model δ l in iteration l. If the current cost function is lower than the average residuals of those sample points, the sample points selected in the last three iterations are likely to belong to the same structure and the iteration can stop. (4) To decrease the probability of accident erroneous estimation, steps (1)–(3) are repeated for reinitialization of random hypothesis generation until there is no decreasing of the cost function in consecutive runs. (5) The model with the minimal cost function is selected and refined using all the inliers by least-squares method, and the algorithm output the final slope.

4. Multisensor Fine Registration

The enhanced subpixel PC method can provide accurate and robust translation estimation as local template matching. In this section, an automatic registration method for precisely aligning coarsely coregistered remote sensing images from different sensors is extended based on the enhanced subpixel PC method. The flowchart of the fine registration method is illustrated in Figure 2, which is divided into four steps as follows.
(1)
Interest point extraction. To improve the localization performance in the presence of complicated radiometric conditions, phase-congruency corner detector is applied to detect the interest points on the reference image. According to Equation (4), we can obtain a phase congruency map. The moment analysis is performed on the phase congruency maps with different orientations, and the minimum moment m is given by [59]:
m = 1 2 ( c + a b 2 + ( a c ) 2 ) a = o ( P C ( θ ) cos ( θ ) ) 2 b = 2 o ( P C ( θ ) cos ( θ ) ) ( P C ( θ ) sin ( θ ) ) c = o ( P C ( θ ) sin ( θ ) ) 2
where P C ( θ ) is the phase congruency value determined at orientation angle θ . The minimum moment is equivalent to the cornerness measure. In order to extract the interest points uniformly distributed over the scene, a block-based strategy is adopted [19]. The image is partitioned into s × s nonoverlapping blocks, and the top h points with the largest minimum moment values are regarded as the interest points for each block.
(2)
Tie point matching. The corresponding points on the sensed image are determined by PC-based template matching. A template window is selected surrounding each interest point. The translation parameters between template windows are estimated by the pixel-level PC matching and then refined using the enhanced subpixel PC method. Note that the phase congruency calculated in the last step can be reused in the subpixel PC matching.
(3)
Mismatch elimination. There inevitably exist false matches in the results of tie point matching due to shadow and featureless areas. These mismatched tie points can be filtered by considering two aspects: the similarity measure and geometric consistency. The peak value of PC function provides a measure to assess the correctness of the match. The unreliable measurements with small PC peak values are firstly removed. Then, the residual outliers are eliminated by an iterative consistency check of tie points based on a global transformation [19]. In each iteration of consistency check, a transformation model is estimated using all the tie points with the transformation residuals calculated. The tie point with the largest residual is excluded, and the transformation model is estimated again on the remaining points. The procedure is repeated until the largest residual is less than a given threshold (e.g., 1.5 pixels). The three-order polynomial model is selected in this study since it can better handle the local deformations resulted from sensor error and terrain relief especially for high-resolution images.
(4)
Image warping. With the refined tie points, a transformation model that maps the sensed image to the reference image can be determined. We employ a piecewise linear model that is known to be appropriate for mitigating local geometric distortions between satellite images [60]. This function divides the image into triangular regions by the Delaunay’s triangulation algorithm using all the tie points, and estimates an affine transformation for each triangular region. For warping the regions outside the convex hull of the points, we estimate a global transformation model from the points defining the convex hull [61].

5. Experiments and Discussion

In order to verify the effectiveness of the proposed method, experiments were conducted in two parts, a tie point matching experiment and a fine registration experiment. The tie point matching experiment evaluates the matching performance of the enhanced subpixel PC method, and the fine registration experiment analyzes the alignment performance of the presented registration method based on the enhanced PC matching.

5.1. Tie Point Matching Experiment

5.1.1. Experimental Details

In this experiment, the enhanced subpixel PC method was assessed and compared with other PC methods and area-based matching methods. The block-based phase congruency detector was first applied to extract 400 interest points (top four points in each 10 × 10 nonoverlapping blocks) uniformly distributed over the reference image, whose corresponding points were then determined by template matching. The results obtained from the proposed method were compared with those from five state-of-the-art Fourier-based correlation methods including PC with quadratic fitting (PC_QF), Foroosh’s method [40], upsampled cross correlation (UCC) [42], Hoge’s method [46], and SVD-RANSAC (singular value decomposition-random sample consensus) [49], as well as five other representative area-based matching methods including NCC [25], MI [27], MTM [30], HOPCncc (NCC of the HOPC descriptors) [9] and enhanced correlation coefficient (ECC) [62]. PC_QF, NCC and MI are available in MATLAB; the codes of UCC, MTM, HOPCncc, and ECC are provided by the authors, and the others are our re-implementations. For the Fourier-based correlation methods, the image decomposition algorithm was adopted to mitigate the influence of edge effects. For PC_QF, NCC, MTM, and HOPCncc, the subpixel measurements were obtained by fitting the similarity function using a quadratic polynomial model. Three sizes of template windows (40 × 40, 60 × 60, 80 × 80 pixels) were tested to analyze the matching performance under different template sizes, and the size of search region was set as 20 × 20 pixels.
Three sets of remote sensing image pairs acquired from different satellites and sensors were used. The basic information about these multisensor images are given in Table 1. Each image pair contains a reference image (upper) and a sensed image (lower) captured by different sensors with diverse spatial resolution and imaging data. All these image pairs have been coarsely registered based on the metadata and georeferencing information, and resampled to the same ground sampling distance. For the image pair with large deviation due to the sensor positioning error (e.g., approximate 70 pixels for Data 1), the global translations between images were compensated using the pixel-level PC with inputs of the entire image. Therefore, the test image pairs are free of obvious scale, rotation, and translation differences, but still show significant intensity and textural changes due to various resolution, imaging time, and spectrum.
For each test data, 40–50 evenly distributed check points were manually selected from the reference and sensed images, and a three-order polynomial model can be estimated from the check points. The matching errors of tie points were then measured according to this transformation model, and the correct matches were identified as the tie points with matching errors smaller than a threshold. This threshold was set as 1 pixel for Data 2 and Data 3, and set as 1.5 pixels for Data 1 because of more severe local distortions and higher spatial resolution. The evaluation criteria used in this experiment include the precision and root mean square error (RMSE) of tie points. The precision refers to the correct match rate calculated as the number of correct matches divided by the total number of matches. The RMSE between transformed points and matched points was calculated from both the correct matches and total matches to evaluate the matching accuracy and stability.

5.1.2. Results and Discussion

Figure 3 displays the tie points achieved by the block-based phase congruency detector and the enhanced subpixel PC method. It can be seen that the image pair represents significant radiometric and textural differences. The enhanced subpixel PC method is able to identify enough well-distributed tie points in multisensor remote sensing images, and the locations of tie points correctly correspond to each other. This will be beneficial for the following multisensor fine registration.
Since the proposed method embeds three additional operations, first the performance gain of each individual operation was demonstrated. Besides the baseline of Hoge’s method and the final proposed method, two variants were also evaluated using Data 1: Variant 1 combines Hoge’s method with structural representation; Variant 2 combines Variant 1 with robust model fitting; and the final proposed method further embeds robust masked rank-one matrix approximation. The precisions and RMSEs of the total matches of these four methods are shown in Table 2. It can be seen that the matching performance gradually improves from the baseline method to Variant 1, Variant 2, and the final proposed method by integrating different additional operation. This indicates that the phase congruency-based structural representation, robust masked rank-one matrix approximation and stable robust model fitting are all effective to enhance the matching accuracy and robustness.
The comparative results of different template matching methods in terms of matching precision are shown in Figure 4, Figure 5 and Figure 6 for three test data, respectively, and the RMSEs of the correct matches and total matches of various matching methods with three different template sizes are presented in Table 3. As can be seen from the figures and table, the enhanced subpixel PC method, SVD-RANSAC, and HOPCncc generate the overall best results, achieving higher values of matching precision and lower RMSEs. The performances of other methods are negatively affected by the complicated radiometric and textural changes in multisensor images, which are manifested by more false matches and inferior matching accuracy in the results. With regard to the RMSEs of the correct matches, the proposed method reaches the lowest values for Data 2 and 3, but is not obviously better for Data 1. The possible explanation for Data 1 is due to the severe local distortions. Since the correct matches are identified using the manual check points by thresholding, the RMSEs of the correct matches will be close to the accuracy of check points in the case of severe local distortions, and are less dominated by the accuracy of matching algorithms. Compared with other line fitting-based PC methods, such as Hoge’s, SVD-RANSAC, and other Fourier-based correlation methods, the proposed method improves the accuracy and robustness of subpixel translation estimation by integrating phase congruency-based structural representation, L1-norm-based rank-one matrix approximation with frequency masking and robust model fitting using higher than minimal subset sampling. Based on the resistance of phase congruency to nonlinear intensity difference [9], HOPCncc obtains the comparable results in most cases. In general, the proposed method performs better than HOPCncc method. The improved correct match rate and subpixel capability benefit from the use of pixelwise structure representation and theoretical model based on the translation property of FT. The experimental results demonstrate the superiority and reliability of the proposed method in tie point matching of multisensor remote sensing images.
It can be found that the matching performance of all methods is related to the template sizes. The matching precision and accuracy become worse with the decreasing template sizes due to less structural information for matching. Frequency-based image correlation methods are hypothesized to be more limited in small template sizes following the Heisenberg’s uncertainty principle [20]. In this case, several obviously erroneous measurements exist in the matching results that affect the RMSEs. In addition, the local geometric distortions degrade the matching results. For Data 1 with higher resolution and larger geo-positioning errors, the success match ratio is significantly lower than for the other two datasets. Therefore, a potential refinement point of the proposed method in future work is to mitigate the influence of local geometric deformations, especially in the case of a small template.

5.2. Fine Registration Experiment

5.2.1. Experimental Details

In this experiment, the fine registration method presented was validated and compared with feature-based methods. Besides the above-mentioned HOPCncc and SVD-RANSAC methods, three state-of-the-art feature detectors and descriptors, namely SIFT [14], ORB (oriented features from accelerated segment test and rotated binary robust independent elementary features) [63], and RIFT (radiation-variation insensitive feature transform) [59] were used for comparison. SIFT detects keypoints based on the Difference-of-Gaussian scale space and generates a float-type descriptor for each feature based on the orientation of image gradient. ORB identifies keypoints using an oriented version of FAST corner detector and computes a binary-type feature vector using the rotation-aware BRIEF descriptor. RIFT extracts radiation-robust features based on phase congruency and log-Gabor convolution of different orientations. For the feature-based methods, the nearest neighbor distance ratio strategy [14] and random sample consensus algorithm [58] were adopted to eliminate the outliers in the matched features. For the area-based methods, the fine registration pipeline introduced in Section 4 was adopted, wherein a set of 600–700 evenly distributed tie points were extracted and matched based on the block-based phase congruency detector and the corresponding template matching methods. The piecewise linear transformation model was employed to warp the sensed image according to the tie points obtained by different methods.
As shown in Table 4, three sets of multisensor optical image pairs were tested in this experiment. These images range from very high resolution (submeter) to medium resolution (dozen of meters), and cover different scenes including urban and suburban areas. A temporal difference also exists between reference and sensed images with the maximum gap for more than three years. Similarity, all these image pairs have been preregistered though direct georeferencing and resampled to the same resolution of reference image to remove the obvious rotation, scale, and translation differences.
The distribution quality of tie points and final registration performance were assessed for all four methods. The distribution quality was measured by an index considering the area and shape of the triangles formed by tie points [64], which can be defined as:
D A = i = 1 t ( A i A ¯ 1 ) 2 t 1 , A ¯ = i = 1 t A i t D S = i = 1 t ( S i 1 ) 2 t 1 , S i = 3 max ( J i ) π D Q = D A D S = i = 1 t ( A i A ¯ 1 ) 2 i = 1 t ( S i 1 ) 2 t 1
where t is the number of triangles, A i denotes the area of the i-th triangle, and max ( J i ) denotes the largest internal angle of the i-th triangle. The smaller value of DQ indicates in Equation (14) the better distribution of tie points. To evaluate the quantitative registration performance, 40–50 evenly distributed check points were manually selected between each image pair, and the RMSE and standard deviation (STD) of the check points after registration was calculated.

5.2.2. Results and Discussion

In our registration method, the tie points were matched by the enhanced subpixel PC and filtered by the correlation values and iterative consistency check, and the warped sensed images were generated by the combination of piecewise linear functions and a global transformation. The registration results, including the Delaunay triangulations constructed from the filtered tie points and the chessboard images generated from the reference images and warped sensed images, are shown in Figure 7, and the enlarged subsets corresponding to the sample regions in the third column of Figure 7 are presented in Figure 8. It can be seen that the scenes accord well in two images after fine registration for all test cases with a simple visual inspection of the registration results, which qualitatively confirms the satisfactory registration performance of the presented method based on the enhanced PC matching.
In order to further validate the effectiveness, the number of matches, distribution quality index, RMSE, and STD of check points obtained from different methods are reported in Table 5. From the comparative results, it can be seen that the registration method presented significantly outperforms the other three feature-based methods in terms of distribution quality and registration accuracy. The presented method obtains tie points with the minimum value of DQ index indicating better distribution over the entire image. This is attributed to adopting the block-based strategy and limiting the search range, which is one of the advantages of area-based matching methods. The excellent distribution and high matching accuracy of tie points facilitate a good registration using a nonrigid piecewise linear model. Therefore, the presented method achieves a higher and more uniform registration accuracy with the minimum values of both RMSE and STD in the numerical analysis compared with the other three feature-based methods. Moreover, the presented method using the enhanced subpixel PC matching also obtains consistently lower values of RMSE and STD than the HOPCncc and SVD-RANSAC methods. It is worth noting that the RMSEs of three test data are all less than 1 pixel for our registration method, but grow with the decreasing spatial resolutions. The qualitative and quantitative analyses indicate that the presented method has the capability to offer an automatic and reliable solution to fine registration of multisensor remote sensing images.

6. Conclusions

In this paper, we propose an enhanced subpixel PC method and perform fine registration of multisensor remote sensing images based on this subpixel PC matching. The enhanced subpixel PC method achieves accurate and reliable template matching by adopting phase congruency-based structural representation, L1-norm-based rank-one matrix approximation with masking data, and stable robust model fitting. These operations ensure the calculated phase difference in practice better agree with the theoretical linear model based on the translation property of FT. The fine registration method combines the enhanced subpixel PC matching with block-based phase congruency feature detector, iterative consistency check, and image warping using piecewise linear transformation to precisely coregister the images from different satellites and sensors. Tie point matching and fine registration experiments were conducted, each using three sets of multisensor image pairs. In the tie point matching experiment, the enhanced subpixel PC method outperformed other state-of-the-art PC and area-based methods with a higher correct match rate and better matching accuracy. In the fine registration experiment, the proposed fine registration method outperformed state-of-the-art feature-based methods in terms of distribution quality and registration performance. The promising results indicate that the proposed method is robust and effective for multisensor fine registration.
Local deformation is an impact factor degrading the matching performance. The proposed method may be less effective in the presence of severe relief displacements, which is a common issue for high-resolution image registration. In future work, the proposed method will be refined to mitigate the influence of local deformation and utilize the prior knowledge from digital surface model and shadow map. In addition, this study mainly presents fine registration of multisensor optical remote sensing images, future works will explore the performance in more complicated multimodal applications.

Author Contributions

All authors contributed to this manuscript: conceptualization, Z.Y. and Y.X.; methodology and software, Z.Y., J.K. and J.Y.; experiment and analysis, Z.Y. and W.S.; data curation, W.S. and X.L.; writing—original draft preparation, Z.Y.; writing—review and editing, J.K., S.L. and Y.X.; supervision and funding acquisition, J.K. and X.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the German Research Foundation (DFG) and the Technische Universität Berlin within the funding program Open Access Publishing; the National Key Research and Development Program of China, grant number 2018YFB0505000 and 2018YFB0505400.

Acknowledgments

The authors would like to thank Yuanxin Ye, Jiayuan Li, Manuel Guizar-Sicairos, Georgios Evangelidis, Yacov Hel-Or and VLFeat team for providing their codes and software.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zitova, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef] [Green Version]
  2. Belward, A.S.; Skøien, J.O. Who launched what, when and why; trends in global land-cover observation capacity from civilian earth observation satellites. ISPRS J. Photogramm. Remote Sens. 2015, 103, 115–128. [Google Scholar] [CrossRef]
  3. Klaric, M.N.; Claywell, B.C.; Scott, G.J.; Hudson, N.J.; Sjahputera, O.; Li, Y.; Barratt, S.T.; Keller, J.M.; Davis, C.H. GeoCDX: An automated change detection and exploitation system for high-resolution satellite imagery. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2067–2086. [Google Scholar] [CrossRef]
  4. Chen, S.; Yuan, X.; Yuan, W.; Niu, J.; Xu, F.; Zhang, Y. Matching multi-sensor remote sensing images via an affinity tensor. Remote Sens. 2018, 10, 1104. [Google Scholar] [CrossRef] [Green Version]
  5. Stumpf, A.; Michéa, D.; Malet, J.-P. Improved co-registration of Sentinel-2 and Landsat-8 imagery for earth surface motion measurements. Remote Sens. 2018, 10, 160. [Google Scholar] [CrossRef] [Green Version]
  6. Han, Y.; Choi, J.; Byun, Y.; Kim, Y. Parameter optimization for the extraction of matching points between high-resolution multisensor images in urban areas. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5612–5621. [Google Scholar] [CrossRef]
  7. Gong, M.; Zhao, S.; Jiao, L.; Tian, D.; Wang, S. A novel coarse-to-fine scheme for automatic image registration based on SIFT and mutual information. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4328–4338. [Google Scholar] [CrossRef]
  8. Li, H.; Man, Y. Robust multi-source image registration for optical satellite based on phase information. Photogramm. Eng. Remote Sens. 2016, 82, 865–878. [Google Scholar] [CrossRef]
  9. Ye, Y.; Shan, J.; Bruzzone, L.; Shen, L. Robust registration of multimodal remote sensing images based on structural similarity. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2941–2958. [Google Scholar] [CrossRef]
  10. Scheffler, D.; Hollstein, A.; Diedrich, H.; Segl, K.; Hostert, P. AROSICS: An automated and robust open-source image co-registration software for multi-sensor satellite data. Remote Sens. 2017, 9, 676. [Google Scholar] [CrossRef] [Green Version]
  11. Wong, A.; Clausi, D.A. ARRSI: Automatic registration of remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1483–1493. [Google Scholar] [CrossRef]
  12. Han, Y.; Bovolo, F.; Bruzzone, L. An approach to fine coregistration between very high resolution multispectral images based on registration noise distribution. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6650–6662. [Google Scholar] [CrossRef]
  13. Hu, H.; Ding, Y.; Zhu, Q.; Wu, B.; Xie, L.; Chen, M. Stable least-squares matching for oblique images using bound constrained optimization and a robust loss function. ISPRS J. Photogramm. Remote Sens. 2016, 118, 53–67. [Google Scholar] [CrossRef]
  14. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  15. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  16. Sedaghat, A.; Ebadi, H. Remote sensing image matching based on adaptive binning SIFT descriptor. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5283–5293. [Google Scholar] [CrossRef]
  17. Bellavia, F.; Colombo, C. Is there anything new to say about SIFT matching? Int. J. Comput. Vis. 2020, 128, 1847–1866. [Google Scholar] [CrossRef] [Green Version]
  18. Ye, Z.; Xu, Y.; Chen, H.; Zhu, J.; Tong, X.; Stilla, U. Area-based dense image matching with subpixel accuracy for remote sensing applications: Practical analysis and comparative study. Remote Sens. 2020, 12, 696. [Google Scholar] [CrossRef] [Green Version]
  19. Ye, Y.; Shan, J. A local descriptor based registration method for multispectral remote sensing images with non-linear intensity differences. ISPRS J. Photogramm. Remote Sens. 2014, 90, 83–95. [Google Scholar] [CrossRef]
  20. Tong, X.; Ye, Z.; Xu, Y.; Gao, S.; Xie, H.; Du, Q.; Liu, S.; Xu, X.; Liu, S.; Luan, K.; et al. Image registration with Fourier-based image correlation: A comprehensive review of developments and applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4062–4081. [Google Scholar] [CrossRef]
  21. Reddy, B.S.; Chatterji, B.N. An FFT-based technique for translation, rotation, and scale-invariant image registration. IEEE Trans. Image Process. 1996, 5, 1266–1271. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Yang, H.; Li, X.; Zhao, L.; Chen, S. A novel coarse-to-fine scheme for remote sensing image registration based on SIFT and phase correlation. Remote Sens. 2019, 11, 1833. [Google Scholar] [CrossRef] [Green Version]
  23. Dong, Y.; Jiao, W.; Long, T.; He, G.; Gong, C. An extension of phase correlation-based image registration to estimate similarity transform using multiple polar Fourier transform. Remote Sens. 2018, 10, 1719. [Google Scholar] [CrossRef] [Green Version]
  24. Feng, R.; Du, Q.; Li, X.; Shen, H. Robust registration for remote sensing images by combining and localizing feature- and area-based methods. ISPRS J. Photogramm. Remote Sens. 2019, 151, 15–26. [Google Scholar] [CrossRef]
  25. Lewis, J.P. Fast template matching. In Proceedings of the Vision Interface, Quebec City, QC, Canada, 15–19 May 1995; pp. 120–123. [Google Scholar]
  26. Le Moigne, J.; Netanyahu, N.S.; Eastman, R.D. Image Registration for Remote Sensing; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  27. Viola, P.; Wells, W.M., III. Alignment by maximization of mutual information. Int. J. Comput. Vis. 1997, 24, 137–154. [Google Scholar] [CrossRef]
  28. Wang, F.; Vemuri, B.C. Non-rigid multi-modal image registration using cross-cumulative residual entropy. Int. J. Comput. Vis. 2007, 74, 201–215. [Google Scholar] [CrossRef] [Green Version]
  29. Xu, X.; Li, X.; Liu, X.; Shen, H.; Shi, Q. Multimodal registration of remotely sensed images based on Jeffrey’s divergence. ISPRS J. Photogramm. Remote Sens. 2016, 122, 97–115. [Google Scholar] [CrossRef]
  30. Hel-Or, Y.; Hel-Or, H.; David, E. Matching by tone mapping: Photometric invariant template matching. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 317–330. [Google Scholar] [CrossRef]
  31. Suri, S.; Reinartz, P. Mutual-information-based registration of TerraSAR-X and Ikonos imagery in urban areas. IEEE Trans. Geosci. Remote Sens. 2010, 48, 939–949. [Google Scholar] [CrossRef]
  32. Hasan, M.; Pickering, M.R.; Jia, X. Robust automatic registration of multimodal satellite images using CCRE with partial volume interpolation. IEEE Trans. Geosci. Remote Sens. 2012, 50, 4050–4061. [Google Scholar] [CrossRef]
  33. Li, K.; Zhang, Y.; Zhang, Z.; Lai, G. A coarse-to-fine registration strategy for multi-sensor images with large resolution differences. Remote Sens. 2019, 11, 470. [Google Scholar] [CrossRef] [Green Version]
  34. Chen, S.; Li, X.; Zhao, L.; Yang, H. Medium-low resolution multisource remote sensing image registration based on SIFT and robust regional mutual information. Int. J. Remote Sens. 2018, 39, 3215–3242. [Google Scholar] [CrossRef]
  35. Lee, J.; Cai, X.; Schönlieb, C.; Coomes, D.A. Nonparametric image registration of airborne LiDAR, hyperspectral and photographic imagery of wooded landscapes. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6073–6084. [Google Scholar] [CrossRef]
  36. Fan, J.; Wu, Y.; Li, M.; Liang, W.; Cao, Y. SAR and optical image registration using nonlinear diffusion and phase congruency structural descriptor. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5368–5379. [Google Scholar] [CrossRef]
  37. Hao, M.; Jin, J.; Zhou, M.; Tian, Y.; Shi, W. Robust multisource remote sensing image registration method based on scene shape similarity. Photogramm. Eng. Remote Sens. 2019, 85, 725–736. [Google Scholar] [CrossRef]
  38. Xiong, X.; Xu, Q.; Jin, G.; Zhang, H.; Gao, X. Rank-based local self-similarity descriptor for optical-to-SAR image matching. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  39. Ye, Y.; Bruzzone, L.; Shan, J.; Bovolo, F.; Zhu, Q. Fast and robust matching for multimodal remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9059–9070. [Google Scholar] [CrossRef] [Green Version]
  40. Foroosh, H.; Zerubia, J.B.; Berthod, M. Extension of phase correlation to subpixel registration. IEEE Trans. Image Process. 2002, 11, 188–200. [Google Scholar] [CrossRef] [Green Version]
  41. Alba, A.; Vigueras-Gomez, J.F.; Arce-Santana, E.R.; Aguilar-Ponce, R.M. Phase correlation with sub-pixel accuracy: A comparative study in 1D and 2D. Comput. Vis. Image Understand. 2015, 137, 76–87. [Google Scholar] [CrossRef]
  42. Guizar-Sicairos, M.; Thurman, S.T.; Fienup, J.R. Efficient subpixel image registration algorithms. Opt. Lett. 2008, 33, 156–158. [Google Scholar] [CrossRef] [Green Version]
  43. Skakun, S.; Roger, J.-C.; Vermote, E.F.; Masek, J.G.; Justice, C.O. Automatic sub-pixel co-registration of Landsat-8 Operational Land Imager and Sentinel-2A Multi-Spectral Instrument images using phase correlation and machine learning based mapping. Int. J. Digit. Earth 2017, 10, 1253–1269. [Google Scholar] [CrossRef] [PubMed]
  44. Stone, H.S.; Orchard, M.T.; Chang, E.-C.; Martucci, S.A. A fast direct Fourier-based algorithm for subpixel registration of images. IEEE Trans. Geosci. Remote Sens. 2001, 39, 2235–2243. [Google Scholar] [CrossRef] [Green Version]
  45. Ye, Z.; Tong, X.; Xu, Y.; Gao, S.; Liu, S.; Xie, H.; Lu, W.; Liu, X. An improved subpixel phase correlation method with application in videogrammetric monitoring of shaking table tests. Photogramm. Eng. Remote Sens. 2018, 84, 579–592. [Google Scholar] [CrossRef]
  46. Hoge, W.S. A subspace identification extension to the phase correlation method [MRI application]. IEEE Trans. Med. Imag. 2003, 22, 277–280. [Google Scholar] [CrossRef]
  47. Chen, Z.; Liu, B.; Wang, S.; Liu, E. Efficient subpixel registration for polarization-modulated 3D imaging. Opt. Express 2018, 26, 23040–23050. [Google Scholar] [CrossRef]
  48. Leprince, S.; Barbot, S.; Ayoub, F.; Avouac, J.-P. Automatic and precise orthorectification, coregistration, and subpixel correlation of satellite images, application to ground deformation measurements. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1529–1558. [Google Scholar] [CrossRef] [Green Version]
  49. Tong, X.; Ye, Z.; Xu, Y.; Liu, S.; Li, L.; Xie, H.; Li, T. A novel subpixel phase correlation method using singular value decomposition and unified random sample consensus. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4143–4156. [Google Scholar] [CrossRef]
  50. Kovesi, P. Image features from phase congruency. Videre J. Comp. Vis. Res. 1999, 1, 1–26. [Google Scholar]
  51. Moisan, L. Periodic plus smooth image decomposition. J. Math. Imaging Vis. 2011, 39, 161–179. [Google Scholar] [CrossRef] [Green Version]
  52. Dong, Y.; Jiao, W.; Long, T.; Liu, L.; He, G. Eliminating the effect of image border with image periodic decomposition for phase correlation based remote sensing image registration. Sensors 2019, 19, 2329. [Google Scholar] [CrossRef] [Green Version]
  53. Zheng, Y.; Liu, G.; Sugimoto, S.; Yan, S.; Okutomi, M. Practical low-rank matrix approximation under robust L1-norm. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1410–1417. [Google Scholar]
  54. Tennakoon, R.B.; Bab-Hadiashar, A.; Cao, Z.; Hoseinnezhad, R.; Suter, D. Robust model fitting using higher than minimal subset sampling. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 350–362. [Google Scholar] [CrossRef] [PubMed]
  55. Wan, X.; Liu, J.G.; Li, S.; Yan, H. Phase correlation decomposition: The impact of illumination variation for robust subpixel remotely sensed image matching. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6710–6725. [Google Scholar] [CrossRef]
  56. Ye, Z.; Tong, X.; Zheng, S.; Guo, C.; Gao, S.; Liu, S.; Xu, X.; Jin, Y.; Xie, H.; Liu, S.; et al. Illumination-robust subpixel Fourier-based image correlation methods based on phase congruency. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1995–2008. [Google Scholar] [CrossRef]
  57. Dong, Y.; Long, T.; Jiao, W.; He, G.; Zhang, Z. A novel image registration method based on phase correlation using low-rank matrix factorization with mixture of Gaussian. IEEE Trans. Geosci. Remote Sens. 2018, 56, 446–460. [Google Scholar] [CrossRef]
  58. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  59. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-modal image matching based on radiation-variation insensitive feature transform. IEEE Trans. Image Process. 2020, 29, 3296–3310. [Google Scholar] [CrossRef]
  60. Han, Y.; Kim, T.; Yeom, J. Improved piecewise linear transformation for precise warping of very-high-resolution remote sensing images. Remote Sens. 2019, 11, 2235. [Google Scholar] [CrossRef] [Green Version]
  61. Han, Y.; Byun, Y.; Choi, J.; Han, D.; Kim, Y. Automatic registration of high-resolution images using local properties of features. Photogramm. Eng. Remote Sens. 2012, 78, 211–221. [Google Scholar] [CrossRef]
  62. Evangelidis, G.D.; Psarakis, E.Z. Parametric image alignment using enhanced correlation coefficient maximization. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1858–1865. [Google Scholar] [CrossRef] [Green Version]
  63. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  64. Zhu, Q.; Wu, B.; Xu, Z.-X. Seed point selection method for triangle constrained image matching propagation. IEEE Geosci. Remote Sens. Lett. 2006, 3, 207–211. [Google Scholar] [CrossRef]
Figure 1. Overall workflow of the enhanced subpixel phase correlation method.
Figure 1. Overall workflow of the enhanced subpixel phase correlation method.
Sensors 20 04338 g001
Figure 2. Flowchart of the fine registration method based on the enhanced phase correlation (PC) matching.
Figure 2. Flowchart of the fine registration method based on the enhanced phase correlation (PC) matching.
Sensors 20 04338 g002
Figure 3. Tie points achieved by the enhanced subpixel PC method with the template size of 80 × 80 pixels for three test image pair. (a) Data 1; (b) Data 2; and (c) Data 3.
Figure 3. Tie points achieved by the enhanced subpixel PC method with the template size of 80 × 80 pixels for three test image pair. (a) Data 1; (b) Data 2; and (c) Data 3.
Sensors 20 04338 g003
Figure 4. Precision values of different methods for Data 1. NCC, normalized cross correlation; MI, mutual information; MTM, matching by tone matching; HOPCncc, NCC of the HOPC descriptors; ECC, enhanced correlation coefficient; PC_QF, PC with quadratic fitting; UCC, upsampled cross correlation; SVD-RANSAC, singular value decomposition-random sample consensus.
Figure 4. Precision values of different methods for Data 1. NCC, normalized cross correlation; MI, mutual information; MTM, matching by tone matching; HOPCncc, NCC of the HOPC descriptors; ECC, enhanced correlation coefficient; PC_QF, PC with quadratic fitting; UCC, upsampled cross correlation; SVD-RANSAC, singular value decomposition-random sample consensus.
Sensors 20 04338 g004
Figure 5. Precision values of different methods for Data 2.
Figure 5. Precision values of different methods for Data 2.
Sensors 20 04338 g005
Figure 6. Precision values of different methods for Data 3.
Figure 6. Precision values of different methods for Data 3.
Sensors 20 04338 g006
Figure 7. Registration results of the presented method for three test image pair. (a) Data 1; (b) Data 2; and (c) Data 3.
Figure 7. Registration results of the presented method for three test image pair. (a) Data 1; (b) Data 2; and (c) Data 3.
Sensors 20 04338 g007
Figure 8. Enlarged subsets of the reference images and warped sensed images corresponding to the boxes I, II, and III in Figure 7.
Figure 8. Enlarged subsets of the reference images and warped sensed images corresponding to the boxes I, II, and III in Figure 7.
Sensors 20 04338 g008
Table 1. Basic information about the images used in the tie point matching experiment.
Table 1. Basic information about the images used in the tie point matching experiment.
Data No.Image SourcesSizeSensor ResolutionDateLocation
1ZiYuan-3 PAN1920 × 19802.1 m2012/02Dengfeng, Henan, China
THEOS PAN1990 × 19922 m2011/12
2Sentinel-2 MSI Band 31800 × 180010 m2015/08Munich, Germany
Landsat 8 OLI Band 81805 × 180515 m2014/06
3Mapping-1 PAN1720 × 17205 m2013/05Dengfeng, Henan, China
ZiYuan-3 MUX Band 31725 × 17255.8 m2012/02
Table 2. Matching performance of the baseline, two variants and the proposed method (root mean square error (RMSE) unit: pixels).
Table 2. Matching performance of the baseline, two variants and the proposed method (root mean square error (RMSE) unit: pixels).
CriterionHogeVariant 1Variant 2Proposed
40Precision53.38%56.14%63.16%64.66%
RMSE2.9282.6312.2452.147
60Precision60.9%63.91%68.92%70.43%
RMSE2.4362.1931.8131.693
80Precision63.25%66%70%71%
RMSE2.0321.9121.6411.607
Table 3. RMSEs of the correct matches (CM) and total matches (TM) of various matching methods with three different template sizes (unit: pixels).
Table 3. RMSEs of the correct matches (CM) and total matches (TM) of various matching methods with three different template sizes (unit: pixels).
No.Template SizeNCCMIMTMHOPCnccECCPC_QFForooshUCCHogeSVD-RANSACProposed
Data 140CM0.7560.7750.7620.7880.7590.8130.8000.7490.8020.7830.775
TM3.4773.7053.4462.2664.3443.6954.4084.4212.9282.3792.147
60CM0.7540.7380.7320.7500.7870.7830.7900.7490.7650.7500.755
TM2.9772.9072.2561.8424.1352.8113.0363.5552.4361.8251.693
80CM0.7800.7350.7430.7520.7820.7570.7650.7520.7660.7530.748
TM2.2472.1862.1461.5913.6522.4832.1582.9882.0321.6591.607
Data 240CM0.4040.4290.4080.4240.4060.4310.4650.3910.3760.3920.385
TM3.4352.9702.5970.7323.9032.3402.6493.8751.8080.9140.822
60CM0.4050.4090.4010.4140.4130.4050.4620.3830.3710.3770.369
TM3.1542.7521.7110.4613.3522.1031.8463.7111.4970.7440.558
80CM0.3880.4010.3960.3770.4070.3970.4600.3630.3590.3590.350
TM2.4181.9981.3070.3833.2051.9771.3093.5870.9420.4570.358
Data 340CM0.4250.4690.4660.4920.4470.4560.4500.4210.4090.4340.389
TM1.6081.9833.0620.9981.6192.5402.6312.3452.1131.7601.218
60CM0.4080.4270.4100.4590.4000.4420.4300.3950.3800.4090.381
TM1.0440.9771.6030.5861.1271.9201.0911.5241.2150.6850.702
80CM0.3800.4140.3870.4420.3860.4180.4150.3760.3740.3820.361
TM0.6710.5660.5430.4981.0541.2010.5121.2480.8740.4700.396
Table 4. Basic information about the images used in the fine registration experiment.
Table 4. Basic information about the images used in the fine registration experiment.
Data No.Image SourcesSizeSensor ResolutionDateLocation
1SPOT-5 PAN1750 × 17005 m2013/06Zhangye, Gansu, China
Sentinel-2 MSI Band 31791 × 171610 m2015/08
2GeoEye-1 RGB1040 × 10102 m2010/02Shanghai, China
ZiYuan-3 PAN1044 × 10112.1 m2013/07
3Hongqi-1-H9 PAN2120 × 21400.75 m2020/02Shanghai, China
Google earth2124 × 21481.19 m2019/10
Table 5. Registration performance of different methods. TN, the number of total matches; RN, the number after outlier removal. The unit of RMSE and STD is pixels.
Table 5. Registration performance of different methods. TN, the number of total matches; RN, the number after outlier removal. The unit of RMSE and STD is pixels.
No.CriterionSIFTORBRIFTHOPCnccSVD-RANSACProposed
Data 1RN/TN1538/26891312/2121502/1403669/711657/711662/711
DQ2.6483.9111.42310.8410.8550.852
RMSE0.9180.8981.2270.5270.5200.494
STD0.4710.4220.5710.2900.2840.272
Data 2RN/TN178/865456/1306332/1040498/600486/600495/600
DQ1.6842.4371.0830.8820.8160.821
RMSE1.3611.4801.2200.6910.6860.642
STD0.6950.7970.5720.3410.3470.330
Data 3RN/TN56/84967/565165/907391/713378/713399/713
DQ1.6522.1271.1671.1521.1961.131
RMSE1.8941.8221.5380.7710.8090.711
STD0.7090.8370.6950.3750.3660.351

Share and Cite

MDPI and ACS Style

Ye, Z.; Kang, J.; Yao, J.; Song, W.; Liu, S.; Luo, X.; Xu, Y.; Tong, X. Robust Fine Registration of Multisensor Remote Sensing Images Based on Enhanced Subpixel Phase Correlation. Sensors 2020, 20, 4338. https://0-doi-org.brum.beds.ac.uk/10.3390/s20154338

AMA Style

Ye Z, Kang J, Yao J, Song W, Liu S, Luo X, Xu Y, Tong X. Robust Fine Registration of Multisensor Remote Sensing Images Based on Enhanced Subpixel Phase Correlation. Sensors. 2020; 20(15):4338. https://0-doi-org.brum.beds.ac.uk/10.3390/s20154338

Chicago/Turabian Style

Ye, Zhen, Jian Kang, Jing Yao, Wenping Song, Sicong Liu, Xin Luo, Yusheng Xu, and Xiaohua Tong. 2020. "Robust Fine Registration of Multisensor Remote Sensing Images Based on Enhanced Subpixel Phase Correlation" Sensors 20, no. 15: 4338. https://0-doi-org.brum.beds.ac.uk/10.3390/s20154338

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop