IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 1, JANUARY 2017
3
Remote Sensing Image Registration With Modified
SIFT and Enhanced Feature Matching
Wenping Ma, Zelian Wen, Yue Wu, Licheng Jiao, Senior Member, IEEE,
Maoguo Gong, Senior Member, IEEE, Yafei Zheng, and Liang Liu
Abstract—The scale-invariant feature transform algorithm and
its many variants are widely used in feature-based remote sensing
image registration. However, it may be difficult to find enough
correct correspondences for remote image pairs in some cases that
exhibit a significant difference in intensity mapping. In this letter,
a new gradient definition is introduced to overcome the difference
of image intensity between the remote image pairs. Then, an
enhanced feature matching method by combining the position,
scale, and orientation of each keypoint is introduced to increase
the number of correct correspondences. The proposed algorithm
is tested on multispectral and multisensor remote sensing images.
The experimental results show that the proposed method improves
the matching performance compared with several state-of-the-art
methods in terms of the number of correct correspondences and
aligning accuracy.
Index Terms—Feature matching, image registration, remote
sensing, scale-invariant feature transform (SIFT).
I. INTRODUCTION
I MAGE registration is the process of matching two or more
images of the same scene with different time, different
sensors, and different viewpoints [1]. It is an indispensable part
for many remote sensing tasks, such as change detection, image
fusion, and environmental monitoring.
A number of methods have been proposed for remote sensing
image registration. These methods can be coarsely partitioned
into two categories: intensity-based methods and feature-based
methods [1], [2]. Intensity-based methods use similarity be-
tween pixel intensities to determine the alignment between two
images. Mainly used similarity measures are cross correlation
and mutual information [2]. However, intensity-based methods
suffer from monotonous textures [3], illumination differences,
and a high degree of computational complexity of global op-
timization [4]. Feature-based methods extract salient features
and use the correlation between those features to determine
the optimal alignment. In general, these features include point,
edge, contour, the centroid of a specific region [5], [6], and
so on. Among the feature-based methods, the scale-invariant
feature transform (SIFT) [7] is the classic algorithm. SIFT is
invariant to image scaling and rotation and partially invariant to
change in illumination and camera viewpoint, and it has been
Manuscript received January 7, 2016; revised June 12, 2016 and July 27,
2016; accepted August 9, 2016. Date of publication December 5, 2016; date of
current version December 26, 2016.
The authors are with the Key Laboratory of Intelligent Perception and
Image Understanding of Ministry of Education, International Research Center
for Intelligent Perception and Computation, Xidian University, Xi’an 710071,
China (e-mail: wpma@mail.xidian.edu.cn; zelianwen@foxmail.com).
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/LGRS.2016.2600858
used successfully in registration of visible images. Some other
improvements have been made in feature-based methods, such
as SURF [8], GLOH [9], and BRISK [10]. These improvements
are mainly to improve the computational efficiency. However,
when SIFT is directly applied to remote sensing images, the
number of correct correspondences is not enough to confirm
matching accuracy [4] due to the significant difference in inten-
sity mapping. The intensity mapping may be linear, nonlinear,
and erratic. To overcome the problem, Li et al. [4] has proposed
robust SIFT (R-SIFT), in which the gradient orientation of
each pixel is refined, and more main orientations are assigned
to each keypoint. In the process of feature matching, the
scale–orientation joint restriction criterion is introduced to im-
prove the matching performance. Kupfer et al. [11] proposed a
fast mode-seeking SIFT (MS-SIFT) algorithm that exploits the
scale, orientation, and position information of SIFT features,
followed by the effective removal of imprecise SIFT keypoint
correspondences. Sedaghat et al. [3] proposed the uniform
R-SIFT (UR-SIFT) algorithm. The UR-SIFT method effec-
tively generates enough robust, reliable, and uniformly distrib-
uted aligned keypoints. Gong et al. [2] proposed a coarse-to-fine
scheme for automatic image registration. The coarse results
provide a near-optimal initial solution for the optimizer in the
fine-tuning process.
In this letter, we propose a new gradient definition to over-
come the difference of image intensity between the remote
image pairs. In addition, a robust point matching algorithm that
combines the position, scale, and orientation of each keypoint
to increase the number of correct correspondences is proposed.
This algorithm is inspired by the SIFT algorithm and will
be called PSO-SIFT. We assume a similarity transformation
model, which is widely used in the registration of remote
sensing images. In Section II, the outline of the classical SIFT
algorithm and its limitations on remotely sensed images are
discussed. The proposed algorithm in this letter is presented in
Section III. The experiment results on three different remote
sensing image pairs are illustrated in Section IV. Concluding
remarks are provided in Section V.
II. BACKGROUND
The SIFT-based registration algorithm consists of three main
modules: keypoint detection, descriptor extraction, and key-
point matching. A difference of Gaussian scale space, as an
approximation of the Laplacian of Gaussian, is constructed.
Local extrema in the three dimensions are then selected as can-
didate keypoints. Each keypoint is assigned one or more main
orientations based on a local histogram of gradient orientation.
Then, a 128-element descriptor is assigned to each keypoint.
The obtained SIFT feature comprises four components: local
1545-598X © 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
4
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 1, JANUARY 2017
Fig. 1. Remote sensing image pairs. (a) Landsat-7 ETM+, band 5. (b) Landsat
4-5 TM, band 3. (c) Landsat TM, band 5. (d) Google earth, optical. (e) ALOS-
PALSAR. (f) Landsat ETM+, band 5.
(xi, yi), scale si, orientation θi, and descriptor di. The last
module uses the minimum Euclidean distance on descriptors as
the keypoint matching criterion. More details about SIFT can
be found in [7].
Fig. 1(a) and (b) shows multispectral images that are ac-
quired in two different bands and different sensor devices.
Therefore, they have very different intensity mappings and are
typically more difficult to register. Fig. 2 shows the histograms
of the scale ratio, main orientation difference, horizontal shifts,
and vertical shifts of the matched original SIFT keypoints in
Fig. 1(a) and (b). The keypoints are matched by the ratio
between the Euclidean distance of the nearest neighbor and that
of the second nearest neighbor of corresponding descriptors.
The threshold on the ratio is set to dratio. Detailed parameter
setting of histograms can be found in [11]. We can see that the
four histograms cannot evidently exhibit a single mode except
for the scale ratio histograms, which, due to the difference of
intensity mapping between the multispectral images and the
same areas in multispectral remote images, could have a signifi-
cant nonlinear intensity difference [4]. Such intensity difference
will result in different main orientation difference of correspon-
dences that are expected to be correctly matched. Moreover, the
feature descriptors are not robust to these differences, because
the main orientation of each keypoint is used in the process of
descriptor extraction to ensure rotation invariance.
III. PROPOSED METHOD
A. New Gradient Definition
The significant difference of intensity mapping will result in
different gradient orientations and gradient magnitudes of the
same area between the remote sensing image pairs. Therefore,
the SIFT correspondences that are expected to be correctly
matched will not have a minimum Euclidean distance as the
correspondence computation depends on the descriptor that
is formed by gradient orientations and gradient magnitudes
around the keypoint location. To make the descriptor more
robust to such differences, we propose a new gradient defin-
ition (including orientation and magnitude) for each pixel in
Gaussian scale space. To improve the efficiency of the proposed
Fig. 2. (a)–(d) Histograms of scale ratio, main orientation difference, horizon-
tal shifts, and vertical shifts.
method, the input image pair is not expanded, and we use
the original input image to building the lowest level of the
Gaussian pyramid. First, we compute the gradient magnitude
of the Gaussian scale-space image by means of Sobel filters as
2
G1
σ =
G1
x,σ
2 +
G1
y,σ
(1)
where σ is the scale of Gaussian scale space, and G1
y,σ
denote the horizontal and vertical derivatives of the Gaussian
scale-space image with scale σ, respectively. Then, we define
the proposed gradient orientation and gradient magnitude as
x,σ and G1
R2
σ = arctan
G2
G2
y,σ
x,σ
, G2
σ =
G2
x,σ
2 +
G2
y,σ
2 (2)
x,σ and G2
where G2
y,σ denote the horizontal and vertical deriva-
tives of gradient magnitude image G1
σ of Gaussian scale space,
respectively. All the derivatives are approximated by means of
Sobel filters. Using the Sobel operator, the derivative [12], [13]
is easily computed. Note that the gradients computed by (2) are
used in the process of orientation assignment and descriptor
extraction. In this letter, we do not use a Gaussian weighting
of the gradient magnitudes when computing histograms [14].
Instead of using a square neighborhood and 4 × 4 square sectors
as in the original SIFT descriptor, we use a GLOH-like [9], [14]
circular neighborhood (radius of 12σ) and log-polar sectors
(17 location bins) to create a feature descriptor. A series of
experiments shows that GLOH obtains the best results [9]. Note
that the gradient orientations are quantized in eight bins. This
results in a 136-dimensional descriptor. Fig. 4(a) illustrates the
approach.
Fig. 3 shows the histograms of the scale ratio, main orien-
tation difference, horizontal shifts, and vertical shifts of the
matched keypoints in Fig. 1(a) and (b) after applying the
proposed gradient definition. In contrast to the results in Fig. 2,
these histograms evidently exhibit a single mode except for the
main orientation difference histogram that has two main modes.
The reason why the main orientation difference histogram
has two main modes is that the circumferential angle is not
MA et al.: REMOTE SENSING IMAGE REGISTRATION WITH MODIFIED SIFT AND ENHANCED FEATURE MATCHING
5
B. Enhanced Feature Matching
Similarity transformation consists of three parameters: trans-
lation, scale, and rotation. Under the similarity transformation
model, the correct-match pairs will have the same rotation angle
in space, the same scale ratio, the same horizontal shifts, and the
same vertical shifts in most cases. Hence, we use the inherent
information (i.e., position, scale, and main orientation) of each
keypoint to increase the number of correct correspondences.
1, p
2,
N have been extracted from the reference and sensed
. . . , p
images, respectively. (xi, yi), si, and θi denote the position,
scale, and main orientation of the keypoint pi, respectively, in
i denote the position,
the reference image. (x
i, y
scale, and main orientation of the keypoint p
i, respectively, in
the sensed image. The position transformation error of corre-
spondence pi and p
Two point feature sets P = p1, p2, . . . , pM and P
i is denoted as
i, and θ
ep(i) = (xi, yi) − T ((x
i, y
i), μ) is the similarity transformation model, and
where T ((x
i, y
μ is the transformation model parameter. We also use the scale
error [4] and the relative main orientation error [4] of point pi
and p
i) , μ)
i), s
= p
i as
1 − (r
,
∗
)
s
i
si
es(i) =
eo(i) = abs(Δθi − Δθ
∗
)
∗
∗
and Δθ
where r
denote the mode locations of scale ratio
and main orientation difference between reference and sensed
images, respectively, and Δθi = θi − θ
i denotes the main ori-
entation difference between pi and p
i. Then, we define a more
robust joint distance named position scale orientation Euclidean
distance (PSOED) as
PSOED(i) = (1 + ep(i)) (1 + es(i)) (1 + eo(i)) ED(i)
(4)
where ED(i) denotes the Euclidean distance of the descriptors
corresponding to the keypoints pi and p
i. The PSOED will be
minimized in most cases when the point pairs are correctly
matched. The proposed matching algorithm is given below.
1) Initial matching: The keypoints are matched by the ratio
between the Euclidean distance of the nearest neighbor
and that of the second nearest neighbor of corresponding
descriptors. The threshold on the ratio is set to dratio. Pair
set PP is obtained, and we set up the histograms of scale
ratio, main orientation difference, horizontal shifts, and
vertical shifts. The mode locations r
, and
are obtained from the histograms. The FSC algo-
Δy
rithm [15] is used to calculate the initial transformation
parameter μ from pair set PP.
∗
, Δx
, Δθ
∗
∗
∗
∗
∗
∗
, Δθ
, and Δy
∗
, Δx
2) Rematching: As the main orientation difference his-
togram has two modes, there are only two different
. For each mode
combinations of r
combination, we use PSOED as the distance measure,
and the keypoints are matched by the ratio between the
distance of the nearest neighbor and that of the second
nearest neighbor. The threshold on the ratio is denoted
as dr. Because matching is performed twice, the key-
points in the reference image or the sensed image will
have one or more matching in another image; hence, we
regard the point pair with the minimum PSOED as the
Fig. 3. (a)–(d) Histograms of scale ratio, main orientation difference, horizon-
tal shifts, and vertical shifts.
Fig. 4. Scheme of log-polar sectors and the mode of the main orientation
difference. (a) Log-polar sectors. Parameter R1 is set to 12σ. Ratio of R3 and
R2 to R1 is 0.25 and 0.73, respectively. (b) Two modes of the main orientation
difference in numerical. Single mode of rotation angle in space.
◦
◦
and 180
continuous in −180
[see Fig. 4(b)]. It is necessary
to point out that there is only one main mode in rotation angle.
If we know one of the locations of the two main models, the
location of another mode can be computed as
Δθ
=
Δθ + 360, Δθ ∈ [−360, 0)
Δθ − 360, Δθ ∈ [0, 360)
(3)
denote the locations of two modes of the
where Δθ and Δθ
main orientation difference histogram, respectively. Fig. 4(b) il-
lustrates the two modes of the main orientation difference and
the single mode of rotation angle. The mode of the scale ratio is
easily located at r = 1.022. Two modes of the main orientation
difference are located at Δθ = −89.61 and Δθ
= 270.97. The
modes of horizontal shifts and vertical shifts are obtained at
Δx = −10.96 and Δy = 608.76. Note that the exact location
of the mode is obtained by an appropriate interpolation method
[11]. The relationship between Δθ and Δθ
is very close to
the results of (3). Finally, we employ the fast sample consensus
(FSC) algorithm [15] to obtain the transformation model para-
meters r = 0.99, Δθ = −90, Δx = −9.71, and Δy = 604.75,
respectively. From the above results and analysis, it can be seen
that the new gradient calculation method is better in dealing
with the problem of intensity differences of remote sensing
image pairs.
6
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 1, JANUARY 2017
candidate matching pair. Then, a keypoint pair set PP1 is
obtained.
3) Outlier removal: There will be some false correspon-
dences in PP1. Hence, we first use the method proposed
in MS-SIFT [11] to filter out most of the outliers. Let
1) denote the coordinates of corre-
(x1, y1) and (x
1, y
sponding keypoints in set PP1. The horizontal and vertical
shifts of corresponding keypoints are defined as
Δx1 = x1 − r
∗ sin(Δθ
∗
∗
∗ cos(Δθ
Δy1 = y1 − r
∗
∗
∗ cos(Δθ
∗
∗ sin(Δθ
∗
) − y
1
) + y
1
(x
1
(x
1
))
)) .
(5)
Then, most of the outliers are eliminated according to
the following logical filter [11]:
|Δx1 − Δx
∗| ≥ Δxth, |Δy1 − Δy
∗| ≥ Δyth
(6)
where Δxth and Δyth denote, respectively, the thresholds
of horizontal and vertical differences. The thresholds are
set to the bin widths of corresponding histograms [11].
Finally, we get a keypoint pair set PP2 from PP1, and
the FSC algorithm [15] is used to find correct correspon-
dences from keypoint pair set PP2.
Note that although we need to compute the Euclidean dis-
tance for every keypoint pair in the initial matching step, we
save computational efforts in the rematching step. To display
the distribution of histograms well, dratio is set to 0.9. To get as
many candidate keypoint pairs as possible, dr is set to 0.9 [4].
The FSC algorithm employed in the outlier removal step can
get more correct matches than random sample consensus [16]
in fewer iterations.
IV. EXPERIMENT AND RESULT
A. Test Image Pairs
◦
To evaluate the proposed method, three image pairs are
tested. These image pairs are shown in Fig. 1. The first pair
P-A is multispectral images from the U.S. Geological Survey
project [17], Lat/Long: 69.6/−92.7, 240-m resolution. A seg-
ment with a size of 614 × 611 from band 5 (Sensor: Landsat-7
ETM+, Date: 2000/7/24) was selected as the reference image.
To increase the difficulty of the test data set, a segment with
a size of 614 × 611 from band 3 (Sensor: Landsat 4-5 TM,
Date: 1999/6/28) after a simulated rotation of 90
was selected
as the sensed image. The second pair P-B is two 412 × 300
multisensor images with the reference image obtained from
band 5 of a scene taken by the sensor Landsat-TM in September
1995 (with a spatial resolution of 30 m), and the sensed image
is the optical image obtained from Google earth (with a spatial
resolution of 5 m). The third pair P-C is two 800 × 800
multisensor images with the reference image obtained from
the HH mode (L-band) of a scene taken by the sensor ALOS-
PALSAR on June 5, 2010, in the region of Campbell River in
British Columbia (with an initial spatial resolution of 15 m
resampled to 30 m), and the sensed image is from band 5
(1.55–1.75 μm) of a scene taken by the sensor Landsat-ETM+
on June 26, 1999, in the same region (with a spatial resolution
of 30 m).
B. Evaluation Criterion
1) Matching Accuracy: The accuracy is evaluated by the
root-mean-square error (rmse) criterion [2], [11]. A total of
i)} are manually
N corresponding point pairs {(xi, yi), (x
i, y
selected from the reference and sensed images. The point pairs
are carefully chosen and are refined to reduce the residual as
low as possible [2]. Hence, those point pairs are used as the
reference to test the precision of model parameters. The rmse is
computed according to
1
N
RMSE =
N
i=1
(xi − x
i )2 + (yi − y
i )2
(7)
i ) denotes the transformed coordinates of (x
i, y
where (x
i).
i , y
For each test image pair, the algorithm is executed ten times,
and the average of the ten results is computed as the final result.
2) Keypoint Number: The number of correct correspondences
is used as the criterion to evaluate the robustness of the pro-
posed method [18]. The thresholds for keypoint detection are
adjusted, so the comparison method and the proposed method
have roughly the same number of detected points.
C. Experimental Results
We compare the proposed PSO-SIFT algorithm with the
SIFT-based FSC algorithm, SURF, SAR-SIFT [14], and the
MS-SIFT algorithm. Although the SAR-SIFT algorithm is
specifically designed for SAR images, it shows a good perfor-
mance for other types of remote sensing images. In addition
to SURF, other methods are implemented under MATLAB
R2012a with an Intel Core 2.53-GHz processor and 6 GB of
physical memory. The source code is available at https://github.
com/ZeLianWen/Image-Registration.
The results of matching accuracy, correctly matched keypoint
number, and the average running time for three test pairs are
shown in Table I. Note that the algorithm is executed ten times
and the average of the ten running time as the final running
time. The test pair P-A is acquired in two different bands and
different sensor devices. Due to the different spectral bands,
there are some irregular relations of the intensity mapping
between the remote image pair. Considering the results in
Table I, SIFT, SAR-SIFT, and SURF fail to register the test
pair P-A. However, after we modified the gradient calculation
method, 121 correctly matched keypoints are filtered, and the
subpixel matching results are achieved, which indicate the
robustness of the new gradient definition. Although the MS-
SIFT can accurately match the test pair P-A, it is difficult to
tune its parameters to obtain satisfactory results. The test pairs
P-B and P-C are acquired with different sensor devices, and
there are monotonic relations of the image intensity between the
remote image pairs. SIFT, SURF, SAR-SIFT, and MS-SIFT can
accurately match this image pair, indicating that those methods
are invariant to monotonic intensity transform. The proposed
gradient calculation method also achieves subpixel accuracy. To
verify the robustness of the enhanced feature matching method,
we compare two algorithms that are named as Proposed gradi-
ent + FSC and Proposed gradient + Enhanced matching. It can
be clearly seen that the enhanced matching method effectively
increases the number of correctly matched keypoints. For test
image pair P-A, though the RMSE of enhanced matching is a
MA et al.: REMOTE SENSING IMAGE REGISTRATION WITH MODIFIED SIFT AND ENHANCED FEATURE MATCHING
7
COMPARISON OF RMSE, CORRECTLY MATCHED NUMBER, AND AVERAGE RUN TIME FOR TEST IMAGES
TABLE I
our method reveals better performance than the state-of-the-art
methods in terms of aligning accuracy and correctly matched
number of keypoints in some cases.
REFERENCES
Int. J. Comput. Vis., vol. 60, no. 2, pp. 91–110, Nov. 2004.
[8] H. Bay, A. Ess, and T. Tuytelaars, “SURF: Speeded-up robust features,”
Comput. Vis. Image Understand., vol. 110, no. 3, pp. 346–359, Jun. 2008.
[9] K. Mikolajczyk and C. Schmid, “A performance evaluation of local
descriptors,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 27, no. 10,
pp. 1615–1630, Oct. 2005.
[1] A. Wong and D. A. Clausi, “ARRSI: Automatic registration of remote
sensing images,” IEEE Trans. Geosci. Remote Sens., vol. 45, no. 5,
pp. 1483–1493, May 2007.
[2] M. Gong, S. Zhao, L. Jiao, D. Tian, and S. Wang, “A novel coarse-to-fine
scheme for automatic image registration based on sift and mutual informa-
tion,” IEEE Trans. Geosci. Remote Sens., vol. 52, no. 7, pp. 4328–4338,
Jul. 2014.
[3] A. Sedaghat, M. Mokhtarzade, and H. Ebadi, “Uniform robust scale in-
variant feature matching for optical remote sensing images,” IEEE Trans.
Geosci. Remote Sens., vol. 49, no. 11, pp. 4516–4527, Nov. 2011.
[4] Q. Li, G. Wang, J. Liu, and S. Chen, “Robust scale-invariant feature
matching for remote sensing image registration,” IEEE Geosci. Remote
Sens. Lett., vol. 6, no. 2, pp. 287–291, Apr. 2009.
[5] J. Senthilnath, S. N. Omkar, V. Mani, and T. Karthikeyan, “Multiob-
jective discrete particle swam optimization for multisensor image align-
ment,” IEEE Geosci. Remote Sens. Lett., vol. 10, no. 5, pp. 1095–1099,
Sep. 2013.
[6] J. Senthilnath, N. P. Kalro, and J. A. Benediktsson, “Accurate point match-
ing based on multi-objective genetic algorithm for multi-sensor satellite
imagery,” Appl. Math. Comput., vol. 236, pp. 546–564, 2014.
[7] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,”
[10] S. Leutenegger, M. Chli, and R. Siegwart, “BRISK: Binary robust in-
variant scalable keypoints,” in Proc. IEEE Int. Conf. Comput. Vis., 2011,
pp. 2548–2555.
[11] B. Kufer, N. S. Netanyahu, and I. Shimshoni, “An efficient SIFT-based
mode-seeking algorithm for sub-pixel registration of remotely sensed
images,” IEEE Geosci. Remote Sens. Lett., vol. 12, no. 2, pp. 379–383,
Feb. 2015.
[12] Q. Yang and X. Xiao, “A new registration method base on improved Sobel
and SIFT algorithms,” in Proc. 3rd Int. Conf. Comput. Elect. Eng., 2010,
p. 528.
[13] Y. Li, L. Liu, L. Wang, D. Li, and M. Zhang, “Fast SIFT algorithm
based on Sobel edge detector,” in Proc. 2nd Int. Conf. Consum. Electron,
Commun. Netw., Apr. 2012, pp. 1820–1823.
[14] F. Dellinger, J. Delon, Y. Gousseau, J. Michel, and F. Tupin, “SAR-SIFT:
A SIFT-like algorithm for SAR images,” IEEE Trans. Geosci. Remote
Sens., vol. 53, no. 1, pp. 453–466, Jan. 2015.
[15] Y. Wu, W. Ma, M. Gong, L. Su, and L. Jiao, “A novel point-matching
algorithm based on fast sample consensus for image registration,” IEEE
Geosci. Remote Sens. Lett., vol. 12, no. 1, pp. 43–47, Jan. 2015.
[16] M. A. Fischler and R. C. Bolles, “Random sample consensus: A para-
digm for model fitting with applications to image analysis and automated
cartography,” Commun. ACM, vol. 24, no. 6, pp. 381–395, Jun. 1981.
[17] [Online]. Available: http://glovis.usgs.gov/
[18] Q. Li, S. Qi, Y. Shen, D. Li, H. Zhang, and T. Wang, “Multispectral
image alignment with nonlinear scale-invariant keypoint and enhanced
local feature matrix,” IEEE Geosci. Remote Sens. Lett., vol. 12, no. 7,
pp. 1551–1556, Jul. 2015.
Fig. 5. Matching results of Fig. 1. (a) and (c) Proposed gradient + FSC.
(b) and (d) Proposed gradient + Enhanced matching.
Fig. 6. Checkerboard mosaiced images of the proposed method. (a) Result for
the image pair P-A. (b) Result for the image pair P-B. (c) Result for the image
pair P-C.
little higher than FSC, it also achieves sub-pixel accuracy. The
robustness of the FSC algorithm is also proved. Because of the
realization of SURF based on C++, the efficiency is very high.
The matching results of two image pairs are shown in Fig. 5.
Compared with FSC, the enhanced matching method greatly
increases the correct matches. The checkerboard mosaiced im-
ages are shown in Fig. 6. It can be seen that the edge and region
of two images are precisely overlapped, which demonstrates the
accuracy of our proposed method.
V. CONCLUSION
In this letter, we have proposed a new gradient computation
method that is more robust to complex nonlinear intensity
transform of remote sensing images. In addition, a robust
point matching algorithm that combines the position, scale, and
orientation of each keypoint to increase the number of correct
correspondences has been proposed. Experimental results on
multispectral and multisensor remote sensing images show that