Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Correspondence________________________________________________________________________
Localization Based on DEM Matching Using Multiple
Aerial Image Pairs
Dong-Gyu Sim and Rae-Hong Park
AbstractThis paper proposes the localization algorithm that estimates
translation parameters of an aircraft by comparing the sampled elevation
map recovered from aerial sequence images, and the digital elevation model
(DEM) with the given orientation and altitude parameters obtained from a
gyroscope. It consists of two stages: recovering the sampled elevation map
from multiple aerial image pairs and matching of the relative recovered
elevation map (REM) with the relative DEM. Computer simulations with
three real aerial image sequences show the effectiveness of the proposed
algorithm.
Index TermsAerial sequence image, attitude parameters, digital elevation model, localization, M-estimator, recovered elevation map, stereo
matching, terrain.
I. INTRODUCTION
NAVIGATION guides a vehicle to a spot where one wants to go [1].
For autonomous navigation, it is important to extract accurate navigation parameters of an aircraft such as position and velocity. Using them,
one can adjust a velocity or direction to the desired destination. Many
approaches to navigation parameter extraction have been presented.
Navigation systems can be realized by a number of different approaches such as terrain contour matching (TERCOM) [2], inertial navigation system (INS) [1], global positioning system (GPS), and so on.
The TERCOM navigates by measuring altitude of terrain with a special radar and by matching terrain contours. This system has difficulty
in estimating the position in plain regions where change in elevation is
small, and it can be out of control by external signals. The INS with
an accelerometer and gyroscope has been widely used for navigation.
It does not require any constraint on flying trajectory and is not affected by any external signal, however it has a drawback that the estimation error increases as navigation proceeds. The GPS estimates an
absolute position with more than three satellites, however the signals
from satellites can be disturbed and modified intentionally. Navigation parameters can also be estimated by using computer vision and
image processing techniques [3][10], however, they are not practical
for implementing real-time navigation systems with real aerial image
sequences, because they require a high computational load or an expensive down-looking camera. On the other hand, three-dimensional
(3-D) reconstruction algorithms have been proposed by matching the
digital elevation model (DEM) and the relative recovered elevation map
(REM) [5], [6].
The proposed algorithm reconstructs the wide sampled REM using
multiple image pairs. It estimates the aircraft position using the relative
geographical data (REM) obtained from multiple aerial image pairs and
Manuscript received January 25, 1999; revised September 22, 2001. This
work was supported in part by Agency for Defense Development and by the Automatic Control Research Center, Seoul National University. An early version
of this paper was presented at the IAPR Workshop on Machine Vision Applications 98, Chiba, Japan, in November 1998. The associate editor coordinating
the review of this manuscript and approving it for publication was Dr. Josiane
B. Zerubia.
The authors are with the Department of Electronic Engineering, Sogang University, Seoul 100-611, Korea (e-mail: rhpark@sogang.ac.kr).
Publisher Item Identifier S 1057-7149(02)01315-5.
jM
the relative DEM, where the robust M-estimator is used. It is important to analyze the characteristics of the REM. The rest of the paper
is structured as follows. The proposed algorithm is presented in Section II. Experimental results are shown in Section III to illustrate the
effectiveness of the proposed algorithm. Finally, conclusions are given
in Section IV.
II. PROPOSED IMAGE-BASED LOCALIZATION ALGORITHM
USING THE DEM
The proposed algorithm consists of two stages. At the first stage,
a sampled elevation profile is constructed by computing elevations, at
sample feature points along an arbitrary horizontal line, based on stereo
matching, with point correspondences established between two consecutive aerial images. In the second stage, a matching point in the DEM is
found by searching the location whose local terrain is the most similar
to the REM, where relative REMs and DEMs are employed.
A. First Stage (Construction of a Sampled Elevation Profile)
First, the global feature point FPk01 that has the maximum local
variance is extracted in the previous frame, where a 31 2 31 local
window is used. The global feature point is a very distinct point. So,
mismatching of global features rarely occurs in over 3000 consecutive
images. It is used as a reference point over the whole frame to compute
relative REMs and DEMs. Feature point correspondence is established
by block-based matching, using the normalized correlation coefficient
(NCC) measure. In Fig. 1, FPk01 denotes the global feature point
having the maximum variance at the (k 0 1)th (previous) frame Ik01
in the aerial image sequence. MPk represents the matching point in the
k th (current) frame Ik , corresponding to FPk01 . Pk01 and Pk denote
3-D positions of the optical centers of a camera in the (k 0 1)th and
k th frames, respectively. The difference between these two positions
is denoted by the base line Bk
Bk
= rk R (FPk01 ; Pk01 )
0 k0
r R
(MPk ; Pk )
(1)
rk R (FPk01 ; Pk01 )z
rk0 R (MPk ; Pk )z
0 r0
j
k
=P 01
=P
z;k
(2)
z;k
(3)
R MPjk ; Pk
=B
(4)
to
step
For
to
step
For
For
to
to
Rel REM
step
step
DEM
DEM
DEM
DEM
Rel DEM
;
DEM
DEM ;
Rel REM
IF
Rel DEM
; Store
are satisfied, where rk (rk0 ) represents the distance from the feature
point FPk01 (MPk ) to the zero-elevation ground point, and the subscript z denotes the z coordinate of the 3-D position. The longitude
and latitude of the approximate base line can be estimated using three
equations by a simple regression method with the altitude of an aircraft
obtained from a gyroscope. Even though the estimated base line is not
accurate, global features and shapes of the REM reconstructed with
the approximate base line yields proper matching of the REM with the
DEM.
FPk0 01 , 1 j M , is the j th sample feature point in the (k 0 1)th
frame, where M denotes the total number of sample feature points employed. The sample feature points FPjk01 are selected along a fixed
row with equal intervals. In experiments, they are selected along the
80th row, with 25 pixels apart from each other, where the image size
is 320 2 240. It is assumed that the size of the area overlapped by two
successive images is greater than half the image size. The fixed line is
arbitrarily chosen, for simplicity of implementation, in the upper half
region of the image, where an aircraft is assumed to fly upwards in the
image plane. MPjk of the k th frame are corresponding feature points
detected by the NCC measure. Note that these matching points are detected for (N 0 1) image pairs. A sampled REM is recovered with the
consecutive sets of feature points and matching points, i.e., the REM
at Mkj is calculated by the stereo matching method [6], [9], [11], with
FPkj 01 and Mkj as well as attitude parameters 8, ! , and ' obtained
from a gyroscope and flying altitude from an altimeter. The equality
53
where M represents the total number of feature points, along a line for
each image pair, N signifies the number of image frames used, x and
y denote the search intervals along the x and y directions, respectively.
dx and dy represent the compensated distances with respect to the initial latitude and longitude, respectively. Mk = (Mx;k ; My;k ; Mz;k )
signifies the ground point corresponding to the global feature point
FPk01 , with Mz;k corresponding to the REM at Mk . Similarly, Mkj =
j
j
j
(Mx;k
; My;k
; Mz;k
) is the ground point corresponding to the sample
j
feature point FPk01 in the (k 0 1)th frame and their matching point
MPjk in the k th frame, and (x1 0 x0 )(y1 0 y0 ) denotes the size of the
search area in the DEM. In experiments, DEM(x; y ) represents an elevation at position (x; y ). DEMk and DEMjk denote the elevations at the
global feature point FPk01 and the j th sample feature point FPkj 01 , respectively. Rel DEMjk (Rel REMjk ) signifies the relative DEM (REM)
at Mkj with respect to that at Mk . For computational simplicity, the
error in elevation estimation is assumed to be the same for all feature
points in a frame. So, relative elevations instead of absolute elevations
are used in computing the distance for magnitude-invariant matching.
The relative DEM matching algorithm also employs a robust cost function (x) based on M-estimation, by which some outliers could be rejected [12], [13]. The cost function is defined by
(x) =
x2 ; x <
2; x
(5)
54
TABLE I
ESTIMATION ERROR OF THE CONVENTIONAL AND PROPOSED ALGORITHMS (M)
Fig. 2.
1).
2 240, spot
represent the sample feature points FPjk01 in the previous frame for elevation estimation, and the box ( ) marks denote the matching points
MPjk in the current frame corresponding to FPjk01 . Note that a helicopter flew upwards, and that in the first and last frames only the sample
feature points and the matching points are shown, respectively.
In experiments, 11 sample feature points are used in stereo matching
and they are well corresponded as shown in Fig. 2. In experiments,
654 2 896 DEM is used, and the matching result is shown in Fig. 3,
where the darker (brighter) region represents the lower (higher) terrain
and the black dot () marks denote the sample feature points FPjk01 .
Whereas the asterisk (3) mark represents the real position in the last
frame which is obtained by manually matching the aerial image and a
map, the circle () and diamond (5) marks signify the positions estimated by the conventional [5] and proposed algorithms, respectively.
The DEM is generated by interpolating the contours that are manually
drawn with 25 m interval. In the case of the first test image sequence,
REFERENCES
[1] C.-F. Lin, Advanced Control Systems Design. Englewood Cliffs, NJ:
Prentice-Hall, 1994.
[2] J. P. Golden, Terrain contour matching (TERCOM): A cruise missile
guidance aid, in Proc. Image Processing Missile Guidance, vol. 238,
San Diego, CA, July/Aug. 1980, pp. 1018.
[3] S. J. Merhav and Y. Bresler, On-line vehicle motion estimation from
visual terrain information Part I: Recursive image registration, IEEE
Trans. Aerosp. Electron. Syst., vol. AES-22, pp. 583587, Sept. 1986.
[4] Y. Bresler and S. J. Merhav, On-line vehicle motion estimation from
visual terrain information Part II: Ground velocity and position estimation, IEEE Trans. Aerosp. Electron. Syst., vol. AES-22, pp. 588604,
Sept. 1986.
[5] J. J. Rodrguez and J. K. Aggarwal, Matching aerial images to 3-D
terrain maps, IEEE Trans. Pattern Anal. Machine Intell., vol. PAMI-12,
pp. 11381149, Dec. 1990.
[6] M. S. Kang, R.-H. Park, and K.-H. Lee, Recovering an elevation map
by stereo modeling of the aerial image sequence, Opt. Eng., vol. 33, no.
11, pp. 37933802, Nov. 1994.
[7] P. Robertson and J. M. Brady, Adaptive image analysis for aerial
surveillance, IEEE Intell. Syst. Mag., vol. 14, pp. 3036, Mar. 1999.
55
[8] T. Misu, T. Hashimoto, and K. Ninomiya, Optical guidance for autonomous landing of spacecraft, IEEE Trans. Aerosp. Electron. Syst.,
vol. 35, pp. 459473, Apr. 1999.
[9] D.-G. Sim, S.-Y. Jeong, D.-H. Lee, R.-H. Park, R.-C. Kim, S. U. Lee,
and I. C. Kim, Hybrid estimation of navigation parameters from aerial
image sequence, IEEE Trans. Image Processing, vol. 8, pp. 429435,
Mar. 1999.
[10] M.-G. Lee, J.-H. Park, S. B. Pan, R.-C. Kim, K.-S. Kim, R.-H. Park,
S. U. Lee, and I. C. Kim, Implementation of the image-based absolute
position compensation algorithm in the navigation parameter extraction
system, in Proc. 9th Int. Conf. Signal Processing Applications Technology, Toronto, ON , Canada, Sept. 1998, pp. 16031607.
[11] W.-H. Liao and J. K. Aggarwal, Cooperative matching paradigm for
the analysis of stereo image sequence, Int. J. Imag. Syst. Technol., vol.
9, pp. 192200, Apr. 1998.
[12] I. Pitas and A. N. Venetsanopoulos, Nonlinear Digital Filters: Principles
and Applications. Norwell, MA: Kluwer, 1990.
[13] P. Meer, D. Mintz, and A. Rosenfeld, Robust regression methods for
computer vision: A review, Int. J. Comput. Vis., vol. 6, no. 1, pp. 5970,
Apr. 1990.