IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 4, APRIL 2015

2137

Aerial Image Registration for Tracking Michael E. Linger and A. Ardeshir Goshtasby

Abstract—To facilitate the tracking of moving targets in a video, the relation between the camera and the scene is kept fixed by registering the video frames at the ground level. When the camera capturing the video is fixed with respect to the scene, detected motion will represent the target motion. However, when a camera in motion is used to capture the video, image registration at ground level is required to separate camera motion from target motion. An image registration method is introduced that is capable of registering images from different views of a 3-D scene in the presence of occlusion. The proposed method is capable of withstanding considerable occlusion and homogeneous areas in images. The only requirement of the method is for the ground to be locally flat and sufficient ground cover be visible in the frames being registered. Experimental results of 17 videos from the Brown University data set demonstrate robustness of the method in registering consecutive frames in videos covering various urban and suburban scenes. Additional experimental results are presented demonstrating the suitability of the method in registering images captured from different views of hilly and coastal scenes. Index Terms—Aerial video, homography, image registration, multiview images, tracking.

I. I NTRODUCTION

T

O DETECT and track moving targets in a video captured by a moving camera, frames in the video are registered to keep the relation between the camera and the scene fixed. Registration of images from a moving camera is challenging particularly when 3-D structures, such as buildings, are present in the scene. Scene points visible in one view may be occluded from another view by the structures. A method for registering multiview images in the presence of occlusion is introduced. Images (a) and (b) in Fig. 1 show two frames in an aerial video of a scene containing a parking structure and surrounding buildings captured by a moving camera. Although individual parking lots represent a plane, not all parking lots represent the same plane. Registering images (a) and (b) in Fig. 1 by a single homography and finding absolute intensity differences of registered images, the result shown in Fig. 1(c) is obtained. Homography is the transformation relating images of a planar scene that have been obtained from different views. Areas where registration is accurate appear dark, while misregistered areas appear bright. When a single homography is used to register the images, although cars in some lots align well, cars in other lots misalign because they do not belong to the plane the homography was calculated for. A rectangular area covering adjacent parking lots in Fig. 1(c) is zoomed in Fig. 1(e) for better viewing. Although the cars

Manuscript received May 24, 2014; revised July 30, 2014 and August 19, 2014; accepted September 3, 2014. The authors are with the Department of Computer Science and Engineering, Wright State University, Dayton, OH 45435 USA. Digital Object Identifier 10.1109/TGRS.2014.2356177

Fig. 1. (a) and (b) Two images representing consecutive frames in an aerial video of a parking structure captured by a moving camera. These images are courtesy of the Brown University Computer Vision Group. (c) Absolute intensity differences of images when registered by a global homography. The rectangular area within the difference image is zoomed in (e) for better viewing. (d) Registration using a combination of global and local homographies, with the same rectangular area zoomed in (f).

in the left parking lot are aligned well, the cars in the right parking lot are misaligned. When registering the images by the method proposed in this paper, the result shown in Fig. 1(d) is obtained, with the same rectangular area zoomed in Fig. 1(f). The cars in the left parking lot as well as in the right parking lot are now aligned well because different homographies are used to register different parts of the images. Dim blobs in Fig. 1(f) are mostly due to intensity difference between shiny cars when viewed from different angles. The brighter blobs represent poles in the scene that are displaced in the images. The brighter blobs also represent moving cars that are displaced from one image to another. In the proposed method, the global and local transformations are separated from each other. The global transformation describes the relation between the camera positions capturing the images and the scene, and the local transformations characterize the scene structure at various resolutions. As the images are reduced in size (resolution), the global geometric difference between the images remains unchanged while local geometric differences between the images reduce. By reducing the resolution of the images sufficiently so that local geometric differences between the images become negligible, the global geometric difference between the images is estimated. If images at the lowest resolution can be considered images of an approximately flat scene, a global homography will bring the images into approximate alignment.

0196-2892 © 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

2138

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 4, APRIL 2015

In the proposed method, a homography is used to relate the images globally and approximately align them. Since corresponding points in approximately aligned images fall near each other, this step will speed up the search for landmark correspondences that are needed to register the images locally. Fig. 1(c) and (d) demonstrate the difference between the initial registration using a global homography and the final registration using a combination of global and local homographies. The global and local homographies are determined by following a coarse-to-fine strategy. The global homography is determined by reducing the resolution (size) of the images sufficiently so that local geometric differences between the images become negligible. After finding the global relation between the images, the resolution of the images is gradually increased while adding more local homographies to the process. Registration at the highest resolution is achieved by using the global homography obtained at the lowest resolution and a sequence of homographies obtained at higher resolutions to account for local geometric differences between the images from coarse to fine. In the following sections, after reviewing related work, steps in the proposed method are detailed. Next, experimental results using various multiview aerial images are presented and evaluated. Finally, obtained results are summarized, and concluding remarks are made. II. R ELATED W ORK Automatic image registration is a required capability in target tracking. Various automatic image registration methods have been developed throughout the years [1]–[3]. Specialized methods for registering high-altitude aerial images where occlusion is not significant have been proposed also [4]–[6]. This paper describes a method for registering low-altitude aerial images where occlusion can be significant. Image registration for target tracking has been attempted before. Cheng and Menq [7] solved the case of a flat scene and translational motion. The global displacement between consecutive frames was determined by registering the frames by cross-correlation. Jackovitz et al. [8] used phased correlation to globally align high-altitude aerial images to facilitate target tracking. Mei and Porikli [9] registered images globally by an affine transformation to stabilize the camera with respect to the scene. They then detected the moving targets and tracked them using a factorial hidden Markov model. Ayala et al. [10] segmented each video frame into regions and tracked the regions from frame to frame. A region that had a motion inconsistent with the motion of surrounding regions was considered a target. In this manner, background motion caused by camera motion was removed, retaining target motion. Jackson and Goshtasby [11] registered frames in a video using a single homography to facilitate target tracking. This method is applicable to videos where the ground level across the scene can be modeled by a plane. In the proposed method, the ground level across the scene can change in slope and height. The only requirement is for the ground to be locally planar. A hierarchical method for registering multiview images is developed. Hierarchical methods for image registration have been developed before [12]–[15]. These methods track corresponding landmarks in images from low to high resolution.

Fig. 2.

Flow of computations in the proposed hierarchical image registration.

Once corresponding landmarks in the highest resolution images are determined, a globally defined transformation function such as thin-plate spline [16], [17] is used to register the images. In the proposed hierarchical method, rather than the thinplate spline, a weighted sum of local and global homographies is used to register the images. Homography is used for local registration because it represents the true relation between images of a planar patch when viewed from different angles. Since our objective is to register images at ground level and the ground is locally flat, we use a homography to register corresponding neighborhoods in the images. Details of the proposed method follow. III. P ROPOSED R EGISTRATION The flow of computations in the proposed image registration is shown in Fig. 2. First, a global homography is found to approximately align the images. This is achieved by reducing the resolution of the images sufficiently so that local geometric differences between the images become small and negligible. After registering the images at the lowest resolution, the resolution of the images is increased by a factor of two while dividing each image into four equal blocks. A local homography is then calculated to register corresponding blocks in the images. The resolution of the images is increased by a factor of two in this manner while dividing each image block into four new blocks until the blocks in the highest resolution images are obtained. From the global homography obtained at the lowest resolution and local homographies obtained at the intermediate and highest resolutions, an overall transformation is calculated to register the images. These steps are described in detail in the following sections. A. Finding the Global Homography To register images from different views of a 3-D scene, a combination of global and local homographies is used. The global homography accounts for the view-angle difference between the cameras capturing the images and is independent of the scene content. The local homographies are scene dependent and account for local geometric differences between the images captured from different views of the scene. To estimate the global homography, local geometric differences between the images are reduced by reducing the resolution

LINGER AND GOSHTASBY: AERIAL IMAGE REGISTRATION FOR TRACKING

(size) of the images. Experiments with multiview aerial images of urban and suburban scenes of 30-cm/pixel resolution suggest that blocks of 128 × 128 pixels contain sufficient corresponding landmarks to estimate homography parameters to register corresponding blocks in the images. Therefore, if D is the diameter of the reference image, the images must be reduced in size by a factor of 2m , with m to satisfy 128 ≤

D < 256. 2m

(1)

For instance, if the reference image is of dimensions 1280 × 720, D = 1340, and we find m = 3 from relation (1). Therefore, the lowest resolution reference image used to calculate the global homography is of dimensions 160 × 90. The dimensions of the sensed image are reduced by the same amount (23 ). The reference and sensed images at this resolution are the lowest resolution images used in the proposed hierarchical registration. Note that, if (x, y) and (X, Y ) represent the coordinates of a scene point in the reference and sensed images at the lowest resolution, the global homography to register the lowest resolution images can be written as X=

ax + by + c gx + hy + 1

(2)

Y =

dx + ey + f . gx + hy + 1

(3)

x and X represent the image columns in the reference and sensed images, respectively, increasing from left to right, and y and Y represent image rows, increasing from bottom to top. If we scale the images by a factor of s to obtain new images with coordinates (x , y  ) and (X  , Y  ), from x = sx

y  = sy

X  = sX

Y  = sY

(4)

we can find x, y, X, Y in terms of x , y  , X  , Y  and substitute them into (2) and (3) to obtain relations between (x , y  ) and (X  , Y  ) ax + by  + cs X =  g    h   s x + s y +1

(5)

dx + ey  + f s Y  = g  h  . s x + s y +1

(6)

Equations (5) and (6) show that, once the global homography is found, it can be used to register the same images when scaled by s simply by replacing c with cs, f with f s, g with g/s, and h with h/s. Therefore, once the global homography with parameters a−h is found to register the images at the lowest resolution, the same images at a higher resolution can be globally registered (although approximately) by scaling the parameters of the homography, as shown by (5) and (6). Equations (2) and (3) have eight unknown parameters, requiring four or more corresponding landmarks in the images to find them. To have a fully automatic method, it is required that four or more corresponding landmarks be determined in the images. Landmark correspondence is achieved in two steps. First, a set of landmarks is detected in each image, and then, correspondence is established between the two sets of landmarks. The details of these steps are provided next.

2139

1) Landmark Detection: Landmarks are locally unique points in images, facilitating the search for corresponding points in the images. A great number of landmark detection methods have been developed throughout the years [18], [19]. We are interested in detectors that are independent of the orientation of an image. Two widely used detectors are Harris corners [20] and the scale-invariant feature transform (SIFT) points [21]. The former detects corners, and the latter detects centers of blobs. Depending on the type of images available, one may use one or the other detector. If the images contain buildings and other man-made structures where corners are abundant, the corner detector is more appropriate, while if the images represent terrain scenes where man-made structures are absent, the blob detector is more appropriate. If no information about the images is available, one may use one detector as the default and, if that fails, use the alternate detector. In our work, since the focus is on images of urban and suburban scenes, the Harris detector is used as the default, and the SIFT detector is used as the alternate. 2) Landmark Features: Once unique landmarks are detected in each image, there is a need to find correspondence between the landmarks. To expedite the correspondence process, a number of features are extracted from the surroundings of each landmark. The vector of features associated with a landmark is known as a descriptor. The SIFT landmark detector, as well as many other detectors, has associating descriptors that characterize various properties of the neighborhood of a landmark. The SIFT detector associates a vector of 128 features, describing the intensity gradient properties of the neighborhood of a landmark. If the Harris corner detector is used, features must be calculated for each corner. To make the features invariant of an image’s orientation, five invariant features are calculated at a neighborhood of radius 12 pixels centered at the landmark. The radius 12 is determined experimentally to achieve the highest speed while capturing sufficient local geometric information about the pattern centered at the landmark when working with 30-cm/pixel images. The features used are as follows: 1) a Hu invariant moment of order two; 2) a Hu invariant moment of order three; 3) a complex moment of order two; 4) the Laplacian of Gaussian of standard deviation of 3 pixels; and 5) the center contrast feature at the landmark. Methods for calculating these features are described in [22, Ch. 4]. Independent of whether a landmark represents a corner or the center of a blob, these five features are calculated and associated with the landmark. Therefore, in our implementation, the landmarks detected by both Harris and SIFT detectors use these invariant features when determining the correspondence between the landmarks. The algorithm to establish correspondence between two sets of landmarks using these features is described next. 3) Landmark Correspondence: As the resolution of the images is reduced, local geometric differences between the images decrease while the global relation between the images remains the same. The coarse-to-fine strategy is, therefore, to reduce local geometric differences between the images sufficiently so that the images can be globally registered with a homography. Once this global homography is determined, the resolution of the images can be gradually increased, and the neighborhoods can be tracked from low to high resolution. The correspondences

2140

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 4, APRIL 2015

obtained at a resolution are used to refine the transformations obtained at one level lower resolution. After finding the features of the landmarks, each feature is normalized to have a mean of zero and a variance of one over all landmarks in an image. Then, putative correspondence is established between landmarks in the images using the following rules. 1) For each landmark in the reference image, the landmark in the sensed image having the closest feature vector, measured with Euclidean distance, is identified, and is considered its correspondence. 2) If a landmark in the sensed image ends up corresponding to more than one landmark in the reference image, all those correspondences are considered ambiguous and are discarded. 3) The remaining correspondences, which are unique, are considered putative correspondences. Some of the putative correspondences could be wrong due to the fact that a landmark in the reference image may be occluded in the sensed image. Incorrect correspondences, known as outliers, must be identified and removed before finding the transformation parameters. The process of eliminating the outliers and finding the parameters of a homography by least squares using a random sample and consensus (RANSAC) algorithm has been described by Hartley and Zisserman [23], and Chen and Suter [24]. This step not only identifies the outliers but also determines the correspondence between the inliers and finds the parameters of the global homography defined by (2) and (3) by least squares. Since the scale ratio between images at the lowest and images at a higher resolution is known, the global transformation can be used to approximately align images at any resolution using (5) and (6). By gradually going from low to high resolution, the global relation between the images is adapted more and more to local geometric differences between the images by adding more local homographies to the process. This coarse-to-fine adaptation is detailed next. B. Coarse-to-Fine Adaptation Consider the reference image hierarchy shown in Fig. 3. At the bottom is the highest resolution image, showing the original image. At the top is the image in the lowest resolution, which is used to determine the global homography. The figure shows a case where the coarse-to-fine hierarchy has three levels, representing the reference image at three resolutions. In this hierarchy, image resolution (size) decreases by a factor of two by going up the hierarchy by one level. Therefore, if the reference image at the highest resolution (bottom level or level 0) is of dimensions m × n, the image above it at level 1 will be of dimensions m1 × n1 , where m1 = m/2 and n1 = n/2. The image at the top level (level 2) will be of dimensions m2 × n2 , where m2 = m1 /2 = m/22 and n2 = n1 /2 = n/22 . Given images at level 0, images at level 1 are created by averaging 2 × 2 windows in the images at level 0 to obtain the intensity at a pixel at level 1. More specifically, the intensity of pixel (x, y) in the reference image at level i + 1 is obtained from the average of intensities at (2x, 2y), (2x + 1, 2y), (2x, 2y + 1), (2x + 1, 2y + 1) in the reference image at

Fig. 3.

Image hierarchy showing images at three resolutions (levels).

level i for i ≥ 0. Alternatively, the image may be smoothed with an appropriate Gaussian filter and resampled at every other row and column to produce a lower resolution image. Although Gaussian smoothing is generally preferred over 2 × 2 averaging, experimental results show that no significant gain is made in matching when using Gaussian smoothing over 2 × 2 averaging. Therefore, for speed reasons, in this work, 2 × 2 averaging is chosen over Gaussian smoothing when reducing the resolution of an image by a factor of two. The sensed image follows the same hierarchy. Therefore, if an image hierarchy with L levels is created for the reference image, an image hierarchy with L levels is created for the sensed image. Once images at the top level are registered with the homography defined by (2) and (3), (5) and (6) can be used to determine the parameters of the homography to globally align images at lower levels. Alignment at each level by the global homography will be approximate, as shown in the example in Fig. 1(c). Except for the top level where only a homography is found to relate the images, at lower levels, a combination of local and global homographies is used to register the images. In Fig. 3, the images at level 1 are subdivided into four equal blocks, with each block being the size of the image at level 2. Knowing the global homography at level 2, the global homography at level 1 will be estimated from (5) and (6). This will bring images at level 1 into approximate alignment. Next, a homography to relate corresponding reference and sensed blocks is determined the same way a homography was determined to register images at level 2 (the top level). Note that the homography defined by (2) and (3) can be written in matrix form in homogeneous coordinates by ⎡ ⎤ ⎡ ⎤⎡ ⎤ XW a b c x ⎣ Y W ⎦ = ⎣d e f ⎦⎣y ⎦. (7) W g h 1 1 If corresponding points in the reference and sensed images at the top level in homogeneous coordinates are pL = [xL yL 1]t and PL = [XL YL 1]t , where t denotes transpose operation, relation (7) can be written as PL =

1 HL pL WL

(8)

LINGER AND GOSHTASBY: AERIAL IMAGE REGISTRATION FOR TRACKING

where HL is the 3 × 3 homography matrix registering the images at the top level. Note that, once the parameters of the global homography are determined, for each point pL = [xL yL 1]t in the reference image at the top level, WL = (gxL + hyL + 1) can be determined and used to calculate the coordinates of the corresponding point PL = [XL YL 1]t in the sensed image at the top level. Having found the global homography at the top level, the global homography at one lower level can be calculated from (5) and (6) by letting s = 2. The obtained homography will approximately establish correspondence between points in images at level L − 1 PL−1 ≈

1 WL−1

HL−1 pL−1 .

(9)

After finding the homography to approximately align the images at level L − 1, a new homography is calculated for corresponding blocks in the reference and sensed images as a local adjustment to the approximate homography. Doing so, four new homographies will be obtained for the four corresponding blocks. Denoting the four local homographies by H00 , H01 , H10 , and H11 and combining the global homography and the local homography obtained for block 00 (lower left block), we obtain (1/W00 )(1/WL−1 )H00 HL−1 as the refined transformation to register corresponding lower left blocks in the reference and sensed images at level L − 1. Therefore PL−1 =

1 1 H00 HL−1 pL−1 W00 WL−1

1 00 00 HL−1 pL−1 WL−1

00 = fL−1 pL−1

same time increases the overlap between corresponding blocks as we go down the hierarchy. C. Overall Transformation Function In the preceding sections, a coarse-to-fine strategy for finding a homography at corresponding local neighborhoods (image blocks) was described. In the image hierarchy shown in Fig. 3, the images at the bottom level are subdivided into a grid of 4 × 4 equal blocks. A homography determined for a block at level 0 shows the relation between corresponding blocks in the original reference and sensed images. If image blocks are registered individually, gaps may appear between adjacent blocks. In order to make the overall transformation continuous and smooth, local homographies are blended together by a weighted mean approach. If reference and sensed images at level 0 (the bottom level) are subdivided into M × N blocks, and the combined homography found for block ij at level 0 is f0ij , the overall transformation is defined by F=

(12)

00 00 where WL−1 = W00 WL−1 , H00 L−1 = H00 HL−1 , and fL−1 = 00 00 00 (1/WL−1 )HL−1 . fL−1 is the combined transformation for block 00 at level L − 1. Similarly, the combined transformation to register blocks 01, 10, and 11 in the images can be determined. Going from a block at level 0 < l < L to blocks below it at level l − 1 is the same as going from the image at top level (L) to subimages/blocks at level (L − 1). Therefore, once the procedure to use registration information at level L to register blocks at level L − 1 is worked out, the same procedure can be followed to use registration information at a block at level 0 < l < L to register four blocks below it at level l − 1, and the process can be repeated in this manner until blocks at the bottom level (highest resolution) are registered. Note that, as we go down the hierarchy, further adjustments are made to the homography calculated at the top level, adapting the combined homography to geometric differences between corresponding neighborhoods (blocks) in the images. As we go down the hierarchy, the geometry of the sensed image is changed more locally to resemble that of the reference image, bringing the images into closer alignment. This process at the

wij f0ij

(13)

where wij is the weight of f0ij and is a function of x and y. More precisely, (13) can be written as Fxy =

(11)

M −1 N −1   i=0 j=0

(10)

represents correspondence between pixels in the lower left blocks in the images at level L − 1. Since the product of two homographies is a new homography, the aforementioned relation can be written as PL−1 =

2141

M −1 N −1  

wij (x, y)f0ij .

(14)

i=0 j=0

The transformation defined by relation (14) depends on point (x, y) in the reference image. The right-hand side is a weighted sum of homographies, which are 3 × 3 matrices, resulting in a 3 × 3 matrix, which is a new homography. For a point p = [x y 1]t in homogeneous coordinates in the reference image, using relation (14), we can obtain the corresponding point in the sensed image from Fxy p =

M −1 N −1  

wij (x, y)f0ij p.

(15)

i=0 j=0

The right-hand side in (15) can be considered a weighted sum of points, each obtained by a homography of a block near point (x, y). The right-hand side of (15) when evaluated represents the corresponding point P = [X Y 1]t in the sensed image. f0ij shows the homography computed at block ij in the bottom level. The farther the center of block ij is from point (x, y), the smaller the influence of f0ij will be on the overall transformation computed at (x, y). The weight functions are selected in such a way that the sum of all weight functions everywhere in the reference image domain becomes one, with each weight function monotonically decreasing from the center of the block it is associated with. This property ensures that the overall transformation will closely follow the local transformation at a block because the influence of distant transformations on the block becomes small and negligible. If rational weight functions are used, it is easy to ensure that the sum of weights everywhere in the image domain becomes

2142

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 4, APRIL 2015

one. Consider associating the inverse distance radial function centered at (xij , yij ) with the ijth block 1 (16) rij (x, y) = 1 . [(x − xij )2 + (y − yij )2 ] 2 Then, rational weight functions defined by wij (x, y) =

rij (x, y) M −1 −1 N

(17)

rij (x, y)

i=0 j=0

for i = 1, . . . , M − 1 and j = 0, . . . , N − 1 will guarantee a sum of one everywhere in the (reference) image domain. If rij is taken to be a Gaussian of height of one centered at (xij , yij )

(x − xij )2 + (y − yij )2 rij (x, y) = exp − (18) 2σ 2 the obtained rational weights will have a sum of one again everywhere in the image domain. Arsigny et al. [25] used the inverse distance function of (16), first proposed by Shepard [26], to represent the rational weights. Since the inverse distance function of (16) is fixed, it cannot be adjusted to the size of a block. In this paper, rational Gaussian weights are used because the standard deviation σ of the weight functions can be used as a free parameter to adapt the registration to images with different block sizes. Suitable values for σ are found to be between a half to one-fourth the radius of the circle inscribed in a block. Selecting σ in this manner ensures that the influence of a homography found at a block diminishes a few blocks away. The relation between the overall transformation Fxy and corresponding points in the images can be written as P = Fxy p.

(19)

For each point p in the reference image, (19) finds the corresponding point in the sensed image, establishing correspondence between individual points in the images and making it possible to resample the sensed image to the geometry of the reference image. The overall transformation function Fxy is a weighted sum of homographies, each registering corresponding local neighborhoods (blocks) in the images with the weights dependent on (x, y) or p. Each such homography itself is obtained from a sequence of homographies, the first being the global homography to approximately align the images globally and a sequence of local homographies accounting for geometric differences between corresponding neighborhoods (blocks) in the images as resolution is increased from the lowest to the highest. It is important to note that the transformation at the bottom level contains all homographies calculated at the upper levels. Therefore, resampling is performed using the original sensed image, resulting in a sharp resampled image. Also, note that, although the overall transformation is defined globally, when dealing with very large images, the overall transformation at a point p can be calculated using the local homographies of a small number of surrounding blocks because influence of local homographies centered at blocks farther than a few standard deviations to the point under consideration becomes smaller than half a pixel and can be ignored when using nearest neighbor resampling.

Fig. 4. (a) and (b) Aerial images of an airport scene with occluded and homogeneous regions. These images are of dimensions 1280 × 720 and are courtesy of the Brown University Computer Vision Group. (c) Absolute intensity differences of registered images at ground level by the proposed method. (d) Close-up view of the difference image within a window containing an airplane.

D. Handling Occluded and Homogeneous Regions Images from different views of a 3-D scene often contain occluded regions along depth discontinuities. Since locations of depth discontinuities are not known beforehand, landmarks detected at occluded regions cannot be identified. Such landmarks do not produce correspondences when used in RANSAC. Homogeneous blocks also do not produce landmarks and consequently do not produce correspondences. When the images are subdivided into small blocks, some blocks may completely fall on occluded or homogeneous regions. Such blocks produce insufficient or no correspondences to enable estimation of the homography parameters. For example, consider the images of an airport scene shown by (a) and (b) in Fig. 4. Landmarks detected in the images represent corners of airplanes and markings on the ground. The asphalt and the grass areas hardly produce any landmarks. The images at the lowest resolution are registered using landmarks of the airplanes and markings on the ground because both types of landmarks satisfy the homography constraint. As resolution of the images is increased, landmarks that belong to the airplanes, which are above the ground, do not satisfy the homography constraint and gradually drop from the registration process. Consequently, objects above the ground appear misregistered due to image disparity caused by view-angle difference between the images, as shown in images (c) and (d) in Fig. 4. Generally, higher disparities are obtained for scene points that are higher above the ground. If sufficient correspondences are not obtained at a block at level l in the image hierarchy, or if average absolute intensity difference between registered blocks is greater than that determined at level l + 1, the transformation obtained at level l + 1 is used to register the blocks at level l. In this manner, once the homography at the top level is determined, determination of homographies at all blocks and at all levels is guaranteed. Discussions above focused on registration of individual video frames. After registering the frames, there is sometimes a need to create a mosaic of the area covered by the video. Even when individual frames register accurately, when registration

LINGER AND GOSHTASBY: AERIAL IMAGE REGISTRATION FOR TRACKING

2143

TABLE I D EPENDENCE OF THE P ROPOSED M ETHOD ON S CENE C ONTENT AND V IEW-A NGLE D IFFERENCE B ETWEEN C ONSECUTIVE F RAMES . T HE N UMBER OF F RAMES N IN A V IDEO I S S HOWN IN THE S ECOND ROW, AND THE M AXIMUM D IFFERENCE IN F RAME N UMBERS k IN A V IDEO S EQUENCE T HAT CAN BE R EGISTERED W ITHOUT B REAKING THE R EGISTRATION C HAIN I S S HOWN IN THE T HIRD ROW

Fig. 5. This figure shows a frame from each of the 17 videos used in this study. The videos are numbered from left to right and from top to bottom. Frames in the videos are of dimensions 1280 × 720 pixels.

is performed over a long sequence, very small registration errors can accumulate and become large, causing a video that is captured along a closed circular path to produce a mosaic that does not close. Molina and Zhu [27] discuss means to avoid drifts caused by accumulation of registration inaccuracies and ensure that frames in a video captured along a closed circular path register and produce a seamless closed circular mosaic. IV. R ESULTS To evaluate the robustness and performance of the proposed registration on various types of images, the Providence Aerial Multiview data set prepared by the Brown University Computer Vision Group is used. The data set contains 17 aerial videos captured from a helicopter over the city of Providence, RI, and its surroundings in a nearly circular path and at about the same altitude to produce a 30-cm/pixel resolution at ground level. The helicopter makes a large circular path that can be considered locally straight. No information about variation in speed of the helicopter during acquisition of the videos is available. A frame from each video is shown in Fig. 5. Both color and grayscale videos are available except video number 13, which comes only in gray scale. We have chosen the grayscale videos in this study. When registering frames in a color video, either the grayscale version of the frames should be used to detect the landmarks or a detector that uses color information [28], [29] should be used to find the landmarks. The videos represent from nearly flat scenes to scenes containing high-rise buildings. If a video contains N frames, we register frame i to frame i + k, skipping k − 1 frames. There-

fore, frame 0 is registered to frame k, frame 1 is registered to frame k + 1, and so on until frame N − k − 1 is registered to frame N − 1. For a video containing N frames and skipping k − 1 frames (k ≥ 1) while registering the frames, there will be N − k − 1 registration cases. For each video, first, no frames are skipped when registering the frames (k = 1). If the process succeeds for all frames, k is incremented by one, and the process is repeated until the process fails to register at least one pair of frames in the sequence. This happens when the images at the top level (lowest resolution) fail to register, thereby failing registration at lower levels. When two frames in a sequence register correctly, average absolute intensity difference between corresponding landmarks is rather small, but when two frames do not register correctly, this absolute intensity difference is relatively large. In our implementation, when average absolute intensity difference at corresponding landmarks in a registration increases by a factor greater than two when compared to the preceding registration, the registration is considered incorrect. Gonçalves et al. [30] have proposed a number of other measures that can be used to identify registration failure. The results obtained for the 17 videos are shown in Table I. Consecutive frames from a video containing tall buildings have more occlusion and geometric differences than consecutive frames in a video of a relatively flat scene. As the speed of the platform increases, view-angle difference between consecutive frames increases. Considering that larger view-angle differences and taller structures increase occlusion and geometric differences between consecutive frames, it can be concluded that a smaller parameter k for a video in Table I is a clue that at least a part of the video contains either taller structures or is obtained with a higher platform speed, or both. When the images are approximately aligned by the global homography, buildings that are above the ground appear displaced with respect to each other in the images. As resolution is increased, this displacement, known as disparity, increases. Therefore, after the global alignment, corresponding blocks in the images at lower levels may not contain the same scene parts due to displacement of scene structures in the images. Consequently, landmarks detected in corresponding blocks may not produce sufficient correspondences to locally register the blocks. Such blocks are registered using the homography obtained at one higher level, which itself could be coming from a level above it, and this can go up all the way to the top level. To demonstrate this phenomenon, consider frames 100 and 102 of video 13, as shown in (a) and (b) in Fig. 6. Absolute intensity differences of images after registration are shown in

2144

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 4, APRIL 2015

Fig. 6. (a) and (b) Frames 100 and 102 of video 13. (c) Registration of the frames at ground level by the proposed method. The area within the rectangular window is zoomed in (d) for better viewing. Cars in motion appear as bright blobs in the difference image. The arrows show motion of cars from frame 100 to frame 102.

Fig. 6(c). Ground areas visible in both images are registered with high precision using information at the highest resolution in images, while image areas belonging to the buildings are registered only approximately using the global homography obtained at the top level. A small area of Fig. 6(c) at ground level is zoomed in Fig. 6(d) for better viewing. Motion of cars from frame 100 to frame 102 is shown by arrows. The ground level after registration and subtraction appears dark. Moving cars appear like bright blobs against the dark ground. Two bright vertical bars in this image represent traffic light structures that are above the ground and displaced from one frame to another. Many of the videos used in the experiments contain a relatively flat ground level. To demonstrate the suitability of the proposed method in the registration of hilly scenes, images (a) and (b) in Fig. 7 are used. The images are obtained from different views and distances of a hilly urban scene. Registration of the images are shown in Fig. 7(c). As long as the ground level in the scene can be considered locally flat, and there are sufficient ground areas visible in both images, the proposed method should be able to register the images. Finally, images of a coastal town shown in (d) and (e) in Fig. 7 with considerable homogeneous regions are registered. Lack of landmarks at some blocks or presence of landmarks from waves that do not produce correspondences at higher resolutions prompts the process to use homographies calculated at lower resolutions to register some of the blocks at the bottom level. The process uses high-resolution information, where available, to register some blocks with high precision and uses lower resolution information to approximately register homogeneous blocks and blocks that do not produce sufficient correspondences. Computation time for registering two images of dimensions 1280 × 720 varies between 1.5 and 2.5 s on a Windows PC with an Intel Core i7 processor. More time is spent on image pairs with blocks containing high rises that are displaced with respect to each other. For blocks with displaced contents, considerable time is spent searching for corresponding points that do not exist. The breakout of the computation times is as follows: 20% for landmark selection and feature extraction in both images, 5% for finding putative correspondences using the five invari-

Fig. 7. (a) and (b) Images from different views of a hilly urban scene and (c) registration of the images. (d) and (e) Images of a coastal town with considerable homogeneous regions and presence of waves that move irregularly from one image to another and (f) registration of the images. Images (a), (b), (d), and (e) are of dimensions 1510 × 877, 1130 × 1076, 1844 × 2271, and 2155 × 2292, respectively, and are courtesy of Image Registration and Fusion Systems.

ant features of the landmarks, 25% for finding the landmark correspondences by RANSAC, and 50% for estimating the individual homographies, calculating the overall transformation function, and resampling the sensed image to the geometry of the reference image. V. S UMMARY AND C ONCLUSION A method for registering multiview aerial images to facilitate target tracking was described. The method has the following characteristics. 1) It registers images at the ground level using the homography constraint in landmark correspondences, thereby avoiding landmarks that belong to buildings and moving targets when determining the registration parameters. 2) It defines an overall transformation in terms of a global transformation that depends on the camera positions with respect to the scene and a number of local transformations that characterize the contents of the scene. 3) It registers images captured with different camera orientations by using rotationally invariant landmark features. 4) It registers images at the ground level even when the ground changes in slope and elevation. This is achieved by using different homographies to register different image neighborhoods/blocks. 5) As long as the images cover a continuous and smooth ground and there are sufficient correspondences at ground level, it can find a continuous and smooth transformation to register the images at ground level. 6) The proposed method is particularly useful when there is a need to track moving targets in an urban or suburban scene. The ability to register images at ground level makes it possible to remove camera motion and keep only target motion in registered images. 7) Limited ground areas visible to both images limit usage of the method. If sufficient ground areas do not appear in the images due to occlusion or lack of contrast, sufficient correspondences required to determine the global homography may not be possible to find, failing the registration.

LINGER AND GOSHTASBY: AERIAL IMAGE REGISTRATION FOR TRACKING

8) The proposed method cannot register images captured by different sensors of a scene due to the inability to find corresponding landmarks in such images. ACKNOWLEDGMENT The authors would like to thank the reviewers for their insightful comments. The authors would also like to thank the Brown University Computer Vision Group and Image Registration and Fusion Systems for making available the videos and images used in this study. The editorial assistance of L. Stephens in preparation of this manuscript is also greatly appreciated. R EFERENCES [1] H. Gonçalves, L. Corte-Real, and J. A. Gonçalves, “Automatic image registration through image segmentation and SIFT,” IEEE Trans. Geosci. Remote Sens., vol. 49, no. 7, pp. 2589–2600, Jul. 2011. [2] T. Kim and Y.-J. Im, “Automatic satellite image registration by combination of matching and random sample consensus,” IEEE Trans. Geosci. Remote Sens., vol. 41, no. 5, pp. 1111–1117, May 2011. [3] A. Wong and D. A. Clausi, “ARRSI: Automatic registration of remotesensing images,” IEEE Trans. Geosci. Remote Sens., vol. 45, no. 5, pp. 1483–1493, May 2007. [4] X. Fan, H. Rhody, and E. Saber, “A spatial-feature-enhanced MMI algorithm for multimodal airborne image registration,” IEEE Trans. Geosci. Remote Sens., vol. 48, no. 6, pp. 2580–2589, Jun. 2010. [5] Z. Liu, J. An, and Y. Jing, “A simple and robust feature point matching algorithm based on restricted spatial order constraints for aerial image registration,” IEEE Trans. Geosci. Remote Sens., vol. 50, no. 2, pp. 514– 527, Feb. 2012. [6] J. Ma, J. C.-W. Chan, and F. Canters, “Fully automatic subpixel image registration of multiangle CHRIS/Proba data,” IEEE Trans. Geosci. Remote Sens., vol. 48, no. 7, pp. 2829–2839, Jul. 2010. [7] P. Cheng and C.-H. Menq, “Real-time continuous image registration enabling ultraprecise 2-D motion tracking,” IEEE Trans. Image Process., vol. 22, no. 5, pp. 2081–2090, May 2013. [8] K. Jackovitz, V. Asari, E. Balster, J. Vasquez, and P. Hytla, “Registration of region of interest for object tracking applications in wide area motion imagery,” in Proc. AIPR, 2012, pp. 1–8. [9] X. Mei and F. Porikli, “Joint tracking and video registration by factorial hidden Markov models,” in Proc. IEEE Int. Conf. Acoust., Speech Signal Process., 2008, pp. 973–976. [10] I. L. Ayala, D. A. Orton, J. B. Larson, and D. F. Elliott, “Moving target tracking using symbolic registration,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 4, no. 5, pp. 515–520, Sep. 1982. [11] B. P. Jackson and A. A. Goshtasby, “Registering aerial video images using the projective constraint,” IEEE Trans. Image Process., vol. 19, no. 3, pp. 795–804, Mar. 2010. [12] M. Gong, S. Zhao, L. Jiao, D. Tian, and S. Wang, “A novel coarse-tofine scheme for automatic image registration based on SIFT and mutual information,” IEEE Trans. Geosci. Remote Sens., vol. 52, no. 7, pp. 4328– 4338, Jul. 2014. [13] B. P. Jackson and A. A. Goshtasby, “Adaptive registration of very large images,” in Proc. CVPR Workshop Registration Very Large Images, Columbus, OH, Canada, Jun. 2014, pp. 345–350. [14] S. R. Lee, “A coarse-to-fine approach for remote-sensing image registration based on a local method,” Int. J. Smart Sens. Intell. Syst., vol. 3, no. 4, pp. 690–702, Dec. 2010. [15] B. Likar and F. Pernus, “A hierarchical approach to elastic registration based on mutual information,” Image Vis. Comput., vol. 19, no. 1/2, pp. 33–44, Jan. 2001.

2145

[16] A. Goshtasby, “Registration of images with geometric distortions,” IEEE Trans. Geosci. Remote Sens., vol. 26, no. 1, pp. 60–64, Jan. 1988. [17] F. L. Bookstein, “Principal warps: Thin-plate splines and the decomposition of deformations,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 11, no. 6, pp. 567–585, Jun. 1989. [18] S. Gauglitz, T. Höllerer, and M. Turk, “Evaluation of interest point detectors and feature descriptors for visual tracking,” Int. J. Comput. Vis., vol. 94, no. 3, pp. 335–360, Sep. 2011. [19] T. Tyutelaars and K. Mikolajczyk, “Local invariant feature detectors: A survey,” Found. Trends Comput. Graph. Vis., vol. 3, no. 3, pp. 177–280, Jan. 2008. [20] C. Harris and M. Stephens, “A combined corner and edge detector,” in Proc. 4th AVC, 1988, pp. 147–151. [21] D. Lowe, “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vis., vol. 60, no. 2, pp. 91–110, Nov. 2004. [22] A. Goshtasby, Image Registration: Principles, Tools and Methods. New York, NY, USA: Springer-Verlag, 2012. [23] R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, 2nd ed. Cambridge, U.K.: Cambridge Univ. Press, 2003, ch. 4. [24] P. Chen and D. Suter, “Rank constraints for homographies over two views: Revisiting the rank four constraint,” Int. J. Comput. Vis., vol. 81, no. 2, pp. 205–225, Feb. 2009. [25] V. Arsigny, X. Pennec, and N. Ayache, “Polyrigid and polyaffine transformations: A novel geometrical tool to deal with non-rigid deformations—Application to the registration of histologic slices,” Med. Image Anal., vol. 9, no. 6, pp. 507–523, Dec. 2005. [26] D. Shepard, “A two-dimensional interpolation function for irregularly spaced data,” in Proc. 23rd Nat. Conf. ACM, 1968, pp. 517–524. [27] E. Molina and Z. Zhu, “Persistent aerial video registration and fast multiview mosaicing,” IEEE Trans. Image Process., vol. 23, no. 5, pp. 2184– 2192, May 2014. [28] P. Montesinos, V. Gouet, and R. Deriche, “Differential invariants for color images,” in Proc. 14th Int. Conf. Pattern Recogn., 1998, vol. 1, pp. 838–840. [29] J. van de Weijer, T. Gevers, and A. D. Bagdanov, “Boosting color saliency in image feature detection,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 1, pp. 150–156, Jan. 2006. [30] H. Gonçalves, J. A. Gonçalves, and L. Corte-Real, “Measures for an objective evaluation of the geometric correction process quality,” IEEE Geosci. Remote Sens. Lett., vol. 6, no. 2, pp. 292–296, Apr. 2009.

Michael E. Linger received the A.A.S. degree in electronics engineering from the ITT Technical Institute, Dayton, OH, USA, and the B.S. degree in mathematics, the B.S. degree in computer science, and the M.S. degree in computer science from Wright State University, Dayton, OH, where he is currently working toward the Ph.D. degree in the Department of Computer Science and Engineering. His current project involves reconstruction of urban scenes in 3-D using line features in images.

A. Ardeshir Goshtasby received the B.E. degree in electronics engineering from the University of Tokyo, Tokyo, Japan, the M.S. degree in computer science from the University of Kentucky, Lexington, KY, USA, and the Ph.D. degree in computer science from Michigan State University, East Lansing, MI, USA. He is currently a Professor of computer science and engineering with Wright State University, Dayton, OH, USA. His main area of research is image registration and has authored two books in the area: 2-D and 3-D Image Registration for Medical, Remote Sensing and Industrial Application (Wiley, 2005) and Image Registration: Principles, Tools and Methods (Springer, 2012).

Aerial Image Registration For Tracking ieee.pdf

terrain scenes where man-made structures are absent, the blob. detector is more appropriate. If no information about the images. is available, one may use one detector as the default and, if that. fails, use the alternate detector. In our work, since the focus is. on images of urban and suburban scenes, the Harris detector.

947KB Sizes 0 Downloads 201 Views

Recommend Documents

Image Transformation for Object Tracking in High ...
and a wireless data transmitter walks across the field of view of a camera. In case of a ..... Computer Vision, Cambridge Univ. Press, March 2004. [5] S. Park and ...

Image processing in aerial surveillance and reconnaissance: from ...
Top and side view of 3D point cloud of the reconstructed scene (right). 4. ... In Figure 4-3 three different zoom levels for aerial video data are shown. ..... time or for forensic applications, Some image processing techniques such as image enhancem

a niche based genetic algorithm for image registration
Image registration aims to find the unknown set of transformations able to reduce two or more images to ..... of phenotypic similarity measure, as domain-specific.

Removing mismatches for retinal image registration via ...
16 Aug 2016 - 1(d), while Fig. 1 (i) shows the perfect image registration after removing mismatches. Furthermore, vector field interpolation shows the smooth vector field learning from the feature matching. For retinal image registration, however, mu

interpreting ultrasound elastography: image registration ...
brains and prostates due to the availability of anatomical structures that can be readily ... parameters to fit onto the reference point set (considered as data.

Multiresolution Feature-Based Image Registration - CiteSeerX
Jun 23, 2000 - Then, only the set of pixels within a small window centered around i ... takes about 2 seconds (on a Pentium-II 300MHz PC) to align and stitch by ... Picard, “Virtual Bellows: Constructing High-Quality Images for Video,” Proc.

Cyclone Tracking using Multiple Satellite Image Sources
Nov 6, 2009 - California Institute of. Technology. Pasadena, CA 91109. {Anand. .... sures near ocean surface wind speed and direction under all weather and ...

(Aerial) Data for Disaster Response
Fully automated methods based on supervised learn- .... We could add a second quality assurance control by using our ..... A comprehensive analysis of building.

Piksi™ for UAV Aerial Surveying -
... and Ground Control. 6. 4.2 Mission Planning and Camera Configuration. 6. 5.0 Post-Processing Techniques ..... the edge of the mission range. This placement ...

Non-Rigid Image Registration under Non-Deterministic ...
aDepartment of Electrical and Computer Engineering ... ∗*This work was partially supported by the National Science Foundation ... al.10 propose a groupwise, non-registration algorithm to simultaneously register all subjects in a population to.

Localization and registration accuracy in image guided ... - Springer Link
... 9 January 2008 / Accepted: 23 September 2008 / Published online: 28 October 2008. © CARS 2008. Abstract. Purpose To measure and compare the clinical localization ..... operating room, the intraoperative data was recorded with the.

Lucas-Kanade image registration using camera ...
a photograph. We then estimate only the extrinsic parameters for image registration, considering two types of camera motions, 3D rotations and full 3D motions with translations and rotations. As the known ..... [1] Lucas, B. and Kanade, T., “An ite

Medical image registration using machine learning ...
Medical image registration using machine learning-based interest ... experimental results shows an improvement in 3D image registration quality of 18.92% ...

Image Registration by Minimization of Residual ...
plexity of the residual image between the two registered im- ages. This measure produces accurate registration results on both artificial and real-world problems that we have tested, whereas many other state-of-the-art similarity mea- sures have fail

Multi-Modal Medical Image Registration based on ...
(spatial index), s (imaging plane index); so that the 3-D snake function to be optimized is defined as f(v) = Į1ŒvrŒ +. Į2ŒvsŒ+ ȕ1ŒvrrŒ + ȕ2ŒvssŒ + ȕ4ŒvrsŒ+ E, where {Įi} are constants imposing a tension constraint, and {ȕi} are cons

LNCS 4191 - Registration of Microscopic Iris Image ... - Springer Link
Casey Eye Institute, Oregon Health and Science University, USA. {xubosong .... sity variance in element m, and I is the identity matrix. This is equivalent to.

Visual Servoing from Robust Direct Color Image Registration
as an image registration problem. ... article on direct registration methods of color images and ..... (either in the calibrated domain or in the uncalibrated case).

Visual Servoing from Robust Direct Color Image Registration
article on direct registration methods of color images and their integration in ..... related to surfaces. (either in the calibrated domain or in the uncalibrated case).

Image Registration by Minimization of Mapping ...
deformation we put a uniform grid of 6 × 6 control points over the image, randomly .... is a trade-off between the computational load and accuracy of the registration. ... quire storage of eigenvectors, and can be a choice for the large data-set ...

non-rigid biomedical image registration using graph ...
method, we discuss two different biomedical applications in this paper. The first ... biomedical images in a graph cut-based framework, we compare our method with .... typical desktop with Intel(R) Core(TM)2 Duo processor with a speed of 2.8 ...

A local fast marching-based diffusion tensor image registration ...
relatively low resolution of DTI and less advanced alignment techniques in the initial works, global brain registration was also applied to quantitatively ...... Illustration of the local neighborhood and the deformed tensors. (a) The tensors in the 

Localization and registration accuracy in image guided neurosurgery ...
navigation system (Medtronic Inc, USA) on a laptop installed with the ... 9(Pt. 2):637-44. 3. Fitzpatrick, J.M., J.B. West, and C.R. Maurer, Jr., Predicting error in ...

Aerial Filming.pdf
aerial farm photography. aerial filming. aerial filming drone. aerial filming services. aerial footage. Page 3 of 5. Aerial Filming.pdf. Aerial Filming.pdf. Open.