Accuracy improvement and evaluation measures for registration of ...

4 downloads 127 Views 1MB Size Report
Image registration, Intensity-based, Similarity measure, Accuracy assessment, Hyperion, IKONOS. Kabul: 25 Aralık .... bedded into low resolution using spatial domain techniques. .... (Url-3) has four multispectral bands with 4 m spatial reso-.
UCTEA Chamber of Surveying and Cadastre Engineers

Journal of Geodesy and Geoinformation

TMMOB Harita ve Kadastro Mühendisleri Odası

Jeodezi ve Jeoinformasyon Dergisi

Vol.1  Iss.2  pp.97-103  November 2012  Journal No.106  Doi: 10.9733/jgg.251212.1 www.hkmodergi.org

Accuracy improvement and evaluation measures for registration of multisensor remote sensing imagery Deniz Gerçek*, Davut Çeşmeci, Mehmet Kemal Güllü, Alp Ertürk, Sarp Ertürk Kocaeli University Laboratory of Image and Signal Processing (KULIS), Electronics and Telecommunication Engineering Deptartment, Umuttepe Campus, Kocaeli, Turkey

Abstract Accepted: 25 December 2012 Received: 24 September 2012 Pub. Online: 12 March 2013 Volume: 1 Issue: 2 Page: 97 - 103 November 2012

Intensity based image registration methods are widely used in fine geometric registration of multisensor images. Accordingly, for images that are compared through translation of image templates, position where similarity measure is maximized is assumed to indicate best registration. Image registration quality is of crucial importance especially for studies that have high geometric accuracy requirements; e.g. image fusion, change detection, multichannel segmentation, and Digital Terrain Model (DTM) generation. Accuracy of image registration is conventionally evaluated by means of error measures (e.g. RMSE) obtained through comparison of coordinates of control points from the target and the reference / ground truth. However, especially for multisensor images with low spatial resolution component, difficulty in precisely positioning control points inhibits both sub pixel accuracy and evaluation of the registration. In this study, three widespread measures in intensity-based image registration namely, Normalized Cross Correlation (NCC), Mutual Information (MI), and Phase Correlation (PC) are tested for registering images acquired from EO-1 Hyperion and IKONOS sensors. We propose the use of ‘global similarity’ and ‘inverse consistency’ measures for evaluating the performance of these intensity based automated registration methods. Keywords Image registration, Intensity-based, Similarity measure, Accuracy assessment, Hyperion, IKONOS.

Özet Farklı alıcılardan elde edilen uydu görüntülerinin çakıştırılmasında başarım artırma ve değerlendirme ölçütleri Kabul: 25 Aralık 2012 Alındı: 24 Eylül 2012 Web Yayın: 12 Mart 2013 Cilt: 1 Sayı: 2 Sayfa: 97 - 103 Kasım 2012

Farklı uydu alıcılarından alınan görüntülerin hassas biçimde çakıştırılmasında parlaklık temelli yöntemler yaygın olarak kullanılmaktadır. Buna göre, ötelenerek birbiri ile karşılaştırılan görüntü parçalarında, benzerlik ölçütünü enyüksekleyen konum, en iyi çakışma olarak kabul edilir. Çakıştırmanın kalitesi özellikle imge kaynaştırma, değişim tespiti, çok kanallı bölütleme, Sayısal Arazi Modeli (SAM) üretimi vb. çalışmalar için kritik öneme sahiptir. Çakıştırma başarımı genellikle hedef imge ve referans / yer doğrusu üzerinde bulunan ortak nesne koordinatlarının karşılaştırılmasından elde edilen hata ölçütleri ile (ör. RMSE) ile değerlendirilir. Ancak, özellikle farklı alıcılardan elde edilen görüntülerde düşük çözünürlüklü bir bileşen var ise, kontrol noktalarını konumlandırmadaki güçlük, çakıştırmada piksel altı başarımı düşürmekte ve çakıştırma değerlendirmesini zorlaştırmaktadır. Bu çalışmada farklı alıcı karakteristiklerine sahip görüntülerin parlaklık temelli otomatik yöntemlerle çakıştırılmasında yaygın olarak kullanılan üç yöntem; Normalize Çapraz Korelasyon (NCC), Ortak Bilgi (MI) ve Faz Korelasyonu (PC), EO-1 Hyperion ve IKONOS alıcılarından elde edilen görüntüleri çakıştırmak üzere test edilmektedir. Her bir yönteme göre elde edilen çakıştırma sonuçlarının başarımını değerlendirmek üzere, ‘global benzerlik’ ve ‘ters tutarlılık’ ölçütlerinin kullanımı önerilmektedir. Anahtar Sözcükler İmge çakıştırma, Parlaklık temelli, Benzerlik ölçütü, Başarım değerlendirme, Hyperion, IKONOS.

*Corresponding

Author: Tel: +90 (262) 3033392

Fax: +90 (262) 3033003

E-mail: [email protected] (Gerçek D.), [email protected] (Çeşmeci D.), kemalg@ kocaeli.edu.tr (Güllü M.K.), [email protected] (Ertürk A.), [email protected] (Ertürk S.), © 2012 HKMO

98 Accuracy improvement and evaluation measures for registration of multisensor remote sensing imagery

1. Introduction Multi-sensor image analysis has been a major field of interest in remote sensing. Bringing together images from different sensors, views and time periods is essential for extracting relevant information for particular studies. However, multi-modalities/multi-sensor characteristics are usually a challenge in bringing images into a common basis of comparison or aggregation. Registration of images in this context is a fundamental prerequisite. Along with that, particular studies e.g. data/image fusion (Mitianoudis and Stathaki 2007; Yokoya et al. 2012), resolution enhancement (Eismann and Hardie 2008), change detection (Dai and Khorram 1998), stereo surface modeling (Baltsavias et al. 2008) demand higher geometric accuracy standards that may necessitate quantitative evaluations. For such studies, accuracy errors not exceeding 0.1–0.2 of the low-resolution pixel size is expected for image co-registration (Dai and Khorram 1998; Zhukov et al. 1999). Accuracy of image registration is conventionally evaluated by measuring error between control points that both exist in target and the reference images. However especially for the remote sensing images of very different spatial resolution it is barely possible to precisely locate corresponding ground control points both in high and low resolution images. Hence, there is no basic standard for evaluating accuracy. Image registration, so-called image co-registration is defined as geometrically aligning images from different sources, different periods or views (Brown 1992). Registering images from different sensors has been a hot research topic in the literature for over a decade with the advent of sensors with various characteristics. There are numerous remote sensing satellites in mission. GeoEye and DigitalGlobe are the main commercial satellite organizations that provide high spatial resolution multispectral images. Hyperspectral images on the other hand provide high spectral resolution with contiguous and narrow bands in a wide range spectrum. Therefore, they outperform on target detection and classification over multispectral images. However, hyperspectral images have the disadvantage of low spatial resolution due to technical constraints that seem to last for the near future (Thomas et al. 2008). This encourages the research on enhancement of spatial resolution of Hyperspectral images. Hyperspectral and optical multisensor images are also complementary in a way that hyperspectral images are of high spectral and low spatial resolution, and multispectral images are vice versa. Therefore, an information content of high spectral-spatial resolution is anticipated from multisensor image analysis, provided that these two imageries are registered with high accuracy. In this study, image registration with high accuracy is proposed prior to analysis intended for data fusion of Hyperion and IKONOS multispectral imagery. Amount of information that can be gathered from multiple sources is significantly higher when compared to that of datasets would individually produce (Bunting et al. 2010). However there are also challenges in registering multiple-source remotely sensed data, especially when there is big difference in their spatial resolution. In image fusion, ratio between spatial resolution of the high-resolution image and that of the low-resolution multispectral image; so-called ‘spatial resolution ratio’ is known

to influence the quality of the fused image. Experimental results indicate that a spatial resolution ratio of higher than 1/10 is usually desired for optimal multisensor image fusion (Ling et al. 2008). In image fusion approaches that intend information extraction or classification; high resolution content can be embedded into low resolution using spatial domain techniques. These techniques transfer high-resolution information from the high-resolution image to low resolution spectral bands using various deterministic or statistical predictors (Zhukov et al. 1999; Yokoya et al. 2012). High resolution content can also be utilized in the form of objects constructed from high spatial resolution image using object-based techniques (Mitianoudis and Stathaki 2007). Image registration approaches are broadly grouped into two as feature-based and intensity-based (Zitova and Flusser 2003). Feature-based methods require a set of control points that can be detected both in the reference and the target imagery. As this is not quite possible for the images with very different spatial resolution, we have focused on intensity based registration methods that utilize intensity of image templates to match the image pairs. Accuracy of each method is evaluated by means of ‘global similarity’ and ‘inverse consistency’ of transformation. High global similarity between transformed target image and the reference image is assumed to be an indicator of high registration accuracy. Similarly, higher consistency between forward and backward transformation is assumed to produce more accurate results.

2. Image registration methods and evaluation measures Image registration has many potential applications in various fields of research including computer vision, medical imaging and remote sensing. During the last decades, growing amount and diversity of obtained images invoked the research on automated image registration. A complete view of image registration research in two decades and mainstream approaches can be found in Brown (1992), Zitova and Flusser (2003), Wyawahare et al. (2009), and Gruen (2012). Image registration approaches are broadly grouped into two as feature-based and intensity-based (Zitova and Flusser 2003). Feature-based methods are applied through either manual or automated (e.g. SIFT, Harris) detection of control points that tie common features in images from different sensors. Intensity-based approaches rather make use of a similarity measure between the reference and the target and adjust the transformation until the similarity is maximized based on the assumption that the images will be most similar at the correct registration. Intensity-based methods are effective in correcting translation and rotation errors in images; hence they are widely used for medical image registration where rigid transformation is usually satisfactory (Pluim et al. 2003). However, in the case of registering satellite imageries that are usually subject to local geometric distortions, intensity-based approaches tend to fail. To overcome this problem, a widely used technique is to shift area/blocks of target image across the reference image to calculate similarity between target

Deniz Gerçek, Davut Çeşmeci, Mehmet Kemal Güllü, Alp Ertürk, Sarp Ertürk / Vol.1  Iss.2  2012 99

and the reference image templates that is called area-based approach (Inglada and Giros 2004). Similarity measures calculate the similarity of templates of target and reference images, and translate target within a buffer zone across the reference to find the translation where they best match. These translations are then utilized to transform target image to match the reference. Three similarity measures namely Normalized cross-correlation (NCC), Mutual Information (MI), and Phase Correlation (PC) that are widely used in the literature to register images are examined and tested for this study. 2.1 Similarity measures used in intensity-based image registration Normalized Cross-Correlation (NCC) among similarity measures is widely used for registering images of the same modality. Correlation coefficient (r) as formulated in Equation 1 is assumed to be at its maximum for the best match. r

r=

c

∑ ∑ [ ( B1 (i, j ) − µ1 ) *( B2 (i, j ) − µ2 )]

=i 1 =j 1

( B1 (i, j ) − µ1 )

2

( B2 (i, j ) − µ2 )



(1)

2

r is the correlation coefficient and µ is the mean value of related image block ‘B’. Mutual Information (MI) originating from ‘information theory’ is a measure of statistical dependency rather than linear relationship between image pairs as in NCC. Therefore it is particularly suitable for registration of multi-modal/sensor images (Fookes and Bennamoun 2002). MI calculates dependency as the difference between marginal entropies and joint entropy as shown in Equation 2 and is assumed to get its maximum when the images totally match. Normalized MI (NMI) calculated as in Equation 3 is reported to have improvement on registration results compared to MI (Suganya et al. 2010). MI ( B1 , B2 ) = H ( B1 ) + H ( B2 ) − H ( B1 , B2 )

(2)

MI ( B= H (( B1 ) + H ( B2 )) / H ( B1 , B2 ) 1 , B2 )

(3)

Based on Shannon’s theory, marginal entropies of the two random variables and their joint entropy are formulated in Equation 4, and 5 respectively. H ( B1 ) =− ∑ PB1 (a) ⋅ log PB1 (a)

(4)

H ( B1 , B2 ) = ∑ − PB1 , B2 (a, b) ⋅ log PB1 , B2 (a, b)

(5)

a

a ,b

a and b are intensities of corresponding pixels in two images. PG1 (.) describes the probability of a given intensity value being present in the image block ‘B’. Phase Correlation (PC) is a frequency-domain based similarity measure that provides straightforward estimation of translation between two images, which is based on Fourier shift property (Yan and Li 2008). If B1(i,j) and B2(i,j)

represent two image blocks and X1[u,v] and X2[u,v] show the corresponding discrete Fourier transforms (DFTs), the phase correlation is formulated as in Equation 6.   X [u , v] × X 2*[u , v]  S [n] = F −1  1  X1[u , v] × X 2*[u , v]   

(6)

where F-1 represents the inverse discrete Fourier transform. A window function e.g. Hamming is usually applied prior to Fourier transformation to remove edge effects and increase accuracy (Stone et al. 2001; Keller and Averbuch 2007). If the two images are exactly the same, i.e. B1(i,j) = B2(i,j), phase correlation result is obtained as a peak of unity located at (i,j) = 0. However, in our case where the two images to be registered are not identical due to noise, intensity changes, etc., height of the peak of phase correlation is used as a good similarity measure for matching. Phase correlation has another property that motivates its use in image registration. That is, if two images differ merely by translation, than the amount of this translation (or shift) can be obtained by phase correlation as formulated in Equation 7, and 8 respectively. B2 (i, j ) = B1 ( i − t1 , j − t2 )

S [i, j ] = F

−1

(7)

 − t iu − t iv  X1 [ k ] × X1 [ k ] × e 1 e 2  *  X [ k ] × X [ k ] × e −tiu 1 1 

(

(

)

*

)  =δ [i − t , j − t ] (8)  

1

2

2.2. Evaluation measures for image registration Inverse consistency: Consistency of registration can be evaluated by analyzing ‘invertibility’ property of transformation. ‘Forward transformation’ is described as transforming target to match reference, and ‘reverse transformation’ is described as transforming reference to match target image. Ideally ‘forward’ transformation should equal the ‘inverse of the reverse’ transformation producing a consistent correspondence between two images (Christensen et al. 2006). However, this is usually not the case for an actual image registration problem. Inconsistency occurs as a result of the matching criteria’s inability to uniquely describe the correspondence between two images (Christensen and Johnson 2001). Inverse consistency is quantified with an error statistic between a forward and reverse transformation, i.e. RMSE. Global Similarity: Intensity based registration methods aim at optimizing image intensity similarities between the source and the target images. Global similarity assessment is based on the assumption that transformed target image is in good correspondence with the reference image. This correspondence can be quantified by obtaining similarity between images. High similarity is assumed to indicate good registration, although not fully guaranteed (Zitova and Flusser 2003).

100 Accuracy improvement and evaluation measures for registration of multisensor remote sensing imagery

3. Material and the study area Hyperspectral spaceborne imaging in mission today is limited with NASA’s Earth Observing System EO-1 Hyperion sensor (Url-1) and ESA's CHRIS/Proba (Url-2), an alternative research-oriented hyperspectral sensor that provides bands at VNIR range. EO-1 Hyperion has more than 200 spectral bands in VNIR and SWIR range with 30 m spatial resolution. There are numerous commercial spaceborne sensors that provide high spatial resolution images today. In this study, IKONOS multispectral image is utilized. IKONOS imagery (Url-3) has four multispectral bands with 4 m spatial resolution and a panchromatic band with 1 m spatial resolution. In this study Hyperion bands in the visible range and IKONOS R, G, B bands with 4 m resolution is utilized. There is a spatial resolution ratio of 1/7.5 for high resolution IKONOS and low resolution Hyperion imagery. Hyperion bands in the visible range are used as high spectral resolution component of the registration process to obtain a deformation grid for geometric transformation that can be utilized to transform all of the remaining bands of Hyperion image in the NIR and SWIR ranges. Images used in this study date back to 2007 summer and have 5 week temporal span of acquisition. Case area is near Durusu (Terkos Lake), 45 km North-West of Istanbul, and it covers a region of about 5 km2 including agricultural areas, forests, Durusu settlement, and a part of the Black sea as shown in Figure 1. Hyperion and IKONOS images being acquired from different sensors are subject to image multimodalities. These include; different spatial resolution, different spectral resolution, and intensity changes between the imageries due to time span in the growing season. IKONOS image is orthorectified. However, Hyperion Level 1gst image although reported as orthorectified (URL-1) has local distortions that are attributed to the low resolution Digital Elevation Model (DEM) used in orthorectification. Multimodalities of images from two different sensors that entail consideration in our case are given in Table 1.

a

Table 1: List of modalities in Hyperion and IKONOS images used in the study Hyperion

IKONOS

Spectral resolution

bands 8-36 of 242(covering R,G,B)

3 bands (R,G,B, multispectral)

Spatial resolution

30 m

4m

Parallax

Level 1gst (orthorectified with coarse DEM)

orthorectified

Acquisition

08 Jul 2007

30 May 2007

4. Method A pixelwise registration that is based on the similarity of reference and target blocks is proposed. IKONOS image is taken as target and Hyperion image is taken as reference and ground control for accuracy evaluation as well. Although IKONOS has true geometry with orthorectification, it is used as target and transformed to match Hyperion essentially not to alter radiometry of Hyperion that is going to be further used for spectral analysis. Translation at best match per Hyperion grid (30 m) was recorded and used as displacement vectors to construct a deformation grid. Deformation grid - similar to an optical flow - is used to establish a fine sampling of the deformation to be locally minimized. Subsequently, geometric transformation relocates each IKONOS pixel in a distance-weighted manner from four nearest displacement vectors in 30 m span and reassigns pixel values by interpolation. Method is applied to multispectral R, G, B bands of IKONOS and synthetically produced R, G, B substitutes of Hyperion imagery. All of the processing has been implemented in MATLAB. 4.1 Prepocessing 4.1.1 Spectral Normalization

Hyperion imagery nominally has 10 nm bandwidth for each of its bands where IKONOS has broader multispectral response that corresponds to band 8-17, 16-27, and 27-36 of Hyperion R, G and B bands respectively. To make direct ra-

b

Figure 1: (a) RGB composite of IKONOS and (b) Synthetic RGB composite Hyperion image (as mentioned in 4.1)

Deniz Gerçek, Davut Çeşmeci, Mehmet Kemal Güllü, Alp Ertürk, Sarp Ertürk / Vol.1  Iss.2  2012 101

diometric comparison of both imageries, an aggregation that combines portions of the Hyperion bands to emulate Spectral Response Function (SRF) of IKONOS is needed. Gaussian curves that represent SRF of IKONOS sensor at R, G, B channels were used to produce weighted sum of Hyperion bands to produce synthetic Hyperion R, G, and B bands. 4.1.2 Spatial normalization

Hyperion and IKONOS imageries have to be brought into equivalent spatial resolution. IKONOS multispectral image with 4 m pixel size is downsampled to 30 m spatial resolution. Prior to downsampling, to overcome the aliasing effect in the frequency domain, a low-pass filter of window size that is twice the sampling rate is performed. For calculation of similarity masure for each translation, IKONOS image is downsampled to Hyperion resolution. 4.2 Translation of IKONOS imagery 4.2.1 Determining block size

An adequate size for IKONOS blocks is defined for translation. Size of blocks should be large enough to produce meaningful statistics and not smaller than a particular size in order to capture image features. This critical threshold is determined considering spatial autocorrelation in the images based on the idea that adjacent observations of the same phenomenon are correlated (Griffith 2003). Extent to which there exists spatial autocorrelation is assumed to give a measure for feature sizes in the images. An empirical semivariogram that is based on common quantitative index of spatial autocorrelation; Geary Ratio (GR) (Geary 1954) and between-point distances is built. An exponential variogram model fitted to the empirical semivariogram is observed to level out at a particular interpoint distance that is called ‘range’ value. Range value represents distance where points (pixels) in the image are no longer autocorrelated and taken as empirical block size in this study. Range distance is approximately 950 meter that is equivalent to an area of about 31x31 pixels blocks of 30 m pixel size of Hyperion image.

4.2.2 Translation

Blocks were translated 1 m each time within range -45 +45 m. 45 m equals to one and a half pixel of Hyperion imagery, and it covers amount of misregistration between two images as we have observed. Translation is performed in both x, y directions from the origin for each reference pixel (90x90=8100 repetitions for each successive translation) (Figure 2). Downsampling phase in preprocessing step (section 4.1.2) is repeated after each translation. Three similarity measures; NCC, MI (including both MI and NMI), and PC were used to calculate similarity between target and the reference images for each translation. A windowing function is usually applied prior to PC to improve accuracy by reducing edge effects (Stone et al. 2001). For PC, a hamming window that is 2/3 of block size is utilized. 4.3 Construction of deformation grid Translation index at x, y gives displacement direction and amount. Translation at best match is recorded per pixel as displacement vectors. This matrix of displacement vectors, so called optical flow, establishes a deformation grid that is used in geometric transformation (Figure 3). Deformation grid for this study has 30 m span in accordance with Hyperion image for its grids. 4.4 Geometric transformation Last step is geometric transformation that utilizes deformation grid to transform original IKONOS imagery. Accordingly, displacement vectors that are regularly distributed at 30 m span were used to relocate each IKONOS pixel (4 m). For each IKONOS pixel, four nearest displacement vectors are distance weighted using Euclidian metric and a new displacement per pixel is set. This produces a fine deformation grid of 4 m sampling. Pixel values are reassigned according to this deformation grid by bilinear interpolation that produces smoother output compared to nearest neighbor interpolation.

Figure 2: Translation of target block on reference image for each reference pixel

102 Accuracy improvement and evaluation measures for registration of multisensor remote sensing imagery

Figure 3: Local shifts obtained for Ikonos image in a regular grid of 30 m span

5. Experimental Results Three similarity measures that are commonplace in intensity-based image registration; NCC, MI (MI and NMI), and PC are examined for their performance and accuracy in our image registration problem. Accuracy and performance evaluations are made based on ‘inverse consistency’ of the transformation (Christensen et al. 2006) and ‘global similarity’ of the resultant images. In addition to these quantitative measures, performance is also tested in terms of process load that is quantified by time elapsed for the procedures in MATLAB. Inverse Consistency Results: Inverse consistency for this study is reported as mean and root mean squared error (RMSE) between the forward and reverse transformations of the images (Table 2). According to error statistics obtained from inverse consistency assessment, PC with the lowest error score yielded most consistent transformation among registration results based on three similarity measures. NCC follows as the second best and NMI yielded unsatisfactory results with ME and RMSE far higher compared to other two methods. Global Similarity Results: Similarity measures that have been utilized for registration; NCC, MI, and PC, this time are utilized to calculate similarity of the target IKONOS imagery and Hyperion imagery to assess accuracy of each registration. Evaluations were given as correlation coefficient ‘r’ for NCC, an entropy-based measure for NMI, and phase correlation coefficient for PC as depicted in Table 2. r for NCC can get values between 0-1 where 1 describes identical images. Entropy measure for NMI can get values between 0-2 where 2 describes identical images and PC coefficient can

get values between 0-1 where 1 describes identical images. In the columns; NCC, MI and PC are expected to get higher values for good correspondence. Values are summed totals of similarity measures for three bands; R, G and B. According to measures obtained from global similarity assessment, MI yielded poor registration accuracy with lowest similarity between transformed target and the reference image. MI has shown results of poorer quality compared to NMI as previously reported (Pluim et al. 2003; Suganya et al. 2010). Therefore MI results have been excluded. MI measures on the other side as opposed to reported results in the literature about higher performance in comparison to NCC for coping with intensity changes between images (Chen et al. 2003; Roshni and Revathy 2008) yielded very poor results in our study. PC yielded best results for all of the global similarity measures. Process load is measured as the time elapsed for single run of a translation in a workstation and is given in Table 2. Accordingly; NCC has the least process load, where NMI has the highest process load that is about five times the NCC. PC has slightly more processing load compared to NCC.

6. Conclusions In this study, we have proposed the use of ‘Inverse Consistency’ and ‘Global Similarity’ measures for evaluating accuracy of intensity based automated registration of Hyperion and IKONOS images using NCC, MI, and PC as similarity measures. Experimental results although not giving an absolute accuracy measure, present accuracy measures for evaluating efficiency of the registration. Inverse Consistency provides an error measure about geometric correspondence, where Global Similarity gives a similarity measure of intensity correspondence. Global Similarity evaluations are based on the similarity of intensity values in transformed target and the reference. Therefore, difference in intensities among images is a factor that reduces the success of this evaluation to an extent method can cope with intensity changes. As a result, Global Similarity measure is prone to actual changes in intensity values. Invertibility property providing a geometric measure is more robust to these changes and represents method’s strength for accurate geometric transformation. Process load is another critical indicator for efficiency where larger images with many block translations may demand long periods of time of processing. According to overall experimental results, PC show reasonably better evaluation results compared to the other methods. Process load for PC is slightly more than NCC, but this

Table 2: Inverse consistency error statistics and global similarity measures for transformation methods; NCC, NMI, PC. Similarity Measures

NCC NMI PC (2/3 windowing)

Forward-Backward Transformation Difference

Global Similarity

Process Load

RMSE

Mean Error

NCC

NMI

PC

Time (sec)

6.9590

3.2101

2.7218

3.7172

1.6319

*2.6

9.9147

3.9068

2.5056

3.1567

1.4397

13.9

*1.8002

*1.6721

*2.7908

*3.7183

*1.7030

3.46

(* Lowest error measure for inverse consistency, highest similarity measure for global matching, and shortest time for processing)

Deniz Gerçek, Davut Çeşmeci, Mehmet Kemal Güllü, Alp Ertürk, Sarp Ertürk / Vol.1  Iss.2  2012 103

is tolerable regarding the improvement it has made in the results. NCC has slightly worse evaluation results compared to PC. However, NMI yielded quite unsatisfactory results where MI yielded results even worse. Overall scores obtained for MI and NMI in our study in terms of accuracy and process load discourages the use of these methods in intensity based registration of multisensor remote sensing imagery Evaluations provide useful information to decide for the method to employ. However it should be noted that evaluations on accuracy of image registration is a nontrivial task. Inaccuracy might be due to errors that get involved into image registration at some processing stage. Moreover, actual physical differences in the image contents might also be attributed as inaccuracy.

Acknowledgements This study has been supported by The Scientific and Technological Research Council of Turkey (TÜBİTAK) as part of Project: 111E052.

References Baltsavias E., Gruen A., Eisenbeiss H., Zhang L., Waser L.T., (2008), High-quality image matching and automated generation of 3D tree models, International Journal of Remote Sensing, 29(5), 1243–1259. Brown L.G., (1992), A survey of image registration techniques, ACM Computing Surveys, 24(4), 325-376. Bunting P., Labrosse F., Lucas R., (2010), A multi-resolution areabased technique for automatic multi-modal image registration, Image and Vision Computing 28, 1203–1219. Chen H.M., Pramod K.,Varshney P.K., Arora M.K., (2003), Performance of Mutual Information Similarity Measure for Registration of Multitemporal Remote Sensing Images, IEEE Transactions on Geoscience and Remote Sensing, 41( 11), 2445-2454. Christensen G.E., Geng X, Kuhl J.G., Bruss J., Grabowski T.J., Allen J.S, Pirwani I.A., Vannier M.W., Damasio H., (2006), Introduction to the non-rigid image registration evaluation project (NIREP). In:Proceedings of the Third international conference on Biomedical Image Registration, SpringerVerlagBerlin, Heidelberg, pp.128-135. Christensen G.E., Johnson H.J., (2001), Consistent image registration, IEEE Transactions on Medical Imaging, 20(7), 568–582. Dai X., Khorram S., (1998), Effects of image misregistration on the accuracy of remotely sensed change detection, IEEE Transactions on Geoscience and Remote Sensing, 36(5), 1566– 1577. Eismann M.T., Hardie R.C., (2008), Hyperspectral Resolution Enhancement Using High-Resolution Multispectral Imagery With Arbitrary Response Functions, IEEE Transactions on Geoscience and Remote Sensing, 43(3), 455-465. Fookes C., Bennamoun M., (2002), The Use of Mutual Information for Rigid Medical Image Registration: A Review, 2002 IEEE Conference on Systems, Man and Cybernetics, Vol.4,October 6-9, 2002, Hammamet - Tunisia. Geary R.C., (1954), The Contiguity Ratio and Statistical Mapping, The Incorporated Statistician,5(3),115–145. Griffth D.A., (2003), Spatial Autocorrelation and Spatial Filtering, Springer Berlin, Germany.

Gruen A., (2012), Development and status of Image matching in Photogrammetry, The Photogrammetric Record, 27(137), 36–57. Inglada J.,Giros A., (2004), On the Possibility of Automatic Multisensor Image Registration, IEEE Transactions on Geoscience and Remote Sensing, 42(10), 2104-2120. Keller Y., Averbuch A., (2007), A projection-based extension to phase correlation image alignment, Signal Processing, 87(1), 124-133. Mitianoudis N., Stathaki T., (2007), Pixel-based and region-based image fusion schemes using ICA bases, Information Fusion, 8, 131–142. Ling, Y., Ehlers, M., Usery, E.L., Madden, M., (2008), Effects of spatial resolution ratio in image fusion, International Journal of Remote Sensing, 29(7), 2157-2167. Pluim J.P.W., Maintz J.B.A., Viergever M.A., (2003), MutualInformation-Based Registration of Medical Images: A Review, IEEE Transaction on Medical Imaging, 22, 986-1004. Roshni V.S., Revathy DR. K., (2008), Using mutual information and cross correlation as metrics for registration of images, Journal of Theoretical and Applied Information Technology, 4(6), 474-481. Suganya R., Priyadharsini K., Rajaram Dr. S., (2010), Intensity Based Image Registration by Maximization of Mutual Information, International Journal of Computer Applications, 1(20), 1-5. Stone H.S., Orchard M.T., Chang E.-C., Martucci S.A., (2001), A fast direct Fourier-based algorithm for subpixel registration of image, IEEE Transactions on Geoscience and Remote Sensing, 39(10), 2235-2243. Thomas C., T. Ranchin T., Wald L., Chanussot L., (2008), Synthesis of Multispectral Images to High Spatial Resolution A Critical Review of Fusion Methods Based on Remote Sensing Physics, IEEE Transactions in Geoscience and Remote Sensing, 46(5), 1301-1312. Url-1, Hyperion Sensor, USGS EO-1 Website, http://eo1.usgs.gov/, [Accessed November, 2012]. Url-2, CHRIS Proba Sensor, Proba – Earthnet Online website, https://earth.esa.int/web/guest/missions/esa-operational-eomissions/proba, [Accessed November, 2012]. Url-3: GeoEye Satellites, http://www.geoeye.com/CorpSite/ products/earth-imagery/geoeye-satellites.aspx, [Accessed November, 2012]. Wyawahare M.V., Pradeep D., Patil M., Abhyankar H.K., (2009), Image Registration Techniques: An overview, International Journal of Signal Processing, Image Processing and Pattern Recognition, 2(3), 11-28. Yan H., Li J.G., (2008), Robust Phase Correlation based feature matching for image co-registration and DEM generation, The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. XXXVII, Part B7, 17511756. Yokoya N., Yairi T., Iwasaki A., (2012), Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion, IEEE Transactions on Geoscience and Remote Sensing, 50(2), 528-537. Zitova B., Flusser J., (2003), Image registration methods: a survey, Image and Vision Computing, 21, 977-1000. Zhukov B., Oertel D., Lanzl F., Reinhackel G., (1999), UnmixingBased Multisensor Multiresolution Image Fusion, IEEE Transactions in Geoscience and Remote Sensing, 37(3), 12121226.