DOI QR코드

DOI QR Code

Automatic Registration between EO and IR Images of KOMPSAT-3A Using Block-based Image Matching

  • Kang, Hyungseok (Researcher, The 3rd R&D Institute, Agency for Defense Development)
  • Received : 2020.08.04
  • Accepted : 2020.08.18
  • Published : 2020.08.31

Abstract

This paper focuses on automatic image registration between EO (Electro-Optical) and IR (InfraRed) satellite images with different spectral properties using block-based approach and simple preprocessing technique to enhance the performance of feature matching. If unpreprocessed EO and IR images from Kompsat-3A satellite were applied to local feature matching algorithms(Scale Invariant Feature Transform, Speed-Up Robust Feature, etc.), image registration algorithm generally failed because of few detected feature points or mismatched pairs despite of many detected feature points. In this paper, we proposed a new image registration method which improved the performance of feature matching with block-based registration process on 9-divided image and pre-processing technique based on adaptive histogram equalization. The proposed method showed better performance than without our proposed technique on visual inspection and I-RMSE. This study can be used for automatic image registration between various images acquired from different sensors.

Keywords

1. Introduction

Image registration is the process of aligning objective image geometrically to the reference image with the same scene taken from different orientations or different sensors. Image registration is important on the image processing because its result can be used for various fields of image processing such as change detection, time-series analysis, image fusion, pansharpening, image mosaic, etc.(Bentoutou et al., 2005).

In the previous researches, the image registration method is generally classified into two types: the area-based method and the feature-based method. The area-based approach extracts small windows from the input images and quantifies the similarity between two images using with the pixel value or statistical value to apply the image matching process. There are several previous researches using various statistics in the area-based approach, for example, cross-correlation (Li and Zhou, 1995; Li, 1995), sum of squared differences (Irani and Anandan, 1998), and mutual information (Chen et al., 2003). On the other hand, the feature based method detects feature points such as corner and blob in two images. Then, it generates feature descriptors and measures the similarity of descriptors to calculate the transformation matrix between two images(Zheng, 1993; Kim et al., 2007; Kim, 2015; Wu et al., 2015; Zotiva and Flusser, 2003). These two approaches have a problem of mismatching when input images have high noise levels or image properties are different. Accordingly, various approaches to solve these problems have been conducted. Hong and Zhang (2007) proposed a method to improve image registration performance by combining an area-based method with normalized cross-correlation and a feature-based wavelet method. SIFT (Lowe, 1999) is known as shown a good performance in the matching process (Oh and Lee, 2011; Nag, 2017).

Recently, several methods of modifying the SIFT algorithm were proposed, such as ASIFT (Morel and Yu, 2009), PCA-SIFT (Ke and Sukthankar, 2004), and there were other methods such as SURF (Bay et al., 2008) and modified SURF descriptor (Bouchiha and Besbes, 2013). On the other hand, multi-modal studies, which is matching method for the images from different types of sensors, matching images obtained from different types of sensors, have been conducted and used them for image registration. Byun et al. (2017) proposed an image registration method between EO images and synthetic aperture radar (SAR) images using mutual information. Also, image registration using a simple pre-processing technique was proposed for the three types of IR images (LWIR, MWIR, SWIR) by Kim (2017).

KOMPSAT-3, 3A, 5 satellites are currently operating in Korea, and the target images of this study are KOMPSAT-3AEO images and IR images. It provides multi-spectral(EO), MWIR, and panchromatic images and the resolution of each image is 2.2 m, 5.5 m, 0.55 m, respectively (Lee et al., 2019). Because of each detector position in the platform is different, the geometry of each image is not coincide with each other, so image registration process is necessary. Previous image registration methods do not work well because EO images and IR images have different spectral characteristics. Several researches using satellite image from KOMPSAT series except 3A have been conducted (Lee and Oh, 2020; Oh and Lee, 2019; Seo and Eo, 2019). Since KOMPSAT-3A satellite has recently started its mission among the KOMPSAT series currently in operation, few studies on image registration between the EO and IR images using the images obtained in KOMPSAT-3A have been conducted.

The purpose of this research is to propose an automatic image registration between the EO and IR images of KOMPSAT-3A satellite using the feature-based method. Our approach attempted to transform the two images with different spectral properties into similar shapes to enhance the performance of feature matching algorithms(Kim, 2017). In addition, we also proposed the block-based approach to improve the accuracy of image registration for a wide area and analyzed the results ofthe block-based method through the experiments. The key points of this research are simple preprocessing, which is applied for the EO and IR images, and block-based method to improve the accuracy of image registration.

Chapter 2 describes the detailed image registration method of the research, and chapter 3 summarizes the experimental results using actual satellite images. Based on the experiments, chapter 4 summarizes the results of the research.

2. Image Registration

The images used in this study are EO images in the visible region and IR images (Fig. 1). EO image is composed of reflection information in a visible light wavelength, and IR image is composed of temperature information. Since the positions of each sensor on the platform are different, their actual ground location can be different. It can occur to align images at the same area which are taken at different time. However, due to differences ofthe image formation process, it is difficult to obtain good registration results for the different sensorimages by simply applying the feature matching algorithms (Kim, 2017). To solve the above problem, automatic image registration between EO and IRimage was performed as follows: ① image preprocessing, ② block division, ③ feature point extraction and feature matching, ④ removing mismatching and image registration. Details of each process are described in the relevant sections.

OGCSBN_2020_v36n4_545_f0001.png 이미지

Fig. 1. The spectrum of visible and IR wavelength (Courtesy of InfraTec).

1) Image preprocessing

In this step, two images with different spectral characteristics are preprocessed to make similar characteristics of images. Dark pixels are the object with high temperature due to low reflectivity, and light pixels are the object with low temperature due to high reflectivity (Wu et al., 2015). Thus, the phenomenon of contrastreversalmay occurin a part ofthe EO image and the IR image. Contrast reversal decreases the performance of feature matching algorithm due to changing the direction of the gradient. In order to overcome above problem, a simple preprocessing method for the input images was used for feature matching between EO and IR images. Preprocessing was performed as the following equation (Eq. (1)) to maintain local contrast and suppress noise in images over a wide area, such as Kim (2017).

Ipp = abs (ahisteq (Iall) – median (Iall))       (1)

Where Iall means whole image, abs, ahisteq, median means absolute value, adaptive histogram equalization (Pizer et al., 1986) and median value respectively. Adaptive histogram equalization was applied to the images so that the two images had similar brightness values as a whole, and that the contrast wasreversed by taking an absolute distance from the median intensity value. Using this preprocessing, the area where contrast reversal occurred was processed (Fig. 2).

OGCSBN_2020_v36n4_545_f0002.png 이미지

Fig. 2. Correction of contrast reversal using pre-processing (Up : EO, Down : IR).

2) Block division

When the entire image is realigned by a single transformation matrix, it is difficult to match exactly for the whole image because of the height difference among matching areas and non-uniformly distributed matching points. Fig. 3 shows the result of image registration with manually selected ground control points. Ground control points are uniformly distributed in the whole image and the image was warped using 2-D affine transformation.

OGCSBN_2020_v36n4_545_f0003.png 이미지

Fig. 3. Dislocated road of manual processing.

EO and IR images are shown as the form of a chessboard to check visibly the connection status of the straight road, by this process, 2-pixel dislocation can be found. For this reason, if the wide area image registration is performed, inaccurate results may be occurred.

Therefore, we proposed a dividing process by dividing the EO and IR images into nine sub-images. On dividing the image into sub-images, 50-pixel margin was placed at the boundary of the each divided image to remove a blank area due to difference of transformation matrix as shown in Fig. 4. The method of generating each sub-image is as Eq. (2).

OGCSBN_2020_v36n4_545_f0004.png 이미지

Fig. 4. Position of divided images from Eq. (2).

\(\begin{array}{l} \mathrm{I}_{\left(b_{x}, b_{y}\right)}=I&\left(\max \left(1,\left(b_{y}-1\right) * \frac{n_{\text {column }}}{3}-50\right)\right. \\ &\min \left(b_{y} * \frac{n_{\text {column }}}{3}-50, n_{\text {column }}\right) \\ &\max \left(1,\left(b_{x}-1\right) * \frac{n_{\text {row }}}{3}-50\right) \\ &\min \left(b_{x} * \frac{n_{\text {row }}}{3}-50, n_{\text {row }}\right) \end{array}\)       (2)

where ncolumn, nrow means the number of horizontal and vertical pixels of the input image and bx, by∈{1, 2, 3}.

3) Feature extraction and matching

Since this study is a block-based method, image registration is possible for the entire image only if true matching pairs exist in all sub-images. As a result of comparing the three methods of SIFT (Lowe, 2004), SURF (Bay et al., 2008), and BRISK (Leutenegger et al., 2011), true matching pairs were produced stably fromall blocks when we used SIFTalgorithm. Because the EO image and the IR image have different resolution, SIFT algorithm known for scale invariant has been stably produced the true matching pair in all blocks (Mistry and Banarjee, 2017).

The SIFT algorithm is used for generating descriptors by extracting feature points that are invariant to scale and rotation of an image. It shows the robustness of scale change, illumination change, translation,rotation, and some occlusion. The sequence of SIFT algorithm consists offoursteps: ① Scale-spaceExtremaDetection, ② Keypoint Localization, ③ OrientationAssignment, and ④ Keypoint Descriptor.In the process ①, extreme points are detected from DoG (Difference of Gaussian) asthe candidates offeature point afterGaussian pyramid calculation. In the process ②, incorrect candidates are excluded, and the orientation of the detected feature points is assigned in the step ③. If the positions of the feature points determined through the previous step, 128-dimensional SIFT feature point descriptors are generated in the step ④. Comparing the feature descriptor of one image with that of the other image, matching pairs are generated. In this research, we use the spectral angle instead of the Euclidean distance when comparing vector values, because Euclidean distance increases in computational complexity as the vector dimension increases.

4) Removal of mismatched pair and image registration

Mismatched pairs exist among the matching pairs resulting from the above process, and it is necessary to remove them to calculate more accurate transformation matrix. Mismatched pairs can be removed by RANSAC, as the outlier removal method. RANSAC algorithm is a prediction method that repeatedly finds the optimal solution by randomly sampling the minimum data to determine the coefficients of model among all data. In this study, mismatched pairs were more robustly removed using MSAC (Torr et al., 2000), the variation of RANSAC algorithm. The positions of feature points in the two images are input variables, 1000 iterations and 2-pixel of matching threshold are set to remove mismatching pairs.

There are forms oftranslation,rotation,scale, affine, and projective for the transformation of 2D image. In this research, it was assumed that the number of matching pairs existed sufficiently, and since the position difference of the sensors on the platform was not large, image matching was performed by calculating the affine transformation matrix between the two images as the following equation (Eq. (3)).

\(\left[\begin{array}{l} x^{\prime} \\ y^{\prime} \\ 1 \end{array}\right]=\left[\begin{array}{lll} a & b & t_{x} \\ c & d & t_{x} \\ 0 & 0 & 1 \end{array}\right]\left[\begin{array}{l} x \\ y \\ 1 \end{array}\right]\)       (3)

Two-dimensional affine transformation matrix is represented using homogeneous coordinate system. Where a, b, c and d are coefficients including scaling, rotation, shearing, tx and ty express parallel movement of horizontal and vertical direction. When three or more matching pairs are detected in order to calculate six unknowns, coefficients of the affine transformation matrix can be calculated using least square method.

3. Experiments

1) Experiment Image

The images used in the experiment are the EO and IR satellite images of KOMPSAT-3A, the Greater Hobby Area in Houston, USA. The EO image has spatial resolution of 2.2 m and the IR images of 5.5 m. The number of pixels of the EO image is 5,850 × 6,090, and the number of pixels of the IR image is 1,840 × 1,810. IR image was transformed with crop and 30 degree rotation to test proposed method on experiments. Although the experiments were performed using satellite images with original resolution, printed IR images were downsampled to 33 m resolution for the security reason.

Since the EO image contains the reflection information of the visible region of the object and the IR image containstemperature information, the image properties are different, the preprocessing was performed so that the two images have similar characteristics. The preprocessed images are divided into 9 blocks having 100-pixel overlapped area at 1/3 and 2/3 of whole image (Fig. 5).

OGCSBN_2020_v36n4_545_f0005.png 이미지

Fig. 5. Division result of preprocessed satellite images (Left : EO, Right : IR).

2) Feature point extraction and matching pair connection

For each block of the input image, feature points were extracted using the SIFT algorithm and feature were matched. The most of feature points extracted from one block is 13,279, which is about 2.8% of the pixels. The matching threshold, which indicates the ratio when finding similarfeature points, was 0.75. The matching pairs contain incorrect matchingsin position because the vector values of the feature descriptors are only compared. Thus, the pixel locations of the matching pair in the two images were assumed 2- dimension affine transform and compared before and after applying MSAC (M-estimator SAmple and Consensus) algorithm. Fig. 6 is the matching result of the (2, 2) center block of whole image and Fig. 7 is the matching result of the (1, 1) edge block of whole image. There are several mismatching pairs before MSAC, but mismatching pairs can be removed after MSAC. Table 1 shows the number of feature points detected on each block and the number of matching pairs before and after MSAC.

OGCSBN_2020_v36n4_545_f0006.png 이미지

Fig. 6. Matching results on block (2,2) (Upper : without MSAC, Lower : with MSAC).

OGCSBN_2020_v36n4_545_f0007.png 이미지

Fig. 7. Matching results on block (1,1) (Upper : without MSAC, Lower : with MSAC).

Table 1. Result of SIFT feature and MSAC algorithm

OGCSBN_2020_v36n4_545_t0001.png 이미지

3) Image registration and accuracy comparison

Image registration was finally performed using the matching pair obtained as a result ofthe previoussteps. By calculating the affine transformationmatrix for each block, the IR image was warped to align to the EO image.I-RMSE (InverseRoot Mean Square Error) was calculated to analyze quantitatively the accuracy of image registration (Table 1).

The accuracy using by the proposed block-based method was compared with the result produced by manually selected ground control points in Fig. 3 because image matching with SIFT algorithm didn’t work due to few detected feature points andmismatched pairs.I-RMSE value of Fig. 3 is about 1.4983 pixel and I-RMSE value of our block-based method is about 0.6389 pixel.

While the affine transformation model cannotreflect some factorssuch asregional altitude difference when image registration is performed on large area asshown in Fig. 3, the proposed block-based algorithm performed on relatively small areas and produced small I-RMSE value.

On the other hand, there is a methods for visually analyzing the results of image registration (Liu and Seipel, 2015; Kim, 2017). Result images are shown as the form of a chessboard for visual inspection if the roads are properly connected or not. By comparing Fig. 8 with Fig. 3, the proposed method hasrepresented more accurate registration result by visual analyzing.

OGCSBN_2020_v36n4_545_f0008.png 이미지

Fig. 8. Result of our image registration algorithm.

4. Conclusion

This paper is a research on the method of automatic image registration of the KOMPSAT-3A EO and IR images. We proposed a feature based method which produced similar properties for the two input images through the preprocessing and the division of 9 subimages for the purpose of applying SIFT algorithm on them.

Because matching points were uniformly distributed on the images, image registration could be performed accurately by the block-basedmethod.The experiments showed performance improvement ofimage registration using with lower I-RMSE value and visual inspection. The proposed approach on this paper could be expanded to utilized for image registration between not only satellite EO and IR images but also among other heterogeneous images.

On the other hand, if matching points are not detected sufficiently in a block or both blocks of the same number do not have common area, the proposed method may have a problem to complete registration as a whole because the block would not be matched accurately. Further research should be continued to improve the versatility ofthe approach by suggesting a solution of the above problem.

References

  1. Bay, H., A. Ess, T. Tuytelaars, and L. Van Gool, 2008. Speeded Up Robust Features (SURF), Computer Vision and Image Understanding, 110(3): 346-359. https://doi.org/10.1016/j.cviu.2007.09.014
  2. Bentoutou, Y., N. Taleb, K. Kpalma, and J. Ronsin, 2005. An Automatic Image Registration for Applications in Remote Sensing, IEEE Transactions on Geoscience and Remote Sensing, 43(9): 2127-2137. https://doi.org/10.1109/TGRS.2005.853187
  3. Bouchiha, R. and K. Besbes, 2013. Automatic Remotesensing Image Registration using SURF, International Journal of Computer Theory and Engineering, 5(1): 88-92. https://doi.org/10.7763/IJCTE.2013.V5.653
  4. Byun, Y., J. Choi, and Y. Han, 2013. An Area-Based Image Fusion Scheme for the Integration of SAR and Optical Satellite Imagery, IEEE Journal of Selected Topics in Applied Earth Observation and Remote Sensing, 6(5): 2212-2220. https://doi.org/10.1109/JSTARS.2013.2272773
  5. Chen, H.M., M.K. Arora, and P.K. Varshney, 2003. Mutual information-based image registration for remote sensing data, International Journal of Remote Sensing, 24(18): 3701-3706. https://doi.org/10.1080/0143116031000117047
  6. Hong, G. and Y. Zhang, 2007. Combination of Featurebased and Area-based Image Registration Technique for High Resolution Remote Sensing Image, Proc. of IEEE International Geoscience and Remote Sensing Symposium, Barcelona, Spain, Jul. 23-27, pp. 377-380.
  7. Irani, M. and P. Anandan, 1998. Robust Multi-Sensor Image Alignment, Proc. of the 6th IEEE International Conference on Computer Vision, Bombay, India, Jan. 4-7, pp. 959-966.
  8. Ke, Y. and R. Sukthankar, 2004. PCA-SIFT: A more distinctive representation for local image descriptors, Proc. of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, D.C., USA, Jun. 29-Jul. 1, vol. 2, pp. 506-513.
  9. Kim, D.S., 2017. Automatic Registration between Multiple IR Images Using Simple Pre-processing Method and Modified Local Features Extraction Algorithm, Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography, 35(6): 485-494 (in Korean with English abstract). https://doi.org/10.7848/ksgpc.2017.35.6.485
  10. Kim, D.S., Y.I. Kim, and Y.D. Eo, 2007. A Study on Automatic Co-registration and Band Selection of Hyperion Hyperspectral Images for Change Detection, Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography, 25(5): 383-392 (in Korean with English abstract).
  11. Kim, K.S., 2015. Survey on Registration Techniques of Visible and Infrared Images, IT CoNvergence PRActice (INPRA), 3(2): 25-35.
  12. Lee, C. and J. Oh, 2020. Rigorous Co-Registration of KOMPSAT-3 Multispectral and Panchromatic Images for Pan-Sharpening Image Fusion, Sensors, 20(7): 2100. https://doi.org/10.3390/s20072100
  13. Lee, K.J., K.Y. Oh, T.B. Chae, and W.J. Lee, 2019. Research Trend in KOMPSAT Series, Korean Journal of Remote Sensing, 35(6-4): 1313-1318 (in Korean with English abstract). https://doi.org/10.7780/kjrs.2019.35.6.4.1
  14. Leutenegger, S., M. Chli, and R. Siegwart, 2011. BRISK: Binary Robust Invariant Scalable Keypoints, Proc. of the IEEE International Conference on Computer Vision, Barcelona, Spain, Nov. 6-13, pp. 2548-2555.
  15. Li, H. and Y.T. Zhou, 1995. Automatic EO/IR Sensor Image Registration, Proc. of the IEEE International Conference on Image Processing, Washington, D.C., USA, Oct. 23-26, vol. 3, pp. 240-243.
  16. Liu, F. and S. Seipel, 2015. Infrared-visible image registration for augmented reality-based thermographic building diagnostics, Visualization in Engineering, 3(16): 1-15. https://doi.org/10.1186/s40327-014-0014-y
  17. Lowe, D.G., 1999. Object Recognition from Local Scale-invariant Features, Proc. of the IEEE International Conference on Computer Vision, Corfu, Greece, Sep. 20-25, vol. 2, pp. 1150-1157.
  18. Mistry, D. and A. Banerjee, 2017. Comparison of feature detection and matching approaches: SIFT and SURF, Global-Research and Development Journal for Engineering, 2(4): 7-13.
  19. Morel, J.M. and G. Yu, 2009. ASIFT: A new framework for fully affine invariant image comparison, SIAM Journal on Imaging Sciences, 2(2): 438-469. https://doi.org/10.1137/080732730
  20. Nag, S., 2017. Image Registration Techniques: A Survey, arXiv preprint arXiv:1712.07540.
  21. Oh, J.H. and C.N. Lee, 2019. Conjugate Point Extraction for High-Resolution Stereo Images Orientation, Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography, 37(2): 55-62. https://doi.org/10.7848/ksgpc.2019.37.2.55
  22. Oh, J. and H. Lee, 2011. A Performance Analysis of the SIFT Matching on Simulated Geospatial Image Differences, Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography, 29(5): 449-457. https://doi.org/10.7848/ksgpc.2011.29.5.449
  23. Pizer, S.M., E.P. Amburn, J.D. Austin, R. Cromartie, A. Geselowitz, T. Greer, and K. Zuideveld, 1987. Adaptive histogram equalization and its variations, Computer Vision, Graphics, and Image Processing, 39(3): 355-368. https://doi.org/10.1016/s0734-189x(87)80186-x
  24. Seo, D.K. and Y.D. Eo, 2019. Local-based Iterative Histogram Matching for Relative Radiometric Normalization, Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography, 37(5): 323-330. https://doi.org/10.7848/ksgpc.2019.37.5.323
  25. Torr, P.H. and A. Zisserman, 2000. MLESAC: A new robust estimator with application to estimating image geometry, Computer Vision and Image Understanding, 78(1): 138-156. https://doi.org/10.1006/cviu.1999.0832
  26. Wu, F., B. Wang, X. Yi, M. Li, J. Hao, H. Qin, and H. Zhou, 2015. Visible and Infrared Image Registration based on Visual Salient Features, Journal of Electronic Imaging, 24(5): 053027. https://doi.org/10.1117/1.JEI.24.5.053027
  27. Zheng, Q., 1993. A Computational Vision Approach to Image Registration, IEEE Transactions on Image Processing, 2(3): 311-326. https://doi.org/10.1109/83.236535
  28. Zotiva, B. and J. Flusser, 2003. Image Registration Method: A Survey, Image and Vision Computing, 21(11): 977-1000. https://doi.org/10.1016/S0262-8856(03)00137-9

Cited by

  1. 최근 다목적실용위성 시리즈 활용 현황 vol.36, pp.6, 2020, https://doi.org/10.7780/kjrs.2020.36.6.2.1
  2. 다목적실용위성 영상 활용 vol.37, pp.6, 2021, https://doi.org/10.7780/kjrs.2021.37.6.3.1