DOI QR코드

DOI QR Code

Three-Dimensional Automatic Target Recognition System Based on Optical Integral Imaging Reconstruction

  • Lee, Min-Chul (Department of Computer Science and Electronics, Kyushu Institute of Technology) ;
  • Inoue, Kotaro (Department of Computer Science and Electronics, Kyushu Institute of Technology) ;
  • Cho, Myungjin (Department of Electrical, Electronic, and Control Engineering, IITC, Hankyong National University)
  • Received : 2015.12.15
  • Accepted : 2016.01.29
  • Published : 2016.03.31

Abstract

In this paper, we present a three-dimensional (3-D) automatic target recognition system based on optical integral imaging reconstruction. In integral imaging, elemental images of the reference and target 3-D objects are obtained through a lenslet array or a camera array. Then, reconstructed 3-D images at various reconstruction depths can be optically generated on the output plane by back-projecting these elemental images onto a display panel. 3-D automatic target recognition can be implemented using computational integral imaging reconstruction and digital nonlinear correlation filters. However, these methods require non-trivial computation time for reconstruction and recognition. Instead, we implement 3-D automatic target recognition using optical cross-correlation between the reconstructed 3-D reference and target images at the same reconstruction depth. Our method depends on an all-optical structure to realize a real-time 3-D automatic target recognition system. In addition, we use a nonlinear correlation filter to improve recognition performance. To prove our proposed method, we carry out the optical experiments and report recognition results.

Keywords

I. INTRODUCTION

Object recognition’s many applications include its use in autonomous vehicles, artificial intelligence, and defense. Within this field, three-dimensional (3-D) automatic target recognition using integral imaging techniques has been an important topic of research [1-9].

Integral imaging, first proposed by G. Lippmann in 1908 [10], employees lenslet or camera arrays to capture and display 3-D images. Each lenslet provides a different perspective on the subject, generating an elemental image for capture. The resulting array of 2-D images, illuminated from behind and projected through another lenslet array, displays a 3-D reconstruction. The technique provides full parallax, continuous viewing points, and relatively high depth resolution without the need for special glasses [10-15], all desirable characteristics.

The mechanism for image reconstruction, itself, can be used to implement forms of auto target recognition. For example, in [4], the authors propose a method for locating the position of target objects in 3-D space. They use the fact that computational integral imaging reconstruction (CIIR) can output a range of 3-D reconstructions, each at a different reconstruction depth [14]. Then, computers apply advanced correlation filters to the range of reconstructed images created by CIIR and complete the automatic target recognition task.

In practice, such methods are difficult to apply in real-time applications. The computation time and resources required for 3-D reconstruction limit its usefulness [6]. To address the limitation imposed by the computational cost of 3-D reconstruction, an all-optical structure for 3-D automatic target recognition offers an alternative.

This paper proposes such an all-optical integral imaging approach. In our method, we optically reconstruct 3-D reference and target images. We then perform automatic target recognition, applying a nonlinear correlation filter to the reference and target images. To assess this proposed method, we discuss optical experiments performed and present recognition results obtained in these experiments.

 

II. REVIEW OF CIIR-TYPE 3-D OBJECT CORRELATION METHOD

Fig. 1 illustrates the principle of the conventional CIIR-type 3-D object correlation method. The process is separated into two main stages: pickup and recognition. In the pickup stage, we assume that zr, the distance between the reference object and the lenslet array, is known, and that the distance between the target and the lenslet array, zo, is unknown or arbitrary. For both reference and target objects, elemental images created by the lenslet array are recorded by the pickup device. In the recognition stage, a CIIR technique reconstructs images of the reference object at known reconstruction depths and the image of the target object at a range of reconstruction depths [5-9]. CIIR can be implemented by computational simulation of ray optics.

Fig. 1.CIIR-type 3-D object correlation method. (a) Storing of reconstructed 3-D image for reference, (b) reconstruction of target images at various depths, followed by cross-correlation with reference.

Fig. 1(a) depicts the reference object method, and Fig. 1(b) the steps in the target object method. Fig. 1(b) also depicts the automatic target recognition step, implemented using cross-correlation between a single, reference 3-D image and the various target 3-D images.

When recognition is implemented with this method, a high correlation peak can be obtained only at z = zr. Thus, this method can extract the depth information of a reconstructed target image that matches the depth of the reconstructed reference image.

While the method works, CIIR is very slow, computationally, and requires substantial computing resources. Therefore, if real-time processing is required, an alternative method may be superior.

 

III. PROPOSED 3-D CORRELATION METHOD BASED ON OPTICAL INTEGRAL IMAGING RECONSTRUCTION

Fig. 2 illustrates our proposed 3-D object correlation method. In the conventional CIIR-type 3-D object correlation method presented above, 3-D images are reconstructed by computer simulation. In our proposed method, an optical display system reconstructs the 3-D images. The all-optical structure allows us to implement automatic target recognition in real-time.

Fig. 2.Structure of the proposed 3-D object correlation method. (a) Storing of reconstructed 3-D image for reference object, (b) reconstruction of target object at all depth planes.

Our system has three main parts: pickup, display, and recognition. For pickup, let us assume that the reference object has known position r(xr, yr, zr) at a distance of zr from the lenslet array and that the target object at t(xo, yo, zo) stands at an unknown (or arbitrary) distance, zo, from the lenslet array. Elemental images for reference and target objects are obtained by the first image sensor (Sensor 1) by way of the lenslet array.

In the display part of the process, 3-D image reconstructtion begins when these elemental images are reproduced on a liquid crystal display (LCD). Light illuminating the LCD passes through the lenslet array, generating a 3-D optical image for capture by the second image sensor (Sensor 2) at the speed of light. This design choice produces a system that is much faster than those used to implement the conventional CIIR-type correlation method. The 3-D, optically-created images have relatively high 3-D resolution (lateral and depth) since all elemental images are superimposed by optical magnification.

For the reference object, Sensor 2 records its reconstructed 3-D image at the known distance zr. By locating this sensor at the focal plane, we can obtain a high-quality image. We call the reconstructed 3-D image of the reference object at known zr-plane R(xr, yr, zr). This image is stored in the computer for use as a reference image.

To record the reconstructed image of the target object, Sensor 2 is positioned at zr, the identical position from which the reference object was recorded, as shown in Fig. 2(a). Recall that, in the pickup stage, the target object was located at the zo-plane. Images reconstructed at the other various reconstruction depths are noticeably out-of-focus or blurred. Only when zo = zr is the best, clearly reconstructed 3-D image obtained. This is a main feature of our 3-D object correlation method. Let T(xo, yo, zr) be the reconstructed 3-D image of the target object as captured by Sensor 2 at zr.

Recognition proceeds by means of a correlation process comparing T(xo, yo, zr) and R(xr, yr, zr) , the 3-D target and reference images reconstructed at a distance of zr from the lenslet array. Correlation results are obtained by means of the following kth-law nonlinear correlation filter

where are the Fourier transform and phase of the reconstructed reference image, R, at the distance zr, are the Fourier transform and phase of the reconstructed target image, T, at the distance zr, and k is the nonlinearity parameter for the correlation filter. When k = 0, Eq. (1) becomes a phase-only filter (POF). When k = 1, Eq. (1) becomes a linear correlation filter.

 

IV. EXPERIMENTAL RESULTS

We carried out optical experiments to assess our proposed 3-D object recognition method. A ‘die’, shown in Fig. 3(a), was used as the reference and target object in this experiment. Its size was 2 cm × 2 cm × 2 cm (W × H × D). When serving as the reference object, the ‘die’ was positioned at (xr, yr, zr) = (0 cm, 0 cm, 6 cm). The experimental setup is depicted in Fig. 3(b). Here, we used a 30 × 30 array of lenslets. Each lenslet measured 1.08 mm × 1.08 mm and was square in shape. The lenslet focal length was 3 mm. Total resolution of the elemental images was 900 × 900 pixels (H × V). Each elemental image had 30 × 30 pixels as the pixel pitch of the LCD panel shown in Fig. 3(c) was 36 μm.

Fig. 3.(a) 3-D object, (b) optical pickup setup, and (c) optical display setup.

A set of recorded elemental images is shown in Fig. 4(a). We used a paper screen and a digital camera to form Sensor 2. The paper screen was placed z = 6 cm from the lenslet array. The digital camera captured images from the paper screen. An example of the images captured in this manner is shown in Fig. 4(b). This reconstructed 3-D reference image was taken at z = 6 cm.

Fig. 4.Reference 3-D object. (a) Elemental images, (b) reconstructed 3-D image at z = 6 cm.

When used as a target object, the ‘die’ was shifted along the z-axis or rotated by an arbitrary angle on the y-axis. Elemental images for each case were recorded using the optical pickup apparatus shown in Fig. 3(b). Then, with a set of elemental images for each case, reconstructed 3-D images at various reconstruction depths could be obtained using the optical display system shown in Fig. 3(c). Examples of reconstructed 3-D images at different reconstruction depths are shown in Fig. 5(a). The figure clearly shows that the 3-D image reconstructed at z = 6 cm exhibits the sharpest focus. Fig. 5(b) provides examples of the impact of target rotation angle on the reconstructed 3-D image. The figures demonstrate that reconstructed 3-D images of target objects depend on object distance and rotation angle.

Fig. 5.Reconstructed 3-D images with (a) various reconstruction depths, (b) various rotation angles.

Finally, we present the cross-correlation between the reference image and the reconstructed 3-D images of the target object at various depths and rotation angles. We computed correlation using the kth-law nonlinear correlation filter described by Eq. (1). Correlation results, with k = 0.5, k = 0.7, and k = 1, are shown in Fig. 6. The highest peaks (Fig. 6(a), (c), and (e)) correspond to z = 6 cm. In this case, the reconstruction plane matches the z-plane in which the target object was located in the pickup phase. Lower peaks were observed when the reconstruction plane and the z-plane during pickup did not match. Fig. 6(b), (d), and (f) display correlation results for z = 10 cm.

Fig. 6.kth-law nonlinear correlation results at z = 6 cm when (a) k = 1, (c) k = 0.7, (e) k = 0.5 at z = 10 cm when (b) k = 1, (d) k = 0.7, (f) k = 0.5.

We repeated the analysis for the other reconstructions of the target object image shown in Fig. 5. Applying the correlation test to the reference and each reconstructed target image produced the results shown in Fig. 7. Thus, we conclude that our proposed method is robust when the reconstruction plane shifts along the z-axis and when the target object is rotated.

Fig. 7.Experimental results of nonlinear correlation under (a) various reconstruction depths, (b) various rotation angles.

 

V. CONCLUSIONS

In this paper, we presented a 3-D automatic target recognition system based on integral imaging and using an all-optical structure to provide real-time processing.

Fig. 8 shows a practical application of our method. The two main processing stages in Fig. 2 may be combined into one imaging system using an auto-collimating screen [16]. In this system, the reconstructed 3-D images are generated at light speed and displayed on a spatial light modulator (SLM) to convert these images into coherent images. Then, a joint transform correlator (JTC) could serve as a correlation filter. This design may produce a real-time, compact 3-D object recognition system.

Fig. 8.Practical system concept of real-time 3-D automatic target recognition using our proposed method.

In future work, we will investigate practical real-time systems that use our proposed method.

References

  1. O. Matoba, E. Tajahuerce, and B. Javidi, “Real-time three-dimensional object recognition with multiple perspectives imaging,” Applied Optics, vol. 40, no. 20, pp. 3318-3325, 2001. https://doi.org/10.1364/AO.40.003318
  2. Y. Frauel and B. Javidi, “Digital three-dimensional image correlation by use of computer-reconstructed integral imaging,” Applied Optics, vol. 41, no. 26, pp. 5488-5496, 2002. https://doi.org/10.1364/AO.41.005488
  3. J. H. Park, J. Kim, and B. Lee, “Three-dimensional optical correlator using a sub-image array,” Optics Express, vol. 13, no. 13, pp. 5116-5126, 2005. https://doi.org/10.1364/OPEX.13.005116
  4. J. S. Park, D. C. Hwang, D. H. Shin, and E. S. Kim, “Resolution-enhanced three-dimensional image correlator using computationally reconstructed integral images,” Optics Communications, vol. 276, no. 1, pp. 72-79, 2007. https://doi.org/10.1016/j.optcom.2007.04.007
  5. Y. Piao, M. Zhang, J. J. Lee, D. Shin, and B. G. Lee, “Orthoscopic integral imaging display by use of the computational method based on lenslet model,” Optics and Lasers in Engineering, vol. 52, pp. 184-188, 2014. https://doi.org/10.1016/j.optlaseng.2013.06.012
  6. D. H. Shin and H. Yoo, “Image quality enhancement in 3D computational integral imaging by use of interpolation methods,” Optics Express, vol. 15, no. 19, pp. 12039-12049, 2007. https://doi.org/10.1364/OE.15.012039
  7. H. Yoo and D. Shin, "Fast computational integral imaging reconstruction method using a fractional delay filter," Japanese Journal of Applied Physics, vol. 49, article ID. 022503, 2010.
  8. J. Y. Jang, M. Cho, and E. S. Kim, "3D image reconstruction with controllable spatial filtering based on correlation of multiple periodic functions in computational integral imaging," Chinese Optics Letters. vol. 13, no. 3, article ID. 031101, 2015.
  9. J. Y. Jang and M. Cho, “Fast computational integral imaging reconstruction by combined use of spatial filtering and rearrangement of elemental image pixels,” Optics and Lasers in Engineering, vol. 75, pp. 57-62, 2015. https://doi.org/10.1016/j.optlaseng.2015.06.007
  10. G. Lippmann, “La photographie integrale,” Académie des Sciences. vol. 146, pp. 446-451, 1908.
  11. F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Three-dimensional video system based on integral photography,” Optical Engineering, vol. 38, no. 6, pp. 1072-1077, 1999. https://doi.org/10.1117/1.602152
  12. J. S. Jang and B. Javidi, “Improved viewing resolution of three-dimensional integral imaging by use of nonstationary micro-optics,” Optics Letters, vol. 27, no. 5, pp. 324-326, 2002. https://doi.org/10.1364/OL.27.000324
  13. B. Lee, S. Y. Jung, S. W. Min, and J. H. Park, “Three-dimensional display by use of integral photography with dynamically variable image planes,” Optics Letters, vol. 26, no. 19, pp. 1481-1482, 2001. https://doi.org/10.1364/OL.26.001481
  14. M. Martinez-Corral, B. Javidi, R. Martinez-Cuenca, and G. Saavedra, “Multifacet structure of observed reconstructed integral images,” Journal of the Optical Society of America A, vol. 22, no. 4, pp. 597-603, 2005. https://doi.org/10.1364/JOSAA.22.000597
  15. M. Cho, A. Mahalanobis, and B. Javidi, “3D passive photon counting automatic target recognition using advanced correlation filters,” Optics Letters, vol. 36, no. 6, pp. 861-863, 2011. https://doi.org/10.1364/OL.36.000861
  16. N. Davies, M. McCormick, and L. Yang, “Three-dimensional imaging system: a new development,” Applied Optics, vol. 27, no. 21, pp. 4520-4528, 1988. https://doi.org/10.1364/AO.27.004520

Cited by

  1. New-generation hybrid guidance system based on infrared and millimeter waves vol.33, pp.7, 2018, https://doi.org/10.1109/MAES.2018.170022
  2. 실시간 초음파 영상에서 노이즈 개선을 위한 GPU 기반의 필터 알고리즘 vol.19, pp.6, 2016, https://doi.org/10.9728/dcs.2018.19.6.1207
  3. Free-view Pixels of Elemental Image Rearrangement Technique (FPERT) vol.17, pp.1, 2016, https://doi.org/10.6109/jicce.2019.17.1.60