DOI QR코드

DOI QR Code

Fast Object Classification Using Texture and Color Information for Video Surveillance Applications

비디오 감시 응용을 위한 텍스쳐와 컬러 정보를 이용한 고속 물체 인식

  • Islam, Mohammad Khairul (Dept. of Information & Telecommunication Engineering, Korea Aerospace University) ;
  • Jahan, Farah (Dept. of Information & Telecommunication Engineering, Korea Aerospace University) ;
  • Min, Jae-Hong (Dept. of Information & Telecommunication Engineering, Korea Aerospace University) ;
  • Baek, Joong-Hwan (Dept. of Information & Telecommunication Engineering, Korea Aerospace University)
  • Received : 2011.01.31
  • Accepted : 2011.02.28
  • Published : 2011.02.28

Abstract

In this paper, we propose a fast object classification method based on texture and color information for video surveillance. We take the advantage of local patches by extracting SURF and color histogram from images. SURF gives intensity content information and color information strengthens distinctiveness by providing links to patch content. We achieve the advantages of fast computation of SURF as well as color cues of objects. We use Bag of Word models to generate global descriptors of a region of interest (ROI) or an image using the local features, and Na$\ddot{i}$ve Bayes model for classifying the global descriptor. In this paper, we also investigate discriminative descriptor named Scale Invariant Feature Transform (SIFT). Our experiment result for 4 classes of the objects shows 95.75% of classification rate.

본 논문에서는 텍스쳐와 컬러 정보를 기반으로 비디오 감시를 위한 빠른 물체 분류 방법을 제안한다. 영상들로부터 SURF와 색 히스토그램의 국부적 패치들을 추출하여 그들의 장점을 이용한다. SURF는 명암 내용 정보를 제공하고 색 정보는 패치에 대한 특이성을 증강시킨다. SURF의 빠른 계산뿐만 아니라 객체의 색 정보를 활용한다. 국부적 특징을 이용하여 관심 영역 혹은 영상의 전역적 서술자를 생성하기 위해 Bag of Word 모델을 이용하고, 전역적 서술자를 분류하기 위해 Na$\ddot{i}$ve Bayes 모델을 이용한다. 또한 본 논문에서는 판별적인 기술자인 SIFT도 성능 분석한다. 네 종류의 객체에 대한 실험결과 95.75%의 인식률을 보였다.

Keywords

References

  1. S. Belongie, J. Malik, and J. Puzicha, "Shape Matching and Object Recognition Using Shape Context," IEEE Transaction on Pattern Analysis and Machine Intelligence, vol. 24, no. 4, pp. 509-522, April 2002. https://doi.org/10.1109/34.993558
  2. S. Ullman, "High-level vision: Object recognition and visual recognition", MIT Press, 1996.
  3. D.G. Lowe, "Distinctive image features from scale-invariant keypoints," International Journal of Computer Vision, vol. 60, no. 2, pp. 91-110, 2004.
  4. T. Kadir, A. Zisserman, and M. Brady, "An affine invariant salient region detector," Proc. of European Conference on Computer Vision, Prague, Czech Republic, pp. 228-241, 2004.
  5. Zhi-Gang Fan, Jilin Li, Bo Wu, and Yadong Wu, "Local patterns constrained image histograms for image retrieval", IEEE International Conference on Image Processing (ICIP), San Diego, CA, pp. 941 - 944, December 12, 2008.
  6. J. Sivic and A. Zisserman, "Video google: A text retrieval approach to object matching in videos," IEEE International Conference on Computer Vision (ICCV), pp. 1470-1477, Oct. 2003.
  7. D. Nistier and H. Stewenius, "Scalable recognition with a vocabulary tree," IEEE Computer Vision and Pattern Recognition (CVPR), pp. 2161-2168, Jun. 2006.
  8. M. Brown and D.G. Lowe, "Invariant features from interest point groups", British Machine Vision Conference, pp. 656-665, 2002.
  9. P. Domingos, and M. Pazzani, "On the optimality of the simple Bayesian classifier under zero-one loss", Machine Learning, vol. 29, pp. 103-130, 1997. https://doi.org/10.1023/A:1007413511361
  10. P. Cheeseman, and J. Stutz, "Bayesian classification (AutoClass): Theory and results", Advances in knowledge discovery and data mining, AAAI Press, Menlo Park, CA, pp. 153-180, 1996.