• 제목/요약/키워드: Object-based Image Classification

검색결과 242건 처리시간 0.026초

Hardware Accelerated Design on Bag of Words Classification Algorithm

  • Lee, Chang-yong;Lee, Ji-yong;Lee, Yong-hwan
    • Journal of Platform Technology
    • /
    • 제6권4호
    • /
    • pp.26-33
    • /
    • 2018
  • In this paper, we propose an image retrieval algorithm for real-time processing and design it as hardware. The proposed method is based on the classification of BoWs(Bag of Words) algorithm and proposes an image search algorithm using bit stream. K-fold cross validation is used for the verification of the algorithm. Data is classified into seven classes, each class has seven images and a total of 49 images are tested. The test has two kinds of accuracy measurement and speed measurement. The accuracy of the image classification was 86.2% for the BoWs algorithm and 83.7% the proposed hardware-accelerated software implementation algorithm, and the BoWs algorithm was 2.5% higher. The image retrieval processing speed of BoWs is 7.89s and our algorithm is 1.55s. Our algorithm is 5.09 times faster than BoWs algorithm. The algorithm is largely divided into software and hardware parts. In the software structure, C-language is used. The Scale Invariant Feature Transform algorithm is used to extract feature points that are invariant to size and rotation from the image. Bit streams are generated from the extracted feature point. In the hardware architecture, the proposed image retrieval algorithm is written in Verilog HDL and designed and verified by FPGA and Design Compiler. The generated bit streams are stored, the clustering step is performed, and a searcher image databases or an input image databases are generated and matched. Using the proposed algorithm, we can improve convenience and satisfaction of the user in terms of speed if we search using database matching method which represents each object.

내용기반으로한 이미지 검색에서 이미지 객체들의 외형특징추출 (Feature Extraction of Shape of Image Objects in Content-based Image Retrieval)

  • 조준서
    • 정보처리학회논문지B
    • /
    • 제10B권7호
    • /
    • pp.823-828
    • /
    • 2003
  • 이 논문의 주요 목적은 내용을 기반으로 하는 이미지 검색에서 이미지 객체의 외형특징을 추출하는 방법을 제시하는 것이다. 대부분의 실질적인 객체들의 외형은 불규칙적이고, 이러한 객체를 수치화하기위한 일반적인 방법은 없다. 특히 전자 카타로그들은 상품들을 나타내는 많은 이미지를 포함하고 있다. 이 논문에서는 이미지 전체가 아닌 이미지내의 개별 객체들을 기반으로 특징을 추출하는 방법을 제시한다. 왜냐하면 제시된 방법은 한 이미지내에서 RLC lines을 사용하여 각 객체들의 외형을 기반으로하는 방법을 사용하기 때문이다. 실험결과는 일반적으로 가장 많이 사용하는 특징인 Texture와 비교를 했고 제시된 외형을 나타내는 변수들이 전자카타로그의 이미지 객체들을 뚜렷하게 나타냈고, 보다 정확하게 객체들을 분류하고 구별하였다.

위성영상을 이용한 개발과 미개발 지역의 구분을 위한 탐색적 방법 (Investigating Ways of Developed and Undeveloped Features from Satellite Images -Balancing Coastal Development and Preservation-)

  • 양병윤
    • 한국측량학회지
    • /
    • 제30권2호
    • /
    • pp.189-197
    • /
    • 2012
  • This research attempted to find possibilities of the practical use utilizing geospatial methods for the balanced promotion of sustainable coastal development and preservation through a case study of Jekyll Island, one of Georgia's barrier islands. In response, this research provided ways for practical use in sustainable development and preservation plans. First this research thoroughly investigated the 1996 master plan of Jekyll Island and tried to recalculate developed and undeveloped areas. Second, new estimations for developed areas were investigated through field survey. Third, this research proposed the use of the satellite images with different levels of spatial resolutions and tested different classification schemes to find possibilities for practical use. For these purposes, first, we classified developed and undeveloped features by manual digitization using an aerial photo image with 0.5m spatial resolution. Second, a Landsat 7 ETM+ and a QuickBird satellite images with mid- and high-levels of spatial resolutions were applied to identify developed and undeveloped areas using both the National Land Cover Data (NLCD) and the Coastal Change Analysis Program (CCAP) classification schemes. Also, GEOBIA (Geographic Object-Based Image Analysis) was conducted to accurately identify developed and undeveloped areas.

딥러닝 기술을 이용한 3차원 객체 추적 기술 리뷰 (A Review of 3D Object Tracking Methods Using Deep Learning)

  • 박한훈
    • 융합신호처리학회논문지
    • /
    • 제22권1호
    • /
    • pp.30-37
    • /
    • 2021
  • 카메라 영상을 이용한 3차원 객체 추적 기술은 증강현실 응용 분야를 위한 핵심 기술이다. 영상 분류, 객체 검출, 영상 분할과 같은 컴퓨터 비전 작업에서 CNN(Convolutional Neural Network)의 인상적인 성공에 자극 받아, 3D 객체 추적을 위한 최근의 연구는 딥러닝(deep learning)을 활용하는 데 초점을 맞추고 있다. 본 논문은 이러한 딥러닝을 활용한 3차원 객체 추적 방법들을 살펴본다. 딥러닝을 활용한 3차원 객체 추적을 위한 주요 방법들을 설명하고, 향후 연구 방향에 대해 논의한다.

Covariance-based Recognition Using Machine Learning Model

  • Osman, Hassab Elgawi
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송공학회 2009년도 IWAIT
    • /
    • pp.223-228
    • /
    • 2009
  • We propose an on-line machine learning approach for object recognition, where new images are continuously added and the recognition decision is made without delay. Random forest (RF) classifier has been extensively used as a generative model for classification and regression applications. We extend this technique for the task of building incremental component-based detector. First we employ object descriptor model based on bag of covariance matrices, to represent an object region then run our on-line RF learner to select object descriptors and to learn an object classifier. Experiments of the object recognition are provided to verify the effectiveness of the proposed approach. Results demonstrate that the propose model yields in object recognition performance comparable to the benchmark standard RF, AdaBoost, and SVM classifiers.

  • PDF

다중 센서를 사용한 주행 환경에서의 객체 검출 및 분류 방법 (A New Object Region Detection and Classification Method using Multiple Sensors on the Driving Environment)

  • 김정언;강행봉
    • 한국멀티미디어학회논문지
    • /
    • 제20권8호
    • /
    • pp.1271-1281
    • /
    • 2017
  • It is essential to collect and analyze target information around the vehicle for autonomous driving of the vehicle. Based on the analysis, environmental information such as location and direction should be analyzed in real time to control the vehicle. In particular, obstruction or cutting of objects in the image must be handled to provide accurate information about the vehicle environment and to facilitate safe operation. In this paper, we propose a method to simultaneously generate 2D and 3D bounding box proposals using LiDAR Edge generated by filtering LiDAR sensor information. We classify the classes of each proposal by connecting them with Region-based Fully-Covolutional Networks (R-FCN), which is an object classifier based on Deep Learning, which uses two-dimensional images as inputs. Each 3D box is rearranged by using the class label and the subcategory information of each class to finally complete the 3D bounding box corresponding to the object. Because 3D bounding boxes are created in 3D space, object information such as space coordinates and object size can be obtained at once, and 2D bounding boxes associated with 3D boxes do not have problems such as occlusion.

Classification of Objects using CNN-Based Vision and Lidar Fusion in Autonomous Vehicle Environment

  • G.komali ;A.Sri Nagesh
    • International Journal of Computer Science & Network Security
    • /
    • 제23권11호
    • /
    • pp.67-72
    • /
    • 2023
  • In the past decade, Autonomous Vehicle Systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. The fusion of light detection and ranging (LiDAR) and camera data in real-time is known to be a crucial process in many applications, such as in autonomous driving, industrial automation and robotics. Especially in the case of autonomous vehicles, the efficient fusion of data from these two types of sensors is important to enabling the depth of objects as well as the classification of objects at short and long distances. This paper presents classification of objects using CNN based vision and Light Detection and Ranging (LIDAR) fusion in autonomous vehicles in the environment. This method is based on convolutional neural network (CNN) and image up sampling theory. By creating a point cloud of LIDAR data up sampling and converting into pixel-level depth information, depth information is connected with Red Green Blue data and fed into a deep CNN. The proposed method can obtain informative feature representation for object classification in autonomous vehicle environment using the integrated vision and LIDAR data. This method is adopted to guarantee both object classification accuracy and minimal loss. Experimental results show the effectiveness and efficiency of presented approach for objects classification.

Efficient Object-based Image Retrieval Method using Color Features from Salient Regions

  • An, Jaehyun;Lee, Sang Hwa;Cho, Nam Ik
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제6권4호
    • /
    • pp.229-236
    • /
    • 2017
  • This paper presents an efficient object-based color image-retrieval algorithm that is suitable for the classification and retrieval of images from small to mid-scale datasets, such as images in PCs, tablets, phones, and cameras. The proposed method first finds salient regions by using regional feature vectors, and also finds several dominant colors in each region. Then, each salient region is partitioned into small sub-blocks, which are assigned 1 or 0 with respect to the number of pixels corresponding to a dominant color in the sub-block. This gives a binary map for the dominant color, and this process is repeated for the predefined number of dominant colors. Finally, we have several binary maps, each of which corresponds to a dominant color in a salient region. Hence, the binary maps represent the spatial distribution of the dominant colors in the salient region, and the union (OR operation) of the maps can describe the approximate shapes of salient objects. Also proposed in this paper is a matching method that uses these binary maps and which needs very few computations, because most operations are binary. Experiments on widely used color image databases show that the proposed method performs better than state-of-the-art and previous color-based methods.

Comparison of estimating vegetation index for outdoor free-range pig production using convolutional neural networks

  • Sang-Hyon OH;Hee-Mun Park;Jin-Hyun Park
    • Journal of Animal Science and Technology
    • /
    • 제65권6호
    • /
    • pp.1254-1269
    • /
    • 2023
  • This study aims to predict the change in corn share according to the grazing of 20 gestational sows in a mature corn field by taking images with a camera-equipped unmanned air vehicle (UAV). Deep learning based on convolutional neural networks (CNNs) has been verified for its performance in various areas. It has also demonstrated high recognition accuracy and detection time in agricultural applications such as pest and disease diagnosis and prediction. A large amount of data is required to train CNNs effectively. Still, since UAVs capture only a limited number of images, we propose a data augmentation method that can effectively increase data. And most occupancy prediction predicts occupancy by designing a CNN-based object detector for an image and counting the number of recognized objects or calculating the number of pixels occupied by an object. These methods require complex occupancy rate calculations; the accuracy depends on whether the object features of interest are visible in the image. However, in this study, CNN is not approached as a corn object detection and classification problem but as a function approximation and regression problem so that the occupancy rate of corn objects in an image can be represented as the CNN output. The proposed method effectively estimates occupancy for a limited number of cornfield photos, shows excellent prediction accuracy, and confirms the potential and scalability of deep learning.

드론영상에서 구조요청자 자동추출 방안: 도심지역 촬영영상을 중심으로 (Automatic Extraction of Rescue Requests from Drone Images: Focused on Urban Area Images)

  • 박창민
    • 디지털산업정보학회논문지
    • /
    • 제15권3호
    • /
    • pp.37-44
    • /
    • 2019
  • In this study, we propose the automatic extraction method of Rescue Requests from Drone Images. A central object is extracted from each image by using central object extraction method[7] before classification. A central object in an images are defined as a set of regions that is lined around center of the image and has significant texture distribution against its surrounding. In this case of artificial objects, edge of straight line is often found, and texture is regular and directive. However, natural object's case is not. Such characteristics are extracted using Edge direction histogram energy and texture Gabor energy. The Edge direction histogram energy calculated based on the direction of only non-circular edges. The texture Gabor energy is calculated based on the 24-dimension Gebor filter bank. Maximum and minimum energy along direction in Gabor filter dictionary is selected. Finally, the extracted rescue requestor object areas using the dominant features of the objects. Through experiments, we obtain accuracy of more than 75% for extraction method using each features.