• Title/Summary/Keyword: object detection and classification

Search Result 296, Processing Time 0.031 seconds

A selective sparse coding based fast super-resolution method for a side-scan sonar image (선택적 sparse coding 기반 측면주사 소나 영상의 고속 초해상도 복원 알고리즘)

  • Park, Jaihyun;Yang, Cheoljong;Ku, Bonwha;Lee, Seungho;Kim, Seongil;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.37 no.1
    • /
    • pp.12-20
    • /
    • 2018
  • Efforts have been made to reconstruct low-resolution underwater images to high-resolution ones by using the image SR (Super-Resolution) method, all to improve efficiency when acquiring side-scan sonar images. As side-scan sonar images are similar with the optical images with respect to exploiting 2-dimensional signals, conventional image restoration methods for optical images can be considered as a solution. One of the most typical super-resolution methods for optical image is a sparse coding and there are studies for verifying applicability of sparse coding method for underwater images by analyzing sparsity of underwater images. Sparse coding is a method that obtains recovered signal from input signal by linear combination of dictionary and sparse coefficients. However, it requires huge computational load to accurately estimate sparse coefficients. In this study, a sparse coding based underwater image super-resolution method is applied while a selective reconstruction method for object region is suggested to reduce the processing time. For this method, this paper proposes an edge detection and object and non object region classification method for underwater images and combine it with sparse coding based image super-resolution method. Effectiveness of the proposed method is verified by reducing the processing time for image reconstruction over 32 % while preserving same level of PSNR (Peak Signal-to-Noise Ratio) compared with conventional method.

Gait Recognition Algorithm Based on Feature Fusion of GEI Dynamic Region and Gabor Wavelets

  • Huang, Jun;Wang, Xiuhui;Wang, Jun
    • Journal of Information Processing Systems
    • /
    • v.14 no.4
    • /
    • pp.892-903
    • /
    • 2018
  • The paper proposes a novel gait recognition algorithm based on feature fusion of gait energy image (GEI) dynamic region and Gabor, which consists of four steps. First, the gait contour images are extracted through the object detection, binarization and morphological process. Secondly, features of GEI at different angles and Gabor features with multiple orientations are extracted from the dynamic part of GEI, respectively. Then averaging method is adopted to fuse features of GEI dynamic region with features of Gabor wavelets on feature layer and the feature space dimension is reduced by an improved Kernel Principal Component Analysis (KPCA). Finally, the vectors of feature fusion are input into the support vector machine (SVM) based on multi classification to realize the classification and recognition of gait. The primary contributions of the paper are: a novel gait recognition algorithm based on based on feature fusion of GEI and Gabor is proposed; an improved KPCA method is used to reduce the feature matrix dimension; a SVM is employed to identify the gait sequences. The experimental results suggest that the proposed algorithm yields over 90% of correct classification rate, which testify that the method can identify better different human gait and get better recognized effect than other existing algorithms.

A Study on Automatic Vehicle Extraction within Drone Image Bounding Box Using Unsupervised SVM Classification Technique (무감독 SVM 분류 기법을 통한 드론 영상 경계 박스 내 차량 자동 추출 연구)

  • Junho Yeom
    • Land and Housing Review
    • /
    • v.14 no.4
    • /
    • pp.95-102
    • /
    • 2023
  • Numerous investigations have explored the integration of machine leaning algorithms with high-resolution drone image for object detection in urban settings. However, a prevalent limitation in vehicle extraction studies involves the reliance on bounding boxes rather than instance segmentation. This limitation hinders the precise determination of vehicle direction and exact boundaries. Instance segmentation, while providing detailed object boundaries, necessitates labour intensive labelling for individual objects, prompting the need for research on automating unsupervised instance segmentation in vehicle extraction. In this study, a novel approach was proposed for vehicle extraction utilizing unsupervised SVM classification applied to vehicle bounding boxes in drone images. The method aims to address the challenges associated with bounding box-based approaches and provide a more accurate representation of vehicle boundaries. The study showed promising results, demonstrating an 89% accuracy in vehicle extraction. Notably, the proposed technique proved effective even when dealing with significant variations in spectral characteristics within the vehicles. This research contributes to advancing the field by offering a viable solution for automatic and unsupervised instance segmentation in the context of vehicle extraction from image.

Quantitative Evaluations of Deep Learning Models for Rapid Building Damage Detection in Disaster Areas (재난지역에서의 신속한 건물 피해 정도 감지를 위한 딥러닝 모델의 정량 평가)

  • Ser, Junho;Yang, Byungyun
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.40 no.5
    • /
    • pp.381-391
    • /
    • 2022
  • This paper is intended to find one of the prevailing deep learning models that are a type of AI (Artificial Intelligence) that helps rapidly detect damaged buildings where disasters occur. The models selected are SSD-512, RetinaNet, and YOLOv3 which are widely used in object detection in recent years. These models are based on one-stage detector networks that are suitable for rapid object detection. These are often used for object detection due to their advantages in structure and high speed but not for damaged building detection in disaster management. In this study, we first trained each of the algorithms on xBD dataset that provides the post-disaster imagery with damage classification labels. Next, the three models are quantitatively evaluated with the mAP(mean Average Precision) and the FPS (Frames Per Second). The mAP of YOLOv3 is recorded at 34.39%, and the FPS reached 46. The mAP of RetinaNet recorded 36.06%, which is 1.67% higher than YOLOv3, but the FPS is one-third of YOLOv3. SSD-512 received significantly lower values than the results of YOLOv3 on two quantitative indicators. In a disaster situation, a rapid and precise investigation of damaged buildings is essential for effective disaster response. Accordingly, it is expected that the results obtained through this study can be effectively used for the rapid response in disaster management.

Key Point Extraction from LiDAR Data for 3D Modeling (3차원 모델링을 위한 라이다 데이터로부터 특징점 추출 방법)

  • Lee, Dae Geon;Lee, Dong-Cheon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.34 no.5
    • /
    • pp.479-493
    • /
    • 2016
  • LiDAR(Light Detection and Ranging) data acquired from ALS(Airborne Laser Scanner) has been intensively utilized to reconstruct object models. Especially, researches for 3D modeling from LiDAR data have been performed to establish high quality spatial information such as precise 3D city models and true orthoimages efficiently. To reconstruct object models from irregularly distributed LiDAR point clouds, sensor calibration, noise removal, filtering to separate objects from ground surfaces are required as pre-processing. Classification and segmentation based on geometric homogeneity of the features, grouping and representation of the segmented surfaces, topological analysis of the surface patches for modeling, and accuracy assessment are accompanied by modeling procedure. While many modeling methods are based on the segmentation process, this paper proposed to extract key points directly for building modeling without segmentation. The method was applied to simulated and real data sets with various roof shapes. The results demonstrate feasibility of the proposed method through the accuracy analysis.

A Study on the Characteristics of Linear Smoothing Algorithm for Image-Based Object Detection of Water Friendly Facilities in River (영상 기반의 하천 친수시설 추출을 위한 선형 평활화 알고리즘 특성 연구)

  • Im, Yun Seong;Kim, Seo Jun;Kim, Chang Sung;Kim, Seong Jun
    • Ecology and Resilient Infrastructure
    • /
    • v.8 no.4
    • /
    • pp.266-272
    • /
    • 2021
  • Water friendly space refers to a place designated to plan and manage spaces for residents Water friendly activities. Efficient management of river Water friendly parks requires automated GIS data and DB construction of the water friendly facilities. Object-based classification using drone images or aerial images is attracting attention as an efficient means to acquire 3D spatial information in the country. To remove the miscellaneous image included in the extracted outline, a linear simplification of the outline is required, and it is difficult to apply manually, so various automation methods have been developed to overcome this, and among them, the most widely studied and utilized is the linear simplification method. In this study, the suitability of linear simplification algorithms such as Douglas-Peucker, Visvalingam-Whyatt, and Bend-simplify algorithms for the geometric shape of hydrophilic facilities was determined.

Improving Weak Classifiers by Using Discriminant Function in Selecting Threshold Values (판별 함수를 이용한 문턱치 선정에 의한 약분류기 개선)

  • Shyam, Adhikari;Yoo, Hyeon-Joong;Kim, Hyong-Suk
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.12
    • /
    • pp.84-90
    • /
    • 2010
  • In this paper, we propose a quadratic discriminant analysis based approach for improving the discriminating strength of weak classifiers based on simple Haar-like features that were used in the Viola-Jones object detection framework. Viola and Jones built a strong classifier using a boosted ensemble of weak classifiers. However, their single threshold (or decision boundary) based weak classifier is sub-optimal and too weak for efficient discrimination between object class and background. A quadratic discriminant analysis based approach is presented which leads to hyper-quadric boundary between the object class and background class, thus realizing multiple thresholds based weak classifiers. Experiments carried out for car detection using 1000 positive and 3000 negative images for training, and 500 positive and 500 negative images for testing show that our method yields higher classification performance with fewer classifiers than single threshold based weak classifiers.

Machine learning application for predicting the strawberry harvesting time

  • Yang, Mi-Hye;Nam, Won-Ho;Kim, Taegon;Lee, Kwanho;Kim, Younghwa
    • Korean Journal of Agricultural Science
    • /
    • v.46 no.2
    • /
    • pp.381-393
    • /
    • 2019
  • A smart farm is a system that combines information and communication technology (ICT), internet of things (IoT), and agricultural technology that enable a farm to operate with minimal labor and to automatically control of a greenhouse environment. Machine learning based on recently data-driven techniques has emerged with big data technologies and high-performance computing to create opportunities to quantify data intensive processes in agricultural operational environments. This paper presents research on the application of machine learning technology to diagnose the growth status of crops and predicting the harvest time of strawberries in a greenhouse according to image processing techniques. To classify the growth stages of the strawberries, we used object inference and detection with machine learning model based on deep learning neural networks and TensorFlow. The classification accuracy was compared based on the training data volume and training epoch. As a result, it was able to classify with an accuracy of over 90% with 200 training images and 8,000 training steps. The detection and classification of the strawberry maturities could be identified with an accuracy of over 90% at the mature and over mature stages of the strawberries. Concurrently, the experimental results are promising, and they show that this approach can be applied to develop a machine learning model for predicting the strawberry harvesting time and can be used to provide key decision support information to both farmers and policy makers about optimal harvest times and harvest planning.

Implementation of a Classification System for Dog Behaviors using YOLI-based Object Detection and a Node.js Server (YOLO 기반 개체 검출과 Node.js 서버를 이용한 반려견 행동 분류 시스템 구현)

  • Jo, Yong-Hwa;Lee, Hyuek-Jae;Kim, Young-Hun
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.21 no.1
    • /
    • pp.29-37
    • /
    • 2020
  • This paper implements a method of extracting an object about a dog through real-time image analysis and classifying dog behaviors from the extracted images. The Darknet YOLO was used to detect dog objects, and the Teachable Machine provided by Google was used to classify behavior patterns from the extracted images. The trained Teachable Machine is saved in Google Drive and can be used by ml5.js implemented on a node.js server. By implementing an interactive web server using a socket.io module on the node.js server, the classified results are transmitted to the user's smart phone or PC in real time so that it can be checked anytime, anywhere.

DCNN Optimization Using Multi-Resolution Image Fusion

  • Alshehri, Abdullah A.;Lutz, Adam;Ezekiel, Soundararajan;Pearlstein, Larry;Conlen, John
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.11
    • /
    • pp.4290-4309
    • /
    • 2020
  • In recent years, advancements in machine learning capabilities have allowed it to see widespread adoption for tasks such as object detection, image classification, and anomaly detection. However, despite their promise, a limitation lies in the fact that a network's performance quality is based on the data which it receives. A well-trained network will still have poor performance if the subsequent data supplied to it contains artifacts, out of focus regions, or other visual distortions. Under normal circumstances, images of the same scene captured from differing points of focus, angles, or modalities must be separately analysed by the network, despite possibly containing overlapping information such as in the case of images of the same scene captured from different angles, or irrelevant information such as images captured from infrared sensors which can capture thermal information well but not topographical details. This factor can potentially add significantly to the computational time and resources required to utilize the network without providing any additional benefit. In this study, we plan to explore using image fusion techniques to assemble multiple images of the same scene into a single image that retains the most salient key features of the individual source images while discarding overlapping or irrelevant data that does not provide any benefit to the network. Utilizing this image fusion step before inputting a dataset into the network, the number of images would be significantly reduced with the potential to improve the classification performance accuracy by enhancing images while discarding irrelevant and overlapping regions.