• Title/Summary/Keyword: YOLOv5s

Search Result 47, Processing Time 0.018 seconds

Development of Personal Mobility Safety Assistants using Object Detection based on Deep Learning (딥러닝 기반 객체 인식을 활용한 퍼스널 모빌리티 안전 보조 시스템 개발)

  • Kwak, Hyeon-Seo;Kim, Min-Young;Jeon, Ji-Yong;Jeong, Eun-Hye;Kim, Ju-Yeop;Hyeon, So-Dam;Jeong, Jin-Woo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.3
    • /
    • pp.486-489
    • /
    • 2021
  • Recently, the demand for the use of personal mobility vehicles, such as an electric kickboard, is increasing explosively because of its high portability and usability. However, the number of traffic accidents caused by personal mobility vehicles has also increased rapidly in recent years. To address the issues regarding the driver's safety, we propose a novel approach that can monitor context information around personal mobility vehicles using deep learning-based object detection and smartphone captured videos. In the proposed framework, a smartphone is attached to a personal mobility device and a front or rear view is recorded to detect an approaching object that may affect the driver's safety. Through the detection results using YOLOv5 model, we report the preliminary results and validated the feasibility of the proposed approach.

Deep-Learning-Based Water Shield Automation System by Predicting River Overflow and Vehicle Flooding Possibility (하천 범람 및 차량 침수 가능성 예측을 통한 딥러닝 기반 차수막 자동화 시스템)

  • Seung-Jae Ham;Min-Su Kang;Seong-Woo Jeong;Joonhyuk Yoo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.3
    • /
    • pp.133-139
    • /
    • 2023
  • This paper proposes a two-stage Water Shield Automation System (WSAS) to predict the possibility of river overflow and vehicle flooding due to sudden rainfall. The WSAS uses a two-stage Deep Neural Network (DNN) model. First, a river overflow prediction module is designed with LSTM to decide whether the river is flooded by predicting the river's water level rise. Second, a vehicle flooding prediction module predicts flooding of underground parking lots by detecting flooded tires with YOLOv5 from CCTV images. Finally, the WSAS automatically installs the water barrier whenever the river overflow and vehicle flooding events happen in the underground parking lots. The only constraint to implementing is that collecting training data for flooded vehicle tires is challenging. This paper exploits the Image C&S data augmentation technique to synthesize flooded tire images. Experimental results validate the superiority of WSAS by showing that the river overflow prediction module can reduce RMSE by three times compared with the previous method, and the vehicle flooding detection module can increase mAP by 20% compared with the naive detection method, respectively.

A Study on the Dataset Construction and Model Application for Detecting Surgical Gauze in C-Arm Imaging Using Artificial Intelligence (인공지능을 활용한 C-Arm에서 수술용 거즈 검출을 위한 데이터셋 구축 및 검출모델 적용에 관한 연구)

  • Kim, Jin Yeop;Hwang, Ho Seong;Lee, Joo Byung;Choi, Yong Jin;Lee, Kang Seok;Kim, Ho Chul
    • Journal of Biomedical Engineering Research
    • /
    • v.43 no.4
    • /
    • pp.290-297
    • /
    • 2022
  • During surgery, Surgical instruments are often left behind due to accidents. Most of these are surgical gauze, so radioactive non-permeable gauze (X-ray gauze) is used for preventing of accidents which gauze is left in the body. This gauze is divided into wire and pad type. If it is confirmed that the gauze remains in the body, gauze must be detected by radiologist's reading by imaging using a mobile X-ray device. But most of operating rooms are not equipped with a mobile X-ray device, but equipped C-Arm equipment, which is of poorer quality than mobile X-ray equipment and furthermore it takes time to read them. In this study, Use C-Arm equipment to acquire gauze image for detection and Build dataset using artificial intelligence and select a detection model to Assist with the relatively low image quality and the reading of radiology specialists. mAP@50 and detection time are used as indicators for performance evaluation. The result is that two-class gauze detection dataset is more accurate and YOLOv5 model mAP@50 is 93.4% and detection time is 11.7 ms.

A Study on Traffic Vulnerable Detection Using Object Detection-Based Ensemble and YOLOv5

  • Hyun-Do Lee;Sun-Gu Kim;Seung-Chae Na;Ji-Yul Ham;Chanhee Kwak
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.1
    • /
    • pp.61-68
    • /
    • 2024
  • Despite the continuous efforts to mitigate pedestrian accidents at crosswalks, the problem persist. Vulnerable groups, including the elderly and disabled individuals are at a risk of being involved in traffic incidents. This paper proposes the implementation of object detection algorithm using the YOLO v5 model specifically for pedestrians using assistive devices like wheelchairs and crutches. For this research, data was collected and utilized through image crawling, Roboflow, and Mobility Aids datasets, which comprise of wheelchair users, crutch users, and pedestrians. Data augmentation techniques were applied to improve the model's generalization performance. Additionally, ensemble techniques were utilized to mitigate type 2 errors, resulting in 96% recall rate. This demonstrates that employing ensemble methods with a single YOLO model to target transportation-disadvantaged individuals can yield accurate detection performance without overlooking crucial objects.

Field Applicability Study of Hull Crack Detection Based on Artificial Intelligence (인공지능 기반 선체 균열 탐지 현장 적용성 연구)

  • Song, Sang-ho;Lee, Gap-heon;Han, Ki-min;Jang, Hwa-sup
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.59 no.4
    • /
    • pp.192-199
    • /
    • 2022
  • With the advent of autonomous ships, it is emerging as one of the very important issues not only to operate with a minimum crew or unmanned ships, but also to secure the safety of ships to prevent marine accidents. On-site inspection of the hull is mainly performed by the inspector's visual inspection, and video information is recorded using a small camera if necessary. However, due to the shortage of inspection personnel, time and space constraints, and the pandemic situation, the necessity of introducing an automated inspection system using artificial intelligence and remote inspection is becoming more important. Furthermore, research on hardware and software that enables the automated inspection system to operate normally even under the harsh environmental conditions of a ship is absolutely necessary. For automated inspection systems, it is important to review artificial intelligence technologies and equipment that can perform a variety of hull failure detection and classification. To address this, it is important to classify the hull failure. Based on various guidelines and expert opinions, we divided them into 6 types(Crack, Corrosion, Pitting, Deformation, Indent, Others). It was decided to apply object detection technology to cracks of hull failure. After that, YOLOv5 was decided as an artificial intelligence model suitable for survey and a common hull crack dataset was trained. Based on the performance results, it aims to present the possibility of applying artificial intelligence in the field by determining and testing the equipment required for survey.

Estimating vegetation index for outdoor free-range pig production using YOLO

  • Sang-Hyon Oh;Hee-Mun Park;Jin-Hyun Park
    • Journal of Animal Science and Technology
    • /
    • v.65 no.3
    • /
    • pp.638-651
    • /
    • 2023
  • The objective of this study was to quantitatively estimate the level of grazing area damage in outdoor free-range pig production using a Unmanned Aerial Vehicles (UAV) with an RGB image sensor. Ten corn field images were captured by a UAV over approximately two weeks, during which gestating sows were allowed to graze freely on the corn field measuring 100 × 50 m2. The images were corrected to a bird's-eye view, and then divided into 32 segments and sequentially inputted into the YOLOv4 detector to detect the corn images according to their condition. The 43 raw training images selected randomly out of 320 segmented images were flipped to create 86 images, and then these images were further augmented by rotating them in 5-degree increments to create a total of 6,192 images. The increased 6,192 images are further augmented by applying three random color transformations to each image, resulting in 24,768 datasets. The occupancy rate of corn in the field was estimated efficiently using You Only Look Once (YOLO). As of the first day of observation (day 2), it was evident that almost all the corn had disappeared by the ninth day. When grazing 20 sows in a 50 × 100 m2 cornfield (250 m2/sow), it appears that the animals should be rotated to other grazing areas to protect the cover crop after at least five days. In agricultural technology, most of the research using machine and deep learning is related to the detection of fruits and pests, and research on other application fields is needed. In addition, large-scale image data collected by experts in the field are required as training data to apply deep learning. If the data required for deep learning is insufficient, a large number of data augmentation is required.

Video Analysis System for Action and Emotion Detection by Object with Hierarchical Clustering based Re-ID (계층적 군집화 기반 Re-ID를 활용한 객체별 행동 및 표정 검출용 영상 분석 시스템)

  • Lee, Sang-Hyun;Yang, Seong-Hun;Oh, Seung-Jin;Kang, Jinbeom
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.89-106
    • /
    • 2022
  • Recently, the amount of video data collected from smartphones, CCTVs, black boxes, and high-definition cameras has increased rapidly. According to the increasing video data, the requirements for analysis and utilization are increasing. Due to the lack of skilled manpower to analyze videos in many industries, machine learning and artificial intelligence are actively used to assist manpower. In this situation, the demand for various computer vision technologies such as object detection and tracking, action detection, emotion detection, and Re-ID also increased rapidly. However, the object detection and tracking technology has many difficulties that degrade performance, such as re-appearance after the object's departure from the video recording location, and occlusion. Accordingly, action and emotion detection models based on object detection and tracking models also have difficulties in extracting data for each object. In addition, deep learning architectures consist of various models suffer from performance degradation due to bottlenects and lack of optimization. In this study, we propose an video analysis system consists of YOLOv5 based DeepSORT object tracking model, SlowFast based action recognition model, Torchreid based Re-ID model, and AWS Rekognition which is emotion recognition service. Proposed model uses single-linkage hierarchical clustering based Re-ID and some processing method which maximize hardware throughput. It has higher accuracy than the performance of the re-identification model using simple metrics, near real-time processing performance, and prevents tracking failure due to object departure and re-emergence, occlusion, etc. By continuously linking the action and facial emotion detection results of each object to the same object, it is possible to efficiently analyze videos. The re-identification model extracts a feature vector from the bounding box of object image detected by the object tracking model for each frame, and applies the single-linkage hierarchical clustering from the past frame using the extracted feature vectors to identify the same object that failed to track. Through the above process, it is possible to re-track the same object that has failed to tracking in the case of re-appearance or occlusion after leaving the video location. As a result, action and facial emotion detection results of the newly recognized object due to the tracking fails can be linked to those of the object that appeared in the past. On the other hand, as a way to improve processing performance, we introduce Bounding Box Queue by Object and Feature Queue method that can reduce RAM memory requirements while maximizing GPU memory throughput. Also we introduce the IoF(Intersection over Face) algorithm that allows facial emotion recognized through AWS Rekognition to be linked with object tracking information. The academic significance of this study is that the two-stage re-identification model can have real-time performance even in a high-cost environment that performs action and facial emotion detection according to processing techniques without reducing the accuracy by using simple metrics to achieve real-time performance. The practical implication of this study is that in various industrial fields that require action and facial emotion detection but have many difficulties due to the fails in object tracking can analyze videos effectively through proposed model. Proposed model which has high accuracy of retrace and processing performance can be used in various fields such as intelligent monitoring, observation services and behavioral or psychological analysis services where the integration of tracking information and extracted metadata creates greate industrial and business value. In the future, in order to measure the object tracking performance more precisely, there is a need to conduct an experiment using the MOT Challenge dataset, which is data used by many international conferences. We will investigate the problem that the IoF algorithm cannot solve to develop an additional complementary algorithm. In addition, we plan to conduct additional research to apply this model to various fields' dataset related to intelligent video analysis.