• Title/Summary/Keyword: yolo

Search Result 393, Processing Time 0.028 seconds

Human Detection using Real-virtual Augmented Dataset

  • Jongmin, Lee;Yongwan, Kim;Jinsung, Choi;Ki-Hong, Kim;Daehwan, Kim
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.1
    • /
    • pp.98-102
    • /
    • 2023
  • This paper presents a study on how augmenting semi-synthetic image data improves the performance of human detection algorithms. In the field of object detection, securing a high-quality data set plays the most important role in training deep learning algorithms. Recently, the acquisition of real image data has become time consuming and expensive; therefore, research using synthesized data has been conducted. Synthetic data haves the advantage of being able to generate a vast amount of data and accurately label it. However, the utility of synthetic data in human detection has not yet been demonstrated. Therefore, we use You Only Look Once (YOLO), the object detection algorithm most commonly used, to experimentally analyze the effect of synthetic data augmentation on human detection performance. As a result of training YOLO using the Penn-Fudan dataset, it was shown that the YOLO network model trained on a dataset augmented with synthetic data provided high-performance results in terms of the Precision-Recall Curve and F1-Confidence Curve.

A Beverage Can Recognition System Based on Deep Learning for the Visually Impaired (시각장애인을 위한 딥러닝 기반 음료수 캔 인식 시스템)

  • Lee Chanbee;Sim Suhyun;Kim Sunhee
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.19 no.1
    • /
    • pp.119-127
    • /
    • 2023
  • Recently, deep learning has been used in the development of various institutional devices and services to help the visually impaired people in their daily lives. This is because not only are there few products and facility guides written in braille, but less than 10% of the visually impaired can use braille. In this paper, we propose a system that recognizes beverage cans in real time and outputs the beverage can name with sound for the convenience of the visually impaired. Five commercially available beverage cans were selected, and a CNN model and a YOLO model were designed to recognize the beverage cans. After augmenting the image data, model training was performed. The accuracy of the proposed CNN model and YOLO model is 91.2% and 90.8%, respectively. For practical verification, a system was built by attaching a camera and speaker to a Raspberry Pi. In the system, the YOLO model was applied. It was confirmed that beverage cans were recognized and output as sound in real time in various environments.

Automatic Notification System of Expiration Date Based on YOLO and OCR algorithm for Blind Person (시각 장애우를 위한 YOLO와 OCR 알고리즘 기반의 유통기한 자동 알림 시스템)

  • Kim, Min-Soo;Moon, Mi-kyung;Han, Chang-hee
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.07a
    • /
    • pp.697-698
    • /
    • 2021
  • 본 논문에서는 시각 장애우의 식품 안전성 증진을 위해 광학 문자 인식 (optical character recognition, OCR) 및 실시간 객체 인식 (you only look once, YOLO) 알고리즘에 기반한 식품의 유통기한 자동 알림 시스템을 제안한다. 제안하는 시스템은 1) 스마트폰 카메라를 통해 실시간으로 입력되는 영상에서 YOLO 알고리즘을 활용하여 유통기한으로 예측되는 이미지 영역을 검출하고, 2) 검출된 영역에서 OCR 알고리즘을 활용하여 유통기한 데이터를 추출하며, 3) 최종 추출된 유통기한 데이터를 음성으로 변환하여 시각 장애우에게 전달한다. 개발된 시스템은 유통기한 정보를 추출해서 사용자에게 전달하기까지 평균 약 7초 이내의 빠른 응답 속도를 보였으며, 62.8%의 객체 인식 정확도와 93.6%의 문자 인식 정확도를 보였다. 이러한 결과들은 제안하는 시스템을 시각 장애우들이 실용적으로 활용할 수 있다는 가능성을 보여준다.

  • PDF

Design and Implementation of Finger Direction Detection Algorithm in YOLO Environment (YOLO 환경에서 손가락 방향감지 알고리즘 설계 및 구현)

  • Lee, Cheol Min;Thar, Min Htet;Lee, Dong Myung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.28-30
    • /
    • 2021
  • In this paper, an algorithm that detects the user's finger direction using the YOLO (You Only Look Once) library was proposed. The processing stage of the proposed finger direction detection algorithm consists of a learning data management stage, a data learning stage, and a finger direction detection stage. As a result of the experiment, it was found that the distance between the camera and the finger had a very large influence on the accuracy of detecting the direction of the finger. We plan to apply this function to Turtlebot3 after improving the accuracy and reliability of the proposed algorithm in the future.

  • PDF

An Design and Implementation of YOLO-based Fruit Classification Model. (YOLO 기반 과실 분류 모델 설계 및 구현)

  • Koo, Bon-Geun;Jeong, Da-Un;Kim, Ji-Young;Choi, Ji-Won;Park, Jang-Woo;Cho, Young-Yun;Shin, Chang-Sun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.47-50
    • /
    • 2021
  • 일반적으로, 과실 재배 농가에서의 과실 분류 작업은 많은 노동력을 필요로 한다. 최근 코로나19 방역조치로 외국인 근로자 입국 제한으로 인해 농가에서는 인력 수급에 어려움을 겪고 있다. 본 연구에서는 이러한 농가 피해 상황을 해소하고 수급 문제를 해결하기 위해 YOLO 기반 과실 분류 모델설계 및 구현을 소개한다. 본 모델을 기반으로 여러 과실 중 사과에 적용하여 인력이 많이 동원되는 선별과정을 YOLO의 객체 인식을 통해 해결하고 적은 비용으로 효율성 있는 분류 모델을 구축한다.

YOLO models based Bounding-Box Ensemble Method for Patient Detection In Homecare Place Images (조호환경 내 환자 탐지를 위한 YOLO 모델 기반 바운딩 박스 앙상블 기법)

  • Park, Junhwi;Kim, Beomjun;Kim, Inki;Gwak, Jeonghwan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.11a
    • /
    • pp.562-564
    • /
    • 2022
  • 조호환경이란 환자의 지속적인 추적 및 관찰이 필요한 환경으로써, 병원 입원실, 요양원 등을 의미한다. 조호환경 내 환자의 이상 증세가 발생하는 시간 및 이상 증세의 종류는 예측할 수 없기에 인력을 통한 상시 관리는 필수적이다. 또한, 환자의 이상 증세 발견 시간은 발병 시점부터의 소요 시간이 생사와 즉결되기에 빠른 발견이 매우 중요하다. 하지만, 인력을 통한 상시 관리는 많은 경제적 비용을 수반하기에 독거 노인, 빈민층 등 요양 비용을 충당하지 못하는 환자들이 수혜받는 것은 어려우며, 인력을 통해 이루어지기 때문에 이상 증세 발병 즉시 발견에 한계를 가진다. 즉, 기존까지 조호환경 내 환자 관리 방식은 경제적 비용과 이상 증세 발병 즉시 발견에 한계를 가진다는 문제점을 가진다. 따라서 본 논문은 YOLO 모델의 조호환경 내 환자 탐지 성능 비교 및 바운딩 박스 앙상블 기법을 제안한다. 이를 통해, 딥러닝 모델을 통한 환자 상시 관리가 이루어지기에 높은 경제적 비용문제를 해소할 수 있다. 또한, YOLO 모델 바운딩 박스 앙상블 기법 WBF를 통해 폐색이 짙은 조호환경 영상 데이터 내에 객체 탐지 영역 정확도 향상 방법을 연구하였다.

ANALYSIS OF THE FLOOR PLAN DATASET WITH YOLO V5

  • MYUNGHYUN JUNG;MINJUNG GIM;SEUNGHWAN YANG
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.27 no.4
    • /
    • pp.311-323
    • /
    • 2023
  • This paper introduces the industrial problem, the solution, and the results of the research conducted with Define Inc. The client company wanted to improve the performance of an object detection model on the floor plan dataset. To solve the problem, we analyzed the operational principles, advantages, and disadvantages of the existing object detection model, identified the characteristics of the floor plan dataset, and proposed to use of YOLO v5 as an appropriate object detection model for training the dataset. We compared the performance of the existing model and the proposed model using mAP@60, and verified the object detection results with real test data, and found that the performance increase of mAP@60 was 0.08 higher with a 25% shorter inference time. We also found that the training time of the proposed YOLO v5 was 71% shorter than the existing model because it has a simpler structure. In this paper, we have shown that the object detection model for the floor plan dataset can achieve better performance while reducing the training time. We expect that it will be useful for solving other industrial problems related to object detection in the future. We also believe that this result can be extended to study object recognition in 3D floor plan dataset.

Development a Meal Support System for the Visually Impaired Using YOLO Algorithm (YOLO알고리즘을 활용한 시각장애인용 식사보조 시스템 개발)

  • Lee, Gun-Ho;Moon, Mi-Kyeong
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.5
    • /
    • pp.1001-1010
    • /
    • 2021
  • Normal people are not deeply aware of their dependence on sight when eating. However, since the visually impaired do not know what kind of food is on the table, the assistant next to them holds the blind spoon and explains the position of the food in a clockwise direction, front and rear, left and right, etc. In this paper, we describe the development of a meal assistance system that recognizes each food image and announces the name of the food by voice when a visually impaired person looks at their table using a smartphone camera. This system extracts the food on which the spoon is placed through the YOLO model that has learned the image of food and tableware (spoon), recognizes what the food is, and notifies it by voice. Through this system, it is expected that the visually impaired will be able to eat without the help of a meal assistant, thereby increasing their self-reliance and satisfaction.

Implementation of Drowsy Prevention System Using Arduino and YOLO (아두이노와 YOLO를 이용한 졸음 방지 시스템 구현)

  • Lee, Hyun-Ae;Shin, Seong-Yoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.7
    • /
    • pp.917-922
    • /
    • 2021
  • In modern society, deaths and property damage due to drowsiness occur every year enormously. Methods to reduce such damage are being studied a lot in all walks of life, and research on preventing drowsy driving is particularly active in automobiles. In this paper, as an Arduino-based water gun firing system that learns open and closed eyes using YOLO, we propose a drowsy prevention system that fires a water gun when the duration of the closed eye exceeds a certain time. This system can be applied and used in various fields, but especially when applied to a car, it is not necessary to purchase expensive specifications and if you pay a little attention, you can reduce accidents caused by drowsy driving by 100% at a very low cost. In addition, it can be said that it is an independent system that overcomes different specifications for each company.

Implementation of a Classification System for Dog Behaviors using YOLI-based Object Detection and a Node.js Server (YOLO 기반 개체 검출과 Node.js 서버를 이용한 반려견 행동 분류 시스템 구현)

  • Jo, Yong-Hwa;Lee, Hyuek-Jae;Kim, Young-Hun
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.21 no.1
    • /
    • pp.29-37
    • /
    • 2020
  • This paper implements a method of extracting an object about a dog through real-time image analysis and classifying dog behaviors from the extracted images. The Darknet YOLO was used to detect dog objects, and the Teachable Machine provided by Google was used to classify behavior patterns from the extracted images. The trained Teachable Machine is saved in Google Drive and can be used by ml5.js implemented on a node.js server. By implementing an interactive web server using a socket.io module on the node.js server, the classified results are transmitted to the user's smart phone or PC in real time so that it can be checked anytime, anywhere.