• Title/Summary/Keyword: yolo

Search Result 387, Processing Time 0.023 seconds

Object Tracking Method using Deep Learning and Kalman Filter (딥 러닝 및 칼만 필터를 이용한 객체 추적 방법)

  • Kim, Gicheol;Son, Sohee;Kim, Minseop;Jeon, Jinwoo;Lee, Injae;Cha, Jihun;Choi, Haechul
    • Journal of Broadcast Engineering
    • /
    • v.24 no.3
    • /
    • pp.495-505
    • /
    • 2019
  • Typical algorithms of deep learning include CNN(Convolutional Neural Networks), which are mainly used for image recognition, and RNN(Recurrent Neural Networks), which are used mainly for speech recognition and natural language processing. Among them, CNN is able to learn from filters that generate feature maps with algorithms that automatically learn features from data, making it mainstream with excellent performance in image recognition. Since then, various algorithms such as R-CNN and others have appeared in object detection to improve performance of CNN, and algorithms such as YOLO(You Only Look Once) and SSD(Single Shot Multi-box Detector) have been proposed recently. However, since these deep learning-based detection algorithms determine the success of the detection in the still images, stable object tracking and detection in the video requires separate tracking capabilities. Therefore, this paper proposes a method of combining Kalman filters into deep learning-based detection networks for improved object tracking and detection performance in the video. The detection network used YOLO v2, which is capable of real-time processing, and the proposed method resulted in 7.7% IoU performance improvement over the existing YOLO v2 network and 20 fps processing speed in FHD images.

Detection and Identification of Moving Objects at Busy Traffic Road based on YOLO v4 (YOLO v4 기반 혼잡도로에서의 움직이는 물체 검출 및 식별)

  • Li, Qiutan;Ding, Xilong;Wang, Xufei;Chen, Le;Son, Jinku;Song, Jeong-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.1
    • /
    • pp.141-148
    • /
    • 2021
  • In some intersections or busy traffic roads, there are more pedestrians in a specific period of time, and there are many traffic accidents caused by road congestion. Especially at the intersection where there are schools nearby, it is particularly important to protect the traffic safety of students in busy hours. In the past, when designing traffic lights, the safety of pedestrians was seldom taken into account, and the identification of motor vehicles and traffic optimization were mostly studied. How to keep the road smooth as far as possible under the premise of ensuring the safety of pedestrians, especially students, will be the key research direction of this paper. This paper will focus on person, motorcycle, bicycle, car and bus recognition research. Through investigation and comparison, this paper proposes to use YOLO v4 network to identify the location and quantity of objects. YOLO v4 has the characteristics of strong ability of small target recognition, high precision and fast processing speed, and sets the data acquisition object to train and test the image set. Using the statistics of the accuracy rate, error rate and omission rate of the target in the video, the network trained in this paper can accurately and effectively identify persons, motorcycles, bicycles, cars and buses in the moving images.

Fruit's Defective Area Detection Using Yolo V4 Deep Learning Intelligent Technology (Yolo V4 딥러닝 지능기술을 이용한 과일 불량 부위 검출)

  • Choi, Han Suk
    • Smart Media Journal
    • /
    • v.11 no.4
    • /
    • pp.46-55
    • /
    • 2022
  • It is very important to first detect and remove defective fruits with scratches or bruised areas in the automatic fruit quality screening system. This paper proposes a method of detecting defective areas in fruits using the latest artificial intelligence technology, the Yolo V4 deep learning model in order to overcome the limitations of the method of detecting fruit's defective areas using the existing image processing techniques. In this study, a total of 2,400 defective fruits, including 1,000 defective apples and 1,400 defective fruits with scratch or decayed areas, were learned using the Yolo V4 deep learning model and experiments were conducted to detect defective areas. As a result of the performance test, the precision of apples is 0.80, recall is 0.76, IoU is 69.92% and mAP is 65.27%. The precision of pears is 0.86, recall is 0.81, IoU is 70.54% and mAP is 68.75%. The method proposed in this study can dramatically improve the performance of the existing automatic fruit quality screening system by accurately selecting fruits with defective areas in real time rather than using the existing image processing techniques.

YOLO-based Traffic Signal Detection for Identifying the Violation of Motorbike Riders (YOLO 기반의 교통 신호등 인식을 통한 오토바이 운전자의 신호 위반 여부 확인)

  • Wahyutama, Aria Bisma;Hwang, Mintae
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.141-143
    • /
    • 2022
  • This paper presented a new technology to identify traffic violations of motorbike riders by detecting the traffic signal using You Only Look Once (YOLO) object detection. The hardware module that is mounted on the front of the motorbike consists of Raspberry Pi with a camera to run the YOLO object detection, a GPS module to acquire the motorcycle's coordinate, and a LoRa communication module to send the data to a cloud DB. The main goal of the software is to determine whether a motorbike has violated a traffic signal. This paper proposes a function to recognize the red traffic signal colour with its movement inside the camera angle and determine that the traffic signal violation happens if the traffic signal is moving to the right direction (the rider turns left) or moving to the top direction (the riders goes straight). Furthermore, if a motorbike rider is violated the signal, the rider's personal information (name, mobile phone number, etc), the snapshot of the violation situation, rider's location, and date/time will be sent to a cloud DB. The violation information will be delivered to the driver's smartphone as a push notification and the local police station to be used for issuing violation tickets, which is expected to prevent motorbike riders from violating traffic signals.

  • PDF

GPT-enabled SNS Sentence writing support system Based on Image Object and Meta Information (이미지 객체 및 메타정보 기반 GPT 활용 SNS 문장 작성 보조 시스템)

  • Dong-Hee Lee;Mikyeong Moon;Bong-Jun, Choi
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.24 no.3
    • /
    • pp.160-165
    • /
    • 2023
  • In this study, we propose an SNS sentence writing assistance system that utilizes YOLO and GPT to assist users in writing texts with images, such as SNS. We utilize the YOLO model to extract objects from images inserted during writing, and also extract meta-information such as GPS information and creation time information, and use them as prompt values for GPT. To use the YOLO model, we trained it on form image data, and the mAP score of the model is about 0.25 on average. GPT was trained on 1,000 blog text data with the topic of 'restaurant reviews', and the model trained in this study was used to generate sentences with two types of keywords extracted from the images. A survey was conducted to evaluate the practicality of the generated sentences, and a closed-ended survey was conducted to clearly analyze the survey results. There were three evaluation items for the questionnaire by providing the inserted image and keyword sentences. The results showed that the keywords in the images generated meaningful sentences. Through this study, we found that the accuracy of image-based sentence generation depends on the relationship between image keywords and GPT learning contents.

Vehicle Acceleration and Vehicle Spacing Calculation Method Used YOLO (YOLO기법을 사용한 차량가속도 및 차두거리 산출방법)

  • Jeong-won Gil;Jae-seong Hwang;Jae-Kyung Kwon;Choul-ki Lee
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.23 no.1
    • /
    • pp.82-96
    • /
    • 2024
  • While analyzing traffic flow, speed, traffic volume, and density are important macroscopic indicators, and acceleration and spacing are the important microscopic indicators. The speed and traffic volume can be collected with the currently installed traffic information collection devices. However, acceleration and spacing data are necessary for safety and autonomous driving but cannot be collected using the current traffic information collection devices. 'You Look Only Once'(YOLO), an object recognition technique, has excellent accuracy and real-time performance and is used in various fields, including the transportation field. In this study, to measure acceleration and spacing using YOLO, we developed a model that measures acceleration and spacing through changes in vehicle speed at each interval and the differences in the travel time between vehicles by setting the measurement intervals closely. It was confirmed that the range of acceleration and spacing is different depending on the traffic characteristics of each point, and a comparative analysis was performed according to the reference distance and screen angle to secure the measurement rate. The measurement interval was 20m, and the closer the angle was to a right angle, the higher the measurement rate. These results will contribute to the analysis of safety by intersection and the domestic vehicle behavior model.

Real-time Human Detection under Omni-dir ectional Camera based on CNN with Unified Detection and AGMM for Visual Surveillance

  • Nguyen, Thanh Binh;Nguyen, Van Tuan;Chung, Sun-Tae;Cho, Seongwon
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.8
    • /
    • pp.1345-1360
    • /
    • 2016
  • In this paper, we propose a new real-time human detection under omni-directional cameras for visual surveillance purpose, based on CNN with unified detection and AGMM. Compared to CNN-based state-of-the-art object detection methods. YOLO model-based object detection method boasts of very fast object detection, but with less accuracy. The proposed method adapts the unified detecting CNN of YOLO model so as to be intensified by the additional foreground contextual information obtained from pre-stage AGMM. Increased computational time incurred by additional AGMM processing is compensated by speed-up gain obtained from utilizing 2-D input data consisting of grey-level image data and foreground context information instead of 3-D color input data. Through various experiments, it is shown that the proposed method performs better with respect to accuracy and more robust to environment changes than YOLO model-based human detection method, but with the similar processing speeds to that of YOLO model-based one. Thus, it can be successfully employed for embedded surveillance application.

Recession and YOLO: The Influence of Negative Perception of Economic Situation on Present-Biased Preference (경기 불황과 욜로(YOLO): 지각된 부정적 경제 상황이 소비자의 현재에 편향된 선호에 미치는 영향)

  • Jung, Bohee;Jeong, Hyewook
    • Journal of Digital Convergence
    • /
    • v.19 no.11
    • /
    • pp.135-144
    • /
    • 2021
  • This study investigates the underlying mechanism of YOLO in millennial consumer, especially the influence of perceived economic recession on the present-biased preference. In addition, it was attempted to expand the implicit theory by proposing the individual's entity belief as a mediator for the effect of perceived economic situation on consumers' present-biased seeking behavior. In three experimental studies, undergraduate students who both highly primed and measured negative economic situation showed more favorable attitudes towards present-biased persuasive message and related products. The results of this research provides practical implication for marketers especially in the current situation experiencing economic slowdown due to low economic growth and COVID 19.

YOLO based Optical Music Recognition and Virtual Reality Content Creation Method (YOLO 기반의 광학 음악 인식 기술 및 가상현실 콘텐츠 제작 방법)

  • Oh, Kyeongmin;Hong, Yoseop;Baek, Geonyeong;Chun, Chanjun
    • Smart Media Journal
    • /
    • v.10 no.4
    • /
    • pp.80-90
    • /
    • 2021
  • Using optical music recognition technology based on deep learning, we propose to apply the results derived to VR games. To detect the music objects in the music sheet, the deep learning model used YOLO v5, and Hough transform was employed to detect undetected objects, modifying the size of the staff. It analyzes and uses BPM, maximum number of combos, and musical notes in VR games using output result files, and prevents the backlog of notes through Object Pooling technology for resource management. In this paper, VR games can be produced with music elements derived from optical music recognition technology to expand the utilization of optical music recognition along with providing VR contents.

Development of an Efficient 3D Object Recognition Algorithm for Robotic Grasping in Cluttered Environments (혼재된 환경에서의 효율적 로봇 파지를 위한 3차원 물체 인식 알고리즘 개발)

  • Song, Dongwoon;Yi, Jae-Bong;Yi, Seung-Joon
    • The Journal of Korea Robotics Society
    • /
    • v.17 no.3
    • /
    • pp.255-263
    • /
    • 2022
  • 3D object detection pipelines often incorporate RGB-based object detection methods such as YOLO, which detects the object classes and bounding boxes from the RGB image. However, in complex environments where objects are heavily cluttered, bounding box approaches may show degraded performance due to the overlapping bounding boxes. Mask based methods such as Mask R-CNN can handle such situation better thanks to their detailed object masks, but they require much longer time for data preparation compared to bounding box-based approaches. In this paper, we present a 3D object recognition pipeline which uses either the YOLO or Mask R-CNN real-time object detection algorithm, K-nearest clustering algorithm, mask reduction algorithm and finally Principal Component Analysis (PCA) alg orithm to efficiently detect 3D poses of objects in a complex environment. Furthermore, we also present an improved YOLO based 3D object detection algorithm that uses a prioritized heightmap clustering algorithm to handle overlapping bounding boxes. The suggested algorithms have successfully been used at the Artificial-Intelligence Robot Challenge (ARC) 2021 competition with excellent results.