• Title/Summary/Keyword: Pose Detection

Search Result 281, Processing Time 0.029 seconds

A Study on Falling Detection of Workers in the Underground Utility Tunnel using Dual Deep Learning Techniques (이중 딥러닝 기법을 활용한 지하공동구 작업자의 쓰러짐 검출 연구)

  • Jeongsoo Kim;Sangmi Park;Changhee Hong
    • Journal of the Society of Disaster Information
    • /
    • v.19 no.3
    • /
    • pp.498-509
    • /
    • 2023
  • Purpose: This paper proposes a method detecting the falling of a maintenance worker in the underground utility tunnel, by applying deep learning techniques using CCTV video, and evaluates the applicability of the proposed method to the worker monitoring of the utility tunnel. Method: Each rule was designed to detect the falling of a maintenance worker by using the inference results from pre-trained YOLOv5 and OpenPose models, respectively. The rules were then integrally applied to detect worker falls within the utility tunnel. Result: Although the worker presence and falling were detected by the proposed model, the inference results were dependent on both the distance between the worker and CCTV and the falling direction of the worker. Additionally, the falling detection system using YOLOv5 shows superior performance, due to its lower dependence on distance and fall direction, compared to the OpenPose-based. Consequently, results from the fall detection using the integrated dual deep learning model were dependent on the YOLOv5 detection performance. Conclusion: The proposed hybrid model shows detecting an abnormal worker in the utility tunnel but the improvement of the model was meaningless compared to the single model based YOLOv5 due to severe differences in detection performance between each deep learning model

Triangle Method for Fast Face Detection on the Wild

  • Malikovich, Karimov Madjit;Akhmatovich, Tashev Komil;ugli, Islomov Shahboz Zokir;Nizomovich, Mavlonov Obid
    • Journal of Multimedia Information System
    • /
    • v.5 no.1
    • /
    • pp.15-20
    • /
    • 2018
  • There are a lot of problems in the face detection area. One of them is detecting faces by facial features and reducing number of the false negatives and positions. This paper is directed to solve this problem by the proposed triangle method. Also, this paper explans cascades, Haar-like features, AdaBoost, HOG. We propose a scheme using 12-net, 24-net, 48-net to scan images and improve efficiency. Using triangle method for frontal pose, B and B1 methods for other poses in neural networks are proposed.

Pose Invariant View-Based Enhanced Fisher Linear Discriminant Models for Face Recognition

  • Lee, Sung-Oh;Park, Gwi-Tae
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.101.2-101
    • /
    • 2001
  • This paper proposes a novel face recognition algorithm to recognize human face robustly under various conditions, such as changes of pose, illumination, and expression, etc. at indoor environments. A conventional automatic face recognition system consists of the detection and the recognition part. Generally, the detection part is dominant over the other part in the estimating whole recognition rate. So, in this paper, we suggest the view-specific eigenface method as preprocessor to estimate various poses of the face in the input image. Then, we apply the Enhanced FLD Models (EFM) to the result of it, twice. Because, the EFM recognizes human face, and reduces the error of standardization effectively. To deal with view-varying problem, we build one basis vector set for each view individually. Finally, the dimensionalities of ...

  • PDF

Object Recognition and Pose Estimation Based on Deep Learning for Visual Servoing (비주얼 서보잉을 위한 딥러닝 기반 물체 인식 및 자세 추정)

  • Cho, Jaemin;Kang, Sang Seung;Kim, Kye Kyung
    • The Journal of Korea Robotics Society
    • /
    • v.14 no.1
    • /
    • pp.1-7
    • /
    • 2019
  • Recently, smart factories have attracted much attention as a result of the 4th Industrial Revolution. Existing factory automation technologies are generally designed for simple repetition without using vision sensors. Even small object assemblies are still dependent on manual work. To satisfy the needs for replacing the existing system with new technology such as bin picking and visual servoing, precision and real-time application should be core. Therefore in our work we focused on the core elements by using deep learning algorithm to detect and classify the target object for real-time and analyzing the object features. We chose YOLO CNN which is capable of real-time working and combining the two tasks as mentioned above though there are lots of good deep learning algorithms such as Mask R-CNN and Fast R-CNN. Then through the line and inside features extracted from target object, we can obtain final outline and estimate object posture.

A Dangerous Situation Recognition System Using Human Behavior Analysis (인간 행동 분석을 이용한 위험 상황 인식 시스템 구현)

  • Park, Jun-Tae;Han, Kyu-Phil;Park, Yang-Woo
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.3
    • /
    • pp.345-354
    • /
    • 2021
  • Recently, deep learning-based image recognition systems have been adopted to various surveillance environments, but most of them are still picture-type object recognition methods, which are insufficient for the long term temporal analysis and high-dimensional situation management. Therefore, we propose a method recognizing the specific dangerous situation generated by human in real-time, and utilizing deep learning-based object analysis techniques. The proposed method uses deep learning-based object detection and tracking algorithms in order to recognize the situations such as 'trespassing', 'loitering', and so on. In addition, human's joint pose data are extracted and analyzed for the emergent awareness function such as 'falling down' to notify not only in the security but also in the emergency environmental utilizations.

Skeleton Model-Based Unsafe Behaviors Detection at a Construction Site Scaffold

  • Nguyen, Truong Linh;Tran, Si Van-Tien;Bao, Quy Lan;Lee, Doyeob;Oh, Myoungho;Park, Chansik
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.361-369
    • /
    • 2022
  • Unsafe actions and behaviors of workers cause most accidents at construction sites. Nowadays, occupational safety is a top priority at construction sites. However, this problem often requires money and effort from investors or construction owners. Therefore, decreasing the accidents rates of workers and saving monitoring costs for contractors is necessary at construction sites. This study proposes an unsafe behavior detection method based on a skeleton model to classify three common unsafe behaviors on the scaffold: climbing, jumping, and running. First, the OpenPose method is used to obtain the workers' key points. Second, all skeleton datasets are aggregated from the temporary size. Third, the key point dataset becomes the input of the action classification model. The method is effective, with an accuracy rate of 89.6% precision and 90.5% recall of unsafe actions correctly detected in the experiment.

  • PDF

The Object 3D Pose Recognition Using Stereo Camera (스테레오 카메라를 이용한 물체의 3D 포즈 인식)

  • Yoo, Sung-Hoon;Kang, Hyo-Seok;Cho, Young-Wan;Kim, Eun-Tai;Park, Mig-Non
    • Proceedings of the IEEK Conference
    • /
    • 2008.06a
    • /
    • pp.1123-1124
    • /
    • 2008
  • In this paper, we develop a program that recognition of the object 3D pose using stereo camera. In order to detect the object, this paper is applied to canny edge detection algorithm and also used stereo camera to get the 3D point about the object and applied to recognize the pose of the object using iterative closest point(ICP) algorithm.

  • PDF

Augmented Reality Service Based on Object Pose Prediction Using PnP Algorithm

  • Kim, In-Seon;Jung, Tae-Won;Jung, Kye-Dong
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.295-301
    • /
    • 2021
  • Digital media technology is gradually developing with the development of convergence quaternary industrial technology and mobile devices. The combination of deep learning and augmented reality can provide more convenient and lively services through the interaction of 3D virtual images with the real world. We combine deep learning-based pose prediction with augmented reality technology. We predict the eight vertices of the bounding box of the object in the image. Using the predicted eight vertices(x,y), eight vertices(x,y,z) of 3D mesh, and the intrinsic parameter of the smartphone camera, we compute the external parameters of the camera through the PnP algorithm. We calculate the distance to the object and the degree of rotation of the object using the external parameter and apply to AR content. Our method provides services in a web environment, making it highly accessible to users and easy to maintain the system. As we provide augmented reality services using consumers' smartphone cameras, we can apply them to various business fields.

Estimation of a Gaze Point in 3D Coordinates using Human Head Pose (휴먼 헤드포즈 정보를 이용한 3차원 공간 내 응시점 추정)

  • Shin, Chae-Rim;Yun, Sang-Seok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.177-179
    • /
    • 2021
  • This paper proposes a method of estimating location of a target point at which an interactive robot gazes in an indoor space. RGB images are extracted from low-cost web-cams, user head pose is obtained from the face detection (Openface) module, and geometric configurations are applied to estimate the user's gaze direction in the 3D space. The coordinates of the target point at which the user stares are finally measured through the correlation between the estimated gaze direction and the plane on the table plane.

  • PDF

A Method for Body Keypoint Localization based on Object Detection using the RGB-D information (RGB-D 정보를 이용한 객체 탐지 기반의 신체 키포인트 검출 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.6
    • /
    • pp.85-92
    • /
    • 2017
  • Recently, in the field of video surveillance, a Deep Learning based learning method has been applied to a method of detecting a moving person in a video and analyzing the behavior of a detected person. The human activity recognition, which is one of the fields this intelligent image analysis technology, detects the object and goes through the process of detecting the body keypoint to recognize the behavior of the detected object. In this paper, we propose a method for Body Keypoint Localization based on Object Detection using RGB-D information. First, the moving object is segmented and detected from the background using color information and depth information generated by the two cameras. The input image generated by rescaling the detected object region using RGB-D information is applied to Convolutional Pose Machines for one person's pose estimation. CPM are used to generate Belief Maps for 14 body parts per person and to detect body keypoints based on Belief Maps. This method provides an accurate region for objects to detect keypoints an can be extended from single Body Keypoint Localization to multiple Body Keypoint Localization through the integration of individual Body Keypoint Localization. In the future, it is possible to generate a model for human pose estimation using the detected keypoints and contribute to the field of human activity recognition.