• Title/Summary/Keyword: Pose Recognition

Search Result 278, Processing Time 0.031 seconds

Speech-Recognition Drone Camera System using OpenPose (OpenPose를 활용한 음성인식기반 드론제어 촬영시스템)

  • Cho, Yu-Jin;Kim, Se-Hyun;Kwon, Ye-Rim;Jung, Soon-Ho
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.11a
    • /
    • pp.1056-1059
    • /
    • 2020
  • 최근 드론과 1인 미디어 시장의 성장으로, 영상 촬영 분야에서의 드론 산업이 활발하게 발전되고 있다. 본 논문에서는 딥러닝 기반 다중 객체 인식 기술인 Openpose를 활용하여 인물촬영을 위한 음성인식 드론 제어 시스템을 제안한다. 해당 시스템은 자연어 처리된 음성명령어를 통해 드론이 각 촬영 객체에 대한 회전, 초점변화 등 실제 영상촬영기법에 사용되는 다수의 동작을 수행할 수 있도록 한다. 최종적으로 96.2%의 정확도로 음성명령에 따라 동작을 수행하는 것을 확인할 수 있다. 이는 누구나 전문적 지식이나 경험 없이 음성만으로 쉽게 드론을 제어할 수 있을 것으로 기대된다.

A Decision Tree based Real-time Hand Gesture Recognition Method using Kinect

  • Chang, Guochao;Park, Jaewan;Oh, Chimin;Lee, Chilwoo
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.12
    • /
    • pp.1393-1402
    • /
    • 2013
  • Hand gesture is one of the most popular communication methods in everyday life. In human-computer interaction applications, hand gesture recognition provides a natural way of communication between humans and computers. There are mainly two methods of hand gesture recognition: glove-based method and vision-based method. In this paper, we propose a vision-based hand gesture recognition method using Kinect. By using the depth information is efficient and robust to achieve the hand detection process. The finger labeling makes the system achieve pose classification according to the finger name and the relationship between each fingers. It also make the classification more effective and accutate. Two kinds of gesture sets can be recognized by our system. According to the experiment, the average accuracy of American Sign Language(ASL) number gesture set is 94.33%, and that of general gestures set is 95.01%. Since our system runs in real-time and has a high recognition rate, we can embed it into various applications.

A study on hand gesture recognition using 3D hand feature (3차원 손 특징을 이용한 손 동작 인식에 관한 연구)

  • Bae Cheol-Soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.4
    • /
    • pp.674-679
    • /
    • 2006
  • In this paper a gesture recognition system using 3D feature data is described. The system relies on a novel 3D sensor that generates a dense range mage of the scene. The main novelty of the proposed system, with respect to other 3D gesture recognition techniques, is the capability for robust recognition of complex hand postures such as those encountered in sign language alphabets. This is achieved by explicitly employing 3D hand features. Moreover, the proposed approach does not rely on colour information, and guarantees robust segmentation of the hand under various illumination conditions, and content of the scene. Several novel 3D image analysis algorithms are presented covering the complete processing chain: 3D image acquisition, arm segmentation, hand -forearm segmentation, hand pose estimation, 3D feature extraction, and gesture classification. The proposed system is tested in an application scenario involving the recognition of sign-language postures.

Feature Variance and Adaptive classifier for Efficient Face Recognition (효과적인 얼굴 인식을 위한 특징 분포 및 적응적 인식기)

  • Dawadi, Pankaj Raj;Nam, Mi Young;Rhee, Phill Kyu
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2007.11a
    • /
    • pp.34-37
    • /
    • 2007
  • Face recognition is still a challenging problem in pattern recognition field which is affected by different factors such as facial expression, illumination, pose etc. The facial feature such as eyes, nose, and mouth constitute a complete face. Mouth feature of face is under the undesirable effect of facial expression as many factors contribute the low performance. We proposed a new approach for face recognition under facial expression applying two cascaded classifiers to improve recognition rate. All facial expression images are treated by general purpose classifier at first stage. All rejected images (applying threshold) are used for adaptation using GA for improvement in recognition rate. We apply Gabor Wavelet as a general classifier and Gabor wavelet with Genetic Algorithm for adaptation under expression variance to solve this issue. We have designed, implemented and demonstrated our proposed approach addressing this issue. FERET face image dataset have been chosen for training and testing and we have achieved a very good success.

3D object recognition using the CAD model and stereo vision

  • Kim, Sung-Il;Choi, Sung-Jun;Won, Sang-Chul
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.669-672
    • /
    • 2003
  • 3D object recognition is difficult but important in computer vision. The important thing is to understand about the relationship between a geometric structure in three dimensions and its image projection. Most 3D recognition systems construct models either manually or by training the pose and orientation of the objects. But both approaches are not satisfactory. In this paper, we focus on a commercial CAD model as a third type of model building for vision. The models are expressed in Initial Graphics Exchanges Specification(IGES) output and reconstructed in a pinhole camera coordinate.

  • PDF

HSFE Network and Fusion Model based Dynamic Hand Gesture Recognition

  • Tai, Do Nhu;Na, In Seop;Kim, Soo Hyung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3924-3940
    • /
    • 2020
  • Dynamic hand gesture recognition(d-HGR) plays an important role in human-computer interaction(HCI) system. With the growth of hand-pose estimation as well as 3D depth sensors, depth, and the hand-skeleton dataset is proposed to bring much research in depth and 3D hand skeleton approaches. However, it is still a challenging problem due to the low resolution, higher complexity, and self-occlusion. In this paper, we propose a hand-shape feature extraction(HSFE) network to produce robust hand-shapes. We build a hand-shape model, and hand-skeleton based on LSTM to exploit the temporal information from hand-shape and motion changes. Fusion between two models brings the best accuracy in dynamic hand gesture (DHG) dataset.

Realtime Face Recognition by Analysis of Feature Information (특징정보 분석을 통한 실시간 얼굴인식)

  • Chung, Jae-Mo;Bae, Hyun;Kim, Sung-Shin
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2001.12a
    • /
    • pp.299-302
    • /
    • 2001
  • The statistical analysis of the feature extraction and the neural networks are proposed to recognize a human face. In the preprocessing step, the normalized skin color map with Gaussian functions is employed to extract the region of face candidate. The feature information in the region of the face candidate is used to detect the face region. In the recognition step, as a tested, the 120 images of 10 persons are trained by the backpropagation algorithm. The images of each person are obtained from the various direction, pose, and facial expression. Input variables of the neural networks are the geometrical feature information and the feature information that comes from the eigenface spaces. The simulation results of$.$10 persons show that the proposed method yields high recognition rates.

  • PDF

Edge Line Information based Underwater Landmark for UUV

  • Yu, Son-Cheol;Kang, Dong-Joung;Kim, Jae-Soo
    • International Journal of Ocean System Engineering
    • /
    • v.1 no.2
    • /
    • pp.68-75
    • /
    • 2011
  • This paper addresses an underwater landmark for updating UUV positioning information. A method is proposed in which the landmark's cubic shape and edge are recognized. The reliability, installation load, and management of landmark design were taken into consideration in order to assess practical applications of the landmark. Landmark recognition was based on topological features. The straight line recognition confirmed the landmark's location and enabled an UUV to accurately estimated its underwater position with respect to the landmark. An efficient recognition method is proposed, which provides real-time processing with limited UUV computing power. An underwater experiment was conducted in order to evaluate the proposed method's reliability and accuracy.

Development of a visual-data processing system for a polyhedral object recognition by the projection of laser ring beam (다면체 물체 인식을 위한 환상레이져 빔 투사형 시각 정보 처리 시스템 개발)

  • 김종형;조용철;조형석
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1988.10a
    • /
    • pp.428-432
    • /
    • 1988
  • In this study, some issues on 3- dimentional object recognition and pose determination are discussed. The method employs a laser projector which projects a cyliderical light beam on the object plane where it produces a bright ring pattern. The picture is then taken by a T.V camera. The ring pattern is mathmetically the ellipse of which the geometrical parameters have the 3-dimentional feature of the object plane. This paper gives the mathematical aspects of 3-dimentional recognition method and shows experimentally the variations of ellipse parameters as the spatial deviation of the plane object.

  • PDF

Periocular Recognition Using uMLBP and Attribute Features

  • Ali, Zahid;Park, Unsang;Nang, Jongho;Park, Jeong-Seon;Hong, Taehwa;Park, Sungjoo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.12
    • /
    • pp.6133-6151
    • /
    • 2017
  • The field of periocular biometrics has gained wide attention as an alternative or supplemental means to conventional biometric traits such as the iris or the face. Periocular biometrics provide intermediate resolution between the iris and the face, which enables it to support both. We have developed a periocular recognition system by using uniform Multiscale Local Binary Pattern (uMLBP) and attribute features. The proposed system has been evaluated in terms of major factors that need to be considered on a mobile platform (e.g., distance and facial pose) to assess the feasibility of the use of periocular biometrics on mobile devices. Experimental results showed 98.7% of rank-1 identification accuracy on a subset of the Face Recognition Grand Challenge (FRGC) database, which is the best performance among similar studies.