• Title/Summary/Keyword: 얼굴모델

Search Result 602, Processing Time 0.024 seconds

Method of Biological Information Analysis Based-on Object Contextual (대상객체 맥락 기반 생체정보 분석방법)

  • Kim, Kyung-jun;Kim, Ju-yeon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.41-43
    • /
    • 2022
  • In order to prevent and block infectious diseases caused by the recent COVID-19 pandemic, non-contact biometric information acquisition and analysis technology is attracting attention. The invasive and attached biometric information acquisition method accurately has the advantage of measuring biometric information, but has a risk of increasing contagious diseases due to the close contact. To solve these problems, the non-contact method of extracting biometric information such as human fingerprints, faces, iris, veins, voice, and signatures with automated devices is increasing in various industries as data processing speed increases and recognition accuracy increases. However, although the accuracy of the non-contact biometric data acquisition technology is improved, the non-contact method is greatly influenced by the surrounding environment of the object to be measured, which is resulting in distortion of measurement information and poor accuracy. In this paper, we propose a context-based bio-signal modeling technique for the interpretation of personalized information (image, signal, etc.) for bio-information analysis. Context-based biometric information modeling techniques present a model that considers contextual and user information in biometric information measurement in order to improve performance. The proposed model analyzes signal information based on the feature probability distribution through context-based signal analysis that can maximize the predicted value probability.

  • PDF

Implementation of Markerless Augmented Reality with Deformable Object Simulation (변형물체 시뮬레이션을 활용한 비 마커기반 증강현실 시스템 구현)

  • Sung, Nak-Jun;Choi, Yoo-Joo;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.17 no.4
    • /
    • pp.35-42
    • /
    • 2016
  • Recently many researches have been focused on the use of the markerless augmented reality system using face, foot, and hand of user's body to alleviate many disadvantages of the marker based augmented reality system. In addition, most existing augmented reality systems have been utilized rigid objects since they just desire to insert and to basic interaction with virtual object in the augmented reality system. In this paper, unlike restricted marker based augmented reality system with rigid objects that is based in display, we designed and implemented the markerless augmented reality system using deformable objects to apply various fields for interactive situations with a user. Generally, deformable objects can be implemented with mass-spring modeling and the finite element modeling. Mass-spring model can provide a real time simulation and finite element model can achieve more accurate simulation result in physical and mathematical view. In this paper, the proposed markerless augmented reality system utilize the mass-spring model using tetraheadron structure to provide real-time simulation result. To provide plausible simulated interaction result with deformable objects, the proposed method detects and tracks users hand with Kinect SDK and calculates the external force which is applied to the object on hand based on the position change of hand. Based on these force, 4th order Runge-Kutta Integration is applied to compute the next position of the deformable object. In addition, to prevent the generation of excessive external force by hand movement that can provide the natural behavior of deformable object, we set up the threshold value and applied this value when the hand movement is over this threshold. Each experimental test has been repeated 5 times and we analyzed the experimental result based on the computational cost of simulation. We believe that the proposed markerless augmented reality system with deformable objects can overcome the weakness of traditional marker based augmented reality system with rigid object that are not suitable to apply to other various fields including healthcare and education area.

Development of a prototype simulator for dental education (치의학 교육을 위한 프로토타입 시뮬레이터의 개발)

  • Mi-El Kim;Jaehoon Sim;Aein Mon;Myung-Joo Kim;Young-Seok Park;Ho-Beom Kwon;Jaeheung Park
    • The Journal of Korean Academy of Prosthodontics
    • /
    • v.61 no.4
    • /
    • pp.257-267
    • /
    • 2023
  • Purpose. The purpose of the study was to fabricate a prototype robotic simulator for dental education, to test whether it could simulate mandibular movements, and to assess the possibility of the stimulator responding to stimuli during dental practice. Materials and methods. A virtual simulator model was developed based on segmentation of the hard tissues using cone-beam computed tomography (CBCT) data. The simulator frame was 3D printed using polylactic acid (PLA) material, and dentiforms and silicone face skin were also inserted. Servo actuators were used to control the movements of the simulator, and the simulator's response to dental stimuli was created by pressure and water level sensors. A water level test was performed to determine the specific threshold of the water level sensor. The mandibular movements and mandibular range of motion of the simulator were tested through computer simulation and the actual model. Results. The prototype robotic simulator consisted of an operational unit, an upper body with an electric device, a head with a temporomandibular joint (TMJ) and dentiforms. The TMJ of the simulator was capable of driving two degrees of freedom, implementing rotational and translational movements. In the water level test, the specific threshold of the water level sensor was 10.35 ml. The mandibular range of motion of the simulator was 50 mm in both computer simulation and the actual model. Conclusion. Although further advancements are still required to improve its efficiency and stability, the upper-body prototype simulator has the potential to be useful in dental practice education.

A Study on Fast Iris Detection for Iris Recognition in Mobile Phone (휴대폰에서의 홍채인식을 위한 고속 홍채검출에 관한 연구)

  • Park Hyun-Ae;Park Kang-Ryoung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.2 s.308
    • /
    • pp.19-29
    • /
    • 2006
  • As the security of personal information is becoming more important in mobile phones, we are starting to apply iris recognition technology to these devices. In conventional iris recognition, magnified iris images are required. For that, it has been necessary to use large magnified zoom & focus lens camera to capture images, but due to the requirement about low size and cost of mobile phones, the zoom & focus lens are difficult to be used. However, with rapid developments and multimedia convergence trends in mobile phones, more and more companies have built mega-pixel cameras into their mobile phones. These devices make it possible to capture a magnified iris image without zoom & focus lens. Although facial images are captured far away from the user using a mega-pixel camera, the captured iris region possesses sufficient pixel information for iris recognition. However, in this case, the eye region should be detected for accurate iris recognition in facial images. So, we propose a new fast iris detection method, which is appropriate for mobile phones based on corneal specular reflection. To detect specular reflection robustly, we propose the theoretical background of estimating the size and brightness of specular reflection based on eye, camera and illuminator models. In addition, we use the successive On/Off scheme of the illuminator to detect the optical/motion blurring and sunlight effect on input image. Experimental results show that total processing time(detecting iris region) is on average 65ms on a Samsung SCH-S2300 (with 150MHz ARM 9 CPU) mobile phone. The rate of correct iris detection is 99% (about indoor images) and 98.5% (about outdoor images).

A Development of Framework for Selecting Labor Attendance Management System Considering Condition of Construction Site (건설 현장 특성을 고려한 출역관리시스템 선정 프레임워크 개발)

  • Kim, Seong-Ah;Chin, Sang-Yoon;Jang, Moon-Seok;Jung, Choong-Won;Choi, Cheol-Ho
    • Korean Journal of Construction Engineering and Management
    • /
    • v.16 no.4
    • /
    • pp.60-69
    • /
    • 2015
  • Labor attendance management has traditionally been carried out by writing a table for checking an attendance of labor, which requires a lot of time and effort. As electronic devices with additions such as barcodes, Quick Response codes, and Radio Frequency Identification(RFID) have been developed, however, automated labor attendance management systems have appeared. Recently, various types of labor recognition devices converged with biometrics (fingerprint, vein, face recognition, etc.) have been released. However, although these devices can be used to check attendance automatically, there is insufficient guidance when it comes to selecting the appropriate labor attendance management system for construction sites. Therefore, this study proposed a decision framework to determine which labor attendance management system would be suitable for a construction site and to select the labor recognition device. This study investigated different labor recognition devices, focusing on how they worked, and tested the performance of devices and their usability for construction labor attendance management. The test results showed that RFID is most suitable when verifying the attendance of many laborers over a short period of time. The devices for hand vein and fingerprint recognition did not function when there was a foreign material such as cement or paint on the laborer's hand, except for a deformed finger. Reflecting these test results, this study suggested a framework for selecting a labor attendance system and recognition device; this is expected to contribute to the development of more efficient labor management systems.

Autonomous Mobile Robot System Using Adaptive Spatial Coordinates Detection Scheme based on Stereo Camera (스테레오 카메라 기반의 적응적인 공간좌표 검출 기법을 이용한 자율 이동로봇 시스템)

  • Ko Jung-Hwan;Kim Sung-Il;Kim Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.1C
    • /
    • pp.26-35
    • /
    • 2006
  • In this paper, an automatic mobile robot system for a intelligent path planning using the detection scheme of the spatial coordinates based on stereo camera is proposed. In the proposed system, face area of a moving person is detected from a left image among the stereo image pairs by using the YCbCr color model and its center coordinates are computed by using the centroid method and then using these data, the stereo camera embedded on the mobile robot can be controlled for tracking the moving target in real-time. Moreover, using the disparity map obtained from the left and right images captured by the tracking-controlled stereo camera system and the perspective transformation between a 3-D scene and an image plane, depth information can be detected. Finally, based-on the analysis of these calculated coordinates, a mobile robot system is derived as a intelligent path planning and a estimation. From some experiments on robot driving with 240 frames of the stereo images, it is analyzed that error ratio between the calculated and measured values of the distance between the mobile robot and the objects, and relative distance between the other objects is found to be very low value of $2.19\%$ and $1.52\%$ on average, respectably.

Active Object Tracking System based on Stereo Vision (스테레오 비젼 기반의 능동형 물체 추적 시스템)

  • Ko, Jung-Hwan
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.4
    • /
    • pp.159-166
    • /
    • 2016
  • In this paper, an active object tracking system basing on the pan/tilt-embedded stereo camera system is suggested and implemented. In the proposed system, once the face area of a target is detected from the input stereo image by using a YCbCr color model and phase-type correlation scheme and then, using this data as well as the geometric information of the tracking system, the distance and 3D information of the target are effectively extracted in real-time. Basing on these extracted data the pan/tilted-embedded stereo camera system is adaptively controlled and as a result, the proposed system can track the target adaptively under the various circumstance of the target. From some experiments using 480 frames of the test input stereo image, it is analyzed that a standard variation between the measured and computed the estimated target's height and an error ratio between the measured and computed 3D coordinate values of the target is also kept to be very low value of 1.03 and 1.18% on average, respectively. From these good experimental results a possibility of implementing a new real-time intelligent stereo target tracking and surveillance system using the proposed scheme is finally suggested.

A New Height Estimation Scheme Using Geometric Information of Stereo Camera based on Pan/tilt control (팬/틸트 제어기반의 스데레오 카메라의 기하학적 정보를 이용한 새로운 높이 추정기법)

  • Ko Jung-Hwan;Kim Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.2C
    • /
    • pp.156-165
    • /
    • 2006
  • In this paper, a new intelligent moving target tracking and surveillance system basing on the pan/tilt-embedded stereo camera system is suggested and implemented. In the proposed system, once the face area of a target is detected from the input stereo image by using a YCbCr color model and phase-type correlation scheme and then, using this data as well as the geometric information of the tracking system, the distance and 3D information of the target are effectively extracted in real-time. Basing on these extracted data the pan/tilted-imbedded stereo camera system is adaptively controlled and as a result, the proposed system can track the target adaptively under the various circumstance of the target. From some experiments using 480 frames of the test input stereo image, it is analyzed that a standard variation between the measured and computed the estimated target's height and an error ratio between the measured and computed 3D coordinate values of the target is also kept to be very low value of 1.03 and 1.18$\%$ on average, respectively. From these good experimental results a possibility of implementing a new real-time intelligent stereo target tracking and surveillance system using the proposed scheme is finally suggested.

Artificial Intelligence for Assistance of Facial Expression Practice Using Emotion Classification (감정 분류를 이용한 표정 연습 보조 인공지능)

  • Dong-Kyu, Kim;So Hwa, Lee;Jae Hwan, Bong
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.6
    • /
    • pp.1137-1144
    • /
    • 2022
  • In this study, an artificial intelligence(AI) was developed to help with facial expression practice in order to express emotions. The developed AI used multimodal inputs consisting of sentences and facial images for deep neural networks (DNNs). The DNNs calculated similarities between the emotions predicted by the sentences and the emotions predicted by facial images. The user practiced facial expressions based on the situation given by sentences, and the AI provided the user with numerical feedback based on the similarity between the emotion predicted by sentence and the emotion predicted by facial expression. ResNet34 structure was trained on FER2013 public data to predict emotions from facial images. To predict emotions in sentences, KoBERT model was trained in transfer learning manner using the conversational speech dataset for emotion classification opened to the public by AIHub. The DNN that predicts emotions from the facial images demonstrated 65% accuracy, which is comparable to human emotional classification ability. The DNN that predicts emotions from the sentences achieved 90% accuracy. The performance of the developed AI was evaluated through experiments with changing facial expressions in which an ordinary person was participated.

Gaze and Gender Dynamics of Portrait of a Lady on Fire (<타오르는 여인의 초상>의 시선과 젠더의 동학)

  • Kwon, Eunsun
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.3
    • /
    • pp.399-404
    • /
    • 2022
  • Céline Sciamma is a leading female director of her time. She is constantly making works on the subject of women. It can be said that it has pioneered a 'female narrative'. In particular, it shows a unique perspective in dealing with the complexities surrounding the formation of a girl's sexual identity and the influence of peer culture on individuals. In Portrait of a Lady on Fire, 'The Handmaiden', 'The Painter', and 'The Maid' show the solidarity of women without hierarchies in a liberated space where the patriarch is absent, showing a feminist interpretation of classics and a self-reflecting consciousness of women's historical records. Based on the setting of drawing, this film explores the genderization of gaze, gaze and objectification, issues of seeing and power, etc. constructed throughout the history of visual arts. The portrait work of Heloise and Marianne in the film deconstructs the formula of 'a male painter as a viewer and a female model who gazes at herself through the male gaze' and turns the gaze into an interactive process of giving and receiving. The process of exchanging gazes is sealed with a nude painting in which Marianne's face is superimposed on Heloise's body. This overturns the nude painting as the dominant form of engraving the male audience's position in the image through the stylization of the viewing method in the history of Western painting with the erotic gaze of a lesbian in an instant.