• Title/Summary/Keyword: facial video

Search Result 126, Processing Time 0.025 seconds

Using Ensemble Learning Algorithm and AI Facial Expression Recognition, Healing Service Tailored to User's Emotion (앙상블 학습 알고리즘과 인공지능 표정 인식 기술을 활용한 사용자 감정 맞춤 힐링 서비스)

  • Yang, seong-yeon;Hong, Dahye;Moon, Jaehyun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.11a
    • /
    • pp.818-820
    • /
    • 2022
  • The keyword 'healing' is essential to the competitive society and culture of Koreans. In addition, as the time at home increases due to COVID-19, the demand for indoor healing services has increased. Therefore, this thesis analyzes the user's facial expression so that people can receive various 'customized' healing services indoors, and based on this, provides lighting, ASMR, video recommendation service, and facial expression recording service.The user's expression was analyzed by applying the ensemble algorithm to the expression prediction results of various CNN models after extracting only the face through object detection from the image taken by the user.

A Study on Facial Expression Acting in Genre Drama - with Focus on K-Drama Voice2 - (장르 드라마에서의 표정연기연구 - 드라마 '보이스2'를 중심으로 -)

  • Oh, Youn-Hong
    • Journal of Korea Entertainment Industry Association
    • /
    • v.13 no.8
    • /
    • pp.313-323
    • /
    • 2019
  • For the actors on video, facial expression acting can easily become 'forced facial expression' or 'over-acting'. Also, if self-restraint is emphasized too much, then it becomes 'flat acting' with insufficient emotions. By bringing forth questions in regard to such facial expression acting methods, this study analyzed the facial expression acting of the actors in genre dramas with strong commercial aspects. In conclusion, the facial expression acting methods of the actors in genre dramas were being conducted in a typical way. This means that in visual conventions of video acting, the aesthetic standard has become the important standard in the facial expression acting of the actors. In genre dramas, the emotions of the characters are often revealed in close-up shots. Within the close-up shot, the most important expressive medium in a 'zoomed-in face' is the 'pupil of the eye', and emotions are mostly expressed through the movements of the eye and muscles around it. The second most important expressive medium is the 'mouth'. The differences in the degree of opening and closing the mouth convey diverse emotions along with the expression of the 'eye'. In addition, tensions in the facial muscles greatly hinder the expression of emotions, and the movement of facial muscles must be minimized to prevent excessive wrinkles from forming on the surface of the face. Facial expressions are not completed just with the movement of the muscles. Ultimately, the movement of the muscle is the result of emotions. Facial expression acting takes place after having emotional feelings. For this, the actor needs to go through the process of 'personalization' of a character, such as 'emotional memory', 'concentration' and 'relaxation' which are psychological acting techniques of Stanislavsky. Also, the characteristics of close-up shots that visually reveal the 'inner world' should be recognized. In addition, it was discovered that the facial expression acting is the reaction acting that provides the important point in the unfolding of narratives, and that the method of facial expression and the size of the shots required for the actors are different depending on the roles of main and supporting characters.

Comparison Analysis of Four Face Swapping Models for Interactive Media Platform COX (인터랙티브 미디어 플랫폼 콕스에 제공될 4가지 얼굴 변형 기술의 비교분석)

  • Jeon, Ho-Beom;Ko, Hyun-kwan;Lee, Seon-Gyeong;Song, Bok-Deuk;Kim, Chae-Kyu;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.5
    • /
    • pp.535-546
    • /
    • 2019
  • Recently, there have been a lot of researches on the whole face replacement system, but it is not easy to obtain stable results due to various attitudes, angles and facial diversity. To produce a natural synthesis result when replacing the face shown in the video image, technologies such as face area detection, feature extraction, face alignment, face area segmentation, 3D attitude adjustment and facial transposition should all operate at a precise level. And each technology must be able to be interdependently combined. The results of our analysis show that the difficulty of implementing the technology and contribution to the system in facial replacement technology has increased in facial feature point extraction and facial alignment technology. On the other hand, the difficulty of the facial transposition technique and the three-dimensional posture adjustment technique were low, but showed the need for development. In this paper, we propose four facial replacement models such as 2-D Faceswap, OpenPose, Deekfake, and Cycle GAN, which are suitable for the Cox platform. These models have the following features; i.e. these models include a suitable model for front face pose image conversion, face pose image with active body movement, and face movement with right and left side by 15 degrees, Generative Adversarial Network.

Development of a Web-based Presentation Attitude Correction Program Centered on Analyzing Facial Features of Videos through Coordinate Calculation (좌표계산을 통해 동영상의 안면 특징점 분석을 중심으로 한 웹 기반 발표 태도 교정 프로그램 개발)

  • Kwon, Kihyeon;An, Suho;Park, Chan Jung
    • The Journal of the Korea Contents Association
    • /
    • v.22 no.2
    • /
    • pp.10-21
    • /
    • 2022
  • In order to improve formal presentation attitudes such as presentation of job interviews and presentation of project results at the company, there are few automated methods other than observation by colleagues or professors. In previous studies, it was reported that the speaker's stable speech and gaze processing affect the delivery power in the presentation. Also, there are studies that show that proper feedback on one's presentation has the effect of increasing the presenter's ability to present. In this paper, considering the positive aspects of correction, we developed a program that intelligently corrects the wrong presentation habits and attitudes of college students through facial analysis of videos and analyzed the proposed program's performance. The proposed program was developed through web-based verification of the use of redundant words and facial recognition and textualization of the presentation contents. To this end, an artificial intelligence model for classification was developed, and after extracting the video object, facial feature points were recognized based on the coordinates. Then, using 4000 facial data, the performance of the algorithm in this paper was compared and analyzed with the case of facial recognition using a Teachable Machine. Use the program to help presenters by correcting their presentation attitude.

Hardware Implementation for Stabilization of Detected Face Area (검출된 얼굴 영역 안정화를 위한 하드웨어 구현)

  • Cho, Ho-Sang;Jang, Kyoung-Hoon;Kang, Hyun-Jung;Kang, Bong-Soon
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.13 no.2
    • /
    • pp.77-82
    • /
    • 2012
  • This paper presents a hardware-implemented face regions stabilization algorithm that stabilizes facial regions using the locations and sizes of human faces found by a face detection system. Face detection algorithms extract facial features or patterns determining the presence of a face from a video source and detect faces via a classifier trained on example faces. But face detection results has big variations in the detected locations and sizes of faces by slight shaking. To address this problem, the high frequency reduce filter that reduces variations in the detected face regions by taking into account the face range information between the current and previous video frames are implemented in addition to center distance comparison and zooming operations.

Audio and Video Bimodal Emotion Recognition in Social Networks Based on Improved AlexNet Network and Attention Mechanism

  • Liu, Min;Tang, Jun
    • Journal of Information Processing Systems
    • /
    • v.17 no.4
    • /
    • pp.754-771
    • /
    • 2021
  • In the task of continuous dimension emotion recognition, the parts that highlight the emotional expression are not the same in each mode, and the influences of different modes on the emotional state is also different. Therefore, this paper studies the fusion of the two most important modes in emotional recognition (voice and visual expression), and proposes a two-mode dual-modal emotion recognition method combined with the attention mechanism of the improved AlexNet network. After a simple preprocessing of the audio signal and the video signal, respectively, the first step is to use the prior knowledge to realize the extraction of audio characteristics. Then, facial expression features are extracted by the improved AlexNet network. Finally, the multimodal attention mechanism is used to fuse facial expression features and audio features, and the improved loss function is used to optimize the modal missing problem, so as to improve the robustness of the model and the performance of emotion recognition. The experimental results show that the concordance coefficient of the proposed model in the two dimensions of arousal and valence (concordance correlation coefficient) were 0.729 and 0.718, respectively, which are superior to several comparative algorithms.

A Hybrid Approach of Efficient Facial Feature Detection and Tracking for Real-time Face Direction Estimation (실시간 얼굴 방향성 추정을 위한 효율적인 얼굴 특성 검출과 추적의 결합방법)

  • Kim, Woonggi;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.117-124
    • /
    • 2013
  • In this paper, we present a new method which efficiently estimates a face direction from a sequences of input video images in real time fashion. For this work, the proposed method performs detecting the facial region and major facial features such as both eyes, nose and mouth by using the Haar-like feature, which is relatively not sensitive against light variation, from the detected facial area. Then, it becomes able to track the feature points from every frame using optical flow in real time fashion, and determine the direction of the face based on the feature points tracked. Further, in order to prevent the erroneously recognizing the false positions of the facial features when if the coordinates of the features are lost during the tracking by using optical flow, the proposed method determines the validity of locations of the facial features using the template matching of detected facial features in real time. Depending on the correlation rate of re-considering the detection of the features by the template matching, the face direction estimation process is divided into detecting the facial features again or tracking features while determining the direction of the face. The template matching initially saves the location information of 4 facial features such as the left and right eye, the end of nose and mouse in facial feature detection phase and reevaluated these information when the similarity measure between the stored information and the traced facial information by optical flow is exceed a certain level of threshold by detecting the new facial features from the input image. The proposed approach automatically combines the phase of detecting facial features and the phase of tracking features reciprocally and enables to estimate face pose stably in a real-time fashion. From the experiment, we can prove that the proposed method efficiently estimates face direction.

A PERSONAL AUTHENTICATION FROM VIDEO USING HANDHELD CAMERA BY PARAMETRIC EIGENSPACE METHOD

  • Morizumi, Yusuke;Matsuo, Kenji;Kubota, Akira;Hatori, Yoshinori
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.628-631
    • /
    • 2009
  • In this paper, we proposed a new authentication method using video that was taken during moving a hand-held camera in front of the face. The proposed method extracted individuality from the obtained image sequences using the parametric eigenspace scheme. Changes of facial appearance through authentication trials draw continuous tracks in the low dimensional igenspace. The similarity between their continuous tracks are calculated by DP-matching to verify their identities. Experimental results confirmed that different motions and persons change the shapes of continuous tracks, so the proposed method could identify the person.

  • PDF

A Study of Multiple Video Players as Media Art -Focusing on (다중 동영상 플레이어를 이용한 미디어아트로써의 확장성 연구 -작품 <희로애락(喜怒哀樂)>을 중심으로)

  • Na, Jung-Jo;Lee, Jee-Hee;Song, Pil-Jae;Lee, Myung-Hak;Fan, Bai;Kim, Kyu-Jung
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02b
    • /
    • pp.115-120
    • /
    • 2008
  • It provides technical analyze and conceptual idea by oneself through Media Art One's feeling expresses multiple video players by using 3 dimensional graphic techniques Multiple video players work in interlocking. Spectators join in clicking mouse on the screen and the working of multiplevideo players realize with computer software. It has a chance, which spectators get a sensitive stimulation. Each player express human's facial expression - happiness, anger, sadness, excitement. Players are able to accelerateaccording to computer system. It seems that artistic messages send through cooperation and connection of multiple players.

  • PDF

Korean Emotional Speech and Facial Expression Database for Emotional Audio-Visual Speech Generation (대화 영상 생성을 위한 한국어 감정음성 및 얼굴 표정 데이터베이스)

  • Baek, Ji-Young;Kim, Sera;Lee, Seok-Pil
    • Journal of Internet Computing and Services
    • /
    • v.23 no.2
    • /
    • pp.71-77
    • /
    • 2022
  • In this paper, a database is collected for extending the speech synthesis model to a model that synthesizes speech according to emotions and generating facial expressions. The database is divided into male and female data, and consists of emotional speech and facial expressions. Two professional actors of different genders speak sentences in Korean. Sentences are divided into four emotions: happiness, sadness, anger, and neutrality. Each actor plays about 3300 sentences per emotion. A total of 26468 sentences collected by filming this are not overlap and contain expression similar to the corresponding emotion. Since building a high-quality database is important for the performance of future research, the database is assessed on emotional category, intensity, and genuineness. In order to find out the accuracy according to the modality of data, the database is divided into audio-video data, audio data, and video data.