• 제목/요약/키워드: Image to Speech

검색결과 188건 처리시간 0.026초

대학수업에서 교수의 이미지메이킹이 학습자의 수업만족 및 수업몰입에 미치는 영향 (The effect of professor's image-making on college student's class satisfaction and class commitment)

  • 정혜림;박선주
    • 한국의상디자인학회지
    • /
    • 제23권3호
    • /
    • pp.73-85
    • /
    • 2021
  • The purpose of this study is to understand the influence of the professor's image making (internal, external, social image) perceived by college students on instructional outcomes. The influence of the professor's image making on class satisfaction and class commitment was analyzed, and the mediating effect of class satisfaction and the relationship between class satisfaction and class commitment in the relationship between image making and class commitment was considered. First, it was found that the external image and social image of the professor had a significant effect on class satisfaction. The level of interpersonal relations, such as communication, manners, manners, and intimacy as well as the management of external expressions, clothing style, makeup, hair, gestures, postures, attitudes, voices, speech, and speech speed brings satisfaction to the class. Second, it was found that the professor's inner image, outer image, and social image had a significant effect on class commitment. In order to satisfy the students' immersion in class, professors are required to manage internal, external, and social images. Third, it was found that class satisfaction had a significant effect on class commitment. If the class satisfaction is high, it means that class immersion also increases. Fourth, as for the social image of a professor, it was found that class satisfaction had a completely mediating effect in the relationship between class commitment, and the external image of a professor was found to have a partial mediating effect in class satisfaction in the relationship between class commitment. It was found that the social image of professors perceived by college students improve class satisfaction, and this improves class satisfaction further enhances class immersion.

Robust Lip Extraction and Tracking of the Mouth Region

  • Min, Duk-Soo;Kim, Jin-Young;Park, Seung-Ho;Kim, Ki-Jung
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2000년도 ITC-CSCC -2
    • /
    • pp.927-930
    • /
    • 2000
  • Visual features of lip area play an important role in the visual speech information. We are concerned about correct lip area as region of interest (ROI). In this paper, we propose a robust and fast method for locating the mouth corners. Also, we define a region of interest at mouth during speech. A method, which we have used, only uses the horizontal and vertical image operators at mouth area. This searching is performed by fitting the ROI-template to image with illumination control. Most of the lip extraction algorithms are dependent on luminosity of image. We just used the binary image where the variable threshold is applied. The variable threshold varies to illumination condition. In order to control those variations, the gray-tone is converted to binary image by threshold, which is obtained through Multiple Linear Regression Analysis (MLRA) about divided 2D special region. Thus we obtained the region of interest at mouth area, which is the robust extraction about illumination. A region of interest is automatically extracted.

  • PDF

Interactive Rehabilitation Support System for Dementia Patients

  • Kim, Sung-Ill
    • 융합신호처리학회논문지
    • /
    • 제11권3호
    • /
    • pp.221-225
    • /
    • 2010
  • This paper presents the preliminary study of an interactive rehabilitation support system for both dementia patients and their caregivers, the goal of which is to improve the quality of life(QOL) of the patients suffering from dementia through virtual interaction. To achieve the virtual interaction, three kinds of recognition modules for speech, facial image and pen-mouse gesture are studied. The results of both practical tests and questionnaire surveys show that the proposed system had to be further improved, especially in both speech recognition and user interface for real-world applications. The surveys also revealed that the pen-mouse gesture recognition, as one of possible interactive aids, show us a probability to support weakness of speech recognition.

Implementation of Extracting Specific Information by Sniffing Voice Packet in VoIP

  • Lee, Dong-Geon;Choi, WoongChul
    • International journal of advanced smart convergence
    • /
    • 제9권4호
    • /
    • pp.209-214
    • /
    • 2020
  • VoIP technology has been widely used for exchanging voice or image data through IP networks. VoIP technology, often called Internet Telephony, sends and receives voice data over the RTP protocol during the session. However, there is an exposition risk in the voice data in VoIP using the RTP protocol, where the RTP protocol does not have a specification for encryption of the original data. We implement programs that can extract meaningful information from the user's dialogue. The meaningful information means the information that the program user wants to obtain. In order to do that, our implementation has two parts. One is the client part, which inputs the keyword of the information that the user wants to obtain, and the other is the server part, which sniffs and performs the speech recognition process. We use the Google Speech API from Google Cloud, which uses machine learning in the speech recognition process. Finally, we discuss the usability and the limitations of the implementation with the example.

YCbCr 농도 대비를 이용한 입술특징 추출 (Lip Feature Extraction using Contrast of YCbCr)

  • 김우성;민경원;고한석
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2006년도 하계종합학술대회
    • /
    • pp.259-260
    • /
    • 2006
  • Since audio speech recognition is affected by noise in real environment, visual speech recognition is used to support speech recognition. For the visual speech recognition, this paper suggests the extraction of lip-feature using two types of image segmentation and reduced ASM. Input images are transformed to YCbCr based images and lips are segmented using the contrast of Y/Cb/Cr between lip and face. Subsequently, lip-shape model trained by PCA is placed on segmented lip region and then lip features are extracted using ASM.

  • PDF

이미지 사전과 동사기반 문장 생성 규칙을 활용한 보완대체 의사소통 시스템 구현 (Implementation of Augmentative and Alternative Communication System Using Image Dictionary and Verbal based Sentence Generation Rule)

  • 류제;한광록
    • 정보처리학회논문지B
    • /
    • 제13B권5호
    • /
    • pp.569-578
    • /
    • 2006
  • 본 논문에서는 언어장애인이 쉽게 인식할 수 있는 이미지들을 이용한 보완대체 의사소통 시스템의 구현에 관하여 연구하였다. 특히 보완대체 의사소통 도구의 휴대성 및 이동성과 보다 유연한 형태의 의사소통 시스템 구현에 초점을 맞추었다. 이동성과 휴대성을 위하여 PDA와 같은 모바일 기기에서 운용될 수 있는 시스템을 구현하여 사용 장소의 제약에서 벗어나 여러 장소에서 일반인과 다름없는 의사소통을 할 수 있도록 하였으며, 용량이 큰 이미지 데이터의 저장 공간 한계를 극복하기 위하여 유선 또는 무선 인터넷 환경에서 클라이언트/서버 형태의 보완대체 의사소통 시스템을 설계하였다. 또한 사용자의 원활한 의사소통이 가능하도록 동사를 기준으로 하여 동사에 대응하는 명사들을 하위 범주화하여 이미지 사전을 구축하였다. 이를 위하여 문장을 구성하는데 가장 중요한 역할을 하는 품사인 동사에 초점을 맞추어 동사의 유형에 따라 생성되는 문장의 유형을 정규화 하였다.

라이프케어를 위한 이미지메이킹 프로그램 교육의 필요성과 활성화 방안 (A Study on the Educational Necessity and Activation Plan of Image Making Program for Life Care)

  • 윤희
    • 한국엔터테인먼트산업학회논문지
    • /
    • 제14권7호
    • /
    • pp.429-437
    • /
    • 2020
  • 본 연구는 국내 전문대학의 이미지메이킹 프로그램에 관련된 교과과정의 필요성과 교육내용 및 활성화 방안에 대한 실증분석을 통하여 취업을 앞둔 대학생들의 면접지도와 예비 직장인에게 있어서 대인 관계능력 향상을 위한 방안으로 대학 교과과정의 이미지메이킹 프로그램 교육 개발을 위한 기초자료를 제공하고자 시도되었고, 이미지메이킹 교과목을 수강한 광주·전남지역 전문대학교 재학생 400명을 대상으로 설문조사를 실시하였다. 수집된 자료는 SPSS 21.0 프로그램을 사용하여 분석하였다. 연구 결과는 다음과 같다. 첫째, 졸업 후 취업과 예비 직장인의 이미지 관리를 위해 이미지메이킹 프로그램 교과과정이 필요하고, 이미지를 객관적으로 파악하기 위해 다른 사람의 도움이 필요하다고 나타났다. 이미지메이킹 프로그램을 위한 적합한 강의실은 일반 강의실, 시청각 및 멀티미디어실, 전용 강의실 순으로 나타났다. 둘째, 이미지메이킹 프로그램 교육내용의 중요도는 태도(행동), 예절 및 인사법, 표정, 화법, 대인 관계법, 의복, 헤어스타일, 메이크업 순으로 나타났고, 교육내용의 중요 요소는 표정, 메이크업, 헤어스타일, 태도(행동), 대인 관계법, 화법, 의복, 예절 및 인사법 순으로 나타났다. 셋째, 이미지메이킹 프로그램 교육의 영향력 정도는 취업에 미치는 영향력, 대인관계에 미치는 영향력, 생활에 미치는 영향력 순으로 나타났다. 넷째, 이미지메이킹 프로그램 교육 활성화 방안을 분석한 결과 적절한 이미지메이킹 프로그램 교육시점은 2학년부터이고, 교육시간은 한 학기 주 1회, 교과과정 구분은 교양이나 교양선택 과정을 희망하는 것으로 볼 수 있다. 이미지메이킹 프로그램 관련 교과내용은 예절 및 인사법, 메이크업 및 코디 방법, 취업 설명회, 기능 자격 취득을 위한 교육, 국내 기업체 연수 순으로 나타나 예절 및 인사법을 가장 희망하는 것으로 볼 수 있다. 그리고, 이미지메이킹 프로그램 관련 지도자는 전공교수를 희망하고, 교육방법은 시청각 및 멀티미디어를 활용한 매체를 희망한다. 이미지메이킹 프로그램 관련 교육은 화법이나 목소리, 커뮤니케이션에 관해 분석하는 교육, 어울리는 헤어스타일 및 메이크업을 분석하고 실습할 수 있는 교육, 기업의 면접에 관한 교육, 워킹이나 자세교정에 관한 교육 순으로 나타났다. 이상의 결과 본 연구는 이미지메이킹 프로그램 교과과정의 활용방안에 대한 기초자료를 제공하였다.

디지탈 내시경 데이터 management system의 개발 (Development of Digital Endoscopic Data Management System)

  • 송철규;이상민;이영묵;김원기
    • 대한의용생체공학회:학술대회논문집
    • /
    • 대한의용생체공학회 1996년도 추계학술대회
    • /
    • pp.304-306
    • /
    • 1996
  • Endoscopy has become a crucial diagnostic and theraputic procedure in clinical areas. Over the past three years, we have developed a computerized system to record and store clinical data pertaining to endoscopic surgery of laparascopic cholesystectomy, peviscopic endometriosis, and surgical arthroscopy. In this study, we are developed computer system, which is composed of frame grabber, sound board, VCR control board, LAN card and EDMS(endoscopic data management software). Also, computer system has controled over peripheral instruments as a color video printer, video cassette recorder, and endoscopic input/output signals(image and doctor's speech). Also, we are developed one body system of camels control unit including an endoscopic miniature camera and light source. Our system offer unsurpassed image quality in terms of resolution and color fidelity. Digital endoscopic data management system is based on open architecture and a set of widely available industry standards, namely: windows 3.1 as a operating system, TCP/IP as a network protocol and a time sequence based database that handles both an image and drctor's speech synchronized with endoscopic image.

  • PDF

Pathological Vibratory patterns of the Vocal Folds Observed by the High Speed Digital Imaging System

  • Niimi, Seiji
    • 대한음성언어의학회:학술대회논문집
    • /
    • 대한음성언어의학회 1998년도 제10회 학술대회 심포지움
    • /
    • pp.208-209
    • /
    • 1998
  • It is generally known that many cases of pathological rough voice are characterized not by simple random perturbations but by quasi-periodic perturbations in the speech wave. However, there are few studies on the characteristics of perturbations in vocal fold vibrations associated with this type of voice. We have been conducting studies of pathological vocal fold vibration using a high-speed digital image recording system developed by our institute, Compared to the ordinary high-speed-motion picture system, the present system is compact and simple to operate and thus, it suited for pathological data collection. (omitted)

  • PDF

RNN을 이용한 Expressive Talking Head from Speech의 합성 (Synthesis of Expressive Talking Heads from Speech with Recurrent Neural Network)

  • 사쿠라이 류헤이;심바 타이키;야마조에 히로타케;이주호
    • 로봇학회논문지
    • /
    • 제13권1호
    • /
    • pp.16-25
    • /
    • 2018
  • The talking head (TH) indicates an utterance face animation generated based on text and voice input. In this paper, we propose the generation method of TH with facial expression and intonation by speech input only. The problem of generating TH from speech can be regarded as a regression problem from the acoustic feature sequence to the facial code sequence which is a low dimensional vector representation that can efficiently encode and decode a face image. This regression was modeled by bidirectional RNN and trained by using SAVEE database of the front utterance face animation database as training data. The proposed method is able to generate TH with facial expression and intonation TH by using acoustic features such as MFCC, dynamic elements of MFCC, energy, and F0. According to the experiments, the configuration of the BLSTM layer of the first and second layers of bidirectional RNN was able to predict the face code best. For the evaluation, a questionnaire survey was conducted for 62 persons who watched TH animations, generated by the proposed method and the previous method. As a result, 77% of the respondents answered that the proposed method generated TH, which matches well with the speech.