• Title/Summary/Keyword: One-syllable

Search Result 148, Processing Time 0.027 seconds

Synchronizationof Synthetic Facial Image Sequences and Synthetic Speech for Virtual Reality (가상현실을 위한 합성얼굴 동영상과 합성음성의 동기구현)

  • 최장석;이기영
    • Journal of the Korean Institute of Telematics and Electronics S
    • /
    • v.35S no.7
    • /
    • pp.95-102
    • /
    • 1998
  • This paper proposes a synchronization method of synthetic facial iamge sequences and synthetic speech. The LP-PSOLA synthesizes the speech for each demi-syllable. We provide the 3,040 demi-syllables for unlimited synthesis of the Korean speech. For synthesis of the Facial image sequences, the paper defines the total 11 fundermental patterns for the lip shapes of the Korean consonants and vowels. The fundermental lip shapes allow us to pronounce all Korean sentences. Image synthesis method assigns the fundermental lip shapes to the key frames according to the initial, the middle and the final sound of each syllable in korean input text. The method interpolates the naturally changing lip shapes in inbetween frames. The number of the inbetween frames is estimated from the duration time of each syllable of the synthetic speech. The estimation accomplishes synchronization of the facial image sequences and speech. In speech synthesis, disk memory is required to store 3,040 demi-syllable. In synthesis of the facial image sequences, however, the disk memory is required to store only one image, because all frames are synthesized from the neutral face. Above method realizes synchronization of system which can real the Korean sentences with the synthetic speech and the synthetic facial iage sequences.

  • PDF

Study on Extraction of Headwords for Compilation of 「Donguibogam Dictionary」 - Based on Corpus-based Analysis - (『동의보감사전』 편찬을 위한 표제어 추출에 관한 연구 - 코퍼스 분석방법을 바탕으로 -)

  • Jung, Ji-Hun;Kim, Do-Hoon;Kim, Dong-Ryul
    • The Journal of Korean Medical History
    • /
    • v.29 no.1
    • /
    • pp.47-54
    • /
    • 2016
  • This article attempts to extract headwords for complication of "Donguibogam Dictionary" with Corpus-based Analysis. The computerized original text of Donguibogam is changed into a text file by a program 'EM Editor'. Chinese characters of high frequency of exposure among Chinese characters of Donguibogam are extracted by a Corpus-based analytical program 'AntConc'. Two-syllable, three-syllable, four-syllable, and five-syllable words including each Chinese characters of high frequency are extracted through n-cluster, one of functions of AntConc. Lastly, The output that is meaningful as a word is sorted. As a result, words that often appear in Donguibogam can be sorted in this article, and the names of books, medical herbs, disease symptoms, and prescriptions often appear especially. This way to extract headwords by this Corpus-based Analysis can suggest better headwords list for "Donguibogam Dictionary" in the future.

한국어 리듬의 음성학적 연구

  • LEE H.B.
    • MALSORI
    • /
    • no.4
    • /
    • pp.31-48
    • /
    • 1982
  • This paper describes the rhythmic structure of the Korean standard speech of Seoul in terms of what the writer calls 'Speech Segmentv as the basic unit of the rhythm. The speech segment consists of a 'Nucleus'( a stressed syllable ) with or without one or more weak syllable(s) . The nucleus is always long and the weak syllables are short except the last syllable of the speech rhythm, which may be realized nearly as Long as the nuclear syllable.

  • PDF

An Experimental Phonetic Study on the Rhythm of Daegu and Standard Korean --Focusing on Duration-- (대구말과 표준말 리듬의 실험음성학적 비교연구 --길이(duration)를 중심으로--)

  • Jo Un-Il
    • MALSORI
    • /
    • no.27_28
    • /
    • pp.89-109
    • /
    • 1994
  • This thesis compares the duration aspect of the Daegu tongue with that of standard Korean. In the former study on the rhythm of standard Korean, one of the purposes of the study was to compare it with dialects. This thesis is the first attempt to do that. For this purpose, this thesis proceeds as follows. After Introduction, Chapter 2 surveys the former study. Chapter 3 deals with the materials, method and results of the experiment. Chapter 4 analyzes and interprets the results of the experiment, In Conclusion, the most Prominent fact is that the results of the experiment fall short of Daegu tongue speakers' expectations. The Daegu tongue is generally considered as a 'tone language.' And as Daegu tongue speakers sensitively recognize pitch, they think that they quickly say the syllables between the Pitch stressed syllables, whereas standard Korean speakers say those syllables relatively slowly, But in this experiment, which deals with only duration ignoring Pitch, their assumption is proved to be false.

  • PDF

Pronunciation error types and sentence intelligibility of Korean EFL learners (영어 학습자의 발음 오류 유형과 발화 명료도의 관계 연구)

  • Kim, Hyun-Jin
    • English Language & Literature Teaching
    • /
    • v.10 no.3
    • /
    • pp.159-175
    • /
    • 2004
  • This paper investigated the types of errors on English pronunciation and intelligibility of Korean EFL students, and the relationship between the pronunciation accuracy and intelligibility. Thirty one students were evaluated by six English native speakers in terms of overall intelligibility and accuracy In five areas such as nuclear stress, word stress, syllable structure, consonants and vowels. According to the findings of the study, pronunciation errors were made by the subjects more frequently In word stress than any other area of pronunciation accuracy. The Pearson correlation analysis showed that intelligibility was related with word stress, syllable structure, consonants and vowels, and the stepwise multiple regression analysis indicated that, among the above five areas of pronunciation accuracy, word stress best accounted for the intelligibility of a given sentence. In the conclusion, the importance of teaching pronunciation of in those five areas with a special focus on word stress was emphasized m terms of intelligibility.

  • PDF

A Study on the Speaker Adaptation of a Continuous Speech Recognition using HMM (HMM을 이용한 연속 음성 인식의 화자적응화에 관한 연구)

  • Kim, Sang-Bum;Lee, Young-Jae;Koh, Si-Young;Hur, Kang-In
    • The Journal of the Acoustical Society of Korea
    • /
    • v.15 no.4
    • /
    • pp.5-11
    • /
    • 1996
  • In this study, the method of speaker adaptation for uttered sentence using syllable unit hmm is proposed. Segmentation of syllable unit for sentence is performed automatically by concatenation of syllable unit hmm and viterbi segmentation. Speaker adaptation is performed using MAPE(Maximum A Posteriori Probabillity Estimation) which can adapt any small amount of adaptation speech data and add one sequentially. For newspaper editorial continuous speech, the recognition rates of adaptation of HMM was 71.8% which is approximately 37% improvement over that of unadapted HMM

  • PDF

A Study of Syllable Maximum Repetition Rate for Stimuli, Age and Sex (검사어 및 성별.연령에 따른 음절 최대 반복 속도에 대한 연구)

  • 최홍식;차정민;심현섭
    • Journal of the Korean Society of Laryngology, Phoniatrics and Logopedics
    • /
    • v.12 no.1
    • /
    • pp.55-60
    • /
    • 2001
  • Background and Objectives : Syllable Maximum Repetition Rate(MRR) is ability to repeat rapidly the articulators and is assessed for oromechanism function as one of the MPT. MRR is measured by rate(counts/sec), also simultaneously considered accuracy and consistency. The objective of the present was to examine stimuli effects and age and sex differences for MRR. Materials and Method : This study was participated 60 normal males and females(1 : 1) who were divided into two groups young(<40 years old) and old($\geq$40 years old). Stimuli were $/{P^=}a/,/{t^=}a/,/{k^=}a/,/{P^h}a/,/{t^h}a/,/{k^h}a/,/{P^=}{at^=}{ak^=}a/$ for, manner(tense and aspirated) of articulation, $/{p^h}{at^h}{ak^h}a/,/{t^h}{ap^h}{ak^h}a/$ for the effect of the order of syllable, glide /u-i/ for coordination of lip and tongue, interrupted vowel /i/ for laryngeal function. Results : There were little differences in two age groups and sex and manner of articulation for MRR tasks. The fastest average MRR of the single syllable included in this study was $/{t^=}a/$. significant differences existed between MRR for ${p^h}{at^h}{ak^h}a/ and /{t^h}{ap^h}{ak^h}a/$, which suggested that MRR was affected by the order of the syllables. MRR for interrupted vowel /i/ was about 2 counts/sec slower than average rate of 1 syllables. Conclusion : From these results, the order of the syllables was a crucial variable in MRR. rather than age or sex. There were no difference age, sex and manner of rate. The interrupted vowel repetition rate was slightly slow used and can provide basic information to assess the speech mechanism and can be useful to develop effective stimuli to differentiate the disordered group from normal.

  • PDF

Development of a Lipsync Algorithm Based on Audio-visual Corpus (시청각 코퍼스 기반의 립싱크 알고리듬 개발)

  • 김진영;하영민;이화숙
    • The Journal of the Acoustical Society of Korea
    • /
    • v.20 no.3
    • /
    • pp.63-69
    • /
    • 2001
  • A corpus-based lip sync algorithm for synthesizing natural face animation is proposed in this paper. To get the lip parameters, some marks were attached some marks to the speaker's face, and the marks' positions were extracted with some Image processing methods. Also, the spoken utterances were labeled with HTK and prosodic information (duration, pitch and intensity) were analyzed. An audio-visual corpus was constructed by combining the speech and image information. The basic unit used in our approach is syllable unit. Based on this Audio-visual corpus, lip information represented by mark's positions was synthesized. That is. the best syllable units are selected from the audio-visual corpus and each visual information of selected syllable units are concatenated. There are two processes to obtain the best units. One is to select the N-best candidates for each syllable. The other is to select the best smooth unit sequences, which is done by Viterbi decoding algorithm. For these process, the two distance proposed between syllable units. They are a phonetic environment distance measure and a prosody distance measure. Computer simulation results showed that our proposed algorithm had good performances. Especially, it was shown that pitch and intensity information is also important as like duration information in lip sync.

  • PDF

Segmenting and Classifying Korean Words based on Syllables Using Instance-Based Learning (사례기반 학습을 이용한 음절기반 한국어 단어 분리 및 범주 결정)

  • Kim, Jae-Hoon;Lee, Kong-Joo
    • The KIPS Transactions:PartB
    • /
    • v.10B no.1
    • /
    • pp.47-56
    • /
    • 2003
  • Korean delimits words by white-space like English, but words In Korean Is a little different in structure from those in English. Words in English generally consist of one word, but those in Korean are composed of one word and/or morpheme or more. Because of this difference, a word between white-spaces is called an Eojeol in Korean. We propose a method for segmenting and classifying Korean words and/or morphemes based on syllables using an instance-based learning. In this paper, elements of feature sets for the instance-based learning are one previous syllable, one current syllable, two next syllables, a final consonant of the current syllable, and two previous categories. Our method shows more than 97% of the F-measure of word segmentation using ETRI corpus and KAIST corpus.

Phoneme Frequency of 3 to 8-year-old Korean Children (3세${\sim}$8세 아동의 자유 발화 분석을 바탕으로 한 한국어 말소리의 빈도 관련 정보)

  • Sin, Ji-Yeong
    • Proceedings of the KSPS conference
    • /
    • 2005.04a
    • /
    • pp.15-19
    • /
    • 2005
  • The aim of this study is to provide some information on frequencies of occurrence for units of Korean phonemes and syllables analysing spontaneous speech spoken by 3 to 8-year-old Korean children. 49 Korean Children(7${\sim}$10 children for each age) were employed as subjects for this study. Speech data were recorded and phonemically transcribed. 120 utterances for each child were selected for analysis except one child whose data were only 91 utterances. The data size of the present study were 5,971 utterances, 5,1554 syllables, and 105491 phonemes. Among 19 consonants, /n/ showed highest frequency rate of these four conson ants were over 50% for all age groups. Among 18 vowels, /a/ was the most frequent one and /i/ and / ${\wedge}$ were the second and third respectively. The frequency rate of these four consonants were over 50% for all age groups. Frequently occurring syllable types were a part of grammatical word in most cases. Only 5${\sim}$6% of syllable types covered 50% of speech.

  • PDF