• Title/Summary/Keyword: 음절 수

Search Result 318, Processing Time 0.021 seconds

Authorship Attribution in Korean Using Frequency Profiles (빈도 정보를 이용한 한국어 저자 판별)

  • Han, Na-Rae
    • Korean Journal of Cognitive Science
    • /
    • v.20 no.2
    • /
    • pp.225-241
    • /
    • 2009
  • This paper presents an authorship attribution study in Korean conducted on a corpus of newspaper column texts. Based on the data set consisting of a total of 160 columns written by four columnists of Chosun Daily, the approach utilizes relative frequencies of various lexical units in Korean such as fully inflected words, morphemes, syllables and their bigrams in an attempt to establish authorship of a blind text selected from the set. Among these various lexical units, "the morpheme" is found to be most effective in predicting who among the four potential candidates authored a text, reporting accuracies of over 93%. The results indicate that quantitative and statistical techniques in authorship attribution and computational stylistics can be successfully applied to Korean texts.

  • PDF

Tonal Characteristics Based on Intonation Pattern of the Korean Emotion Words (감정단어 발화 시 억양 패턴을 반영한 멜로디 특성)

  • Yi, Soo Yon;Oh, Jeahyuk;Chong, Hyun Ju
    • Journal of Music and Human Behavior
    • /
    • v.13 no.2
    • /
    • pp.67-83
    • /
    • 2016
  • This study investigated the tonal characteristics in Korean emotion words by analyzing the pitch patterns transformed from word utterance. Participants were 30 women, ages 19-23. Each participant was instructed to talk about their emotional experiences using 4-syllable target words. A total of 180 utterances were analyzed in terms of the frequency of each syllable using the Praat. The data were transformed into meantones based on the semi-tone scale. When emotion words were used in the middle of a sentence, the pitch pattern was transformed to A3-A3-G3-G3 for '즐거워서(joyful)', C4-D4-B3-A3 for '행복해서(happy)', G3-A3-G3-G3 for '억울해서(resentful)', A3-A3-G3-A3 for '불안해서(anxious)', and C4-C4-A3-G3 for '침울해서(frustrated)'. When the emotion words were used at the end of a sentence, the pitch pattern was transformed to G4-G4-F4-F4 for '즐거워요(joyful)', D4-D4-A3-G3 for '행복해요(happy)', G3-G3-G3-A3 and F3-G3-E3-D3 for '억울해요(resentful)', A3-G3-F3-F3 for '불안해요(anxious)', and A3-A3-F3-F3 for '침울해요(frustrated)'. These results indicate the differences in pitch patterns depending on the conveyed emotions and the position of words in a sentence. This study presents the baseline data on the tonal characteristics of emotion words, thereby suggesting how pitch patterns could be utilized when creating a melody during songwriting for emotional expression.

Amplification and Howling Suppression of Telephonic Speech for the Hearing-Impaired Person (난청인을 위한 전화기 음성증폭 및 하울링 억제)

  • Lee, Sang-Min
    • Journal of Biomedical Engineering Research
    • /
    • v.19 no.6
    • /
    • pp.623-629
    • /
    • 1998
  • To provide sufficient sound to the hearing-impaired person(HIP) who have many difficulty in communication with others using general telephone, big amplification is needed. But big amplification can occur howling as a side effect. In this study we developed the new technique of big amplification without howling, manufactured and estimated the new hearing aid telephone. Telephone speech is divided to three frequency band, amplified respectively and fitted to HIP's hearing ability. The telephone speech frequency is monitored by counter in time domain. The counter transfers the sinusoidal sound to rectangular wave using comparator and counts the number of rectangular wave in a certain time period, that is frequency, to monitor the howling. Telephone have microphone and speaker, which are fitted in a rigid structure and frequency band of telephone sound is limited, so howling occurs in the limited frequency band. If the counter notices that howling conditions happen, microprocessor decreases quickly the gain of the related frequency band. The result of test of our new hearing aid telephone showed that we can amplifiy the sound as much as 40dB, which is meaningful level to many HIP, and make HIP increase their perception ability from 20% to 60.8% in 1 syllable test and from 28.9% to 78% in 2 syllable test.

  • PDF

Some notes on the French "e muet" (불어의 "묵음 e (e muet)"에 관한 연구)

  • Lee Jeong-Won
    • MALSORI
    • /
    • no.31_32
    • /
    • pp.173-193
    • /
    • 1996
  • 불어의 "묵음 e(e muet)"에 대한 정의를 내리기는 매우 까다롭다. 불어에서 "e"가 "묵음 e(e muet)"로 불리우는 이유는 "e"가 흔히 탈락되기 때문이다. 현재 "e muet"는 다음 발화체에서 볼 수 있듯이 열린음절에서만 나타난다. "Je/le/re/de/man/de/ce/re/por/ta/ge/." [omitted](나는 그 리포트를 다시 요구한다. : 이 경우 실제 발화시 schwa 삭제 규칙이 적용된다.) 둘째, 접두사에 나타나는 "e muet"는 s의 중자음 앞에서 s가 유성음, [z]로 발음되는 것을 막기 위해 쓰인다. "ressembler[omitted](닮다); ressentir[omitted](느끼다)" 같은 경우, 셋째, 몇몇 낱말의 경우 고어의 철자가 약화되어 "e muet"로 발음이 되고 있다. "monsieur[$m{\partial}sj{\emptyset}$](미스터), faisan[$f{\partial}z{\tilde{a}}$](꿩), faisait[$f{\partial}z{\varepsilon}$]("하다"동사의 3인칭 단수 반과거형)"등. 또 과거 문법학자들은 이를 "여성형의 E"로 불렀는데, 이는 형태론적으로 낱말의 여성형을 남성형과 구분짓기 위해 사용되고 있기 때문이기도 하다. 예를 들어, "$aim{\acute{e}}-aim{\acute{e}}e$"(발음은 둘 다 [${\varepsilon}me$]로 동일하다 : 사랑받다)의 경우. 현대불어의 구어체어서 "e muet"는 어말자음을 발음하기 위해 쓰이고 있다. 예를 들어, "pote[pot](단짝)-pot[po](항아리)". 이러한 "e muet"는 발음상으로 지역적, 개인적 및 문맥적 상황에 따라 그 음색 자체가 매우 불안정하며 여러 가지 음가(열린 ${\ae}$ 또는 닫힌 ${\O}$)로 나타난다. 예를 들어 "seul[$s{\ae}l$](홀로), ceux[$s{\O}$](이것들)"에서와 같이 발음되며, 또한 원칙적으로 schwa, [${\partial}$]로 발음이 되는 "Je[$\Im\partial$]"와 "le[$l{\partial}$]"의 경우, Paris 지역에서는 "Je sais[${\Im}{\ae}{\;}s{\Im}$](나는 안다); Prends-le[$pr{\tilde{a}}{\;}l{\ae}$](그것을 집어라)"로 발음을 하는 한편, 프랑스 북부 지방세서는 동일한 발화체를 [${\ae}$]대신에 [${\o}$]로 발음한다. 실제로 언어학적 측면에서 고려되는 "e muet"는 schwa로 나타나는 "Je[$\Im\partial$]"와 "le[$l{\partial}$]"의 경우인데, 불어 음운론에서는 schwa에 의해 대립되는 낱말짝이 없기 때문에 schwa를 음소로 인정할 것인가에 대해 논란이 있다. 그러나 불어에서 schwark 음운론적 역할을 한다는 사실은 다음과 같은 예에서 찾아 볼 수 있다. 첫째, 발음상으로 동사의 변화형에서 "porte[$p{\jmath}rte$](들다: 현재형), porte[$p{\jmath}rte$](과거분사형), porta[$p{\jmath}rte$](단순과거형)"등이 대립되며, 이휘 "Porto[$p{\jmath}rte$](포르토)"와도 대립된다. 둘째, 어휘적 대립 "le haut[$l{\partial}o$](위)/l'eau[lo](물)"와 형태론적 대립 "le[$l{\partial}$](정관사, 남성단수)/les[le](정관사, 복수)"등에서 "묵음 e"는 분명히 음운론적 역할을 하고 있다. 본 논문에서는 이와 같이 음색이 복잡하게 나타나는 "e muet"의 문제를 리듬단위, 문맥적 분포 및 음절모형 측면, 즉 음성학 및 음운론적 측면에서 다양하게 분석하여 그 본질을 규명해 보고 "e muet"탈락현상을 TCG(Theorie de Charme et de Gouvernement) 측면에서 새롭게 해석해 보았다.

  • PDF

Visualization of Korean Speech Based on the Distance of Acoustic Features (음성특징의 거리에 기반한 한국어 발음의 시각화)

  • Pok, Gou-Chol
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.13 no.3
    • /
    • pp.197-205
    • /
    • 2020
  • Korean language has the characteristics that the pronunciation of phoneme units such as vowels and consonants are fixed and the pronunciation associated with a notation does not change, so that foreign learners can approach rather easily Korean language. However, when one pronounces words, phrases, or sentences, the pronunciation changes in a manner of a wide variation and complexity at the boundaries of syllables, and the association of notation and pronunciation does not hold any more. Consequently, it is very difficult for foreign learners to study Korean standard pronunciations. Despite these difficulties, it is believed that systematic analysis of pronunciation errors for Korean words is possible according to the advantageous observations that the relationship between Korean notations and pronunciations can be described as a set of firm rules without exceptions unlike other languages including English. In this paper, we propose a visualization framework which shows the differences between standard pronunciations and erratic ones as quantitative measures on the computer screen. Previous researches only show color representation and 3D graphics of speech properties, or an animated view of changing shapes of lips and mouth cavity. Moreover, the features used in the analysis are only point data such as the average of a speech range. In this study, we propose a method which can directly use the time-series data instead of using summary or distorted data. This was realized by using the deep learning-based technique which combines Self-organizing map, variational autoencoder model, and Markov model, and we achieved a superior performance enhancement compared to the method using the point-based data.

A review on the method of coined words by Korean and Chinese characters (한·중 인물지칭 신어 조어방식에 관한 고찰 - 2017년과 2018년을 중심으로 -)

  • Wang, Yan
    • Journal of Convergence for Information Technology
    • /
    • v.12 no.3
    • /
    • pp.178-185
    • /
    • 2022
  • This study compared and analyzed the characteristics of new words by classifying 197 newly coined Korean and Chinese characters in 2017 and 2018 into single, compound, derivative, abbreviated, and hybrid words according to the coined method. In the case of a single language, Korean is all words borrowed from Chinese and English. However, no monolingual language appeared in Chinese. In the case of compound words, the format of the Chinese synthesis method was much more diverse and the generative power was stronger than that of Korea. In the case of derivatives, there are not many prefixes in both countries, and Korean suffixes have the strongest productivity of Chinese suffixes and weak productivity of foreign and native suffixes. Korean foreign language suffixes were characterized by relatively more appearance than Chinese. In the case of abbreviations, it can be seen that the productivity of dark syllables is stronger for Korean abbreviations, and the productivity of empty syllables is stronger for Chinese abbreviations. In the case of mixed languages, the hybrid form of Korean was much more diverse than that of Chinese. Through this study, it will be possible to help Chinese Korean learners understand the process of forming a new language, and to develop their ability to guess the meaning of Korean words while learning a new language.

Segmenting Korean Nominal Compounds with an Unknown Morpheme Using Back-off Statistics (백오프 통계정보를 이용한 미등록어 포함 복합명사의 분해)

  • Park, Jae-Han;Kim, Myoung-Sun;Rho, Dae-Wook;Ra, Dong-Yul
    • Annual Conference on Human and Language Technology
    • /
    • 2004.10d
    • /
    • pp.65-72
    • /
    • 2004
  • 본 논문에서는 백오프 통계 정보를 이용하여 일반적인 복합명사 뿐만 아니라 외래어 미등록어를 포함한 복합명사도 잘 분해하는 방법을 제안한다. 본 시스템은 입력으로 형태소분석기가 내주는 많은 분석 후보들을 받는다. 단음절 명사를 포함한 분석 후보도 포함되므로 입력 분석 후보의 수는 대단히 많게 된다. 본 모듈의 주요 작업은 이 중에서 가장 좋은 분석후보를 선택하는 것이 된다. 미등록어가 포함된 경우 이에 부합되는 분석 후보를 잘 선택하는 시스템의 개발을 목표로 한다. 이를 위해서 본 시스템에서 사용하는 주요 정보는 단어간 어휘 바이그램 통계정보이다. 또한 외래어 미등록어의 인식 정확성을 높이기 위해 음절 바이그램 정보도 이용한다. 통계정보는 대량의 품사 태깅 말뭉치에서 추출하였다. 데이터 부족 문제를 해소하기 위해서 우리는 백오프(back-off) 평탄화(smoothing) 기법을 이용하였다. 미등록어가 포함된 복합명사의 분석 후보의 수를 줄이기 위한 기술도 연구하였다.

  • PDF

Automatic Music Transcription System Using SIDE (SIDE를 이용한 자동 음악 채보 시스템)

  • Hyoung, A-Young;Lee, Joon-Whoan
    • The KIPS Transactions:PartB
    • /
    • v.16B no.2
    • /
    • pp.141-150
    • /
    • 2009
  • This paper proposes a system that can automatically write singing voices to music notes. First, the system uses Stabilized Diffusion Equation(SIDE) to divide the song to a series of syllabic parts based on pitch detection. By the song segmentation, our method can recognize the sound length of each fragment through clustering based on genetic algorithm. Moreover, this study introduces a concept called 'Relative Interval' so as to recognize interval based on pitch of singer. And it also adopted measure extraction algorithm using pause data to implement the higher precision of song transcription. By the experiments using 16 nursery songs, it is shown that the measure recognition rate is 91.5% and DMOS score reaches 3.82. These findings demonstrate effectiveness of system performance.

Detecting and Interpreting Terms: Focusing Korean Medical Terms (전문용어 탐지와 해석 모델: 한국어 의학용어 중심으로 )

  • Haram-Yeom;Jae-Hoon Kim
    • Annual Conference on Human and Language Technology
    • /
    • 2022.10a
    • /
    • pp.407-411
    • /
    • 2022
  • 최근 COVID-19로 인해 대중의 의학 분야 관심이 증가하고 있다. 대부분의 의학문서는 전문용어인 의학용어로 구성되어 있어 대중이 이를 보고 이해하기에 어려움이 있다. 의학용어를 쉬운 뜻으로 풀이하는 모델을 이용한다면 대중이 의학 문서를 쉽게 이해할 수 있을 것이다. 이런 문제를 완화하기 위해서 본 논문에서는 Transformer 기반 번역 모델을 이용한 의학용어 탐지 및 해석 모델을 제안한다. 번역 모델에 적용하기 위해 병렬말뭉치가 필요하다. 본 논문에서는 다음과 같은 방법으로 병렬말뭉치를 구축한다: 1) 의학용어 사전을 구축한다. 2) 의학 드라마의 자막으로부터 의학용어를 찾아서 그 뜻풀이로 대체한다. 3) 원자막과 뜻풀이가 포함된 자막을 나란히 배열한다. 구축된 병렬말뭉치를 이용해서 Transformer 번역모델에 적용하여 전문용어를 찾아서 해석하는 모델을 구축한다. 각 문장은 음절 단위로 나뉘어 사전학습 된 KoCharELECTRA를 이용해서 임베딩한다. 제안된 모델은 약 69.3%의 어절단위 BLEU 점수를 보였다. 제안된 의학용어 해석기를 통해 대중이 의학문서를 좀 더 쉽게 접근할 수 있을 것이다.

  • PDF

Correlation of acoustic features and electrophysiological outcomes of stimuli at the level of auditory brainstem (자극음의 음향적 특성과 청각 뇌간에서의 전기생리학적 반응의 상관성)

  • Chun, Hyungi;Han, Woojae
    • The Journal of the Acoustical Society of Korea
    • /
    • v.35 no.1
    • /
    • pp.63-73
    • /
    • 2016
  • It is widely acknowledged that the human auditory system is organized tonotopically and people generally listen to sounds as a function of frequency distribution through the auditory system. However, it is still unclear how acoustic features of speech sounds are indicated to the human brain in terms of speech perception. Thus, the purpose of this study is to investigate whether two sounds with similar high-frequency characteristics in the acoustic analysis show similar results at the level of auditory brainstem. Thirty three young adults with normal hearing participated in the study. As stimuli, two Korean monosyllables (i.e., /ja/ and /cha/) and four frequencies of toneburst (i.e., 500, 1000, 2000, and 4000 Hz) were used to elicit the auditory brainstem response (ABR). Measures of monosyllable and toneburst were highly replicable and the wave V of waveform was detectable in all subjects. In the results of Pearson correlation analysis, the /ja/ syllable had a high correlation with 4000 Hz of toneburst which means that its acoustic characteristics (i.e., 3671~5384 Hz) showed the same results in the brainstem. However, the /cha/ syllable had a high correlation with 1000 and 2000 Hz of toneburst although it has acoustical distribution of 3362~5412 Hz. We concluded that there was disagreement between acoustic features and physiology outcomes at the auditory brainstem level. This finding suggests that an acoustical-perceptual mapping study is needed to scrutinize human speech perception.