• Title/Summary/Keyword: Consonants

Search Result 457, Processing Time 0.021 seconds

ACT-R Predictive Model of Korean Text Entry on Touchscreen

  • Lim, Soo-Yong;Jo, Seong-Sik;Myung, Ro-Hae;Kim, Sang-Hyeob;Jang, Eun-Hye;Park, Byoung-Jun
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.2
    • /
    • pp.291-298
    • /
    • 2012
  • Objective: The aim of this study is to predict Korean text entry on touchscreens using ACT-R cognitive architecture. Background: Touchscreen application in devices such as satellite navigation devices, PDAs, mobile phones, etc. has been increasing, and the market size is expanding. Accordingly, there is an increasing interest to develop and evaluate the interface to enhance the user experience and increase satisfaction in the touchscreen environment. Method: In this study, Korean text entry performance in the touchscreen environment was analyzed using ACT-R. The ACT-R model considering the characteristics of the Korean language which is composed of vowels and consonants was established. Further, this study analyzed if the prediction of Korean text entry is possible through the ACT-R cognitive model. Results: In the analysis results, no significant difference on performance time between model prediction and empirical data was found. Conclusion: The proposed model can predict the accurate physical movement time as well as cognitive processing time. Application: This study is useful in conducting model-based evaluation on the text entry interface of the touchscreen and enabled quantitative and effective evaluation on the diverse types of Korean text input interfaces through the cognitive models.

Speech Synthesis Based on CVC Speech Segments Extracted from Continuous Speech (연속 음성으로부터 추출한 CVC 음성세그먼트 기반의 음성합성)

  • 김재홍;조관선;이철희
    • The Journal of the Acoustical Society of Korea
    • /
    • v.18 no.7
    • /
    • pp.10-16
    • /
    • 1999
  • In this paper, we propose a concatenation-based speech synthesizer using CVC(consonant-vowel-consonant) speech segments extracted from an undesigned continuous speech corpus. Natural synthetic speech can be generated by a proper modelling of coarticulation effects between phonemes and the use of natural prosodic variations. In general, CVC synthesis unit shows smaller acoustic degradation of speech quality since concatenation points are located in the consonant region and it can properly model the coarticulation of vowels that are effected by surrounding consonants. In this paper, we analyze the characteristics and the number of required synthesis units of 4 types of speech synthesis methods that use CVC synthesis units. Furthermore, we compare the speech quality of the 4 types and propose a new synthesis method based on the most promising type in terms of speech quality and implementability. Then we implement the method using the speech corpus and synthesize various examples. The CVC speech segments that are not in the speech corpus are substituted by demonstrate speech segments. Experiments demonstrate that CVC speech segments extracted from about 100 Mbytes continuous speech corpus can produce high quality synthetic speech.

  • PDF

The Study on Automatic Speech Recognizer Utilizing Mobile Platform on Korean EFL Learners' Pronunciation Development (자동음성인식 기술을 이용한 모바일 기반 발음 교수법과 영어 학습자의 발음 향상에 관한 연구)

  • Park, A Young
    • Journal of Digital Contents Society
    • /
    • v.18 no.6
    • /
    • pp.1101-1107
    • /
    • 2017
  • This study explored the effect of ASR-based pronunciation instruction, using a mobile platform, on EFL learners' pronunciation development. Particularly, this quasi-experimental study focused on whether using mobile ASR, which provides voice-to-text feedback, can enhance the perception and production of target English consonants minimal pairs (V-B, R-L, and G-Z) of Korean EFL learners. Three intact classes of 117 Korean university students were assigned to three groups: a) ASR Group: ASR-based pronunciation instruction providing textual feedback by the mobile ASR; b) Conventional Group: conventional face-to-face pronunciation instruction providing individual oral feedback by the instructor; and the c) Hybrid Group: ASR-based pronunciation instruction plus conventional pronunciation instruction. The ANCOVA results showed that the adjusted mean score for pronunciation production post-test on the Hybrid instruction group (M=82.71, SD =3.3) was significantly higher than the Conventional group (M=62.6, SD =4.05) (p<.05).

A Study on On-line Recognition System of Korean Characters (온라인 한글자소 인식시스템의 구성에 관한 연구)

  • Choi, Seok;Kim, Gil-Jung;Huh, Man-Tak;Lee, Jong-Hyeok;Nam, Ki-Gon;Yoon, Tae-Hoon;Kim, Jae-Chang;Lee, Ryang-Seong
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.30B no.9
    • /
    • pp.94-105
    • /
    • 1993
  • In this paper propose a Koaren character recognition system using a neural network is proposed. This system is a multilayer neural network based on the masking field model which consists of a input layer, four feature extraction layers which extracts type, direction, stroke, and connection features, and an output layer which gives us recognized character codes. First, 4x4 subpatterns of an NxN character pattern stored in the input buffer are applied into the feature extraction layers sequentially. Then, each of feature extraction layers extracts sequentially features such as type, direction, stroke, and connection, respectively. Type features for direction and connection are extracted by the type feature extraction layer, direction features for stroke by the direction feature extraction layer and stroke and connection features for stroke by the direction feature extraction layer and stroke and connection features for the recongnition of character by the stroke and the connection feature extractions layers, respectively. The stroke and connection features are saved in the sequential buffer layer sequentially and using these features the characters are recognized in the output layer. The recognition results of this system by tests with 8 single consonants and 6 single vowels are promising.

  • PDF

Study on regional Distribution and Etymology according to the Type in the World's Tobacco Name (세계 담배이름의 유형에 따른 지역적 분포와 어원에 관한 연구)

  • Jeong, Kee-Taeg
    • Journal of the Korean Society of Tobacco Science
    • /
    • v.37 no.1
    • /
    • pp.8-17
    • /
    • 2015
  • The purpose of this study is to classify the tobacco names in the world, to investigate the regional distribution of the classified type, and to show origin of names according to the type. The names of tobacco used in this study was 50. The type of tobacco names was classified by the presence or absence of nasal sound(morn) on the first syllable, the Number of syllable, and the structure of consonants and vowels of tobacco names. Type I (Dambago) has the nasal sound on the first syllable. The proportion of Type I(Dambago) was 28%. And the rest(Type I~Type V ; 72%) has no nasal sound. Type II(Tabaco) has three syllables, and its proportion was 20%. Type III(Tabac) has the two syllables and the structure of T+vowels+B+vowels. And its proportion was 30%. Type IV(Tutun) has the two syllables and the structure of T+vowels+T+vowels. And its proportion was 12%. Type V(Duhan) has the two syllables and the structure of D+vowels+H(V)+vowels. And its proportion was 10%. The world's most widely distributed type was Type I(Dambago). regional distribution of the world's tobacco names were clustered by the type. 72% of Type I(Dambago) was distributed in Asia. The etymology of Type I(Dambago) was only 14% Tambaku and the other is not yet known. The etymology of Type I(Dambago) seems to be derived from the Haitian Tambaku(meaning a tobacco pipe). 88% of Type II(Tabaco) and III(Tabac) were distributed in Europe. The etymology of Type II(Tabaco) and Type III(Tabac) were 84% Spanish "Tabaco". 100% of Type IV(Tutun) and V(Duhan) were distributed in Europe. The etymology of Type IV(Tutun) and Type V(Duhan) were 100% Turkish tutun and duhan, respectively. This finding suggests that the etymology of Type I(Dambago) is certainly may be Haitian "Tambaku(meaning a tobacco pipe)".

  • PDF

A Study on Korean Phoneme Classification using Recursive Least-Square Algorithm (Recursive Least-Square 알고리즘을 이용한 한국어 음소분류에 관한 연구)

  • Kim, Hoe-Rin;Lee, Hwang-Su;Un, Jong-Gwan
    • The Journal of the Acoustical Society of Korea
    • /
    • v.6 no.3
    • /
    • pp.60-67
    • /
    • 1987
  • In this paper, a phoneme classification method for Korean speech recognition has been proposed and its performance has been studied. The phoneme classification has been done based on the phonemic features extracted by the prewindowed recursive least-square (PRLS) algorithm that is a kind of adaptive filter algorithms. Applying the PRLS algorithm to input speech signal, precise detection of phoneme boundaries has been made, Reference patterns of Korean phonemes have been generated by the ordinery vector quantization (VQ) of feature vectors obtained manualy from prototype regions of each phoneme. In order to obtain the performance of the proposed phoneme classification method, the method has been tested using spoken names of seven Korean cities which have eleven different consonants and eight different vowels. In the speaker-dependent phoneme classification, the accuracy is about $85\%$ considering simple phonemic rules of Korean language, while the accuracy of the speaker-independent case is far less than that of the speaker-dependent case.

  • PDF

A Study on Multi-Pulse Speech Coding Method by Using V/S/TSIUVC (V/S/TSIUVC를 이용한 멀티펄스 음성부호화 방식에 관한 연구)

  • Lee See-Woo
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.9
    • /
    • pp.1233-1239
    • /
    • 2004
  • In a speech coding system using excitation source of voiced and unvoiced, it would be involved a distortion of speech qualify in case coexist with a voiced and an unvoiced consonants in a frame. This paper present a new multi-pulse coding method by using V/S/TSIUVC switching, individual pitch pulses and TSIUVC approximation-synthesis method in order to restrict a distortion of speech quality. The TSIUVC is extracted by using the zero crossing rate and individual pitch pulse. And the TSIUVC extraction rate was 91% for female voice and 96.2% for male voice respectively. The important thing is that the frequency information of 0.347kHz below and 2.813kHz above can be made with high quality synthesis waveform within TSIUVC. I evaluate the MPC use V/UV and the FBD-MPC use V/S/TSIUVC. As a result, I knew that synthesis speech of the FBD-MPC was better in speech quality than synthesis speech of the MPC.

  • PDF

The Sensitivity Analysis for Customer Feedback on Social Media (소셜 미디어 상 고객피드백을 위한 감성분석)

  • Song, Eun-Jee
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.4
    • /
    • pp.780-786
    • /
    • 2015
  • Social media, such as Social Network Service include a lot of spontaneous opinions from customers, so recent companies collect and analyze information about customer feedback by using the system that analyzes Big Data on social media in order to efficiently operate businesses. However, it is difficult to analyze data collected from online sites accurately with existing morpheme analyzer because those data have spacing errors and spelling errors. In addition, many online sentences are short and do not include enough meanings which will be selected, so established meaning selection methods, such as mutual information, chi-square statistic are not able to practice Emotional Classification. In order to solve such problems, this paper suggests a module that can revise the meanings by using initial consonants/vowels and phase pattern dictionary and meaning selection method that uses priority of word class in a sentence. On the basis of word class extracted by morpheme analyzer, these new mechanisms would separate and analyze predicate and substantive, establish properties Database which is subordinate to relevant word class, and extract positive/negative emotions by using accumulated properties Database.

Articulation error of children with adenoid hypertrophy

  • Eom, Tae-Hoon;Jang, Eun-Sil;Kim, Young-Hoon;Chung, Seung-Yun;Lee, In-Goo
    • Clinical and Experimental Pediatrics
    • /
    • v.57 no.7
    • /
    • pp.323-328
    • /
    • 2014
  • Purpose: Adenoid hypertrophy is a physical alteration that may affect speech, and a speech disorder can have other negative effects on a child's life. Airway obstruction leads to constricted oral breathing and causes postural alterations of several oro-facial structures, including the mouth, tongue, and hyoid bone. The postural modifications may affect several aspects of speech production. Methods: In this study, we compared articulation errors in 19 children with adenoid hypertrophy (subject group) to those of 33 children with functional articulation disorders independent of anatomical problems (control group). Results: The mean age of the subject group was significantly higher (P=0.016). Substitution was more frequent in the subject group (P=0.003; odds ratio [OR], 1.80; 95% confidence interval [CI], 1.23- 2.62), while omission was less frequent (P<0.001; OR, 0.43; 95% CI, 0.27-0.67). Articulation errors were significantly less frequent in the palatal affricative in the subject group (P=0.047; OR, 0.25; 95% CI, 0.07-0.92). The number of articulation errors in other consonants was not different between the two groups. Nasalization and aspiration were significantly more frequent in the subject group (P=0.007 and 0.014; OR, 14.77 and 0.014; 95% CI, [1.62-135.04] and NA, respectively). Otherwise, there were no differences between the two groups. Conclusion: We identified the characteristics of articulation errors in children with adenoid hypertrophy, but our data did not show the relationship between adenoid hypertrophy and oral motor function that has been observed in previous studies. The association between adenoid hypertrophy and oral motor function remains doubtful.

Finger-Touch based Hangul Input Interface for Usability Enhancement among Visually Impaired Individuals (시각 장애인의 입력 편의성 향상을 위한 손가락 터치 기반의 한글 입력 인터페이스)

  • Kang, Seung-Shik;Choi, Yoon-Seung
    • Journal of KIISE
    • /
    • v.43 no.11
    • /
    • pp.1307-1314
    • /
    • 2016
  • Virtual Hangul keyboards like Chun-Ji-In, Narat-Gul, and QWERTY are based on eyesight recognition, in which input letter positions are fixed in the smartphone environment. The input method of a fixed-position style is not very convenient for visually impaired individuals. In order to resolve the issue of inconvenience of the Hangul input system, we propose a new paradigm of the finger-touch based Hangul input system that does not need eyesight recognition of input buttons. For the convenience of learning the touch-motion based keyboard, finger touches are designed by considering the shape and frequencies of Hangul vowels and consonants together with the preference of fingers. The base position is decided by the first touch of the screen, and the finger-touch keyboard is used in the same way for all the other touch-style devices, regardless of the differences in size and operation system. In this input method, unique finger-touch motions are assigned for Hangul letters that significantly reduce the input errors.