• Title/Summary/Keyword: Consonants

Search Result 457, Processing Time 0.025 seconds

A Study on Processing of Speech Recognition Korean Words (한글 단어의 음성 인식 처리에 관한 연구)

  • Nam, Kihun
    • The Journal of the Convergence on Culture Technology
    • /
    • v.5 no.4
    • /
    • pp.407-412
    • /
    • 2019
  • In this paper, we propose a technique for processing of speech recognition in korean words. Speech recognition is a technology that converts acoustic signals from sensors such as microphones into words or sentences. Most foreign languages have less difficulty in speech recognition. On the other hand, korean consists of vowels and bottom consonants, so it is inappropriate to use the letters obtained from the voice synthesis system. That improving the conventional structure speech recognition can the correct words recognition. In order to solve this problem, a new algorithm was added to the existing speech recognition structure to increase the speech recognition rate. Perform the preprocessing process of the word and then token the results. After combining the result processed in the Levenshtein distance algorithm and the hashing algorithm, the normalized words is output through the consonant comparison algorithm. The final result word is compared with the standardized table and output if it exists, registered in the table dose not exists. The experimental environment was developed by using a smartphone application. The proposed structure shows that the recognition rate is improved by 2% in standard language and 7% in dialect.

Preliminary study of Korean Electro-palatography (EPG) for Articulation Treatment of Persons with Communication Disorders (의사소통장애인의 조음치료를 위한 한국형 전자구개도의 구현)

  • Woo, Seong Tak;Park, Young Bin;Oh, Da Hee;Ha, Ji-wan
    • Journal of Sensor Science and Technology
    • /
    • v.28 no.5
    • /
    • pp.299-304
    • /
    • 2019
  • Recently, the development of rehabilitation medical technology has resulted in an increased interest in speech therapy equipment. In particular, research on articulation therapy for communication disorders is being actively conducted. Existing methods for the diagnosis and treatment of speech disorders have many limitations, such as traditional tactile perception tests and methods based on empirical judgment of speech therapists. Moreover, the position and tension of the tongue are key factors of speech disorders with regards to articulation. This is a very important factor in the distinction of Korean characters such as lax, fortis, and aspirated consonants. In this study, we proposed a Korean electropalatography (EPG) system to easily measure and monitor the position and tension of the tongue in articulation treatment and diagnosis. In the proposed EPG system, a sensor was fabricated using an AgCl electrode and biocompatible silicon. Furthermore, the measured signal was analyzed by implementing the bio-signal processing module and monitoring program. In particular, the bio-signal was measured by inserting it into the palatal from an experimental control group. As a result, it was confirmed that it could be applied to clinical treatment in speech therapy.

OnDot: Braille Training System for the Blind (시각장애인을 위한 점자 교육 시스템)

  • Kim, Hak-Jin;Moon, Jun-Hyeok;Song, Min-Uk;Lee, Se-Min;Kong, Ki-sok
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.6
    • /
    • pp.41-50
    • /
    • 2020
  • This paper deals with the Braille Education System which complements the shortcomings of the existing Braille Learning Products. An application dedicated to the blind is configured to perform full functions through touch gestures and voice guidance for user convenience. Braille kit is produced for educational purposes through Arduino and 3D printing. The system supports the following functions. First, the learning of the most basic braille, such as initial consonants, final consonant, vowels, abbreviations, etc. Second, the ability to check learned braille by solving step quizzes. Third, translation of braille. Through the experiment, the recognition rate of touch gestures and the accuracy of braille expression were confirmed, and in case of translation, the translation was done as intended. The system allows blind people to learn braille efficiently.

Multi-Emotion Recognition Model with Text and Speech Ensemble (텍스트와 음성의 앙상블을 통한 다중 감정인식 모델)

  • Yi, Moung Ho;Lim, Myoung Jin;Shin, Ju Hyun
    • Smart Media Journal
    • /
    • v.11 no.8
    • /
    • pp.65-72
    • /
    • 2022
  • Due to COVID-19, the importance of non-face-to-face counseling is increasing as the face-to-face counseling method has progressed to non-face-to-face counseling. The advantage of non-face-to-face counseling is that it can be consulted online anytime, anywhere and is safe from COVID-19. However, it is difficult to understand the client's mind because it is difficult to communicate with non-verbal expressions. Therefore, it is important to recognize emotions by accurately analyzing text and voice in order to understand the client's mind well during non-face-to-face counseling. Therefore, in this paper, text data is vectorized using FastText after separating consonants, and voice data is vectorized by extracting features using Log Mel Spectrogram and MFCC respectively. We propose a multi-emotion recognition model that recognizes five emotions using vectorized data using an LSTM model. Multi-emotion recognition is calculated using RMSE. As a result of the experiment, the RMSE of the proposed model was 0.2174, which was the lowest error compared to the model using text and voice data, respectively.

Development of Youngwoo Keyboard for Dysarthria Patients (구음장애 환자용 영우글자판 개발)

  • Yun, Hyun-Jin;Kim, Young-Chul
    • Journal of Practical Engineering Education
    • /
    • v.14 no.2
    • /
    • pp.341-350
    • /
    • 2022
  • In this paper, we propseded the Youngwoo keyboard to help the communication disorders experienced by patients with dysarthria. Existing methods for communication by patients with dysarthria include a letter board with consonants and vowels, a mouse using eye tracking, and communication assistance applications. However, there was a delay of more than 5 minutes per sentence, so there were many difficulties in communication. However, unlike the existing method, the proposed method has been improved so that the patient with dysarthria can communicate by completing the sentences with eye movements. In the conclusion and future work of this paper, the educational use value for Yeongwoo keyboard was analyzed, and it was shown that this training can be of great help to families with ALS patients. The proposed keyboard will be able to improve the difficulties of families by facilitating communication of patients with dysarthria.

The Influence of Non-Linear Frequency Compression on the Perception of Speech and Music in Patients with High Frequency Hearing Loss

  • Ahn, Jungmin;Choi, Ji Eun;Kang, Ju Yong;Choi, Ik Joon;Lee, Myung-Chul;Lee, Byeong-Cheol;Hong, Sung Hwa;Moon, Il Joon
    • Korean Journal of Audiology
    • /
    • v.25 no.2
    • /
    • pp.80-88
    • /
    • 2021
  • Background and Objectives: Non-linear frequency compression (NLFC) technology compresses and shifts higher frequencies into a lower frequency area that has better residual hearing. Because consonants are uttered in the high-frequency area, NLFC could provide better speech understanding. The aim of this study was to investigate the clinical effectiveness of NLFC technology on the perception of speech and music in patients with high-frequency hearing loss. Subjects and Methods: Twelve participants with high-frequency hearing loss were tested in a counter-balanced order, and had two weeks of daily experience with NLFC set on/off prior to testing. Performance was repeatedly evaluated with consonant tests in quiet and noise environments, speech perception in noise, music perception and acceptableness of sound quality rating tasks. Additionally, two questionnaires (the Abbreviated Profile of Hearing Aid Benefit and the Korean version of the International Outcome Inventory-Hearing Aids) were administered. Results: Consonant and speech perception improved with hearing aids (NLFC on/off conditions), but there was no significant difference between NLFC on and off states. Music perception performances revealed no notable difference among unaided and NLFC on and off states. The benefits and satisfaction ratings between NLFC on and off conditions were also not significantly different, based on questionnaires, however great individual variability preferences were noted. Conclusions: Speech perception as well as music perception both in quiet and noise environments was similar between NLFC on and off states, indicating that real world benefits from NLFC technology may be limited in Korean adult hearing aid users.

The Influence of Non-Linear Frequency Compression on the Perception of Speech and Music in Patients with High Frequency Hearing Loss

  • Ahn, Jungmin;Choi, Ji Eun;Kang, Ju Yong;Choi, Ik Joon;Lee, Myung-Chul;Lee, Byeong-Cheol;Hong, Sung Hwa;Moon, Il Joon
    • Journal of Audiology & Otology
    • /
    • v.25 no.2
    • /
    • pp.80-88
    • /
    • 2021
  • Background and Objectives: Non-linear frequency compression (NLFC) technology compresses and shifts higher frequencies into a lower frequency area that has better residual hearing. Because consonants are uttered in the high-frequency area, NLFC could provide better speech understanding. The aim of this study was to investigate the clinical effectiveness of NLFC technology on the perception of speech and music in patients with high-frequency hearing loss. Subjects and Methods: Twelve participants with high-frequency hearing loss were tested in a counter-balanced order, and had two weeks of daily experience with NLFC set on/off prior to testing. Performance was repeatedly evaluated with consonant tests in quiet and noise environments, speech perception in noise, music perception and acceptableness of sound quality rating tasks. Additionally, two questionnaires (the Abbreviated Profile of Hearing Aid Benefit and the Korean version of the International Outcome Inventory-Hearing Aids) were administered. Results: Consonant and speech perception improved with hearing aids (NLFC on/off conditions), but there was no significant difference between NLFC on and off states. Music perception performances revealed no notable difference among unaided and NLFC on and off states. The benefits and satisfaction ratings between NLFC on and off conditions were also not significantly different, based on questionnaires, however great individual variability preferences were noted. Conclusions: Speech perception as well as music perception both in quiet and noise environments was similar between NLFC on and off states, indicating that real world benefits from NLFC technology may be limited in Korean adult hearing aid users.

Sub-modality of Mental Images to Make lines Alive (대사를 생명력 있게 만드는 멘탈 이미지의 하위양식)

  • Choi, Jung-Sun
    • Journal of Korea Entertainment Industry Association
    • /
    • v.13 no.4
    • /
    • pp.119-129
    • /
    • 2019
  • Traditional speech training in acting education focused on the technical aspects of expressing the lines such as finding long/short syllables in the word, exercising articulation of consonants and vowels, and practicing diction etc. There was a limit on this education to transform written words to vivid verbal words. The lines become live when the actor sees the concrete mental images hidden in the words while speaking the lines. I will bring the knowledge of cognitive brain science and NLP(Neural Linguistic Programming) to investigate what mental images are and why mental images are fundamental elements of thought and emotion. In addition to that, I will examine how the muscles of the body react in the process of visualization of delicate mental images (subordinate form) and how to use the responsive muscles to express speaking materials such as intensity, pause, pitch, intonation etc. Conclusion, I will enumerate the obstacles encountered by actors in the course of practicing mental images, and suggest 'activation of breathing' as a thesis of the follow-up paper to eliminate those obstacles. This process, I intend to make mental images to be the concrete and practical information that can be applied to speak the dialogue in the play.

A study on reading and writing and congnitive processing from multicultural in elementary (다문화가정 초등학생의 읽기, 쓰기와 인지처리능력 연구)

  • Park, Soon-Gil;Cho, Jeung-Ryeul;Kim, Eun-Hee
    • Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology
    • /
    • v.5 no.2
    • /
    • pp.157-165
    • /
    • 2015
  • The purpose of this study is to analyse literacy of children from multicultural backgrounds, and identify cognitive-linguistic predictors that can affect their literacy. First, the higher-grade students showed better cognitive-linguistic variables in reading and writing performance. Second, it has been noted that the predictor variable of reading in children from multicultural backgrounds was homeostasis in visual form, which is a sub-variable of visual perception. This implies that detained characteristics play an important role in reading prerequisite. Therefore it can be said that it is more important to recognise features and clues about the details than reading familiar words. Furthermore, learning consonants and vowels should come first rather than studying letters at the first stages of learning Korean.

Perceptual training on Korean obstruents for Vietnamese learners (베트남 한국어 학습자를 위한 한국어 자음 지각 훈련 연구)

  • Hyosung Hwang
    • Phonetics and Speech Sciences
    • /
    • v.15 no.4
    • /
    • pp.17-26
    • /
    • 2023
  • This study aimed to reveal how Vietnamese adult learners at three different proficiency levels perceive Korean word-initial obstruents and whether errors can be corrected through perceptual training. To this end, 105 Vietnamese beginner, intermediate, and advanced learners were given perceptual training on Korean word-initial. The training materials were created by actively utilizing Korean minimal pairs as natural stimuli recorded by native speakers. Learners in the experimental group performed five 20-40 minute self-directed perceptual training sessions over a period of approximately two weeks, while learners in the control group only participated in the pretest and posttest. The results showed a significant improvement in the perception of sounds that were difficult to distinguish before training, and both beginners and advanced learners benefited from the training. This study confirmed that large-scale perceptual training can play an important role in helping Vietnamese learners learn the appropriate acoustic cues to distinguish different sounds in Korean.