• Title/Summary/Keyword: Korean Language Input System

Search Result 209, Processing Time 0.026 seconds

Prosodic Contour Generation for Korean Text-To-Speech System Using Artificial Neural Networks

  • Lim, Un-Cheon
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.2E
    • /
    • pp.43-50
    • /
    • 2009
  • To get more natural synthetic speech generated by a Korean TTS (Text-To-Speech) system, we have to know all the possible prosodic rules in Korean spoken language. We should find out these rules from linguistic, phonetic information or from real speech. In general, all of these rules should be integrated into a prosody-generation algorithm in a TTS system. But this algorithm cannot cover up all the possible prosodic rules in a language and it is not perfect, so the naturalness of synthesized speech cannot be as good as we expect. ANNs (Artificial Neural Networks) can be trained to learn the prosodic rules in Korean spoken language. To train and test ANNs, we need to prepare the prosodic patterns of all the phonemic segments in a prosodic corpus. A prosodic corpus will include meaningful sentences to represent all the possible prosodic rules. Sentences in the corpus were made by picking up a series of words from the list of PB (phonetically Balanced) isolated words. These sentences in the corpus were read by speakers, recorded, and collected as a speech database. By analyzing recorded real speech, we can extract prosodic pattern about each phoneme, and assign them as target and test patterns for ANNs. ANNs can learn the prosody from natural speech and generate prosodic patterns of the central phonemic segment in phoneme strings as output response of ANNs when phoneme strings of a sentence are given to ANNs as input stimuli.

Implementation of Korean Support System for Language Disorders (언어 장애인을 위한 한국어 지원 시스템의 구현)

  • Choi, J.H.;Choo, K.N.;Woo, Y.S.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.6 no.1
    • /
    • pp.29-35
    • /
    • 2012
  • Most Augmentative and Alternative Communication(AAC) use exclusive equipment or studied desktop, tablet PC based windows. Besides, the preceding study offers proper noun dictionary so, henceforward study has problem to innumerable proper noun processing. This paper suggests a method of proper noun processing using a mobile smart equipment. And via the button with virtual keyboard input method and the errors that can occur is also proposing a complementary way. AAC system to check availability for application on Android has been implemented. Experimental results, depending on user location and selection of proper nouns in the around could produce a sentence is derived.

  • PDF

Study On Design of Encoding Module for PLC input (PIC 입력 인코딩 모듈 설계에 관한 연구)

  • You, Jeong-Bong
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.19 no.3
    • /
    • pp.105-110
    • /
    • 2005
  • Programmable Logic Controller(PLC) is the most widely utilized among many sorts of existing controller for the design of factory automation control system, and study about a PLC language is peformed actively. In this paper, we proposed the encoding module that we increase input points when increased of the input machine which is going to control it in the FA design that used PLC. Input point is going to be extended with 2-3 points in the state that cannot equip input unit in basic base any more. Then an extension base, a power supply, an extension cable and an output nit must be equipped particularly. Then a cost must be added very much. This paper proposed the encoding module which extended an input in a small quantity, and we confirmed a feasibility through a simulation.

Zero-anaphora resolution in Korean based on deep language representation model: BERT

  • Kim, Youngtae;Ra, Dongyul;Lim, Soojong
    • ETRI Journal
    • /
    • v.43 no.2
    • /
    • pp.299-312
    • /
    • 2021
  • It is necessary to achieve high performance in the task of zero anaphora resolution (ZAR) for completely understanding the texts in Korean, Japanese, Chinese, and various other languages. Deep-learning-based models are being employed for building ZAR systems, owing to the success of deep learning in the recent years. However, the objective of building a high-quality ZAR system is far from being achieved even using these models. To enhance the current ZAR techniques, we fine-tuned a pretrained bidirectional encoder representations from transformers (BERT). Notably, BERT is a general language representation model that enables systems to utilize deep bidirectional contextual information in a natural language text. It extensively exploits the attention mechanism based upon the sequence-transduction model Transformer. In our model, classification is simultaneously performed for all the words in the input word sequence to decide whether each word can be an antecedent. We seek end-to-end learning by disallowing any use of hand-crafted or dependency-parsing features. Experimental results show that compared with other models, our approach can significantly improve the performance of ZAR.

The Globalization of the Korean Script from Scriptological View. (훈민정음 확장 가능성에 대한일반문자학적 검토)

  • Yurn, Gyudong
    • Korean Linguistics
    • /
    • v.80
    • /
    • pp.125-150
    • /
    • 2018
  • This paper discusses several prerequisites for recording other languages by using Hunminjeongum (Hangeul). The scientific, systematic, and practical features of Hunminjeongum is undoubtedly obvious. Such a view of Hunminjeongeum is based on its iconicity, featural property and non-linearity. However, the existing conviction that Hunminjeongeum can transcribe all human sounds or that Hunminjeongeum is the only script which is easy for anyone to learn requires a more rigorous consideration. Given the fact that all scripts can be evaluated as equal, without one being more superior than the other, all scripts are optimized for recording the language in its own given linguistic context. Further, the function of a script is not just limited in a simple representation of a language. Adding new letters into the system of Hunminjeongeum or maintaining its current mode of notation can be problematic. First, it is important to solve computer-related issues: the registration of Hunminjeongum into the computer system as well as its output, problems regarding unicode, the new developments required for input devices, keyboards, fonts, and computer programs.

Design and Implementation of a Language Identification System for Handwriting Input Data (필기 입력데이터에 대한 언어식별 시스템의 설계 및 구현)

  • Lim, Chae-Gyun;Kim, Kyu-Ho;Lee, Ki-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.1
    • /
    • pp.63-68
    • /
    • 2010
  • Recently, to accelerate the Ubiquitous generation, the input interface of the mobile machinery and tools are actively being researched. In addition with the existing interfaces such as the keyboard and curser (mouse), other subdivisions including the handwriting, voice, vision, and touch are under research for new interfaces. Especially in the case of small-sized mobile machinery and tools, there is a increasing need for an efficient input interface despite the small screens. This is because, additional installment of other devices are strictly limited due to its size. Previous studies on handwriting recognition have generally been based on either two-dimensional images or algorithms which identify handwritten data inserted through vectors. Futhermore, previous studies have only focused on how to enhance the accuracy of the handwriting recognition algorithms. However, a problem arisen is that when an actual handwriting is inserted, the user must select the classification of their characters (e.g Upper or lower case English, Hangul - Korean alphabet, numbers). To solve the given problem, the current study presents a system which distinguishes different languages by analyzing the form/shape of inserted handwritten characters. The proposed technique has treated the handwritten data as sets of vector units. By analyzing the correlation and directivity of each vector units, a more efficient language distinguishing system has been made possible.

영한자동번역에서의 두단계 영어 전산문법

  • 최승권
    • Language and Information
    • /
    • v.4 no.1
    • /
    • pp.97-109
    • /
    • 2000
  • Application systems of natural language processing such as machine translation system must deal with actual texts including the full range of linguistic phenomena. But it seems to be impossible that the existing grammar covers completely such actual texts because they include disruptive factors such as long sentences, unexpected sentence patterns and erroneous input to obstruct well-formed analysis of a sentence. In order to solve analysis failure due to the disruptive factors or incorrect selection of correct parse tree among forest parse trees, this paper proposes two-level computational grammar which consists of a constraint-based grammar and an error-tolerant grammar. The constraint-based computational grammar is the grammar that gives us the well-formed analysis of English texts. The error-tolerant computational grammar is the grammar that reconstructs a comprehensible whole sentence structure with partially successful parse trees within failed parsing results.

  • PDF

A SDL Hardware Compiler for VLSI Logic Design Automation (VLSI의 논리설계 자동화를 위한 SDL 하드웨어 컴파일러)

  • Cho, Joung Hwee;Chong, Jong Wha
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.23 no.3
    • /
    • pp.327-339
    • /
    • 1986
  • In this paper, a hardware compiler for symbolic description language(SDL) is proposed for logic design automation. Lexical analysis is performed for SDL which describes the behavioral characteristics of a digital system at the register transfer level by the proposed algorithm I. The algorithm I is proposed to get the expressions for the control unit and for the data transfer unit. In order to obtain the network description language(NDL) expressions equivalent to gate-level logic circuits, another algorithm, the the algorithm II, is proposed. Syntax analysis for the data formed by the algorithm I is also Performed using circuit elements such as D Flip-Flop, 2-input AND, OR, and NOT gates. This SDL hardware compiler is implemented in the programming language C(VAX-11/750(UNIX)), and its efficiency is shown by experiments with logic design examples.

  • PDF

Acquirment and Linguistic Expressions of Fuzzy Rules

  • Maebashi, Satoru;Onisawa, Takehsa
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.06a
    • /
    • pp.258-263
    • /
    • 1998
  • Fuzzy rules are often obtained by experts who know an objective system well. fuzzy rules acquired by experts, however, do not express all input-output relations of the system. This paper proposes a method fuzzy rules are expressed in plain language so that the fuzzy rules are understood easily. The proposed method is applied to the control of the distance between cars and running through a crank-typed road, and the validity of the method is confirmed.

  • PDF

A new approach for the saccadic eye movement system simulation (Saccade 안구운동계의 시뮬레이션)

  • 박상희;남문현
    • 전기의세계
    • /
    • v.26 no.1
    • /
    • pp.87-90
    • /
    • 1977
  • Various simulation techniques were developed in the modeling of biological system during the last decades. Mostly analog and hybrid simulation techniques have been used. The authors chose the Digital Analog Simulation (DAS) technique by using the MIMIC language to simulate the saccadic eye movement system performances on the digital computer. There have been various models presented by many investigators after Young & Stark's sampled-data model. The eye movement model chosen by the authors is Robinson's model III which shows the parallel information processing characteristics clearly to the double-step input stimuli. In the process of simulation, the parameter and constants used were based on the neurophysiological data of the human and animals. The analog model blocks were converted to the corresponding MIMIC block diagrams and programmed into the MIMIC statements. The program was run on the CDC Cyber 72-14 computer. The essential input stimulus was double-step of 5 and 10 degrees, and target durations chosen were 50,100 and 150 msec. The results obtained by the DAS technqiue were in good agreement with analog simulation carried out by other investigators as well as with the experimental human saccadic eye movement responses to double-step target movements.

  • PDF