• Title/Summary/Keyword: Speech function

Search Result 696, Processing Time 0.027 seconds

Effects of Cognitive Impairment on Self-reported Hearing Handicap in Older Adults with Early-stage Presbycusis (초기 노인성 난청자에서 인지장애가 일상생활 듣기 어려움에 미치는 영향)

  • Lee, Soo Jung
    • 한국노년학
    • /
    • v.38 no.1
    • /
    • pp.1-14
    • /
    • 2018
  • Everyday hearing handicap caused by presbycusis ultimately reduces quality of life in older adults. The aim of this study was to explore effects of cognitive impairment on self-reported hearing handicap in older adults with early-stage presbycusis. We compared K-HHIE scores between 40 elderly subjects with mild cognitive impairment (MCI) and age- and hearing-threshold matched 40 cognitively normal elderly (CNE) subjects. The results are as follows: 1) The MCI group scored significantly higher than the CNE group on the social/situational and emotional sections, and in total. 2) The MCI group scored significantly higher than the CNE group on all four subscales, and the most significant group difference was on the first subscale relating to interpersonal relationships and social handicaps. 3) Both groups scored highest on the item 8 (problems hearing whispering sounds) and item 15 (problems hearing TV or radio sounds). Besides those two items, the MCI group also scored high on the item 21 (problems hearing in a restaurant), item 6 (problems hearing when attending a party), item 3 (avoiding groups of people), and item 20 (personal or social restrictions). Our findings suggest that, among older adults with early-stage presbycusis, older adults with cognitive impairment tend to report greater everyday hearing handicap than their peers with normal cognitive function. Especially, they show significant problems hearing in background noise or multi-talker situations, which cause social restrictions and social/emotional loneliness.

An Effect for Sequential Information Processing by the Anxiety Level and Temporary Affect Induction (불안수준 및 일시적 유발정서가 서열정보 어휘처리에 미치는 효과)

  • Kim, Choong-Myung
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.4
    • /
    • pp.224-231
    • /
    • 2019
  • The current paper was conducted to unravel the influence of affect induction as a background emotion in the process of cognitive task to judge the degree of sequence in groups with or without anxiety symptoms. Four types of affect induction and two sequential task types were used as within-subject variables, and two types of college students groups classified under the Beck Anxiety Inventory (BAI) as a between-subject variable were selected to determine reaction times involving sequential judgment among the lexical relevance information. DmDx5 was used to present a series of stimuli and elicit a response from subjects. Repeated measured ANOVA analyses revealed that reaction times and error rates were significantly larger with anxiety participants compared to the normal group regardless of affect and task types. Within-subject variable effects found that specific affect type (sorrow condition) and number-related task type showed a more rapid response compared to other affect types and magnitude-related task type, respectively. In sum, these findings confirmed the difference in tendency with reaction time and error rates that varied as a function of accompanying affect types as well as anxiety level and task types suggesting the that underlying background affect plays a major role in processing affect-cognitive association tasks.

Suggestion of a Social Significance Research Model for User Emotion -Focused on Conversational Agent and Communication- (사용자 감정의 사회적 의미 조사 모델 제안 -대화형 에이전트와 커뮤니케이션을 중심으로-)

  • Han, Sang-Wook;Kim, Seung-In
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.3
    • /
    • pp.167-176
    • /
    • 2019
  • The conversational agent, which is at the forefront of the 4th industry, aims to personalize the user-centered focus in the future and holds an important position to have a hub that can be connected to various IoT devices. It is a challenge for interactive agents to recognize the user's emotions and provide the correct interaction to personalization. The study first I looked at emotional definitions and scientific and engineering approaches. Then I recognized through social perspectives what social function and what factors emotions have and how they can be used to understand emotions. Based on this, I explored how users can be discovered emotional social factors in communication. This research has shown that social factors can be found in the user's speech, which can be linked to the social meaning of emotions. Finally, I propose a model to discover social factors in user communication. I hope that this will help designer and researcher to study user-centered design and interaction in designing interactive agents.

Prototype Design and Development of Online Recruitment System Based on Social Media and Video Interview Analysis (소셜미디어 및 면접 영상 분석 기반 온라인 채용지원시스템 프로토타입 설계 및 구현)

  • Cho, Jinhyung;Kang, Hwansoo;Yoo, Woochang;Park, Kyutae
    • Journal of Digital Convergence
    • /
    • v.19 no.3
    • /
    • pp.203-209
    • /
    • 2021
  • In this study, a prototype design model was proposed for developing an online recruitment system through multi-dimensional data crawling and social media analysis, and validates text information and video interview in job application process. This study includes a comparative analysis process through text mining to verify the authenticity of job application paperwork and to effectively hire and allocate workers based on the potential job capability. Based on the prototype system, we conducted performance tests and analyzed the result for key performance indicators such as text mining accuracy and interview STT(speech to text) function recognition rate. If commercialized based on design specifications and prototype development results derived from this study, it may be expected to be utilized as the intelligent online recruitment system technology required in the public and private recruitment markets in the future.

Design of detection method for malicious URL based on Deep Neural Network (뉴럴네트워크 기반에 악성 URL 탐지방법 설계)

  • Kwon, Hyun;Park, Sangjun;Kim, Yongchul
    • Journal of Convergence for Information Technology
    • /
    • v.11 no.5
    • /
    • pp.30-37
    • /
    • 2021
  • Various devices are connected to the Internet, and attacks using the Internet are occurring. Among such attacks, there are attacks that use malicious URLs to make users access to wrong phishing sites or distribute malicious viruses. Therefore, how to detect such malicious URL attacks is one of the important security issues. Among recent deep learning technologies, neural networks are showing good performance in image recognition, speech recognition, and pattern recognition. This neural network can be applied to research that analyzes and detects patterns of malicious URL characteristics. In this paper, performance analysis according to various parameters was performed on a method of detecting malicious URLs using neural networks. In this paper, malicious URL detection performance was analyzed while changing the activation function, learning rate, and neural network structure. The experimental data was crawled by Alexa top 1 million and Whois to build the data, and the machine learning library used TensorFlow. As a result of the experiment, when the number of layers is 4, the learning rate is 0.005, and the number of nodes in each layer is 100, the accuracy of 97.8% and the f1 score of 92.94% are obtained.

Design of CNN-based Braille Conversion and Voice Output Device for the Blind (시각장애인을 위한 CNN 기반의 점자 변환 및 음성 출력 장치 설계)

  • Seung-Bin Park;Bong-Hyun Kim
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.3
    • /
    • pp.87-92
    • /
    • 2023
  • As times develop, information becomes more diverse and methods of obtaining it become more diverse. About 80% of the amount of information gained in life is acquired through the visual sense. However, visually impaired people have limited ability to interpret visual materials. That's why Braille, a text for the blind, appeared. However, the Braille decoding rate of the blind is only 5%, and as the demand of the blind who want various forms of platforms or materials increases over time, development and product production for the blind are taking place. An example of product production is braille books, which seem to have more disadvantages than advantages, and unlike non-disabled people, it is true that access to information is still very difficult. In this paper, we designed a CNN-based Braille conversion and voice output device to make it easier for visually impaired people to obtain information than conventional methods. The device aims to improve the quality of life by allowing books, text images, or handwritten images that are not made in Braille to be converted into Braille through camera recognition, and designing a function that can be converted into voice according to the needs of the blind.

Case Report on NTBC Treatment of Type 1 Tyrosinemia Diagnosed through Newborn Screening (신생아 선별검사를 통해 진단된 1형 타이로신혈증의 NTBC 치료 사례 보고)

  • Ji Eun Jeong;Hwa Young Kim;Jung Min Ko
    • Journal of The Korean Society of Inherited Metabolic disease
    • /
    • v.23 no.2
    • /
    • pp.39-44
    • /
    • 2023
  • Hereditary tyrosinemia type 1 (HT-1) is a metabolic disorder caused by biallelic pathogenic variants in the fumarylacetoacetate hydrolase (FAH) gene, which impairs the function of the FAH enzyme, resulting in the accumulation of tyrosine's toxic metabolites in hepatocytes and renal tubular cells. As a consequence, individuals with HT-1 exhibit symptomatic manifestations. Rapid diagnosis and treatment of HT-1 can prevent short-term death and long-term complications. A 15-day-old boy presented to the outpatient department with elevated levels of tyrosine on his newborn screening tests conducted at the age of 3 and 10 days, respectively. Further blood tests revealed increased levels of alpha-fetoprotein and amino acids including tyrosine and threonine. Urine organic acid tests indicated a significant elevation in tyrosine metabolites, as well as the presence of succinylacetone (SA), which led to the diagnosis of HT-1. Two pathogenic and likely pathogenic variants of FAH compatible with HT-1 were also detected. He began a tyrosine-restricted diet at one month old and received nitisinone (NTBC) at two months old. With continued treatment, the patient's initially elevated AFP level, detection of SA in the urine, and mild hepatomegaly showed improvement. During four years and seven months of treatment, there were no exceptional complications apart from an increase in tyrosine levels and a delay in speech. We report a case of tyrosinemia type 1 detected through newborn screening, treated with dietary restriction and NTBC, with a good prognosis.

  • PDF

A study on Pre-and Post-surgical Patterns of Mandibular Movement and EMG in Skeletal Class III Prognathic Patients who underwent Intraoral Vertical Ramus Osteotomy (하악 전돌증 환자의 구내 하악골 상행지 골절단술전후의 하악골 운동양상 및 저작근 근전도 변화에 관한 연구)

  • Park, Young-Chel;Hwang, Chung-Ju;Yu, Hyung-Seog;Han, Hee-Kyung
    • The korean journal of orthodontics
    • /
    • v.27 no.2
    • /
    • pp.283-296
    • /
    • 1997
  • Stomatognathic system is a complex one that is composed of TMJ, neuromuscular system, teeth and connective tissue, and all its components are doing their parts to maintain their physiological relationships. Mandible, in particular, performs various functions such as mastication, speech, and deglutition, the muscular activities that determine such functions are signalled by numerous types of proprioceptors that exist in periodontal membrane, TMJ, and muscles to be controlled by complicated pathways and mechanics of peripheral and central nervous system. Orthodontic treatment, especially when accompanied by orthognathic surgery, brings dramatic changes of stornatognat is system such as intraoral proprioceptors and muscle activities and thus, changes in patterns of mandibular function result The author tried to analyze changes in patterns of mandibular movement and physiologic activities of surrounding muscles in Skeletal Class III ortlrognathic surgery patients who presently show a great increase in numbers. The purpose of this study was to draw some objective guidelines in evaluating funclierual aspects of orthognathic surgery patients. Mandibular functional analysis using Biopak was performed for skeletal Class III prognathic patients who underwent IVRO(lntraoral Vertical Ramus Osteotmy), and the following results were obtained: 1. Resting EMG was greater in pre-surgical group than the control group, and it showed gradual decrease after the surgery. Clenching EMG of masseter and anterior temporalis of pre-surgical group was smaller than those of control group, they also increased post-surgically, and significant difference was found between pre-surgical and post-surgical(6 months) groups. 2. Resting EMG of anterior ternporalis was greater than that of all the other muscles, but there was no significant difference. Clenching EMG of anterior temporalis and masseter were greater than those of the other muscles with statistical difference. In swallowing, digastric muscle showed the highest EMG with statistical significance. 3. Limited range of mandibular movement was shown in pre-surgical group. Significant increase in maximum mouth opening was observed six months post-surgically, and significant increase in protrusive movement was observed three months post-surgically.

  • PDF

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.