• Title/Summary/Keyword: Nonverbal sound

Search Result 8, Processing Time 0.031 seconds

Nonverbal Expressions in New Media Art -Case Studies about Facial Expressions and Sound (뉴미디어 아트에 나타난 비언어적 표현 -표정과 소리의 사례연구를 중심으로)

  • Yoo, Mi;An, KyoungHee
    • The Journal of the Korea Contents Association
    • /
    • v.19 no.10
    • /
    • pp.146-156
    • /
    • 2019
  • New media art moves out of place and time constraints, sublimates the benefits of technology into art, and presents a new way of communication with the audience. This paper analyses the tendency of nonverbal communication methods by analysing examples of facial expressions and sound used in new media art from early times. As a result, it can be seen that the digital paradigm in the new media art has a nonlinear thinking, which makes a perceptual reduction of immersion and dispersion. The facial expression in new media art made it possible not only to overcome the limit of space and time of various expressions through 'visual distortions, enlargement, and virtualisation', but also to enable new ways of communication to display facial parts combined or separated in the digital environment. The sound in new media art does not stay in auditory sense, but pursues multi-sensory and synesthesia by cooperating with visual and tactile, evolves by revealing characteristics of space expansion and sensibility and interaction of audience.

Interaction Intent Analysis of Multiple Persons using Nonverbal Behavior Features (인간의 비언어적 행동 특징을 이용한 다중 사용자의 상호작용 의도 분석)

  • Yun, Sang-Seok;Kim, Munsang;Choi, Mun-Taek;Song, Jae-Bok
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.8
    • /
    • pp.738-744
    • /
    • 2013
  • According to the cognitive science research, the interaction intent of humans can be estimated through an analysis of the representing behaviors. This paper proposes a novel methodology for reliable intention analysis of humans by applying this approach. To identify the intention, 8 behavioral features are extracted from the 4 characteristics in human-human interaction and we outline a set of core components for nonverbal behavior of humans. These nonverbal behaviors are associated with various recognition modules including multimodal sensors which have each modality with localizing sound source of the speaker in the audition part, recognizing frontal face and facial expression in the vision part, and estimating human trajectories, body pose and leaning, and hand gesture in the spatial part. As a post-processing step, temporal confidential reasoning is utilized to improve the recognition performance and integrated human model is utilized to quantitatively classify the intention from multi-dimensional cues by applying the weight factor. Thus, interactive robots can make informed engagement decision to effectively interact with multiple persons. Experimental results show that the proposed scheme works successfully between human users and a robot in human-robot interaction.

Human Laughter Generation using Hybrid Generative Models

  • Mansouri, Nadia;Lachiri, Zied
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.5
    • /
    • pp.1590-1609
    • /
    • 2021
  • Laughter is one of the most important nonverbal sound that human generates. It is a means for expressing his emotions. The acoustic and contextual features of this specific sound are different from those of speech and many difficulties arise during their modeling process. During this work, we propose an audio laughter generation system based on unsupervised generative models: the autoencoder (AE) and its variants. This procedure is the association of three main sub-process, (1) the analysis which consist of extracting the log magnitude spectrogram from the laughter database, (2) the generative models training, (3) the synthesis stage which incorporate the involvement of an intermediate mechanism: the vocoder. To improve the synthesis quality, we suggest two hybrid models (LSTM-VAE, GRU-VAE and CNN-VAE) that combine the representation learning capacity of variational autoencoder (VAE) with the temporal modelling ability of a long short-term memory RNN (LSTM) and the CNN ability to learn invariant features. To figure out the performance of our proposed audio laughter generation process, objective evaluation (RMSE) and a perceptual audio quality test (listening test) were conducted. According to these evaluation metrics, we can show that the GRU-VAE outperforms the other VAE models.

A Comparative Study of the Effect of Two Analgesic Administration Methods on Post Operative Pain (수술환자에 었어서 마약성 진통제의 자가투여 방법과 근육주사 방법의 효과에 대한 비교연구)

  • 이정화
    • Journal of Korean Academy of Nursing
    • /
    • v.27 no.2
    • /
    • pp.401-410
    • /
    • 1997
  • An acute pain is the common experience following surgery. Pain is a most miserable experience in person and most preoperative patients have fear o! postoperative pain. In nursing, it is very important to understand and relieve the pain of post oprative patients as much as possible. This study was designed to compare the descriptive patterns of pain between group of Patient Controlled Analgesia and group of traditional Muscular Injection in surgcal patients. This information can be utilited as data of understanding nursing care and treatment planning for pain in surgical patients. The subjects in this study were 45 post-hysterectomy patients in Gynecology ward in C. N. U. H., in Taejon. Data was collected from May 12 to June 27. 1996. The instrumants used for this study were subjective Visual Analog Scale, Objedive nonverbal pain scale composed of Facial Apperance. Vocal Sound Change, and Sweating score. and the Melzack's Mcgill pain Qusetionaire. nine Items of Developmental Pain Intensity Scale by Lee En Ok. Analysis of data was done by using S. P. S. S. percentage, t-test, x²-test. ANOVA, and Repeated measure ANOVA. Results were obtained as follows. 1. Hypothesis 1 : There was very highly statistically significant difference in subjective self-report pain score(Visual Analog Scale) between PCA Group and IM Group(P=0.0001). 2. Hypothesis 2 : There was very highly statistically significant difference in muscle strength score (Visual Analog Scale) between PCA Group and IM group(P0.0001). 3. Hypothesis 3 : There was very highly statistically significant difference in facial appearance score between PCA Group and IM group(P=0.0001). 4. Hypothesis 4 : There was very highly statistically significance difference in vocal sound change score between PCA Group and IM group(P=0.0001). 5. Hypothesis 5 : There was no statistically significant difference sweating scores between PCA group and IM group(F=2.50, P=0.1220). But, postoperation time of 12, 24 was statistically difference between two groups(P=0.0001). So, it was partially supported. 6. Hypothesis 6 : There was very highly statistically significant difference in vocabulary pain score between PCA Group and IM group. 7. Hypothesis 7 : There was very highly statistically significant difference in amounts of total analgesic between PCA Group and IM group. There was very highly statistically significant difference in Visual Analog Pain Score, Facial Appearance Score, Vocal Sound Change Score, Vocabulary Score, amounts of total analgesic between PCA group and IM group. So, It is verified to asses of postoperative pain with VAS, Checklist of facia appearance, vocal sound change, and sweating, and Vocabulary Scale.

  • PDF

The Emotional Boundary Decision in a Linear Affect-Expression Space for Effective Robot Behavior Generation (효과적인 로봇 행동 생성을 위한 선형의 정서-표정 공간 내 감정 경계의 결정 -비선형의 제스처 동기화를 위한 정서, 표정 공간의 영역 결정)

  • Jo, Su-Hun;Lee, Hui-Sung;Park, Jeong-Woo;Kim, Min-Gyu;Chung, Myung-Jin
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.540-546
    • /
    • 2008
  • In the near future, robots should be able to understand human's emotional states and exhibit appropriate behaviors accordingly. In Human-Human Interaction, the 93% consist of the speaker's nonverbal communicative behavior. Bodily movements provide information of the quantity of emotion. Latest personal robots can interact with human using multi-modality such as facial expression, gesture, LED, sound, sensors and so on. However, a posture needs a position and an orientation only and in facial expression or gesture, movements are involved. Verbal, vocal, musical, color expressions need time information. Because synchronization among multi-modalities is a key problem, emotion expression needs a systematic approach. On the other hand, at low intensity of surprise, the face could be expressed but the gesture could not be expressed because a gesture is not linear. It is need to decide the emotional boundaries for effective robot behavior generation and synchronization with another expressible method. If it is so, how can we define emotional boundaries? And how can multi-modality be synchronized each other?

  • PDF

Korean first graders' word decoding skills, phonological awareness, rapid automatized naming, and letter knowledge with/without developmental dyslexia (초등 1학년 발달성 난독 아동의 낱말 해독, 음운인식, 빠른 이름대기, 자소 지식)

  • Yang, Yuna;Pae, Soyeong
    • Phonetics and Speech Sciences
    • /
    • v.10 no.2
    • /
    • pp.51-60
    • /
    • 2018
  • This study aims to compare the word decoding skills, phonological awareness (PA), rapid automatized naming (RAN) skills, and letter knowledge of first graders with developmental dyslexia (DD) and those who were typically developing (TD). Eighteen children with DD and eighteen TD children, matched by nonverbal intelligence and discourse ability, participated in the study. Word decoding of Korean language-based reading assessment(Pae et al., 2015) was conducted. Phoneme-grapheme correspondent words were analyzed according to whether the word has meaning, whether the syllable has a final consonant, and the position of the grapheme in the syllable. Letter knowledge asked about the names and sounds of 12 consonants and 6 vowels. The children's PA of word, syllable, body-coda, and phoneme blending was tested. Object and letter RAN was measured in seconds. The decoding difficulty of non-words was more noticeable in the DD group than in the TD one. The TD children read the syllable initial and syllable final position with 99% correctness. Children with DD read with 80% and 82% correctness, respectively. In addition, the DD group had more difficulty in decoding words with two patchims when compared with the TD one. The DD group read only 57% of words with two patchims correctly, while the TD one read 91% correctly. There were significant differences in body-coda PA, phoneme level PA, letter RAN, object RAN, and letter-sound knowledge between the two groups. This study confirms the existence of Korean developmental dyslexics, and the urgent need for the inclusion of a Korean-specific phonics approach in the education system.

Classification of Infant Crying Audio based on 3D Feature-Vector through Audio Data Augmentation

  • JeongHyeon Park;JunHyeok Go;SiUng Kim;Nammee Moon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.9
    • /
    • pp.47-54
    • /
    • 2023
  • Infants utilize crying as a non-verbal means of communication [1]. However, deciphering infant cries presents challenges. Extensive research has been conducted to interpret infant cry audios [2,3]. This paper proposes the classification of infant cries using 3D feature vectors augmented with various audio data techniques. A total of 5 classes (belly pain, burping, discomfort, hungry, tired) are employed in the study dataset. The data is augmented using 5 techniques (Pitch, Tempo, Shift, Mixup-noise, CutMix). Tempo, Shift, and CutMix augmentation techniques demonstrated improved performance. Ultimately, applying effective data augmentation techniques simultaneously resulted in a 17.75% performance enhancement compared to models using single feature vectors and original data.

The Synesthetic Presence and Physical Movement of Nong-ak as Seen Through Affect Theory (정동 이론으로 본 농악의 공감각적 현존과 신체 운동)

  • Kwon, Eun-Young
    • (The) Research of the performance art and culture
    • /
    • no.40
    • /
    • pp.5-35
    • /
    • 2020
  • Affect is intensity and quality that are generated as the physical body senses the outside world. Of experienced affect, notions that are granted meaning and interpretation are emotions. Affect theory distinguishes emotion and affect and by focusing on affect, it provides methods with which to analyze physical body responses and changes and it presents new possibilities to performing arts research that uses the physical body as a medium. Nong-ak is art that concentrates mainly on the occurrence of affect rather than 'representation'. Nong-ak is a performance type in which sound, color, texture, and physical movement overlap and exist in a synesthetic way. Here, physical things such as instruments, props, costumes, and stage devices are gathered together with non-physical things such as rhythm, mood, and atmosphere around human bodies. The physical body is stimulated by these things, displays tendencies that suit performances, and becomes 'the body without an image' as it immerses itself into the performance, acting while displaying 'quasi-corporeality'. The body, which moves automatically as if without consciousness, appears more easily within groups. To transition individuals of everyday life to 'the body without an image', Nong-ak executes the group physical exercise of 'Jinpuri'. Such physical exercise builds up affect by increasing nonverbal communion and communication and brings out the creativity of individuals within mutual trust and a sense of belonging. Affect and emotion stirred up by Nong-ak act as confirmation and affirmation of the existence, vitality, and ability of one's self and groups. Such affirmation recalls Nong-ak as a meaningful and important value from group dimensions and perceives it as a performance form that should be preserved and passed on.