• Title/Summary/Keyword: Voice Interaction

Search Result 157, Processing Time 0.021 seconds

Agent Mobility in Human Robot Interaction

  • Nguyen, To Dong;Oh, Sang-Rok;You, Bum-Jae
    • Proceedings of the KIEE Conference
    • /
    • 2005.07d
    • /
    • pp.2771-2773
    • /
    • 2005
  • In network human-robot interaction, human can access services of a robot system through the network The communication is done by interacting with the distributed sensors via voice, gestures or by using user network access device such as computer, PDA. The service organization and exploration is very important for this distributed system. In this paper we propose a new agent-based framework to integrate partners of this distributed system together and help users to explore the service effectively without complicated configuration. Our system consists of several robots. users and distributed sensors. These partners are connected in a decentralized but centralized control system using agent-based technology. Several experiments are conducted successfully using our framework The experiments show that this framework is good in term of increasing the availability of the system, reducing the time users and robots needs to connect to the network at the same time. The framework also provides some coordination methods for the human robot interaction system.

  • PDF

Effects on the continuous use intention of AI-based voice assistant services: Focusing on the interaction between trust in AI and privacy concerns (인공지능 기반 음성비서 서비스의 지속이용 의도에 미치는 영향: 인공지능에 대한 신뢰와 프라이버시 염려의 상호작용을 중심으로)

  • Jang, Changki;Heo, Deokwon;Sung, WookJoon
    • Informatization Policy
    • /
    • v.30 no.2
    • /
    • pp.22-45
    • /
    • 2023
  • In research on the use of AI-based voice assistant services, problems related to the user's trust and privacy protection arising from the experience of service use are constantly being raised. The purpose of this study was to investigate empirically the effects of individual trust in AI and online privacy concerns on the continued use of AI-based voice assistants, specifically the impact of their interaction. In this study, question items were constructed based on previous studies, with an online survey conducted among 405 respondents. The effect of the user's trust in AI and privacy concerns on the adoption and continuous use intention of AI-based voice assistant services was analyzed using the Heckman selection model. As the main findings of the study, first, AI-based voice assistant service usage behavior was positively influenced by factors that promote technology acceptance, such as perceived usefulness, perceived ease of use, and social influence. Second, trust in AI had no statistically significant effect on AI-based voice assistant service usage behavior but had a positive effect on continuous use intention. Third, the privacy concern level was confirmed to have the effect of suppressing continuous use intention through interaction with trust in AI. These research results suggest the need to strengthen user experience through user opinion collection and action to improve trust in technology and alleviate users' concerns about privacy as governance for realizing digital government. When introducing artificial intelligence-based policy services, it is necessary to disclose transparently the scope of application of artificial intelligence technology through a public deliberation process, and the development of a system that can track and evaluate privacy issues ex-post and an algorithm that considers privacy protection is required.

Object Magnification and Voice Command in Gaze Interface for the Upper Limb Disabled (상지장애인을 위한 시선 인터페이스에서의 객체 확대 및 음성 명령 인터페이스 개발)

  • Park, Joo Hyun;Jo, Se-Ran;Lim, Soon-Bum
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.7
    • /
    • pp.903-912
    • /
    • 2021
  • Eye tracking research for upper limb disabilities is showing an effect in the aspect of device control. However, the reality is that it is not enough to perform web interaction with only eye tracking technology. In the Eye-Voice interface, a previous study, in order to solve the problem that the existing gaze tracking interfaces cause a malfunction of pointer execution, a gaze tracking interface supplemented with a voice command was proposed. In addition, the reduction of the malfunction rate of the pointer was confirmed through a comparison experiment with the existing interface. In this process, the difficulty of pointing due to the small size of the execution object in the web environment was identified as another important problem of malfunction. In this study, we propose an auto-magnification interface of objects so that people with upper extremities can freely click web contents by improving the problem that it was difficult to point and execute due to the high density of execution objects and their arrangements in web pages.

Interface Modeling for Digital Device Control According to Disability Type in Web

  • Park, Joo Hyun;Lee, Jongwoo;Lim, Soon-Bum
    • Journal of Multimedia Information System
    • /
    • v.7 no.4
    • /
    • pp.249-256
    • /
    • 2020
  • Learning methods using various assistive and smart devices have been developed to enable independent learning of the disabled. Pointer control is the most important consideration for the disabled when controlling a device and the contents of an existing graphical user interface (GUI) environment; however, difficulties can be encountered when using a pointer, depending on the disability type; Although there are individual differences depending on the blind, low vision, and upper limb disability, problems arise in the accuracy of object selection and execution in common. A multimodal interface pilot solution is presented that enables people with various disability types to control web interactions more easily. First, we classify web interaction types using digital devices and derive essential web interactions among them. Second, to solve problems that occur when performing web interactions considering the disability type, the necessary technology according to the characteristics of each disability type is presented. Finally, a pilot solution for the multimodal interface for each disability type is proposed. We identified three disability types and developed solutions for each type. We developed a remote-control operation voice interface for blind people and a voice output interface applying the selective focusing technique for low-vision people. Finally, we developed a gaze-tracking and voice-command interface for GUI operations for people with upper-limb disability.

SPACIAL POEM: A New Type of Experimental Visual Interaction in 3D Virtual Environment

  • Choi, Jin-Young
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02b
    • /
    • pp.405-410
    • /
    • 2008
  • There is always a rhythm in our language and speech. As soon as we speech out, even just simple words and voice we make are edited as various emotions and information. Through this process we succeed or fail in our communication, and it becomes a fun communication or a monotonous delivery. Even with the same music, impression of the play can be different according to each musician' s emotion and their understanding. We 'play' our language in the same way as that. However, I think, people are used to the variety, which is, in fact, the variation of a set format covered with hollow variety. People might have been living loosing or limiting their own creative way to express themselves by that hollow variety. SPACIAL POEM started from this point. This is a new type of 'real-time visual interaction' expressing our own creative narrative as real-time visual by playing a musical instrument which is an emotional human behavior. Producing many kinds of sound by playing musical instruments is the same behavior with which we express our emotions through. There are sensors on each hole on the surface of the musical instrument. When you play it, sensors recognize that you have covered the holes. All sensors are connected to a keyboard, which means your playing behavior becomes a typing action on the keyboard. And I programmed the visual of your words to spread out in a virtual 3D space when you play the musical instrument. The behavior when you blow the instrument, to make sounds, changes into the energy that makes you walk ahead continuously in a virtual space. I used a microphone sensor for this. After all by playing musical instrument, we get back the emotion we forgot so far, and my voice is expressed with my own visual language in virtual space.

  • PDF

Development of Smart Mirror System based on the Raspberry Pi (Raspberry Pi를 이용한 스마트 미러 개발)

  • Lin, Zhi-Ming;Kim, Chul-Won
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.2
    • /
    • pp.379-384
    • /
    • 2021
  • With people's continuous research and exploration in the field of artificial intelligence, more relatively mature artificial intelligence technology is applied in people's daily life. Mirrors are the most commonly used daily necessities in life, and it should be applied to artificial intelligence. The research results of this paper show that the smart mirror designed based on the raspberry pi displays weather, temperature, greetings, and has a human-mirror interaction function. The research method of this paper uses the Raspberry pi 3B + as the core controller and Google Assistant as the intelligent control. When connected to the network via Raspberry Pi's own WiFi, the mirror can automatically display and update time, weather and news information features. You can wake up the Google Assistant using keywords, then control the mirror to play music, remind the time, It implements the function of smart mirror voice interaction. Also, all the hardware used in this study is modular assembly. Later, it is convenient for user to assemble by himself later. It is suitable for market promotion at an affordable price.

HomeN manager system based on multimodal context-aware middleware (멀티모달 상황인지 미들웨어 기반의 홈앤(HomeN) 매니저 시스템)

  • Ahn, Se-Yeol;Park, Sung-Chan;Park, Seong-Soo;Koo, Myung-Wan;Jeong, Yeong-Joon;Kim, Myung-Sook
    • Proceedings of the KSPS conference
    • /
    • 2006.11a
    • /
    • pp.120-123
    • /
    • 2006
  • The provision of personalized user interfaces for mobile devices is expected to be used for different devices with a wide variety of capabilities and interaction modalities. In this paper, we implemented a multimodal context-aware middleware incorporating XML-based languages such as XHTML, VoiceXML. SCXML uses parallel states to invoke both XHTML and VoiceXML contents as well as to gather composite multimodal inputs or synchronize inter-modalities through man-machine I/Os. We developed home networking service named "HomeN" based on our middleware framework. It demonstrates that users could maintain multimodal scenarios in a clear, concise and consistent manner under various user's interactions.

  • PDF

Noise Robust Emotion Recognition Feature : Frequency Range of Meaningful Signal (음성의 특정 주파수 범위를 이용한 잡음환경에서의 감정인식)

  • Kim Eun-Ho;Hyun Kyung-Hak;Kwak Yoon-Keun
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.23 no.5 s.182
    • /
    • pp.68-76
    • /
    • 2006
  • The ability to recognize human emotion is one of the hallmarks of human-robot interaction. Hence this paper describes the realization of emotion recognition. For emotion recognition from voice, we propose a new feature called frequency range of meaningful signal. With this feature, we reached average recognition rate of 76% in speaker-dependent. From the experimental results, we confirm the usefulness of the proposed feature. We also define the noise environment and conduct the noise-environment test. In contrast to other features, the proposed feature is robust in a noise-environment.

A Novel Computer Human Interface to Remotely Pick up Moving Human's Voice Clearly by Integrating ]Real-time Face Tracking and Microphones Array

  • Hiroshi Mizoguchi;Takaomi Shigehara;Yoshiyasu Goto;Hidai, Ken-ichi;Taketoshi Mishima
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1998.10a
    • /
    • pp.75-80
    • /
    • 1998
  • This paper proposes a novel computer human interface, named Virtual Wireless Microphone (VWM), which utilizes computer vision and signal processing. It integrates real-time face tracking and sound signal processing. VWM is intended to be used as a speech signal input method for human computer interaction, especially for autonomous intelligent agent that interacts with humans like as digital secretary. Utilizing VWM, the agent can clearly listen human master's voice remotely as if a wireless microphone was put just in front of the master.

  • PDF

Design of Metaverse for Two-Way Video Conferencing Platform Based on Virtual Reality

  • Yoon, Dongeon;Oh, Amsuk
    • Journal of information and communication convergence engineering
    • /
    • v.20 no.3
    • /
    • pp.189-194
    • /
    • 2022
  • As non-face-to-face activities have become commonplace, online video conferencing platforms have become popular collaboration tools. However, existing video conferencing platforms have a structure in which one side unilaterally exchanges information, potentially increase the fatigue of meeting participants. In this study, we designed a video conferencing platform utilizing virtual reality (VR), a metaverse technology, to enable various interactions. A virtual conferencing space and realistic VR video conferencing content authoring tool support system were designed using Meta's Oculus Quest 2 hardware, the Unity engine, and 3D Max software. With the Photon software development kit, voice recognition was designed to perform automatic text translation with the Watson application programming interface, allowing the online video conferencing participants to communicate smoothly even if using different languages. It is expected that the proposed video conferencing platform will enable conference participants to interact and improve their work efficiency.