• Title/Summary/Keyword: voice user interface

Search Result 146, Processing Time 0.048 seconds

The Effects of Multi-Modality on the Use of Smart Phones

  • Lee, Gaeun;Kim, Seongmin;Choe, Jaeho;Jung, Eui Seung
    • Journal of the Ergonomics Society of Korea
    • /
    • v.33 no.3
    • /
    • pp.241-253
    • /
    • 2014
  • Objective: The objective of this study was to examine multi-modal interaction effects of input-mode switching on the use of smart phones. Background: Multi-modal is considered as an efficient alternative for input and output of information in mobile environments. However, there are various limitations in current mobile UI (User Interface) system that overlooks the transition between different modes or the usability of a combination of multi modal uses. Method: A pre-survey determined five representative tasks from smart phone tasks by their functions. The first experiment involved the use of a uni-mode for five single tasks; the second experiment involved the use of a multi-mode for three dual tasks. The dependent variables were user preference and task completion time. The independent variable in the first experiment was the type of modes (i.e., Touch, Pen, or Voice) while the variable in the second experiment was the type of tasks (i.e., internet searching, subway map, memo, gallery, and application store). Results: In the first experiment, there was no difference between the uses of pen and touch devices. However, a specific mode type was preferred depending on the functional characteristics of the tasks. In the second experiment, analysis of results showed that user preference depended on the order and combination of modes. Even with the transition of modes, users preferred the use of multi-modes including voice. Conclusion: The order of combination of modes may affect the usability of multi-modes. Therefore, when designing a multi-modal system, the fact that there are frequent transitions between various mobile contents in different modes should be properly considered. Application: It may be utilized as a user-centered design guideline for mobile multi modal UI system.

Implementation of Java based SIP User Agent Including RTP transmission module (RTP 전송 모듈을 포함한 Java 기반의 SIP User Agent의 구현)

  • 조현규;김영학;장춘서
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.10e
    • /
    • pp.142-144
    • /
    • 2002
  • VoIP(Voice over IP) 시스템을 구현함에 있어서 호설정을 처리하는 여러 프로토콜이 제안되고 있는 가운데 IETF(Internet Engineering Task Force)에서 제안한 SIP(Session Initiation Protocol)는 텍스트 기반의 프로토콜로서 구현과 파싱이 쉬운 등 많은 장점을 가지고 있어 차세대 VoIP의 표준으로 자리잡고 있다. 또한, 뛰어난 확장성을 가지고 있어 다양한 서비스에 적용할 수 있는 호설정 프로토콜이다. 본 논문에서는 SIP를 이용한 VoIP 시스템을 구현함에 있어 주요 구성요소 중 하나인 UA(User Agent)를 2002년 6월에 발표된 새로운 SIP 버전에 맞추어 개발하였다. 본 UA는 플랫폼에 독립적으로 기능을 할 수 있도록 자바(Java)를 사용하여 GUI(Graphical User Interface)환경으로 구현하였다 그리고 RTP(Real-time Transport Protocol) 전송 모듈을 통하여 호설정이 이루어진 후 실제 음성과 화상통신이 이루어지는 부분을 포함하였다.

  • PDF

Improving Eye-gaze Mouse System Using Mouth Open Detection and Pop Up Menu (입 벌림 인식과 팝업 메뉴를 이용한 시선추적 마우스 시스템 성능 개선)

  • Byeon, Ju Yeong;Jung, Keechul
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.12
    • /
    • pp.1454-1463
    • /
    • 2020
  • An important factor in eye-tracking PC interface for general paralyzed patients is the implementation of the mouse interface, for manipulating the GUI. With a successfully implemented mouse interface, users can generate mouse events exactly at the point of their choosing. However, it is difficult to define this interaction in the eye-tracking interface. This problem has been defined as the Midas touch problem and has been a major focus of eye-tracking research. There have been many attempts to solve this problem using blink, voice input, etc. However, it was not suitable for general paralyzed patients because some of them cannot wink or speak. In this paper, we propose a mouth-pop-up, eye-tracking mouse interface that solves the Midas touch problem as well as becoming a suitable interface for general paralyzed patients using a common RGB camera. The interface presented in this paper implements a mouse interface that detects the opening and closing of the mouth to activate a pop-up menu that the user can select the mouse event. After implementation, a performance experiment was conducted. As a result, we found that the number of malfunctions and the time to perform tasks were reduced compared to the existing method.

An Implementation of Travel Information Service Using VoiceXML and GPS (VoiceXML과 GPS를 이용한 여행정보 서비스의 구현)

  • Oh, Jae-Gyu;Kim, Sun-Hyung
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.8 no.6
    • /
    • pp.1443-1448
    • /
    • 2007
  • In this paper, we implement a distributed computing environment-based travel information service that can use web(internet) and speech interface at the same time and can apply location information, using voice and web browser-based VoiceXML and GPS, to escape the limitations of traditional web(internet)-based travel information services. Because of IVR(Interactive Voice Response) of traditional call center has operated to a pre-installation scenario, it takes much a service time and has the inconveniences that must repeat speech recording according to the revised scenarios in case change response contents. However, suggested VoiceXML and GPS-based travel information service system has advantages that reorganization of system setups is easy, because it consists of the method to update server after make individual conversation scenarios by file format(document), and can provide usefully various travel information in environmental restriction conditions such as the back regions environment, according as our prototype find user's present location using GPS information and then provide various travel information service by this information.

  • PDF

The Implementation of an ISDN System-on-a-Chip and communication terminal (ISDN 멀티미디어 통신단말용 시스템-온-칩 및 소프트웨어 구현)

  • 김진태;황대환
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.6 no.3
    • /
    • pp.410-415
    • /
    • 2002
  • This paper describes the implementation of a SoC(System-on-a-Chip) and an ISDN communication terminal by the SoC in ISDN network. The SoC has been developed with the functions of 32-bit ARM7TDMI RISC core processor, network connection with S/T interface, TDM--bus interface and voice codec, user interface. And we also review the developed software structure and the ISDN service protocol procedures which are working on the SoC. And finally this paper describers a structure of an ISDN terminal equipment using the implemented SoC and terminal software.

Design and Implementation of Multimodal Middleware for Mobile Environments (모바일 환경을 위한 멀티모달 미들웨어의 설계 및 구현)

  • Park, Seong-Soo;Ahn, Se-Yeol;Kim, Won-Woo;Koo, Myoung-Wan;Park, Sung-Chan
    • MALSORI
    • /
    • no.60
    • /
    • pp.125-144
    • /
    • 2006
  • W3C announced a standard software architecture for multimodal context-aware middleware that emphasizes modularity and separates structure, contents, and presentation. We implemented a distributed multimodal interface system followed the W3C architecture, based on SCXML. SCXML uses parallel states to invoke both XHTML and VoiceXML contents as well as to gather composite or sequential multimodal inputs through man-machine interactions. We also hire Delivery Context Interface(DCI) module and an external service bundle enabling middleware to support context-awareness services for real world environments. The provision of personalized user interfaces for mobile devices is expected to be used for different devices with a wide variety of capabilities and interaction modalities. We demonstrated the implemented middleware could maintain multimodal scenarios in a clear, concise and consistent manner by some experiments.

  • PDF

Development of an Operational MICOM for efficient Central Control of Public Address System (통합 전관 방송 시스템의 효율적인 중앙 집중 제어를 위한 운용 MICOM 개발)

  • Kim, Jung-Sook
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.6 no.2
    • /
    • pp.325-329
    • /
    • 2011
  • We are developing a minimized integrated public address system that has facilities such as external input, Mic., CD, MP3 and Radio. In this paper, developing an integrated public address system with operational MICOM will make it possible to control the facilities of digital devices centrally. The operational MICOM is composed of 3 layers which are a control layer, a processing layer and a user interface layer. The control layer controls digital devices in the integrated public address and the processing layer is for processing voice data editing, controlling signals of devices and providing status information of the device. The user interface layer displays the status of control devices and the data processing of a user. A remote control is used for the central control.

A Multimodal Interface for Telematics based on Multimodal middleware (미들웨어 기반의 텔레매틱스용 멀티모달 인터페이스)

  • Park, Sung-Chan;Ahn, Se-Yeol;Park, Seong-Soo;Koo, Myoung-Wan
    • Proceedings of the KSPS conference
    • /
    • 2007.05a
    • /
    • pp.41-44
    • /
    • 2007
  • In this paper, we introduce a system in which car navigation scenario is plugged multimodal interface based on multimodal middleware. In map-based system, the combination of speech and pen input/output modalities can offer users better expressive power. To be able to achieve multimodal task in car environments, we have chosen SCXML(State Chart XML), a multimodal authoring language of W3C standard, to control modality components as XHTML, VoiceXML and GPS. In Network Manager, GPS signals from navigation software are converted to EMMA meta language, sent to MultiModal Interaction Runtime Framework(MMI). Not only does MMI handles GPS signals and a user's multimodal I/Os but also it combines them with information of device, user preference and reasoned RDF to give the user intelligent or personalized services. The self-simulation test has shown that middleware accomplish a navigational multimodal task over multiple users in car environments.

  • PDF

A Fuzzy-Neural Network Based Human-Machine Interface for Voice Controlled Robots Trained by a Particle Swarm Optimization

  • Watanabe, Keigo;Chatterjee, Amitava;Pulasinghe, Koliya;Izumi, Kiyotaka;Kiguchi, Kazuo
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.09a
    • /
    • pp.411-414
    • /
    • 2003
  • Particle swarm optimization (PSO) is employed to train fuzzy-neural networks (FNN), which can be employed as an important building block in real life robot systems, controlled by voice-based commands. The FNN is also trained to capture the user spoken directive in the context of the present performance of the robot system. The system has been successfully employed in a real life situation for navigation of a mobile robot.

  • PDF

Intelligent Retrieval System with Interactive Voice Support (대화형 음성 지원을 통한 지능형 검색 시스템)

  • Moon, K.J.;Yoo, Y.S.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.9 no.1
    • /
    • pp.29-35
    • /
    • 2015
  • In this paper, we propose a intelligent retrieval system with interactive voice support. The developed system helps to find misrecognized words by using the relationship between lexical items in a sentence recognition and present the correct vocabulary. In this study, we implement a simulation system that can be proposed to determine the usefulness of the product search assistance system which offers applications. Experimental results were confirmed to correct the wrong speech recognition vocabulary in a simple user interface to help the product search.

  • PDF