• Title/Summary/Keyword: user command

Search Result 253, Processing Time 0.036 seconds

Human-Computer Interaction Based Only on Auditory and Visual Information

  • Sha, Hui;Agah, Arvin
    • Transactions on Control, Automation and Systems Engineering
    • /
    • 제2권4호
    • /
    • pp.285-297
    • /
    • 2000
  • One of the research objectives in the area of multimedia human-computer interaction is the application of artificial intelligence and robotics technologies to the development of computer interfaces. This involves utilizing many forms of media, integrating speed input, natural language, graphics, hand pointing gestures, and other methods for interactive dialogues. Although current human-computer communication methods include computer keyboards, mice, and other traditional devices, the two basic ways by which people communicate with each other are voice and gesture. This paper reports on research focusing on the development of an intelligent multimedia interface system modeled based on the manner in which people communicate. This work explores the interaction between humans and computers based only on the processing of speech(Work uttered by the person) and processing of images(hand pointing gestures). The purpose of the interface is to control a pan/tilt camera to point it to a location specified by the user through utterance of words and pointing of the hand, The systems utilizes another stationary camera to capture images of the users hand and a microphone to capture the users words. Upon processing of the images and sounds, the systems responds by pointing the camera. Initially, the interface uses hand pointing to locate the general position which user is referring to and then the interface uses voice command provided by user to fine-the location, and change the zooming of the camera, if requested. The image of the location is captured by the pan/tilt camera and sent to a color TV monitor to be displayed. This type of system has applications in tele-conferencing and other rmote operations, where the system must respond to users command, in a manner similar to how the user would communicate with another person. The advantage of this approach is the elimination of the traditional input devices that the user must utilize in order to control a pan/tillt camera, replacing them with more "natural" means of interaction. A number of experiments were performed to evaluate the interface system with respect to its accuracy, efficiency, reliability, and limitation.

  • PDF

한국군 전술컴퓨터의 인간공학적 메인버튼 설계 (User-interface Considerations for the Main Button Layout of the Tactical Computer for Korea Army)

  • 백승창;정의승;박성준
    • 대한인간공학회지
    • /
    • 제28권4호
    • /
    • pp.147-154
    • /
    • 2009
  • The tactical computer is currently being developed and installed in armored vehicles and tanks for reinforcement. With the tactical computer, Korea Army will be able to grasp the deployment status of our forces, enemy, and obstacles under varying situations. Furthermore, it makes the exchange of command and tactical intelligence possible. Recent studies showed that the task performance is greatly affected by the user interface. The U.S. Army is now conducting user-centered evaluation tests based on C2 (Command & Control) to develop tactical intelligence machinery and tools. This study aims to classify and regroup subordinate menu functions according to the user-centered task performance for the Korea Army's tactical computer. Also, the research suggests an ergonomically sound layout and size of main touch buttons by considering human factors guidelines for button design. To achieve this goal, eight hierarchical subordinate menu functions are initially drawn through clustering analysis and then each group of menu functions was renamed. Based on the suggested menu structure, new location and size of the buttons were tested in terms of response time, number of error, and subjective preference by comparing them to existing ones. The result showed that the best performance was obtained when the number of buttons or functions was eight to conduct tactical missions. Also, the improved button size and location were suggested through the experiment. It was found in addition that the location and size of the buttons had interactions regarding the user's preference.

VoiceXML을 이용한 VUI 개발에 관한 연구 (A Study on Development of VUI(Voice User Interface) using VoiceXML)

  • 장민석;양운모
    • 한국정보과학회:학술대회논문집
    • /
    • 한국정보과학회 2002년도 봄 학술발표논문집 Vol.29 No.1 (A)
    • /
    • pp.349-351
    • /
    • 2002
  • 한국현재의 컴퓨팅환경은 Text위주의 Command Line상에서의 입출력에서 GUI(Graphic User Interface)환경으로 전환되었다. 이는 사용자에게 좀더 친근한 방법으로의 컴퓨팅환경을 제공하고 있는 것이다. 하지만 아직까지 그러한 환경에 익숙해지기 위해서는 많은 습득시간이 필요하며 또한, 응용프로그램 간의 인터페이싱 기능 등을 익히기 위해서는 추가적인 학습을 통해야 원활한 작업을 수행할 수 있다. 이를 해결하고자 본 연구는 음성인식/ 합성과, 현재 음성마크업 언어인 VoiceXML 등을 통해서 모색해보고자 한다.

  • PDF

VoiceXML을 이용한 VUI 개발에 관한 연구 (A Study on Development of VUI(Voice User Interface) using VoiceXML)

  • 장민석;양운모
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2002년도 춘계학술발표논문집 (하)
    • /
    • pp.1495-1498
    • /
    • 2002
  • 한국현재의 컴퓨팅환경은 Text위주의 Command Line상에서의 입출력에서 GUI(Graphic User Interface)환경으로 전환되었다. 이는 사용자에게 좀더 친근한 방법으로의 컴퓨팅환경을 제공하고 있는 것이다. 하지만 아직까지 그러한 환경에 익숙해지기 위해서는 많은 습득시간이 필요하며 또한 응용프로그램간의 인터페이싱 기능 등을 익히기 위해서는 추가적인 학습을 통해야 원활한 작업을 수행할 수 있다.이를 해결하고자 본 연구는 음성인식/ 합성과, 현재 음성마크업 언어인 VoiceXML 등을 통해서 모색해보고자 한다.

  • PDF

사용자 의존적 인터페이스를 위한 제스처 명령어 습득 ((Learning Gesture Command for User-Dependent Interface))

  • 양선옥;양선옥
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 1997년도 춘계학술대회 학술발표 논문집
    • /
    • pp.215-219
    • /
    • 1997
  • 명령어로 사용되는 핸드 제스처가 사용자에게 더 많은 친밀감을 주기 위해서는 사용자가 자신의 원하는 형태로 제스처를 정의할 수 있어야 한다. 본 논문에서는 카메라를 통해 입력되는 사용자의 핸드 제스처를 명령어로 이용하는 지능적 사용자 인터페이스에 대하여 소개한다. 지능적 사용자 인터페이스는 제스처 명령어로 이용되는 핸드 제스처의 종류를 사용자가 임의대로 정의할 수 있도록 제스처 명령어 습득 모듈을 포함한다.

  • PDF

소프트 컴퓨팅에 의한 지능형 주행 판단 시스템 (A Judgment System for Intelligent Movement Using Soft Computing)

  • 최우경;서재용;김성현;유성욱;전홍태
    • 한국지능시스템학회논문지
    • /
    • 제16권5호
    • /
    • pp.544-549
    • /
    • 2006
  • 본 논문은 인간의 보조 역할을 하기 위해 자율적인 명령을 내리고 사용자가 직접 제어할 수 있는 지능형 주행 판단 시스템(Judgment System for Intelligent Movement; JSIM)에 대한 연구이다. 본 논문에서는 제어 대상은 이동 로봇으로 한정한다. 이동 로봇은 지능형 주행 판단 모듈을 휴대한 사용자에게 영상정보와 초음파 센서 정보를 제공하고 가이드 역할을 수행한다. 그리고 PDA와 센서박스로 구성된 지능형 주행 판단 시스템은 이동로봇으로부터 얻은 정보와 사용자 명령을 입력으로 사용하는 소프트 컴퓨팅 기법을 이용하여 이동로봇의 속도와 방향을 결정하고 다양한 기능을 수행하도록 로봇을 원격으로 제어한다. 본 논문에서는 몸에 착용하고 주변장치들과 통신을 하며 지능적 판단을 할 수 있는 지능형 주행 판단시스템을 구성하고 실제 환경에서 지능적 판단 알고리즘 적용과 이동로봇을 제어하는 시스템을 구현하여 제안한 시스템의 실현 가능성을 검증한다. 지능 알고리즘은 계층적 퍼지 구조와 신경망을 융합한 구조이다.

HMM을 기반으로 한 자율이동로봇의 음성명령 인식시스템의 개발 (Development of Autonomous Mobile Robot with Speech Teaching Command Recognition System Based on Hidden Markov Model)

  • 조현수;박민규;이현정;이민철
    • 제어로봇시스템학회논문지
    • /
    • 제13권8호
    • /
    • pp.726-734
    • /
    • 2007
  • Generally, a mobile robot is moved by original input programs. However, it is very hard for a non-expert to change the program generating the moving path of a mobile robot, because he doesn't know almost the teaching command and operating method for driving the robot. Therefore, the teaching method with speech command for a handicapped person without hands or a non-expert without an expert knowledge to generate the path is required gradually. In this study, for easily teaching the moving path of the autonomous mobile robot, the autonomous mobile robot with the function of speech recognition is developed. The use of human voice as the teaching method provides more convenient user-interface for mobile robot. To implement the teaching function, the designed robot system is composed of three separated control modules, which are speech preprocessing module, DC servo motor control module, and main control module. In this study, we design and implement a speaker dependent isolated word recognition system for creating moving path of an autonomous mobile robot in the unknown environment. The system uses word-level Hidden Markov Models(HMM) for designated command vocabularies to control a mobile robot, and it has postprocessing by neural network according to the condition based on confidence score. As the spectral analysis method, we use a filter-bank analysis model to extract of features of the voice. The proposed word recognition system is tested using 33 Korean words for control of the mobile robot navigation, and we also evaluate the performance of navigation of a mobile robot using only voice command.

A new human-robot interaction method using semantic symbols

  • Park, Sang-Hyun;Hwang, Jung-Hoon;Kwon, Dong-Soo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2004년도 ICCAS
    • /
    • pp.2005-2010
    • /
    • 2004
  • As robots become more prevalent in human daily life, situations requiring interaction between humans and robots will occur more frequently. Therefore, human-robot interaction (HRI) is becoming increasingly important. Although robotics researchers have made many technical developments in their field, intuitive and easy ways for most common users to interact with robots are still lacking. This paper introduces a new approach to enhance human-robot interaction using a semantic symbol language and proposes a method to acquire the intentions of robot users. In the proposed approach, each semantic symbol represents knowledge about either the environment or an action that a robot can perform. Users'intentions are expressed by symbolized multimodal information. To interpret a users'command, a probabilistic approach is used, which is appropriate for interpreting a freestyle user expression or insufficient input information. Therefore, a first-order Markov model is constructed as a probabilistic model, and a questionnaire is conducted to obtain state transition probabilities for this Markov model. Finally, we evaluated our model to show how well it interprets users'commands.

  • PDF

Investigating Smart TV Gesture Interaction Based on Gesture Types and Styles

  • Ahn, Junyoung;Kim, Kyungdoh
    • 대한인간공학회지
    • /
    • 제36권2호
    • /
    • pp.109-121
    • /
    • 2017
  • Objective: This study aims to find suitable types and styles for gesture interaction as remote control on smart TVs. Background: Smart TV is being developed rapidly in the world, and gesture interaction has a wide range of research areas, especially based on vision techniques. However, most studies are focused on the gesture recognition technology. Also, not many previous studies of gestures types and styles on smart TVs were carried out. Therefore, it is necessary to check what users prefer in terms of gesture types and styles for each operation command. Method: We conducted an experiment to extract the target user manipulation commands required for smart TVs and select the corresponding gestures. To do this, we looked at gesture styles people use for every operation command, and checked whether there are any gesture styles they prefer over others. Through these results, this study was carried out with a process selecting smart TV operation commands and gestures. Results: Eighteen TV commands have been used in this study. With agreement level as a basis, we compared the six types of gestures and five styles of gestures for each command. As for gesture type, participants generally preferred a gesture of Path-Moving type. In the case of Pan and Scroll commands, the highest agreement level (1.00) of 18 commands was shown. As for gesture styles, the participants preferred a manipulative style in 11 commands (Next, Previous, Volume up, Volume down, Play, Stop, Zoom in, Zoom out, Pan, Rotate, Scroll). Conclusion: By conducting an analysis on user-preferred gestures, nine gesture commands are proposed for gesture control on smart TVs. Most participants preferred Path-Moving type and Manipulative style gestures based on the actual operations. Application: The results can be applied to a more advanced form of the gestures in the 3D environment, such as a study on VR. The method used in this study will be utilized in various domains.

문맥 및 사용 패턴 정보를 이용한 음성인식의 성능 개선 (Performance Improvement of Speech Recognition Using Context and Usage Pattern Information)

  • 송원문;김명원
    • 정보처리학회논문지B
    • /
    • 제13B권5호
    • /
    • pp.553-560
    • /
    • 2006
  • 최근 음성인식에서는 잡음환경에서 좀 더 신뢰성 있는 결과를 얻기 위해 인식 결과 도출 단계에서 여러 가지 정보의 내용들을 융합하거나 이전 인식 결과의 후처리를 통하여 성능을 향상시키는 방법들이 연구되고 있다. 본 논문에서는 잡음 환경에서의 인식률 하락을 보완하기 위해 개인 모바일 기기를 위한 음성 명령어 인식에서 사용자의 사용패턴과 문맥 정보를 사용하는 방법을 제안한다. 기본 인식 결과를 보정하기 위해서 현재 명령어를 발화하기 이전에 사용자가 사용한 순차적 명령어 패턴을 사용하였다. 또한 문맥 정보를 위해서는 사용중인 기기의 현재 기능과 발화된 명령어간의 연관성을 사용하였다. 실험을 통해 제안한 방법이 기본 인식 시스템에서 발생한 오인식의 약 50%를 수정하였음을 보였으며 이로써 제안한 방법의 타당성을 검증하였다.