• Title/Summary/Keyword: user command

Search Result 253, Processing Time 0.027 seconds

Hand Gesture Segmentation Method using a Wrist-Worn Wearable Device

  • Lee, Dong-Woo;Son, Yong-Ki;Kim, Bae-Sun;Kim, Minkyu;Jeong, Hyun-Tae;Cho, Il-Yeon
    • Journal of the Ergonomics Society of Korea
    • /
    • v.34 no.5
    • /
    • pp.541-548
    • /
    • 2015
  • Objective: We introduce a hand gesture segmentation method using a wrist-worn wearable device which can recognize simple gestures of clenching and unclenching ones' fist. Background: There are many types of smart watches and fitness bands in the markets. And most of them already adopt a gesture interaction to provide ease of use. However, there are many cases in which the malfunction is difficult to distinguish between the user's gesture commands and user's daily life motion. It is needed to develop a simple and clear gesture segmentation method to improve the gesture interaction performance. Method: At first, we defined the gestures of making a fist (start of gesture command) and opening one's fist (end of gesture command) as segmentation gestures to distinguish a gesture. The gestures of clenching and unclenching one's fist are simple and intuitive. And we also designed a single gesture consisting of a set of making a fist, a command gesture, and opening one's fist in order. To detect segmentation gestures at the bottom of the wrist, we used a wrist strap on which an array of infrared sensors (emitters and receivers) were mounted. When a user takes gestures of making a fist and opening one's a fist, this changes the shape of the bottom of the wrist, and simultaneously changes the reflected amount of the infrared light detected by the receiver sensor. Results: An experiment was conducted in order to evaluate gesture segmentation performance. 12 participants took part in the experiment: 10 males, and 2 females with an average age of 38. The recognition rates of the segmentation gestures, clenching and unclenching one's fist, are 99.58% and 100%, respectively. Conclusion: Through the experiment, we have evaluated gesture segmentation performance and its usability. The experimental results show a potential for our suggested segmentation method in the future. Application: The results of this study can be used to develop guidelines to prevent injury in auto workers at mission assembly plants.

Adaptive Strategy Game Engine Using Non-monotonic Reasoning and Inductive Machine Learning (비단조 추론과 귀납적 기계학습 기반 적응형 전략 게임 엔진)

  • Kim, Je-Min;Park, Young-Tack
    • The KIPS Transactions:PartB
    • /
    • v.11B no.1
    • /
    • pp.83-90
    • /
    • 2004
  • Strategic games are missing special qualities of genre these days. Game engines neither reason about behaviors of computer objects nor have learning ability that can prepare countermeasure in variously command user's strategy. This paper suggests a strategic game engine that applies non-monotonic reasoning and inductive machine learning. The engine emphasizes three components -“user behavior monitor”to abstract user's objects behavior,“learning engine”to learn user's strategy,“behavior display handler”to reflect abstracted behavior of computer objects on game. Especially, this paper proposes two layered-structure to apply non-monotonic reasoning and inductive learning to make behaviors of computer objects that learns strategy behaviors of user objects exactly, and corresponds in user's objects. The engine decides actions and strategies of computer objects with created information through inductive learning. Main contribution of this paper is that computer objects command excellent strategies and reveal differentiation with behavior of existing computer objects to apply non-monotonic reasoning and inductive machine learning.

A Study of EURONET'S, Operation (EURONET의 운영에 관한 연구)

  • Lee, Jin Kyoung
    • Journal of the Korean Society for information Management
    • /
    • v.2 no.2
    • /
    • pp.170-178
    • /
    • 1985
  • The aim of this paper is to explain configuration, charge, and common command language of the EURONET that connect the Data Base of Scientific Technical Information in Europe and offer the On-line service to the user.

  • PDF

A research on man-robot cooperative interaction system

  • Ishii, Masaru
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1992.10b
    • /
    • pp.555-557
    • /
    • 1992
  • Recently, realization of an intelligent cooperative interaction system between a man and robot systems is required. In this paper, HyperCard with a voice control is used for above system because of its easy handling and excellent human interfaces. Clicking buttons in the HyperCard by a mouse device or a voice command means controlling each joint of a robot system. Robot teaching operation of grasping a bin and pouring liquid in it into a cup is carried out. This robot teaching method using HyperCard provides a foundation for realizing a user friendly cooperative interaction system.

  • PDF

CSpeech(Version 3.1)

  • Sik, Choe-Hong
    • Proceedings of the KSLP Conference
    • /
    • 1995.11a
    • /
    • pp.141-153
    • /
    • 1995
  • CSpeech is a software package that implements an audio waveform/speech analysis workstation on an IBM Personal Computer or hardware compatible computer. Features include digitizing audio waveforms on single or multiple channels, displaying the digitized waveforms, playing back audio waveforms from selected intervals of sing1e channels, saving and retrieving waveforms from binary format disk files, and analysing audio waveforms for their temporal and spectral properties. The distinguishing characteristics of CSpeech are its support for multiple channels, minimal restrictions on sample rate and waveform duration support fur a variety of hardware configurations, fast graphics display, and its user- extensible menu- based command structure.

  • PDF

Gaze Recognition Interface Development for Smart Wheelchair (지능형 휠체어를 위한 시선 인식 인터페이스 개발)

  • Park, S.H.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.5 no.1
    • /
    • pp.103-110
    • /
    • 2011
  • In this paper, we propose a gaze recognition interface for smart wheelchair. The gaze recognition interface is a user interface which recognize the commands using the gaze recognition and avoid the detected obstacles by sensing the distance through range sensors on the way to driving. Smart wheelchair is composed of gaze recognition and tracking module, user interface module, obstacle detector, motor control module, and range sensor module. The interface in this paper uses a camera with built-in infra red filter and 2 LED light sources to see what direction the pupils turn to and can send command codes to control the system, thus it doesn't need any correction process per each person. The results of the experiment showed that the proposed interface can control the system exactly by recognizing user's gaze direction.

An analysis of the component of Human-Robot Interaction for Intelligent room

  • Park, Jong-Chan;Kwon, Dong-Soo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.2143-2147
    • /
    • 2005
  • Human-Robot interaction (HRI) has recently become one of the most important issues in the field of robotics. Understanding and predicting the intentions of human users is a major difficulty for robotic programs. In this paper we suggest an interaction method allows the robot to execute the human user's desires in an intelligent room-based domain, even when the user does not give a specific command for the action. To achieve this, we constructed a full system architecture of an intelligent room so that the following were present and sequentially interconnected: decision-making based on the Bayesian belief network, responding to human commands, and generating queries to remove ambiguities. The robot obtained all the necessary information from analyzing the user's condition and the environmental state of the room. This information is then used to evaluate the probabilities of the results coming from the output nodes of the Bayesian belief network, which is composed of the nodes that includes several states, and the causal relationships between them. Our study shows that the suggested system and proposed method would improve a robot's ability to understand human commands, intuit human desires, and predict human intentions resulting in a comfortable intelligent room for the human user.

  • PDF

Implementation of Data Storage Media Control and Command(DSM-CC) Core User-to-User Interface for MPEG-2 Bit Stream Transport

  • Park, Seong-Jong;Kim, Yong-Han;Kim, Jae-Woo;Lee, Ho-Jang;Shim, Jae-Kyu;Kim, Jae-D.;Koh, Jong-Seong
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1998.06b
    • /
    • pp.79-84
    • /
    • 1998
  • This paper describes implementation of the core DSM-CC UU interface. It briefly describes the reference model for the DSM-CC and related standards that should be reviewed for the implementation. The Common Object Request Broker Architechture, Revision 2.0 (CORBA 2.0) is sued as a remote procedure call (RPC) scheme for the UU Interface. Entire system was implemented with C++ on Windows NT platforms. The implementation procedure has been decomposed ito two tasks. The first task is to implement the Naming Service for service navigation. The Naming Service is one of the CORBA Services that extend the core CORBA specification. A client GUI is implemented for easy navigation among various services. The second task is to construct multimedia server and client for a Video-on-Demand (VoD) system. MPEG-2 Transport Stream is transported via ATM AAL5 using the Windows Socket 2.2 ATM extension API. A GUI enables the user to navigate the service domain and select a program. After the selection the user can control the MPEG-2 stream with VCR-like buttons.

  • PDF

A mixed-initiative conversational agent for ubiquitous home environments (유비쿼터스 가정환경을 위한 상호주도형 대화 에이전트)

  • Song In-Jee;Hong Jin-Hyuk;Cho Sung-Bae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.15 no.7
    • /
    • pp.834-839
    • /
    • 2005
  • When a great variety of services become available to user through the broadband convergence network in the ubiquitous home environment, an intelligent agent is required to deal with the complexity of services and perceive intension of a user. Different from the old-fashioned command-based user interface for selecting services, conversation enables flexible and rich interactions between human and agents, but diverse expressions of the user's background and context make conversation hard to implement by using either user-initiative or system-initiative methods. To deal with the ambiguity of diverse expressions between user and agents, we have to apply hierarchial bayesian networks for the mixed initiative conversation. Missing information from user's query is analyzed by hierarchial bayesian networks to inference the user's intension so that can be collected through the agent's query. We have implemented this approach in ubiquitous home environment by implementing simulation program.

Design of a Multi-level VHDL Simulator (다층 레벨 VHDL 시뮬레이터의 설계)

  • 이영희;김헌철;황선영
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.30A no.10
    • /
    • pp.67-76
    • /
    • 1993
  • This paper presents the design and implementation of SVSIM (Sogang VHDL SIMulator), a multi-level VHDL simulator, designed for the construction of an integrated VGDL design environment. The internal model of SVSIM is the hierarchical C/DFG which is extended from C/DFG to include the network hierarchy and local/glabal control informations. Hierarchical network is not flattened for simulation, resulting in the reduction of space complexity. The predufined/user-defined types except for the record type and the predefined/user-defined attributes are supported in SVSIM. Algorithmic-level descriptions can be siumlated by the support of recursive procedure/function calls. Input stimuli can be generated by command script in stimuli file or in VHDL source code. Experimential results show SVSIM can be efficiently used for the simulation of the pure behavioral descriptions, structural descriptions or mixture of these.

  • PDF