• Title/Summary/Keyword: Natural language interfaces

Search Result 17, Processing Time 0.026 seconds

Design of Gesture based Interfaces for Controlling GUI Applications (GUI 어플리케이션 제어를 위한 제스처 인터페이스 모델 설계)

  • Park, Ki-Chang;Seo, Seong-Chae;Jeong, Seung-Moon;Kang, Im-Cheol;Kim, Byung-Gi
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.1
    • /
    • pp.55-63
    • /
    • 2013
  • NUI(Natural User Interfaces) has been developed through CLI(Command Line Interfaces) and GUI(Graphical User Interfaces). NUI uses many different input modalities, including multi-touch, motion tracking, voice and stylus. In order to adopt NUI to legacy GUI applications, he/she must add device libraries, modify relevant source code and debug it. In this paper, we propose a gesture-based interface model that can be applied without modification of the existing event-based GUI applications and also present the XML schema for the specification of the model proposed. This paper shows a method of using the proposed model through a prototype.

Application of Natural Language Processing(2):The Natural Language Interface for an Intelligent Geography Tutoring System. (자연어 활용(2):지능형 지리교육 시스템을 위한 자연어 인터 페이스에 관한 연구)

  • 장덕성;김승광
    • Korean Journal of Cognitive Science
    • /
    • v.3 no.2
    • /
    • pp.291-309
    • /
    • 1992
  • The computer manipulation by means of natural language prcessing will be not only helpful to use the computer with more simple and more comfortable but also flexiable to communicate between human and computers.In this paper the natural language interfaces will be applied to an intelligent geography tutoring systems(IGTS),and we will inspect the something to consider in the case of its implementation.Each module of IGTS is connected to the interface module and correspondence with each other for the sake of natural conversation between system and learner.

Natural Language Interface for Composite Web Services (복합 웹 서비스를 위한 자연어 인터페이스)

  • Lim, Jong-Hyun;Lee, Kyong-Ho
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.2
    • /
    • pp.144-156
    • /
    • 2010
  • With the wide spread of Web services in various fields, there is a growing interest in building a composite Web service, however, it is very difficult for ordinary users to specify how to compose services. Therefore, a convenient interface for generating and invoking composite Web services are required. This paper proposes a natural language interface to invoke services. The proposed interface provides a way to describe users' requests for composite Web Services in a natural language. A user with no technical knowledge about Web services can describe requests for composite Web services through the proposed interface. The proposed method extracts a complex workflow and finds appropriate Web services from the requests. Experimental results show that the proposed method extracts a sophisticated workflow from complex sentences with many phrases and control constructs.

Design of Parallel Input Pattern and Synchronization Method for Multimodal Interaction (멀티모달 인터랙션을 위한 사용자 병렬 모달리티 입력방식 및 입력 동기화 방법 설계)

  • Im, Mi-Jeong;Park, Beom
    • Journal of the Ergonomics Society of Korea
    • /
    • v.25 no.2
    • /
    • pp.135-146
    • /
    • 2006
  • Multimodal interfaces are recognition-based technologies that interpret and encode hand gestures, eye-gaze, movement pattern, speech, physical location and other natural human behaviors. Modality is the type of communication channel used for interaction. It also covers the way an idea is expressed or perceived, or the manner in which an action is performed. Multimodal Interfaces are the technologies that constitute multimodal interaction processes which occur consciously or unconsciously while communicating between human and computer. So input/output forms of multimodal interfaces assume different aspects from existing ones. Moreover, different people show different cognitive styles and individual preferences play a role in the selection of one input mode over another. Therefore to develop an effective design of multimodal user interfaces, input/output structure need to be formulated through the research of human cognition. This paper analyzes the characteristics of each human modality and suggests combination types of modalities, dual-coding for formulating multimodal interaction. Then it designs multimodal language and input synchronization method according to the granularity of input synchronization. To effectively guide the development of next-generation multimodal interfaces, substantially cognitive modeling will be needed to understand the temporal and semantic relations between different modalities, their joint functionality, and their overall potential for supporting computation in different forms. This paper is expected that it can show multimodal interface designers how to organize and integrate human input modalities while interacting with multimodal interfaces.

Speaker Detection and Recognition for a Welfare Robot

  • Sugisaka, Masanori;Fan, Xinjian
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.835-838
    • /
    • 2003
  • Computer vision and natural-language dialogue play an important role in friendly human-machine interfaces for service robots. In this paper we describe an integrated face detection and face recognition system for a welfare robot, which has also been combined with the robot's speech interface. Our approach to face detection is to combine neural network (NN) and genetic algorithm (GA): ANN serves as a face filter while GA is used to search the image efficiently. When the face is detected, embedded Hidden Markov Model (EMM) is used to determine its identity. A real-time system has been created by combining the face detection and recognition techniques. When motivated by the speaker's voice commands, it takes an image from the camera, finds the face inside the image and recognizes it. Experiments on an indoor environment with complex backgrounds showed that a recognition rate of more than 88% can be achieved.

  • PDF

High-Performance Korean Morphological Analyzer Using the MapReduce Framework on the GPU

  • Cho, Shi-Won;Lee, Dong-Wook
    • Journal of Electrical Engineering and Technology
    • /
    • v.6 no.4
    • /
    • pp.573-579
    • /
    • 2011
  • To meet the scalability and performance requirements of data analyses, which often involve voluminous data, efficient parallel or concurrent algorithms and frameworks are essential. We present a high-performance Korean morphological analyzer which employs the MapReduce framework on the graphics processing unit (GPU). MapReduce is a programming framework introduced by Google to aid the development of web search applications on a large number of central processing units (CPUs). GPUs are designed as a special-purpose co-processor. Their programming interfaces are typically formulated for graphics applications. Compared to CPUs, GPUs have greater computation power and memory bandwidth; however, GPUs are more difficult to program because of the design of their architectures. The performance of the Korean morphological analyzer using the MapReduce framework on the GPU is evaluated in comparison with the CPU-based model. The proposed Korean Morphological analyzer shows promising scalable performance on distributed computing with the GPU.

Database Interface System with Dialog (대화를 통한 데이타베이스 인터페이스 시스템)

  • Woo, Yo-Seop;Kang, Seok-Hoon
    • The Transactions of the Korea Information Processing Society
    • /
    • v.3 no.3
    • /
    • pp.417-428
    • /
    • 1996
  • In this paper, a database interface system with natural language dialogue is designed and implemented. The system is made up of language analysis, context processing, dialogue processing and DB processing unit. The method for classifying and processing an undefined word in language analysis is proposed. It reduces the dictionary size, which gives difficulties in DB Interface. And the current DB Interfaces dealt with an input utterance independently. But the system in this paper provides a user with the interface environment in which he or she can have a continuous conversation with the system and retrieve DB information. Thus in this paper, speech acts which include user's inattentions well as propositional contents are defined, and user action hierarchical model for library DB retrieval is constructed. And the system uses the defined knowledge to recognize-user's plan, effectively understanding and managing the ongoing dialogue. And the system is implemented in the domain of library database in order to prove the proposed methods in this paper.

  • PDF

Automation of One-Dimensional Finite Element Analysis of a Direct-Connection Spindle System of Machine Tools Using ANSYS (ANSYS를 활용한 공작기계 직결주축 시스템의 1차원 유한요소해석 자동화)

  • Choi, Jin-Woo
    • Journal of the Korean Society of Manufacturing Process Engineers
    • /
    • v.14 no.2
    • /
    • pp.127-133
    • /
    • 2015
  • In this study, an analytical model was developed for one-dimensional finite element analysis (1D FEA) of a spindle system of machine tools and then implemented to automate the FEA as a tool. FEA, with its vibration characteristics such as natural frequencies and modes, was performed using the universal FEA software ANSYS. VBA of EXCEL was used to provide the programming environment for its implementation. This enabled graphic user interfaces (GUIs) to be developed to allow interactions of users with the tool and, in addition, an EXCEL spreadsheet to be linked with the tool for data arrangement. The language of ANSYS was used to develop a code to perform the FEA. It generates an analytical model of the spindle system based on the information at the GUIs and subsequently performs the FEA based on the model. Automation helps identify the near-optimal design of the spindle system with minimum time and efforts.

A Light Weighted Robust Korean Morphological Analyzer for Korean-to-English Mobile Translator (한영 모바일 번역기를 위한 강건하고 경량화된 한국어 형태소 분석기)

  • Yuh, Sang-Hwa
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.2
    • /
    • pp.191-199
    • /
    • 2009
  • In this paper we present a light weighted robust Korean morphological analyzer for mobile devices such as mobile phones, smart phones, and PDA phones. Such mobile devices are not suitable for natural language interfaces for their low CPU performance and memory restriction. In order to overcome the difficulties we propose 1) an online analysis by using Key Event Handler mechanism, 2) and a robust analysis of the Korean sentences with spacing errors without its correction pre-processing. We adapt the proposed Korean analyzer to a Korean-English mobile translator, which shows 5.8% memory usage reduction and 19.0% enhancement of average response time.

Human-Computer Interaction Based Only on Auditory and Visual Information

  • Sha, Hui;Agah, Arvin
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.2 no.4
    • /
    • pp.285-297
    • /
    • 2000
  • One of the research objectives in the area of multimedia human-computer interaction is the application of artificial intelligence and robotics technologies to the development of computer interfaces. This involves utilizing many forms of media, integrating speed input, natural language, graphics, hand pointing gestures, and other methods for interactive dialogues. Although current human-computer communication methods include computer keyboards, mice, and other traditional devices, the two basic ways by which people communicate with each other are voice and gesture. This paper reports on research focusing on the development of an intelligent multimedia interface system modeled based on the manner in which people communicate. This work explores the interaction between humans and computers based only on the processing of speech(Work uttered by the person) and processing of images(hand pointing gestures). The purpose of the interface is to control a pan/tilt camera to point it to a location specified by the user through utterance of words and pointing of the hand, The systems utilizes another stationary camera to capture images of the users hand and a microphone to capture the users words. Upon processing of the images and sounds, the systems responds by pointing the camera. Initially, the interface uses hand pointing to locate the general position which user is referring to and then the interface uses voice command provided by user to fine-the location, and change the zooming of the camera, if requested. The image of the location is captured by the pan/tilt camera and sent to a color TV monitor to be displayed. This type of system has applications in tele-conferencing and other rmote operations, where the system must respond to users command, in a manner similar to how the user would communicate with another person. The advantage of this approach is the elimination of the traditional input devices that the user must utilize in order to control a pan/tillt camera, replacing them with more "natural" means of interaction. A number of experiments were performed to evaluate the interface system with respect to its accuracy, efficiency, reliability, and limitation.

  • PDF