• Title/Summary/Keyword: Voice command

Search Result 94, Processing Time 0.029 seconds

Development of a Work Management System Based on Speech and Speaker Recognition

  • Gaybulayev, Abdulaziz;Yunusov, Jahongir;Kim, Tae-Hyong
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.16 no.3
    • /
    • pp.89-97
    • /
    • 2021
  • Voice interface can not only make daily life more convenient through artificial intelligence speakers but also improve the working environment of the factory. This paper presents a voice-assisted work management system that supports both speech and speaker recognition. This system is able to provide machine control and authorized worker authentication by voice at the same time. We applied two speech recognition methods, Google's Speech application programming interface (API) service, and DeepSpeech speech-to-text engine. For worker identification, the SincNet architecture for speaker recognition was adopted. We implemented a prototype of the work management system that provides voice control with 26 commands and identifies 100 workers by voice. Worker identification using our model was almost perfect, and the command recognition accuracy was 97.0% in Google API after post- processing and 92.0% in our DeepSpeech model.

Object Magnification and Voice Command in Gaze Interface for the Upper Limb Disabled (상지장애인을 위한 시선 인터페이스에서의 객체 확대 및 음성 명령 인터페이스 개발)

  • Park, Joo Hyun;Jo, Se-Ran;Lim, Soon-Bum
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.7
    • /
    • pp.903-912
    • /
    • 2021
  • Eye tracking research for upper limb disabilities is showing an effect in the aspect of device control. However, the reality is that it is not enough to perform web interaction with only eye tracking technology. In the Eye-Voice interface, a previous study, in order to solve the problem that the existing gaze tracking interfaces cause a malfunction of pointer execution, a gaze tracking interface supplemented with a voice command was proposed. In addition, the reduction of the malfunction rate of the pointer was confirmed through a comparison experiment with the existing interface. In this process, the difficulty of pointing due to the small size of the execution object in the web environment was identified as another important problem of malfunction. In this study, we propose an auto-magnification interface of objects so that people with upper extremities can freely click web contents by improving the problem that it was difficult to point and execute due to the high density of execution objects and their arrangements in web pages.

A Concept Study on Improving Command Post Vehicles for Operating Battle Management System (전장관리체계 운용을 위한 전투지휘용 차량의 성능개량 개념)

  • Park, Seung
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.11 no.2
    • /
    • pp.16-22
    • /
    • 2008
  • So far, Korean command post vehicles(CPs) have used voice-only radio communication among the mechanized troop units such MBTs and IFVs, etc. But digital data communication technology should be introduced to improve mission success capabilities and accommodate integrated combat capabilities through increasing SA(situation Awareness), or spatiotemporal synchronization of military operation in battle field environment. Therefore, conventional CPs, which have been operated without digital network, urgently needs to mount a battle management system which furnish with tactical information network, considering that korean new MBTs or new IFVs will be fielded soon. This study suggests some performance enhancement method which might be applicable to CPs which have a role of connecting Battalion unit to Brigade C4I system.

Voice Command-based Prediction and Follow of Human Path of Mobile Robots in AI Space

  • Tae-Seok Jin
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.26 no.2_1
    • /
    • pp.225-230
    • /
    • 2023
  • This research addresses sound command based human tracking problems for autonomous cleaning mobile robot in a networked AI space. To solve the problem, the difference among the traveling times of the sound command to each of three microphones has been used to calculate the distance and orientation of the sound from the cleaning mobile robot, which carries the microphone array. The cross-correlation between two signals has been applied for detecting the time difference between two signals, which provides reliable and precise value of the time difference compared to the conventional methods. To generate the tracking direction to the sound command, fuzzy rules are applied and the results are used to control the cleaning mobile robot in a real-time. Finally the experiment results show that the proposed algorithm works well, even though the mobile robot knows little about the environment.

Control of IEEE1394 digital home appliances using AV/C Command Set

  • Kim, Il-Jung;Park, Jong-An
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.98.2-98
    • /
    • 2001
  • It is necessary to have enough transmission capacity for advanced internet techniques and various digital home appliances networking. Home appliances interface IEEE1394 technique has much wealthy transmission skill. IEEE1394 is using home appliances through various information form image and voice change data real time print out. In this paper, AVC CTS technology and IEEE1394 technology are introduced. Digital Video Camera includes compression format using DV. System composition control is consisted of protocols like IEC-61883 and AV/C command set standard.

  • PDF

A research on man-robot cooperative interaction system

  • Ishii, Masaru
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1992.10b
    • /
    • pp.555-557
    • /
    • 1992
  • Recently, realization of an intelligent cooperative interaction system between a man and robot systems is required. In this paper, HyperCard with a voice control is used for above system because of its easy handling and excellent human interfaces. Clicking buttons in the HyperCard by a mouse device or a voice command means controlling each joint of a robot system. Robot teaching operation of grasping a bin and pouring liquid in it into a cup is carried out. This robot teaching method using HyperCard provides a foundation for realizing a user friendly cooperative interaction system.

  • PDF

A Voice Controlled Service Robot Using Support Vector Machine

  • Kim, Seong-Rock;Park, Jae-Suk;Park, Ju-Hyun;Lee, Suk-Gyu
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1413-1415
    • /
    • 2004
  • This paper proposes a SVM(Support Vector Machine) training algorithm to control a service robot with voice command. The service robot with a stereo vision system and dual manipulators of four degrees of freedom implements a User-Dependent Voice Control System. The training of SVM algorithm that is one of the statistical learning theories leads to a QP(quadratic programming) problem. In this paper, we present an efficient SVM speech recognition scheme especially based on less learning data comparing with conventional approaches. SVM discriminator decides rejection or acceptance of user's extracted voice features by the MFCC(Mel Frequency Cepstrum Coefficient). Among several SVM kernels, the exponential RBF function gives the best classification and the accurate user recognition. The numerical simulation and the experiment verified the usefulness of the proposed algorithm.

  • PDF

A Voice-Annotation Technique in Mobile E-book for Reading-disabled People (독서장애인용 디지털음성도서를 위한 음성 어노테이션 기법)

  • Lee, Kyung-Hee;Lee, Jong-Woo;Lim, Soon-Bum
    • Journal of Digital Contents Society
    • /
    • v.12 no.3
    • /
    • pp.329-337
    • /
    • 2011
  • Digital talking book has been developed to enhance reading experiences for reading-disabled people. In the existing digital talking book, however, annotations can be created only through the screen interfaces. Screen annotation interfaces is of no use for reading-disabled people because they need reader's eyesight. In this paper, we suggest a voice annotation technique can create notes and highlights at any playing time by using hearing sense and voice command. We design a location determination technique that pinpoints where a voice annotation should be placed in the playing sentences. To verify the effectiveness of our voice annotation technique, we implement a prototype in an android platform. We can find out by the black-blindfolded users testing that our system can perfectly locate the exact position that a voice annotation should be placed into.