• 제목/요약/키워드: Human computer

검색결과 4,996건 처리시간 0.03초

MATES/CK 중한기계번역시스템의 구문분석규칙 (Parsing Rules for MATES/CK)

  • 송영미;강원석;김지현;송희정;황금하;최기선
    • 한국정보과학회 언어공학연구회:학술대회논문집(한글 및 한국어 정보처리)
    • /
    • 한국정보과학회언어공학연구회 2000년도 제12회 한글 및 한국어 정보처리 학술대회
    • /
    • pp.337-343
    • /
    • 2000
  • 중한기계번역시스템(MATES/CK)의 구문분석은 1120개의 구문분석규칙과 통계적 정보에 의한 확률기반에 따라 그 문장에 가장 적합한 구문트리를 찾아져 적용되는 방식으로 이루어지고 있다. 기존 구문분석 규칙은 자체에 오류가 많고, 새로운 규칙의 생성도 필요하다. 규칙에 대한 제약조건에도 좀 더 구체적이고 정확성을 높일 수 있는 상태로의 전환이 필요하다. 본 논문에서는 중한기계번역시스템(MATES/CK)의 구문분석의 정확도를 높이기 위하여 구문분석규칙을 수정하는 방법에 관하여 알아보고 그 연구과정을 살펴본다.

  • PDF

개량형 정보표시 화면설계 지침의 일원화 방법론 개발에 관한 연구 (A Study on Development of an Integration Methodology for Design Guideline of Advanced Information Display)

  • 정성해;차우창
    • 대한인간공학회지
    • /
    • 제23권2호
    • /
    • pp.13-24
    • /
    • 2004
  • Human error has brought about accidents more than 50% in system of a large size and complicated expecially in nuclear power plants(NPPs). The technology of Man Machine Interface(MMI) has been changed to the digitalized controls employing computer-based technology. According to this trend. the human factors guidelines are becoming main issue for reliable supports to digitalized information displays. However. the existing human factors guidelines is not enough for advanced information display on NPPs. The purpose of this research is to develop the reliable design and evaluation guidelines for advanced information display in main control room (MCR) of NPPs. In this study. the various general human factors guidelines concerning information display on CRT are integrated on data base management system. unified based on the integration rules. and applied in computer based procedures. The use of the integrated guidelines are expected to evaluate the existing information display on MCR in NPPs from the human factors point of view.

로봇과 인간의 상호작용을 위한 얼굴 표정 인식 및 얼굴 표정 생성 기법 (Recognition and Generation of Facial Expression for Human-Robot Interaction)

  • 정성욱;김도윤;정명진;김도형
    • 제어로봇시스템학회논문지
    • /
    • 제12권3호
    • /
    • pp.255-263
    • /
    • 2006
  • In the last decade, face analysis, e.g. face detection, face recognition, facial expression recognition, is a very lively and expanding research field. As computer animated agents and robots bring a social dimension to human computer interaction, interest in this research field is increasing rapidly. In this paper, we introduce an artificial emotion mimic system which can recognize human facial expressions and also generate the recognized facial expression. In order to recognize human facial expression in real-time, we propose a facial expression classification method that is performed by weak classifiers obtained by using new rectangular feature types. In addition, we make the artificial facial expression using the developed robotic system based on biological observation. Finally, experimental results of facial expression recognition and generation are shown for the validity of our robotic system.

A study on human performance in graphic-aided scheduling tasks

  • 백동현;오상윤;윤완철
    • 한국경영과학회:학술대회논문집
    • /
    • 대한산업공학회/한국경영과학회 1994년도 춘계공동학술대회논문집; 창원대학교; 08월 09일 Apr. 1994
    • /
    • pp.357-363
    • /
    • 1994
  • In many industrial situations the human acts as the primary scheduler since there often exist various constraints and considerations that may not be mathematically or quantitatively defined. For proper design of interactive scheduling systems, how human strategy and performance are affected by the fashion of human-computer interaction at various levels of task complexity should be investigated. In this study, two scheduling experiments were conducted. The first one showed that human schedulers could perform better than simple heuristic rules with each of typical performance measures such as average machine utilization, average tardiness, and maximum tardiness. In experiment 2, the effect of providing computer-generated initial solution was investigated. The results was that in complex problems the subjects performed significantly better when the initial solutions were generated by themselves, evidencing the importance of the continuity of strategic search through the problem.

휴먼-로봇 인터페이스를 위한 TTS의 개발 (Development of TTS for a Human-Robot Interface)

  • 배재현;오영환
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 2006년도 춘계 학술대회 발표논문집
    • /
    • pp.135-138
    • /
    • 2006
  • The communication method between human and robot is one of the important parts for a human-robot interaction. And speech is easy and intuitive communication method for human-being. By using speech as a communication method for robot, we can use robot as familiar way. In this paper, we developed TTS for human-robot interaction. Synthesis algorithms were modified for an efficient utilization of restricted resource in robot. And synthesis database were reconstructed for an efficiency. As a result, we could reduce the computation time with slight degradation of the speech quality.

  • PDF

An Image-Based CAPTCHA Scheme Exploiting Human Appearance Characteristics

  • Kalsoom, Sajida;Ziauddin, Sheikh;Abbasi, Abdul Rehman
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제6권2호
    • /
    • pp.734-750
    • /
    • 2012
  • CAPTCHAs are automated tests that are there to avoid misuse of computing and information resources by bots. Typical text-based CAPTCHAs are proven to be vulnerable against malicious automated programs. In this paper, we present an image-based CAPTCHA scheme using easily identifiable human appearance characteristics that overcomes the weaknesses of current text-based schemes. We propose and evaluate two applications for our scheme involving 25 participants. Both applications use same characteristics but different classes against those characteristics. Application 1 is optimized for security while application 2 is optimized for usability. Experimental evaluation shows promising results having 83% human success rate with Application 2 as compared to 62% with Application 1.

DeepAct: A Deep Neural Network Model for Activity Detection in Untrimmed Videos

  • Song, Yeongtaek;Kim, Incheol
    • Journal of Information Processing Systems
    • /
    • 제14권1호
    • /
    • pp.150-161
    • /
    • 2018
  • We propose a novel deep neural network model for detecting human activities in untrimmed videos. The process of human activity detection in a video involves two steps: a step to extract features that are effective in recognizing human activities in a long untrimmed video, followed by a step to detect human activities from those extracted features. To extract the rich features from video segments that could express unique patterns for each activity, we employ two different convolutional neural network models, C3D and I-ResNet. For detecting human activities from the sequence of extracted feature vectors, we use BLSTM, a bi-directional recurrent neural network model. By conducting experiments with ActivityNet 200, a large-scale benchmark dataset, we show the high performance of the proposed DeepAct model.

Estimating People's Position Using Matrix Decomposition

  • Dao, Thi-Nga;Yoon, Seokhoon
    • International journal of advanced smart convergence
    • /
    • 제8권2호
    • /
    • pp.39-46
    • /
    • 2019
  • Human mobility estimation plays a key factor in a lot of promising applications including location-based recommendation systems, urban planning, and disease outbreak control. We study the human mobility estimation problem in the case where recent locations of a person-of-interest are unknown. Since matrix decomposition is used to perform latent semantic analysis of multi-dimensional data, we propose a human location estimation algorithm based on matrix factorization to reconstruct the human movement patterns through the use of information of persons with correlated movements. Specifically, the optimization problem which minimizes the difference between the reconstructed and actual movement data is first formulated. Then, the gradient descent algorithm is applied to adjust parameters which contribute to reconstructed mobility data. The experiment results show that the proposed framework can be used for the prediction of human location and achieves higher predictive accuracy than a baseline model.

Recognition of the 3-D motion of a human arm with HIGIPS

  • Yao, Feng-Hui;Tamaki, Akikazu;Kato, Kiyoshi
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1991년도 한국자동제어학술회의논문집(국제학술편); KOEX, Seoul; 22-24 Oct. 1991
    • /
    • pp.1724-1729
    • /
    • 1991
  • This paper gives an overview of HIGIPS design concepts and prototype HIGIPS configuration, and discusses its application to recognition of the 3-D motion of a human arm. HIGIPS which employs the combination of pipeline architecture and multiprocessor architecture, is a high-speed, high-performance and low cost N * M multimicroprocessor parallel machine, where N is the number of pipeline stages and M is the number of processors in each stage. The algorithm to recognize the motion of a human arm with a single TV camera was developed on personal computer (NEC PC9801 series). As a constraint condition, some simple ring marks are used. Each joint of the arm is attached with a ring mark to obtain its centroid position when the arm moves. These centroid positions in the three-dimensional space are linked at each of the successive pictures of the moving arm to recover its overall motion. This algorithm takes about 2 seconds to process one image frame on the general-purpose personal computer. This paper mainly discuses how to partition this algorithm and execute on HIGIPS, and shows the speed up. From this application, it is clear that HIGIPS is an efficient machine for image processing and recognizing.

  • PDF

PERSONAL SPACE-BASED MODELING OF RELATIONSHIPS BETWEEN PEOPLE FOR NEW HUMAN-COMPUTER INTERACTION

  • Amaoka, Toshitaka;Laga, Hamid;Saito, Suguru;Nakajima, Masayuki
    • 한국방송∙미디어공학회:학술대회논문집
    • /
    • 한국방송공학회 2009년도 IWAIT
    • /
    • pp.746-750
    • /
    • 2009
  • In this paper we focus on the Personal Space (PS) as a nonverbal communication concept to build a new Human Computer Interaction. The analysis of people positions with respect to their PS gives an idea on the nature of their relationship. We propose to analyze and model the PS using Computer Vision (CV), and visualize it using Computer Graphics. For this purpose, we define the PS based on four parameters: distance between people, their face orientations, age, and gender. We automatically estimate the first two parameters from image sequences using CV technology, while the two other parameters are set manually. Finally, we calculate the two-dimensional relationship of multiple persons and visualize it as 3D contours in real-time. Our method can sense and visualize invisible and unconscious PS distributions and convey the spatial relationship of users by an intuitive visual representation. The results of this paper can be used to Human Computer Interaction in public spaces.

  • PDF