• Title/Summary/Keyword: touch gesture

Search Result 73, Processing Time 0.029 seconds

A Unit Touch Gesture Model of Performance Time Prediction for Mobile Devices

  • Kim, Damee;Myung, Rohae
    • Journal of the Ergonomics Society of Korea
    • /
    • v.35 no.4
    • /
    • pp.277-291
    • /
    • 2016
  • Objective: The aim of this study is to propose a unit touch gesture model, which would be useful to predict the performance time on mobile devices. Background: When estimating usability based on Model-based Evaluation (MBE) in interfaces, the GOMS model measured 'operators' to predict the execution time in the desktop environment. Therefore, this study used the concept of operator in GOMS for touch gestures. Since the touch gestures are comprised of possible unit touch gestures, these unit touch gestures can predict to performance time with unit touch gestures on mobile devices. Method: In order to extract unit touch gestures, manual movements of subjects were recorded in the 120 fps with pixel coordinates. Touch gestures are classified with 'out of range', 'registration', 'continuation' and 'termination' of gesture. Results: As a results, six unit touch gestures were extracted, which are hold down (H), Release (R), Slip (S), Curved-stroke (Cs), Path-stroke (Ps) and Out of range (Or). The movement time predicted by the unit touch gesture model is not significantly different from the participants' execution time. The measured six unit touch gestures can predict movement time of undefined touch gestures like user-defined gestures. Conclusion: In conclusion, touch gestures could be subdivided into six unit touch gestures. Six unit touch gestures can explain almost all the current touch gestures including user-defined gestures. So, this model provided in this study has a high predictive power. The model presented in the study could be utilized to predict the performance time of touch gestures. Application: The unit touch gestures could be simply added up to predict the performance time without measuring the performance time of a new gesture.

Towards Establishing a Touchless Gesture Dictionary based on User Participatory Design

  • Song, Hae-Won;Kim, Huhn
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.4
    • /
    • pp.515-523
    • /
    • 2012
  • Objective: The aim of this study is to investigate users' intuitive stereotypes on non-touch gestures and establish the gesture dictionary that can be applied to gesture-based interaction designs. Background: Recently, the interaction based on non-touch gestures is emerging as an alternative for natural interactions between human and systems. However, in order for non-touch gestures to become a universe interaction method, the studies on what kinds of gestures are intuitive and effective should be prerequisite. Method: In this study, as applicable domains of non-touch gestures, four devices(i.e. TV, Audio, Computer, Car Navigation) and sixteen basic operations(i.e. power on/off, previous/next page, volume up/down, list up/down, zoom in/out, play, cancel, delete, search, mute, save) were drawn from both focus group interview and survey. Then, a user participatory design was performed. The participants were requested to design three gestures suitable to each operation in the devices, and they evaluated intuitiveness, memorability, convenience, and satisfaction of their derived gestures. Through the participatory design, agreement scores, frequencies and planning times of each distinguished gesture were measured. Results: The derived gestures were not different in terms of four devices. However, diverse but common gestures were derived in terms of kinds of operations. In special, manipulative gestures were suitable for all kinds of operations. On the contrary, semantic or descriptive gestures were proper to one-shot operations like power on/off, play, cancel or search. Conclusion: The touchless gesture dictionary was established by mapping intuitive and valuable gestures onto each operation. Application: The dictionary can be applied to interaction designs based on non-touch gestures. Moreover, it will be used as a basic reference for standardizing non-touch gestures.

Implementation of a DI Multi-Touch Display Using an Improved Touch-Points Detection and Gesture Recognition (개선된 터치점 검출과 제스쳐 인식에 의한 DI 멀티터치 디스플레이 구현)

  • Lee, Woo-Beom
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.11 no.1
    • /
    • pp.13-18
    • /
    • 2010
  • Most of the research in the multi-touch area is based on the FTIR(Frustrated Total Internal Re리ection), which is just implemented by using the previous approach. Moreover, there are not the software solutions to improve a performance in the multi touch-blobs detection or the user gesture recognition. Therefore, we implement a multi-touch table-top display that is based on the DI(Diffused Illumination), the improved touch-points detection and user gesture recognition. The proposed method supports a simultaneous transformation multi-touch command for objects in the running application. Also, the system latency time is reduced by the proposed ore-testing method in the multi touch-blobs detection processing. Implemented device is simulated by programming the Flash AS3 application in the TUIO(Tangible User Interface Object) environment that is based on the OSC(Open Sound Control) protocol. As a result, Our system shows the 37% system latency reduction, and is successful in the multi-touch gestures recognition.

RealBook: A Tangible Electronic Book Based on the Interface of TouchFace-V (RealBook: TouchFace-V 인터페이스 기반 실감형 전자책)

  • Song, Dae-Hyeon;Bae, Ki-Tae;Lee, Chil-Woo
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.12
    • /
    • pp.551-559
    • /
    • 2013
  • In this paper, we proposed a tangible RealBook based on the interface of TouchFace-V which is able to recognize multi-touch and hand gesture. The TouchFace-V is applied projection technology on a flat surface such as table, without constraint of space. The system's configuration is addressed installation, calibration, and portability issues that are most existing front-projected vision-based tabletop display. It can provide hand touch and gesture applying computer vision by adopting tracking technology without sensor and traditional input device. The RealBook deals with the combination of each advantage of analog sensibility on texts and multimedia effects of e-book. Also, it provides digitally created stories that would differ in experiences and environments with interacting users' choices on the interface of the book. We proposed e-book that is new concept of electronic book; named RealBook, different from existing and TouchFace-V interface, which can provide more direct viewing, natural and intuitive interactions with hand touch and gesture.

Extracting Flick Operator for Predicting Performance by GOMS Model in Small Touch Screen

  • Choi, Mikyung;Lee, Bong Geun;Oh, Hyungseok;Myung, Rohae
    • Journal of the Ergonomics Society of Korea
    • /
    • v.32 no.2
    • /
    • pp.179-187
    • /
    • 2013
  • Objective: The purpose of this study is to extract GOMS manual operator, except for an experiment with participants. Background: The GOMS model has advantage of rapid modeling which is suitable for the environment of technology development which has a short life cycle products with a fast pace. The GOMS model was originally designed for desktop environment so that it is not adequate for implementing into the latest HCI environment such as small touch screen device. Therefore, this research proposed GOMS manual operator extraction methodology which is excluded experimental method. And flick Gesture was selected to explain application of proposed methodology to extract new operator. Method: Divide into start to final step of hand gesture needed to extract as an operator through gesture task analysis. Then apply the original GOMS operator to each similar step of gesture and modify the operator for implementation stage based on existing Fitts' law research. Steps that are required to move are modified based on the Fitts' law developed in touch screen device. Finally, new operator can be derived from using these stages and a validation experiment, performed to verify the validity of new operator and methodology by comparing human performance. Results: The average movement times of the participants' performance and the operator which is extracted in case study are not different significantly. Also the average of movement times of each type of view study is not different significantly. Conclusion: In conclusion, the result of the proposed methodology for extracting new operator is similar to the result of the experiment with their participants. Furthermore the GOMS model included the operator by the proposed methodology in this research could be applied successfully to predict the user's performance. Application: Using this methodology could be applied to develop new finger gesture in the touch screen. Also this proposed methodology could be applied to evaluate the usability of certain system rapidly including the new finger gesture performance.

Video event control system by recognition of depth touch (깊이 터치를 통한 영상 이벤트 제어 시스템)

  • Lee, Dong-Seok;Kwon, Soon-Kak
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.21 no.1
    • /
    • pp.35-42
    • /
    • 2016
  • Various events of stop, playback, capture, and zoom-in/out in playing video is available in the monitor of a small size such as smart phones. However, if the size of the display increases, then the cost of the touch recognition is increased, thus provision of a touch event is not possible in practice. In this paper, we propose a video event control system that recognizes a touch inexpensively from the depth information, then provides the variety of events of the toggle, the pinch-in / out by the single or multi-touch. The proposed method finds a touch location and the touch path by the depth information from a depth camera, and determines the touch gesture type. This touch interface algorithm is implemented in a small single-board system, and can control the video event by sending the gesture information through the UART communication. Simulation results show that the proposed depth touch method can control the various events in a large screen.

MRF Particle filter-based Multi-Touch Tracking and Gesture Likelihood Estimation (MRF 입자필터 멀티터치 추적 및 제스처 우도 측정)

  • Oh, Chi-Min;Shin, Bok-Suk;Klette, Reinhard;Lee, Chil-Woo
    • Smart Media Journal
    • /
    • v.4 no.1
    • /
    • pp.16-24
    • /
    • 2015
  • In this paper, we propose a method for multi-touch tracking using MRF-based particle filters and gesture likelihood estimation Each touch (of one finger) is considered to be one object. One of frequently occurring issues is the hijacking problem which means that an object tracker can be hijacked by neighboring object. If a predicted particle is close to an adjacent object then the particle's weight should be lowered by analysing the influence of neighboring objects for avoiding hijacking problem. We define a penalty function to lower the weights of those particles. MRF is a graph representation where a node is the location of a target object and an edge describes the adjacent relation of target object. It is easy to utilize MRF as data structure of adjacent objects. Moreover, since MRF graph representation is helpful to analyze multi-touch gestures, we describe how to define gesture likelihoods based on MRF. The experimental results show that the proposed method can avoid the occurrence of hijacking problems and is able to estimate gesture likelihoods with high accuracy.

A Study on Developmental Direction of Interface Design for Gesture Recognition Technology

  • Lee, Dong-Min;Lee, Jeong-Ju
    • Journal of the Ergonomics Society of Korea
    • /
    • v.31 no.4
    • /
    • pp.499-505
    • /
    • 2012
  • Objective: Research on the transformation of interaction between mobile machines and users through analysis on current gesture interface technology development trend. Background: For smooth interaction between machines and users, interface technology has evolved from "command line" to "mouse", and now "touch" and "gesture recognition" have been researched and being used. In the future, the technology is destined to evolve into "multi-modal", the fusion of the visual and auditory senses and "3D multi-modal", where three dimensional virtual world and brain waves are being used. Method: Within the development of computer interface, which follows the evolution of mobile machines, actively researching gesture interface and related technologies' trend and development will be studied comprehensively. Through investigation based on gesture based information gathering techniques, they will be separated in four categories: sensor, touch, visual, and multi-modal gesture interfaces. Each category will be researched through technology trend and existing actual examples. Through this methods, the transformation of mobile machine and human interaction will be studied. Conclusion: Gesture based interface technology realizes intelligent communication skill on interaction relation ship between existing static machines and users. Thus, this technology is important element technology that will transform the interaction between a man and a machine more dynamic. Application: The result of this study may help to develop gesture interface design currently in use.

Effect of Tactile Feedback for Button GUI on Mobile Touch Devices

  • Shin, Heesook;Lim, Jeong-Mook;Lee, Jong-Uk;Lee, Geehyuk;Kyung, Ki-Uk
    • ETRI Journal
    • /
    • v.36 no.6
    • /
    • pp.979-987
    • /
    • 2014
  • This paper describes new tactile feedback patterns and the effect of their input performance for a button GUI activated by a tap gesture on mobile touch devices. Based on an analysis of touch interaction and informal user tests, several tactile feedback patterns were designed. Using these patterns, three user experiments were performed to investigate appropriate tactile feedback patterns and their input performance during interaction with a touch button. The results showed that a tactile pattern responding to each touch and release gesture with a rapid response time and short falling time provides the feeling of physically clicking a button. The suggested tactile feedback pattern has a significantly positive effect on the number of typing errors and typing task completion time compared to the performance when no feedback is provided.

Design of Multi-Finger Flick Interface for Fast File Management on Capacitive-Touch-Sensor Device (정전기식 입력 장치에서의 빠른 파일 관리를 위한 다중 손가락 튕김 인터페이스 설계)

  • Park, Se-Hyun;Park, Tae-Jin;Choy, Yoon-Chul
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.8
    • /
    • pp.1235-1244
    • /
    • 2010
  • Most emerging smart phones support capacitive touch sensors. It renders existing gesture-based interfaces not suitable since they were developed for the resistive touch sensors and pen-based input. Unlike the flick gestures from the existing gesture interfaces, the finger flick gesture used in this paper reduces the workload about half by selecting the target and the command to perform on the target at a single touch input. With the combination with multi-touch interface, it supports various menu commands without having to learn complex gestures, and is suitable for the touch-based devices hence it minimizes input error. This research designs and implements the multi-touch and flick interface to provide an effective file management system on the smart phones with capacitive touch input. The evaluation proves that the suggested interface is superior to the existing methods on the capacitive touch input devices.