• Title/Summary/Keyword: real-time visual tracking

Search Result 109, Processing Time 0.021 seconds

Macro-Micro Manipulation with Visual Tracking and its Application to Wheel Assembly

  • Cho Changhyun;Kang Sungchul;Kim Munsang;Song Jae-Bok
    • International Journal of Control, Automation, and Systems
    • /
    • v.3 no.3
    • /
    • pp.461-468
    • /
    • 2005
  • This paper proposes a wheel-assembly automation system, which assembles a wheel into a hub of a vehicle hung to a moving hanger in a car manufacturing line. A macro-micro manipulator control strategy is introduced to increase the system bandwidth and tracking accuracy to ensure insertion tolerance. A camera is equipped at the newly designed wheel gripper, which is attached at the center of the end-effector of the macro-micro manipulator and is used to measure position error of the hub of the vehicle in real time. The redundancy problem in the macro-micro manipulator is solved without complicated calculation by assigning proper functions to each part so that the macro part tracks the velocity error while the micro part regulates the fine position error. Experimental results indicate that tracking error satisfies the insertion tolerance of assembly $({\pm}1mm)$, and thus it is verified that the proposed system can be applied to the wheel assembly task on a moving hanger in the manufacturing line.

A Study on the Eye-Hand Coordination for Korean Text Entry Interface Development (한글 문자 입력 인터페이스 개발을 위한 눈-손 Coordination에 대한 연구)

  • Kim, Jung-Hwan;Hong, Seung-Kweon;Myung, Ro-Hae
    • Journal of the Ergonomics Society of Korea
    • /
    • v.26 no.2
    • /
    • pp.149-155
    • /
    • 2007
  • Recently, various devices requiring text input such as mobile phone IPTV, PDA and UMPC are emerging. The frequency of text entry for them is also increasing. This study was focused on the evaluation of Korean text entry interface. Various models to evaluate text entry interfaces have been proposed. Most of models were based on human cognitive process for text input. The cognitive process was divided into two components; visual scanning process and finger movement process. The time spent for visual scanning process was modeled as Hick-Hyman law, while the time for finger movement was determined as Fitts' law. There are three questions on the model-based evaluation of text entry interface. Firstly, are human cognitive processes (visual scanning and finger movement) during the entry of text sequentially occurring as the models. Secondly, is it possible to predict real text input time by previous models. Thirdly, does the human cognitive process for text input vary according to users' text entry speed. There was time gap between the real measured text input time and predicted time. The time gap was larger in the case of participants with high speed to enter text. The reason was found out investigating Eye-Hand Coordination during text input process. Differently from an assumption that visual scan on the keyboard is followed by a finger movement, the experienced group performed both visual scanning and finger movement simultaneously. Arrival Lead Time was investigated to measure the extent of time overlapping between two processes. 'Arrival Lead Time' is the interval between the eye fixation on the target button and the button click. In addition to the arrival lead time, it was revealed that the experienced group uses the less number of fixations during text entry than the novice group. This result will contribute to the improvement of evaluation model for text entry interface.

Sector Based Multiple Camera Collaboration for Active Tracking Applications

  • Hong, Sangjin;Kim, Kyungrog;Moon, Nammee
    • Journal of Information Processing Systems
    • /
    • v.13 no.5
    • /
    • pp.1299-1319
    • /
    • 2017
  • This paper presents a scalable multiple camera collaboration strategy for active tracking applications in large areas. The proposed approach is based on distributed mechanism but emulates the master-slave mechanism. The master and slave cameras are not designated but adaptively determined depending on the object dynamic and density distribution. Moreover, the number of cameras emulating the master is not fixed. The collaboration among the cameras utilizes global and local sectors in which the visual correspondences among different cameras are determined. The proposed method combines the local information to construct the global information for emulating the master-slave operations. Based on the global information, the load balancing of active tracking operations is performed to maximize active tracking coverage of the highly dynamic objects. The dynamics of all objects visible in the local camera views are estimated for effective coverage scheduling of the cameras. The active tracking synchronization timing information is chosen to maximize the overall monitoring time for general surveillance operations while minimizing the active tracking miss. The real-time simulation result demonstrates the effectiveness of the proposed method.

Fast Natural Feature Tracking Using Optical Flow (광류를 사용한 빠른 자연특징 추적)

  • Bae, Byung-Jo;Park, Jong-Seung
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.345-354
    • /
    • 2010
  • Visual tracking techniques for Augmented Reality are classified as either a marker tracking approach or a natural feature tracking approach. Marker-based tracking algorithms can be efficiently implemented sufficient to work in real-time on mobile devices. On the other hand, natural feature tracking methods require a lot of computationally expensive procedures. Most previous natural feature tracking methods include heavy feature extraction and pattern matching procedures for each of the input image frame. It is difficult to implement real-time augmented reality applications including the capability of natural feature tracking on low performance devices. The required computational time cost is also in proportion to the number of patterns to be matched. To speed up the natural feature tracking process, we propose a novel fast tracking method based on optical flow. We implemented the proposed method on mobile devices to run in real-time and be appropriately used with mobile augmented reality applications. Moreover, during tracking, we keep up the total number of feature points by inserting new feature points proportional to the number of vanished feature points. Experimental results showed that the proposed method reduces the computational cost and also stabilizes the camera pose estimation results.

The Validation of Spreading Activation Model as Evaluation Methodology of Menu Structure: Eye Tracking Approach (메뉴 구조의 평가 방법론으로서 활성화 확산 모델의 타당성 검증: Eye-Tracking 접근 방법)

  • Park, Jong-Soon;Myung, Ro-Hae
    • Journal of the Ergonomics Society of Korea
    • /
    • v.26 no.2
    • /
    • pp.103-112
    • /
    • 2007
  • This study was designed to validate Spreading Activation Theory (SAT) for an evaluation methodology for menu structure through Eye-Tracking approach. When a visual search is on the way, more eye fixations and time are necessary to visually process complex and vague area. From the aspect of recognition, well-designed menu structures were hypothesized to have fewer numbers of fixations and shorter duration because well-designed menu structures reflecting the users' mental model would be well matched with the product's menu structure, resulting in reducing the number of fixations and duration time. The results show that the shorter reaction times for SAT had significantly fewer numbers of fixation and shorter duration time as the hypothesis for this study stated. In conclusion, SAT was proved to be an effective evaluation methodology for menu structure with the eye tracking equipment. In addition, using SAT instead of the real performance experiment would be useful for designing user-centered systems and convenient information structures because SAT was proven to be the theoretical background for design and evaluation of menu structures.

Adaptive Weight Collaborative Complementary Learning for Robust Visual Tracking

  • Wang, Benxuan;Kong, Jun;Jiang, Min;Shen, Jianyu;Liu, Tianshan;Gu, Xiaofeng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.1
    • /
    • pp.305-326
    • /
    • 2019
  • Discriminative correlation filter (DCF) based tracking algorithms have recently shown impressive performance on benchmark datasets. However, amount of recent researches are vulnerable to heavy occlusions, irregular deformations and so on. In this paper, we intend to solve these problems and handle the contradiction between accuracy and real-time in the framework of tracking-by-detection. Firstly, we propose an innovative strategy to combine the template and color-based models instead of a simple linear superposition and rely on the strengths of both to promote the accuracy. Secondly, to enhance the discriminative power of the learned template model, the spatial regularization is introduced in the learning stage to penalize the objective boundary information corresponding to features in the background. Thirdly, we utilize a discriminative multi-scale estimate method to solve the problem of scale variations. Finally, we research strategies to limit the computational complexity of our tracker. Abundant experiments demonstrate that our tracker performs superiorly against several advanced algorithms on both the OTB2013 and OTB2015 datasets while maintaining the high frame rates.

A Study on HMM-Based Segmentation Method for Traffic Monitoring (HMM 분할에 기반한 교통모니터링에 관한 연구)

  • Hwang, Suen-Ki;Kang, Yong-Seok;Kim, Tae-Woo;Kim, Hyun-Yul;Park, Young-Cheol;Bae, Cheol-Soo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.5 no.1
    • /
    • pp.1-6
    • /
    • 2012
  • In this paper, we propose a HMM(Hidden Markov Model)-based segmentation method to model shadows as well as foreground and background regions. The shadow of moving objects often keeps from visual tracking. We propose an HMM-based segmentation method which classifies each object in real time. In the case of traffic monitoring movies, the effectiveness of the proposed method was proved by experiments.

Multi-mode Kernel Weight-based Object Tracking (멀티모드 커널 가중치 기반 객체 추적)

  • Kim, Eun-Sub;Kim, Yong-Goo;Choi, Yoo-Joo
    • Journal of the Korea Computer Graphics Society
    • /
    • v.21 no.4
    • /
    • pp.11-17
    • /
    • 2015
  • As the needs of real-time visual object tracking are increasing in various kinds of application fields such as surveillance, entertainment, etc., kernel-based mean-shift tracking has received more interests. One of major issues in kernel-based mean-shift tracking is to be robust under partial or full occlusion status. This paper presents a real-time mean-shift tracking which is robust in partial occlusion by applying multi-mode local kernel weight. In the proposed method, a kernel is divided into multiple sub-kernels and each sub-kernel has a kernel weight to be determined according to the location of the sub-kernel. The experimental results show that the proposed method is more stable than the previous methods with multi-mode kernels in partial occlusion circumstance.

Development of Visual Servo Control System for the Tracking and Grabbing of Moving Object (이동 물체 포착을 위한 비젼 서보 제어 시스템 개발)

  • Choi, G.J.;Cho, W.S.;Ahn, D.S.
    • Journal of Power System Engineering
    • /
    • v.6 no.1
    • /
    • pp.96-101
    • /
    • 2002
  • In this paper, we address the problem of controlling an end-effector to track and grab a moving target using the visual servoing technique. A visual servo mechanism based on the image-based servoing principle, is proposed by using visual feedback to control an end-effector without calibrated robot and camera models. Firstly, we consider the control problem as a nonlinear least squares optimization and update the joint angles through the Taylor Series Expansion. And to track a moving target in real time, the Jacobian estimation scheme(Dynamic Broyden's Method) is used to estimate the combined robot and image Jacobian. Using this algorithm, we can drive the objective function value to a neighborhood of zero. To show the effectiveness of the proposed algorithm, simulation results for a six degree of freedom robot are presented.

  • PDF

Region-based Q-learning for intelligent robot systems (지능형 로보트 시스템을 위한 영역기반 Q-learning)

  • Kim, Jae-Hyeon;Seo, Il-Hong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.3 no.4
    • /
    • pp.350-356
    • /
    • 1997
  • It is desirable for autonomous robot systems to possess the ability to behave in a smooth and continuous fashion when interacting with an unknown environment. Although Q-learning requires a lot of memory and time to optimize a series of actions in a continuous state space, it may not be easy to apply the method to such a real environment. In this paper, for continuous state space applications, to solve problem and a triangular type Q-value model\ulcorner This sounds very ackward. What is it you want to solve about the Q-value model. Our learning method can estimate a current Q-value by its relationship with the neighboring states and has the ability to learn its actions similar to that of Q-learning. Thus, our method can enable robots to move smoothly in a real environment. To show the validity of our method, navigation comparison with Q-learning are given and visual tracking simulation results involving an 2-DOF SCARA robot are also presented.

  • PDF