• Title/Summary/Keyword: Visual object

검색결과 1,240건 처리시간 0.251초

움직이는 물체의 안정한 Grasping을 위한 시각추적 시스템 개발 (The development of a visual tracking system for the stable grasping of a moving object)

  • 차인혁;손영갑;한창수
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1996년도 한국자동제어학술회의논문집(국내학술편); 포항공과대학교, 포항; 24-26 Oct. 1996
    • /
    • pp.543-546
    • /
    • 1996
  • We propose a new visual tracking system for grasping which can find grasping points of an unknown polygonal object. We construct the system with the image prediction technique and Extended Kalman Filter algorithm. The Extended Kalman Filter(EKF) based on the SVD can improve the accuracy and processing time for the estimation of the nonlinear state variables. By using it, we can solve the numerical unstability problem that can occur in the visual tracking system based on Kalman filter. The image prediction algorithm can reduce the effect of noise and the image processing time. In the processing of a visual tracking, we can construct the parameterized family and can found the grasping points of unknown object through the geometric properties of the parameterized family.

  • PDF

Resolution-enhanced Reconstruction of 3D Object Using Depth-reversed Elemental Images for Partially Occluded Object Recognitionz

  • Wei, Tan-Chun;Shin, Dong-Hak;Lee, Byung-Gook
    • Journal of the Optical Society of Korea
    • /
    • 제13권1호
    • /
    • pp.139-145
    • /
    • 2009
  • Computational integral imaging (CII) is a new method for 3D imaging and visualization. However, it suffers from seriously poor image quality of the reconstructed image as the reconstructed image plane increases. In this paper, to overcome this problem, we propose a CII method based on a smart pixel mapping (SPM) technique for partially occluded 3D object recognition, in which the object to be recognized is located at far distance from the lenslet array. In the SPM-based CII, the use of SPM moves a far 3D object toward the near lenslet array and then improves the image quality of the reconstructed image. To show the usefulness of the proposed method, we carry out some experiments for occluded objects and present the experimental results.

시각 요소와 시각 변수를 통한 시각 객체 질의어(VOQL)의 개선 (Improving Visual Object Query language (VOQL) by Introducing Visual Elements and visual Variables)

  • 이석균
    • 한국정보처리학회논문지
    • /
    • 제6권6호
    • /
    • pp.1447-1457
    • /
    • 1999
  • 최근 제안된 시각 객체 질의어(VOQL)는 시각 질의어로 스키마 정보가 시각적으로 질의 표현에 포함되므로 복잡한 구조의 데이터에 대한 질의를 효과적으로 표현할 수 있는 객체 지향 데이터베이스 질의어이다. VOQL은 귀납적으로 정의된 시맨턱을 갖는 그래프 기반 언어로 다양한 텍스트 경로식들을 그래프로 간결하게 표현 할 뿐 아니라 복잡한 경로식의 시맨틱을 명확하게 전달한다. 그러나 기존의 VOQL은 모든 속성을 다중 값으로 가정하고 있고, 객체변수의 바인딩 개념을 시각화하고 있지 못하고 있다. 이로 인해 VOQL 질의문의 표현이 직관적이지 못할 뿐 아니라 이론적 확장이 쉽지 않다. 본 논문에서는 이러한 문제를 해결하도록 VOQL을 개선하고 한다. 단일 값과 다중 값을 갖는 속성의 결과를 각각 시각 요소와 서브 블랍을 통해 시각화하고, 시각변수를 도입하여 객체 변수의 바인딩을 명시화하여 질의문의 시맨틱을 직관적이고, 명확하게 하고 있다.

  • PDF

객체행위챠트를 위한 LOTOS 정형기법 기반 시각적 시뮬레이션 도구의 개발 (Development of a Visual Simulation Tool for Object Behavior Chart based on LOTOS Formalism)

  • 이광용;오영배
    • 한국정보과학회논문지:컴퓨팅의 실제 및 레터
    • /
    • 제5권5호
    • /
    • pp.595-610
    • /
    • 1999
  • 본 논문에서는 기존 실시간 객체행위 설계기법에 의해 개발된 객체행위챠트의 설계의미를 검증 확인하기 위한 시각적 시뮬레이션 도구를 제안한다. 이 도구의 특징은 실행 가능한 시뮬레이션 기계 즉, 확장된 유한상태기계 생성에 의한 객체들간의 동적인 상호작용을 시각적으로 시뮬레이션할 수 있게 하며, 실제 시스템 구현에 앞서 시각적 객체행위챠트에 여러 가지 논리적, 시간적 문제들을 검출할 수 있게 하는 것이다. 이를 위해, 이 도구에서는 시각적 객체행위챠트로부터 LOTOS 정형명세를 자동으로 생성한 다음 그 명세를 확장된 유한상태기계로 변환한다. 이 도구는 Visual C++ 버전 4.2로 개발되었으며, 현재 PC 윈도우95 환경에서 수행된다. 그리고, 특별히 시뮬레이션을 위해 LOTOS 정형성을 이용하였는데, 이것은 LOTOS의 통신 프로토콜 표현력이 다른 정형언어들에 비해 우수성이 있기 때문이다. 본 연구는 방법론 기반 시각모델과 정형기법 기반 시뮬레이션 기술의 연결이라는 점과 시각모델의 자동화된 검증 확인 기술의 실현이라는 점에서 기여한다.

시각과 시촉각에 의한 운동 측면에서 본 공간과 시간의 관계성 연구 - 연경당 외부공간을 중심으로 - (A Study on the Relationship of Space and Time in Visual Tactility)

  • 육옥수
    • 건축역사연구
    • /
    • 제20권1호
    • /
    • pp.77-93
    • /
    • 2011
  • Across the culture of Western Europe, dichotomy based on the visual sense has evolved. They believed eyes and ears requiring a distance related in recognition, are more developed than any other human senses in human body. Dominant position, as a condition to using a perspective, the eye has been just concentrated in the development of optical sight. But developed a variety of modern media, highlighting the importance of the other perception, it makes dichotomy to the expansion of perception over the single function of visuality. Recently, Guille Deleuze and Merleau-Ponty try to recover the sense of tactility segregated in skin from body keeping eyes for distance. By the result, the activity can be happened by being connected to the body rather than to eye in the space between the subject and object. From the phase of recognition where the human body tries to identify the object in the space considering a time, it will be changed for the subject to the phase of structure vice versa. Visual tactility is to eliminate the distance between subject and object. If the visual tactility is to erase the distance different from the visual in dichotomy, it will be occurred to having a tension and makes new relationship to work trying to move the subjective point of view in object. Like this evidence in analysis of architecture, it can be easy to find the Korean architecture rather than western architecture in terms of emphasizing the time and space. The fact, architecture of Lee Dynasty had been preserved and consisted basic form and style over the centuries makes us assume that visual tactility was considered as well as the visual sense. This study will be intensive in terms of visual and tactile inherent in the subject and how it is being connected to the movement in the space and time.

A Collaborative Visual Language

  • Kim, Kyung-Deok
    • Journal of information and communication convergence engineering
    • /
    • 제1권2호
    • /
    • pp.74-81
    • /
    • 2003
  • There are many researches on visual languages, but the most of them are difficult to support various collaborative interactions on a distributed multimedia environment. So, this paper suggests a collaborative visual language for interaction between multi-users. The visual language can describe a conceptual model for collaborative interactions between multi-users. Using the visual language, generated visual sentences consist of object icons and interaction operators. An object icon represents a user who is responsible for a collaborative activity, has dynamic attributes of a user, and supports flexible interaction between multi-users. An interaction operator represents an interactive relation between multi-users and supports various collaborative interactions. Merits of the visual language are as follows: supporting of both asynchronous interaction and synchronous interaction, supporting flexible interaction between multi-users according to participation or leave of users, supporting a user oriented modeling, etc. For example, an application to a workflow system for document approval is illustrated. So we could be found that the visual language shows a collaborative interaction.

다중 도메인 데이터 기반 구별적 모델 예측 트레커를 위한 동적 탐색 영역 특징 강화 기법 (Reinforced Feature of Dynamic Search Area for the Discriminative Model Prediction Tracker based on Multi-domain Dataset)

  • 이준하;원홍인;김병학
    • 대한임베디드공학회논문지
    • /
    • 제16권6호
    • /
    • pp.323-330
    • /
    • 2021
  • Visual object tracking is a challenging area of study in the field of computer vision due to many difficult problems, including a fast variation of target shape, occlusion, and arbitrary ground truth object designation. In this paper, we focus on the reinforced feature of the dynamic search area to get better performance than conventional discriminative model prediction trackers on the condition when the accuracy deteriorates since low feature discrimination. We propose a reinforced input feature method shown like the spotlight effect on the dynamic search area of the target tracking. This method can be used to improve performances for deep learning based discriminative model prediction tracker, also various types of trackers which are used to infer the center of the target based on the visual object tracking. The proposed method shows the improved tracking performance than the baseline trackers, achieving a relative gain of 38% quantitative improvement from 0.433 to 0.601 F-score at the visual object tracking evaluation.

Background memory-assisted zero-shot video object segmentation for unmanned aerial and ground vehicles

  • Kimin Yun;Hyung-Il Kim;Kangmin Bae;Jinyoung Moon
    • ETRI Journal
    • /
    • 제45권5호
    • /
    • pp.795-810
    • /
    • 2023
  • Unmanned aerial vehicles (UAV) and ground vehicles (UGV) require advanced video analytics for various tasks, such as moving object detection and segmentation; this has led to increasing demands for these methods. We propose a zero-shot video object segmentation method specifically designed for UAV and UGV applications that focuses on the discovery of moving objects in challenging scenarios. This method employs a background memory model that enables training from sparse annotations along the time axis, utilizing temporal modeling of the background to detect moving objects effectively. The proposed method addresses the limitations of the existing state-of-the-art methods for detecting salient objects within images, regardless of their movements. In particular, our method achieved mean J and F values of 82.7 and 81.2 on the DAVIS'16, respectively. We also conducted extensive ablation studies that highlighted the contributions of various input compositions and combinations of datasets used for training. In future developments, we will integrate the proposed method with additional systems, such as tracking and obstacle avoidance functionalities.

인공지능 객체인식에 관한 파라미터 측정 연구 (A Study On Parameter Measurement for Artificial Intelligence Object Recognition)

  • 최병관
    • 디지털산업정보학회논문지
    • /
    • 제15권3호
    • /
    • pp.15-28
    • /
    • 2019
  • Artificial intelligence is evolving rapidly in the ICT field, smart convergence media system and content industry through the fourth industrial revolution, and it is evolving very rapidly through Big Data. In this paper, we propose a face recognition method based on object recognition based on object recognition through artificial intelligence. In this method, Were experimented and studied through the object recognition technique of artificial intelligence. In the conventional 3D image field, general research on object recognition has been carried out variously, and researches have been conducted on the side effects of visual fatigue and dizziness through 3D image. However, in this study, we tried to solve the problem caused by the quantitative difference between object recognition and object recognition for human factor algorithm that measure visual fatigue through cognitive function, morphological analysis and object recognition. Especially, The new method of computer interaction is presented and the results are shown through experiments.

Object Tracking with Sparse Representation based on HOG and LBP Features

  • Boragule, Abhijeet;Yeo, JungYeon;Lee, GueeSang
    • International Journal of Contents
    • /
    • 제11권3호
    • /
    • pp.47-53
    • /
    • 2015
  • Visual object tracking is a fundamental problem in the field of computer vision, as it needs a proper model to account for drastic appearance changes that are caused by shape, textural, and illumination variations. In this paper, we propose a feature-based visual-object-tracking method with a sparse representation. Generally, most appearance-based models use the gray-scale pixel values of the input image, but this might be insufficient for a description of the target object under a variety of conditions. To obtain the proper information regarding the target object, the following combination of features has been exploited as a corresponding representation: First, the features of the target templates are extracted by using the HOG (histogram of gradient) and LBPs (local binary patterns); secondly, a feature-based sparsity is attained by solving the minimization problems, whereby the target object is represented by the selection of the minimum reconstruction error. The strengths of both features are exploited to enhance the overall performance of the tracker; furthermore, the proposed method is integrated with the particle-filter framework and achieves a promising result in terms of challenging tracking videos.