• Title/Summary/Keyword: Motion segmentation

Search Result 203, Processing Time 0.042 seconds

Moving Object Detection using Clausius Entropy and Adaptive Gaussian Mixture Model (클라우지우스 엔트로피와 적응적 가우시안 혼합 모델을 이용한 움직임 객체 검출)

  • Park, Jong-Hyun;Lee, Gee-Sang;Toan, Nguyen Dinh;Cho, Wan-Hyun;Park, Soon-Young
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.47 no.1
    • /
    • pp.22-29
    • /
    • 2010
  • A real-time detection and tracking of moving objects in video sequences is very important for smart surveillance systems. In this paper, we propose a novel algorithm for the detection of moving objects that is the entropy-based adaptive Gaussian mixture model (AGMM). First, the increment of entropy generally means the increment of complexity, and objects in unstable conditions cause higher entropy variations. Hence, if we apply these properties to the motion segmentation, pixels with large changes in entropy in moments have a higher chance in belonging to moving objects. Therefore, we apply the Clausius entropy theory to convert the pixel value in an image domain into the amount of energy change in an entropy domain. Second, we use an adaptive background subtraction method to detect moving objects. This models entropy variations from backgrounds as a mixture of Gaussians. Experiment results demonstrate that our method can detect motion object effectively and reliably.

A study on Recognition and Preference of Healthy and Functional Textile according to Psychological Comfort of the Silver Generation (실버세대의 심리적 안정감에 따른 건강 기능성 섬유 인지 및 선호에 관한 연구)

  • Seo, Min Nyoung;Koo, Young Seok
    • Fashion & Textile Research Journal
    • /
    • v.16 no.5
    • /
    • pp.811-821
    • /
    • 2014
  • The purposes of this study were to investigate difference on recognition, preference, satisfaction and possession of functional textile clothing in psychological comfort groups according to gender and age of the elderly and suggest optimal physical, mental and healthy functional textiles for the elderly. The data was collected from 262 respondents in their age of over 55, who lived in Busan. SPSS 21.0 was used and t-test, cross tabulation analysis, frequency analysis, and descriptive analysis were performed for analysis. The results are as follows. First, high groups of psychological comfort for both gender and age of the elderly showed higher recognition and preference of functional textiles, especially in the high groups of the new silver generation and women. Second, the high groups of psychological comfort for both gender and age of the elderly showed higher satisfaction of functional textiles. However, it was lower than preference of functional textiles. Last, the possession of major clothing function over all the elderly was comfort, fitting motion adaptability, and health safety function in order. Low groups of psychological comfort possessed fitting motion adaptability textile clothing more than the high groups, and the high groups of psychological comfort possessed comfort and health safety textile clothing more than low groups. In conclusion, this research showed the importance of functional textiles for the elderly in terms of psychological comfort as well as the need for healthy comfort textiles for the advanced activities of daily life.

Moving Object Extraction and Relative Depth Estimation of Backgrould regions in Video Sequences (동영상에서 물체의 추출과 배경영역의 상대적인 깊이 추정)

  • Park Young-Min;Chang Chu-Seok
    • The KIPS Transactions:PartB
    • /
    • v.12B no.3 s.99
    • /
    • pp.247-256
    • /
    • 2005
  • One of the classic research problems in computer vision is that of stereo, i.e., the reconstruction of three dimensional shape from two or more images. This paper deals with the problem of extracting depth information of non-rigid dynamic 3D scenes from general 2D video sequences taken by monocular camera, such as movies, documentaries, and dramas. Depth of the blocks are extracted from the resultant block motions throughout following two steps: (i) calculation of global parameters concerned with camera translations and focal length using the locations of blocks and their motions, (ii) calculation of each block depth relative to average image depth using the global parameters and the location of the block and its motion, Both singular and non-singular cases are experimented with various video sequences. The resultant relative depths and ego-motion object shapes are virtually identical to human vision.

Posture Recognition for a Bi-directional Participatory TV Program based on Face Color Region and Motion Map (시청자 참여형 양방향 TV 방송을 위한 얼굴색 영역 및 모션맵 기반 포스처 인식)

  • Hwang, Sunhee;Lim, Kwangyong;Lee, Suwoong;Yoo, Hoyoung;Byun, Hyeran
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.8
    • /
    • pp.549-554
    • /
    • 2015
  • As intuitive hardware interfaces continue to be developed, it has become more important to recognize the posture of the user. An efficient alternative to adding expensive sensors is to implement computer vision systems. This paper proposes a method to recognize a user's postured in a live broadcast bi-directional participatory TV program. The proposed method first estimates the position of the user's hands by generation a facial color map for the user and a motion map. The posture is then recognized by computing the relative position of the face and the hands. This method exhibited 90% accuracy in an experiment to recognize three defined postures during the live broadcast bi-directional participatory TV program, even when the input images contained a complex background.

Fast information extraction algorithm for object-based MPEG-4 conversion from MPEG-1,2 (MPEG-1,2로부터 객체 기반 MPEG-4 변환을 위한 고속 정보 추출 알고리즘)

  • 양종호;박성욱
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.91-102
    • /
    • 2004
  • In this paper, a fast information extraction algorithm for object-based MPEG-4 application from MPEG-1,2 is proposed. For object-based MPEG-4 conversion, we need to extract such information as object-image, shape-image, macro-block motion vector, and header information from MPEG-1,2 bit-stream. If we use the extracted information, fast conversion for object-based MPEG-4 is possible. The proposed object extraction algerian has two important steps, namely the motion vector extraction from MPEG-1,2 bit-stream and the watershed algerian The algorithm extracts objects using user's assistance in the intra frame and tracks then in the following inter frames. If we have an unsatisfactory result for a fast moving object the user can intervene to connect the segmentation. The proposed algorithm consist of two steps, which are intra frame object extracting processing and inter frame tracking processing. Object extracting process is the step in which user extracts a semantic object directly by using the block classification and watersheds. Object tracking process is the step of the following the object in the subsequent frames. It is based on the boundary fitting method using motion vector, object-mask and modified watersheds. Experimental results show that the proposed method can achieve a fast conversion from the MPEG-1,2 bit-stream to the object-based MPEG-4 input.

Real-Time 2D-to-3D Conversion for 3DTV using Time-Coherent Depth-Map Generation Method

  • Nam, Seung-Woo;Kim, Hye-Sun;Ban, Yun-Ji;Chien, Sung-Il
    • International Journal of Contents
    • /
    • v.10 no.3
    • /
    • pp.9-16
    • /
    • 2014
  • Depth-image-based rendering is generally used in real-time 2D-to-3D conversion for 3DTV. However, inaccurate depth maps cause flickering issues between image frames in a video sequence, resulting in eye fatigue while viewing 3DTV. To resolve this flickering issue, we propose a new 2D-to-3D conversion scheme based on fast and robust depth-map generation from a 2D video sequence. The proposed depth-map generation algorithm divides an input video sequence into several cuts using a color histogram. The initial depth of each cut is assigned based on a hypothesized depth-gradient model. The initial depth map of the current frame is refined using color and motion information. Thereafter, the depth map of the next frame is updated using the difference image to reduce depth flickering. The experimental results confirm that the proposed scheme performs real-time 2D-to-3D conversions effectively and reduces human eye fatigue.

An Improved Cast Shadow Removal in Object Detection (객체검출에서의 개선된 투영 그림자 제거)

  • Nguyen, Thanh Binh;Chung, Sun-Tae;Kim, Yu-Sung;Kim, Jae-Min
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2009.05a
    • /
    • pp.889-894
    • /
    • 2009
  • Accompanied by the rapid development of Computer Vision, Visual surveillance has achieved great evolution with more and more complicated processing. However there are still many problems to be resolved for robust and reliable visual surveillance, and the cast shadow occurring in motion detection process is one of them. Shadow pixels are often misclassified as object pixels so that they cause errors in localization, segmentation, tracking and classification of objects. This paper proposes a novel cast shadow removal method. As opposed to previous conventional methods, which considers pixel properties like intensity properties, color distortion, HSV color system, and etc., the proposed method utilizes observations about edge patterns in the shadow region in the current frame and the corresponding region in the background scene, and applies Laplacian edge detector to the blob regions in the current frame and the background scene. Then, the product of the outcomes of application determines whether the blob pixels in the foreground mask comes from object blob regions or shadow regions. The proposed method is simple but turns out practically very effective for Gaussian Mixture Model, which is verified through experiments.

  • PDF

A Ubiquitous Vision System based on the Identified Contract Net Protocol (Identified Contract Net 프로토콜 기반의 유비쿼터스 시각시스템)

  • Kim, Chi-Ho;You, Bum-Jae;Kim, Hagbae
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.10
    • /
    • pp.620-629
    • /
    • 2005
  • In this paper, a new protocol-based approach was proposed for development of a ubiquitous vision system. It is possible to apply the approach by regarding the ubiquitous vision system as a multiagent system. Thus, each vision sensor can be regarded as an agent (vision agent). Each vision agent independently performs exact segmentation for a target by color and motion information, visual tracking for multiple targets in real-time, and location estimation by a simple perspective transform. Matching problem for the identity of a target during handover between vision agents is solved by the Identified Contract Net (ICN) protocol implemented for the protocol-based approach. The protocol-based approach by the ICN protocol is independent of the number of vision agents and moreover the approach doesn't need calibration and overlapped region between vision agents. Therefore, the ICN protocol raises speed, scalability, and modularity of the system. The protocol-based approach was successfully applied for our ubiquitous vision system and operated well through several experiments.

Graph-based Motion Segmentation using Normalized Cuts (Normalized Cuts을 이용한 그래프 기반의 모션 분할)

  • Yun, Sung-Ju;Park, An-Jin;Jung, Kee-Chul
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2008.06c
    • /
    • pp.522-526
    • /
    • 2008
  • 모션 캡쳐 장비는 사람의 자연스러운 행동이나 동작 정보를 정밀하게 얻기 위해 널리 사용되며, 영화나 게임과 같은 콘텐츠에서 자주 활용되고 있다. 하지만 모션 캡쳐 장비가 고가이기 때문에 한번 입력받은 데이터를 모션별로 분할하고 상황에 맞게 재결합하여 사용할 필요가 있으며, 입력 데이터를 모션별로 분할하는 것은 대부분 수동으로 이루어진다. 이 때문에 캡쳐된 데이터를 자동으로 분할하기 위한 연구들이 다양하게 시도되고 있다. 기존의 연구들은 크게 전역적 특성에 대한 고려없이 이웃하는 프레임만을 고려하는 온라인 방식과 데이터를 전역적으로 고려하나 이웃하는 프레임 사이의 관계를 고려하지 않는 오프라인 방식으로 나누어진다. 본 논문에서는 온라인과 오프라인 방식을 병합한 그래프 기반의 모션 분할 방법을 제안한다. 분할을 위해 먼저 모션데이터를 기반으로 그래프를 생성하며, 그래프는 이웃하는 각 프레임사이의 유사도뿐만 아니라 시간축을 기반으로 일정시간내의 프레임들의 유사도를 모두 고려하였다. 이렇게 생성된 그래프를 분할하기 위해 분할된 모션내의 유사도 합을 최소화하고 각 모션간의 유사도는 최대화할 수 있는 normalized cuts을 이용하였다. 실험에서 제안된 방법은 기존의 오프라인 방식 중 하나인 GMM과 온라인 방식 중 하나인 국부최소값 분할 방법보다 좋은 결과를 보였으며, 이는 각 프레임 사이의 유사도뿐만 아니라 일정시간내의 유사도를 전역적으로 고려하기 때문이다.

  • PDF

Support Vector Machine Based Phoneme Segmentation for Lip Synch Application

  • Lee, Kun-Young;Ko, Han-Seok
    • Speech Sciences
    • /
    • v.11 no.2
    • /
    • pp.193-210
    • /
    • 2004
  • In this paper, we develop a real time lip-synch system that activates 2-D avatar's lip motion in synch with an incoming speech utterance. To realize the 'real time' operation of the system, we contain the processing time by invoking merge and split procedures performing coarse-to-fine phoneme classification. At each stage of phoneme classification, we apply the support vector machine (SVM) to reduce the computational load while retraining the desired accuracy. The coarse-to-fine phoneme classification is accomplished via two stages of feature extraction: first, each speech frame is acoustically analyzed for 3 classes of lip opening using Mel Frequency Cepstral Coefficients (MFCC) as a feature; secondly, each frame is further refined in classification for detailed lip shape using formant information. We implemented the system with 2-D lip animation that shows the effectiveness of the proposed two-stage procedure in accomplishing a real-time lip-synch task. It was observed that the method of using phoneme merging and SVM achieved about twice faster speed in recognition than the method employing the Hidden Markov Model (HMM). A typical latency time per a single frame observed for our method was in the order of 18.22 milliseconds while an HMM method applied under identical conditions resulted about 30.67 milliseconds.

  • PDF