• Title/Summary/Keyword: Video object segmentation

Search Result 141, Processing Time 0.024 seconds

Separation of Occluding Pigs using Deep Learning-based Image Processing Techniques (딥 러닝 기반의 영상처리 기법을 이용한 겹침 돼지 분리)

  • Lee, Hanhaesol;Sa, Jaewon;Shin, Hyunjun;Chung, Youngwha;Park, Daihee;Kim, Hakjae
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.2
    • /
    • pp.136-145
    • /
    • 2019
  • The crowded environment of a domestic pig farm is highly vulnerable to the spread of infectious diseases such as foot-and-mouth disease, and studies have been conducted to automatically analyze behavior of pigs in a crowded pig farm through a video surveillance system using a camera. Although it is required to correctly separate occluding pigs for tracking each individual pigs, extracting the boundaries of the occluding pigs fast and accurately is a challenging issue due to the complicated occlusion patterns such as X shape and T shape. In this study, we propose a fast and accurate method to separate occluding pigs not only by exploiting the characteristics (i.e., one of the fast deep learning-based object detectors) of You Only Look Once, YOLO, but also by overcoming the limitation (i.e., the bounding box-based object detector) of YOLO with the test-time data augmentation of rotation. Experimental results with two-pigs occlusion patterns show that the proposed method can provide better accuracy and processing speed than one of the state-of-the-art widely used deep learning-based segmentation techniques such as Mask R-CNN (i.e., the performance improvement over Mask R-CNN was about 11 times, in terms of the accuracy/processing speed performance metrics).

Video Object Segmentation Method Using Spatio-Temporal Information (시공간 정보를 이용한 동영상 객체 분할 기법)

  • Oh, Hyuk;Choi, Hwan-Soo;Jeong, Dong-Seok
    • Proceedings of the IEEK Conference
    • /
    • 2000.09a
    • /
    • pp.349-352
    • /
    • 2000
  • 영상으로부터 의미있는 객체를 영역화하기 위하여, 움직임에 의한 시간적 정보를 이용하거나, 형태학적(Morphological) 기법과 같이 공간적 정보를 이용하는 방법이 있다. 그러나, 단지 시간적 정보나 공간적 정보만을 이용하는 방법은 그 한계를 가지고 있으며, 본 논문에서는 시공간 정보를 이용하여 분할하는 방법을 채택하였다. 시간적 분할에서는, 두 프레임에서 움직임 정보를 찾아내었던 기존 방법을 보완하여 연속되는 세 프레임을 사용하도록 하였다. 이렇게 하면 움직임이 미세한 영상에 대해서도 객체를 분리해 낼 가능성을 높일 수 있게 된다. 공간적 분할시에는, Watershed 알고리즘을 이용하는 형태학적 분할(Morphological Segmentation)[1][2]을 하게 되는데, 전처리 과정의 단일척도경사(Monoscale Gradient) 대신 다중척도 경사(Multiscale Gradient)[3][4]를 사용하여 미세한 경사는 누그러뜨리고 에지 부분의 경사만을 강조하게 하였다. 또한 개선된 Watershed 알고리즘을 제안하여 기존의 Watershed 알고리즘의 과분할 문제를 보완하였다.

  • PDF

Object Segmentation for Image Transmission Services and Facial Characteristic Detection based on Knowledge (화상전송 서비스를 위한 객체 분할 및 지식 기반 얼굴 특징 검출)

  • Lim, Chun-Hwan;Yang, Hong-Young
    • Journal of the Korean Institute of Telematics and Electronics T
    • /
    • v.36T no.3
    • /
    • pp.26-31
    • /
    • 1999
  • In this paper, we propose a facial characteristic detection algorithm based on knowledge and object segmentation method for image communication. In this algorithm, under the condition of the same lumination and distance from the fixed video camera to human face, we capture input images of 256 $\times$ 256 of gray scale 256 level and then remove the noise using the Gaussian filter. Two images are captured with a video camera, One contains the human face; the other contains only background region without including a face. And then we get a differential image between two images. After removing noise of the differential image by eroding End dilating, divide background image into a facial image. We separate eyes, ears, a nose and a mouth after searching the edge component in the facial image. From simulation results, we have verified the efficiency of the Proposed algorithm.

  • PDF

An Efficient Object Extraction Scheme for Low Depth-of-Field Images (낮은 피사계 심도 영상에서 관심 물체의 효율적인 추출 방법)

  • Park Jung-Woo;Lee Jae-Ho;Kim Chang-Ick
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.9
    • /
    • pp.1139-1149
    • /
    • 2006
  • This paper describes a novel and efficient algorithm, which extracts focused objects from still images with low depth-of-field (DOF). The algorithm unfolds into four modules. In the first module, a HOS map, in which the spatial distribution of the high-frequency components is represented, is obtained from an input low DOF image [1]. The second module finds OOI candidate by using characteristics of the HOS. Since it is possible to contain some holes in the region, the third module detects and fills them. In order to obtain an OOI, the last module gets rid of background pixels in the OOI candidate. The experimental results show that the proposed method is highly useful in various applications, such as image indexing for content-based retrieval from huge amounts of image database, image analysis for digital cameras, and video analysis for virtual reality, immersive video system, photo-realistic video scene generation and video indexing system.

  • PDF

Graph-based Moving Object Detection and Tracking in an H.264/SVC bitstream domain for Video Surveillance (감시 비디오를 위한 H.264/SVC 비트스트림 영역에서의 그래프 기반 움직임 객체 검출 및 추적)

  • Sabirin, Houari;Kim, Munchurl
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2012.07a
    • /
    • pp.298-301
    • /
    • 2012
  • This paper presents a graph-based method of detecting and tracking moving objects in H.264/SVC bitstreams for video surveillance applications that makes use the information from spatial base and enhancement layers of the bitstreams. In the base layer, segmentation of real moving objects are first performed using a spatio-temporal graph by removing false detected objects via graph pruning and graph projection, followed by graph matching to precisely identify the real moving objects over time even under occlusion. For the accurate detection and reliable tracking of moving objects in the enhancement layer, as well as saving computational complexity, the identified block groups of the real moving objects in the base layer are then mapped to the enhancement layer to provide accurate and efficient object detection and tracking in the bitstreams of higher resolution. Experimental results show the proposed method can produce reliable results with low computational complexity in both spatial layers of H.264/SVC test bitstreams.

  • PDF

Moving Object Extraction and Relative Depth Estimation of Backgrould regions in Video Sequences (동영상에서 물체의 추출과 배경영역의 상대적인 깊이 추정)

  • Park Young-Min;Chang Chu-Seok
    • The KIPS Transactions:PartB
    • /
    • v.12B no.3 s.99
    • /
    • pp.247-256
    • /
    • 2005
  • One of the classic research problems in computer vision is that of stereo, i.e., the reconstruction of three dimensional shape from two or more images. This paper deals with the problem of extracting depth information of non-rigid dynamic 3D scenes from general 2D video sequences taken by monocular camera, such as movies, documentaries, and dramas. Depth of the blocks are extracted from the resultant block motions throughout following two steps: (i) calculation of global parameters concerned with camera translations and focal length using the locations of blocks and their motions, (ii) calculation of each block depth relative to average image depth using the global parameters and the location of the block and its motion, Both singular and non-singular cases are experimented with various video sequences. The resultant relative depths and ego-motion object shapes are virtually identical to human vision.

An Improved Method for Detection of Moving Objects in Image Sequences Using Statistical Hypothesis Tests

  • Park, Jae-Gark;Kim, Munchurl;Lee, Myoung-Ho;Ahn, Chei-Teuk
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1998.06b
    • /
    • pp.171-176
    • /
    • 1998
  • This paper resents a spatio-temporal video segmentation method. The algorithm segments each frame of video sequences captured by a static or moving camera into moving objects (foreground) and background using a statistical hypothesis test. In the proposed method, three consecutive image frames are exploited and a hypothesis testing is performed by comparing two means from two consecutive difference images, which results in a T-test. This hypothesis test yields change detection mask that indicates moving areas (foreground) and non-moving areas (background). Moreover, an effective method for extracting object mask form change detection mask is proposed.

  • PDF

Object Segmentation Technique for Implementation of Interactive Video (상호작용 동영상 구현을 위한 객체 분리 제작 기법)

  • Sung, Hyuk-Jae;Kwak, Ho-Young
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2018.07a
    • /
    • pp.116-118
    • /
    • 2018
  • 본 논문에서는 기존의 동영상을 그랩컷(GrabCut) 알고리즘과 유니티3D를 이용하여 상호작용이 가능한 동영상을 제작하는 기법을 제안한다. 그랩컷 알고리즘을 이용하여 동영상에서 재생 프레임 단위로 원하는 객체 영역을 추출하고 흑백의 이미지로 이진화한다. 이진화된 결과물과 원본 동영상을 유니티3D에서 동시에 재생하면서 선택 영역의 이진화 픽셀 정보를 기반으로 사용자의 입력을 감지하는 동영상의 제작이 가능함을 보였다.

  • PDF

Video Retrieval based on Objects Motion Trajectory (객체 이동 궤적 기반 비디오의 검색)

  • 유웅식;이규원;김재곤;김진웅;권오석
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.5B
    • /
    • pp.913-924
    • /
    • 2000
  • This paper proposes an efficient descriptor for objects motion trajectory and a video retrieval algorithm based on objects motion trajectory. The algorithm describes parameters with coefficients of 2-order polynomial for objects motion trajectory after segmentation of the object from the scene. The algorithm also identifies types, intervals, and magnitude of global motion caused by camera motion and indexes them with 6-affine parameters. This paper implements content-based video retrieval using similarity-match between indexed parameters and queried ones for objects motion trajectory. The proposed algorithm will support not only faster retrieval for general videos but efficient operation for unmanned video surveillance system.

  • PDF

Comparisons of Color Spaces for Shadow Elimination (그림자 제거를 위한 색상 공간의 비교)

  • Lee, Gwang-Gook;Uzair, Muhammad;Yoon, Ja-Young;Kim, Jae-Jun;Kim, Whoi-Yul
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.5
    • /
    • pp.610-622
    • /
    • 2008
  • Moving object segmentation is an essential technique for various video surveillance applications. The result of moving object segmentation often contains shadow regions caused by the color difference of shadow pixels. Hence, moving object segmentation is usually followed by a shadow elimination process to remove the false detection results. The common assumption adopted in previous works is that, under the illumination variation, the value of chromaticity components are preserved while the value of intensity component is changed. Hence, color transforms which separates luminance component and chromaticity component are usually utilized to remove shadow pixels. In this paper, various color spaces (YCbCr, HSI, normalized rgb, Yxy, Lab, c1c2c3) are examined to find the most appropriate color space for shadow elimination. So far, there have been some research efforts to compare the influence of various color spaces for shadow elimination. However, previous efforts are somewhat insufficient to compare the color distortions under illumination change in diverse color spaces, since they used a specific shadow elimination scheme or different thresholds for different color spaces. In this paper, to relieve the limitations of previous works, (1) the amount of gradients in shadow boundaries drawn to uniform colored regions are examined only for chromaticity components to compare the color distortion under illumination change and (2) the accuracy of background subtraction are analyzed via RoC curves to compare different color spaces without the problem of threshold level selection. Through experiments on real video sequences, YCbCr and normalized rgb color spaces showed good results for shadow elimination among various color spaces used for the experiments.

  • PDF