• Title/Summary/Keyword: Foreground object segmentation

Search Result 45, Processing Time 0.023 seconds

A Robust Object Detection and Tracking Method using RGB-D Model (RGB-D 모델을 이용한 강건한 객체 탐지 및 추적 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.4
    • /
    • pp.61-67
    • /
    • 2017
  • Recently, CCTV has been combined with areas such as big data, artificial intelligence, and image analysis to detect various abnormal behaviors and to detect and analyze the overall situation of objects such as people. Image analysis research for this intelligent video surveillance function is progressing actively. However, CCTV images using 2D information generally have limitations such as object misrecognition due to lack of topological information. This problem can be solved by adding the depth information of the object created by using two cameras to the image. In this paper, we perform background modeling using Mixture of Gaussian technique and detect whether there are moving objects by segmenting the foreground from the modeled background. In order to perform the depth information-based segmentation using the RGB information-based segmentation results, stereo-based depth maps are generated using two cameras. Next, the RGB-based segmented region is set as a domain for extracting depth information, and depth-based segmentation is performed within the domain. In order to detect the center point of a robustly segmented object and to track the direction, the movement of the object is tracked by applying the CAMShift technique, which is the most basic object tracking method. From the experiments, we prove the efficiency of the proposed object detection and tracking method using the RGB-D model.

Joint Segmentation of Multi-View Images by Region Correspondence (영역 대응을 이용한 다시점 영상 집합의 통합 영역화)

  • Lee, Soo-Chahn;Kwon, Dong-Jin;Yun, Il-Dong;Lee, Sang-Uk
    • Journal of Broadcast Engineering
    • /
    • v.13 no.5
    • /
    • pp.685-695
    • /
    • 2008
  • This paper presents a method to segment the object of interest from a set of multi-view images with minimal user interaction. Specifically, after the user segments an initial image, we first estimate the transformations between foreground and background of the segmented image and the neighboring image, respectively. From these transformations, we obtain regions in the neighboring image that respectively correspond to the foreground and the background of the segmented image. We are then able to segment the neighboring image based on these regions, and iterate this process to segment the whole image set. Transformation of foregrounds are estimated by feature-based registration with free-form deformation, while transformation of backgrounds are estimated by homography constrained to affine transformation. Here, both are based on correspondence point pairs. Segmentation is done by estimating pixel color distributions and defining a shape prior based on the obtained foreground and background regions and applying them to a Markov random field (MRF) energy minimization framework for image segmentation. Experimental results demonstrate the effectiveness of the proposed method.

Confidence-based Background Subtraction Algorithm for Moving Cameras (움직이는 카메라를 위한 신뢰도 기반의 배경 제거 알고리즘)

  • Mun, Hyeok;Lee, Bok Ju;Choi, Young Kyu
    • Journal of the Semiconductor & Display Technology
    • /
    • v.16 no.4
    • /
    • pp.30-35
    • /
    • 2017
  • Moving object segmentation from a nonstationary camera is a difficult problem due to the motion of both camera and the object. In this paper, we propose a new confidence-based background subtraction technique from moving camera. The method is based on clustering of motion vectors and generating adaptive multi-homography from a pair of adjacent video frames. The main innovation concerns the use of confidence images for each foreground and background motion groups. Experimental results revealed that our confidence-based approach robustly detect moving targets in sequences taken by a freely moving camera.

  • PDF

Data Augmentation for Tomato Detection and Pose Estimation (토마토 위치 및 자세 추정을 위한 데이터 증대기법)

  • Jang, Minho;Hwang, Youngbae
    • Journal of Broadcast Engineering
    • /
    • v.27 no.1
    • /
    • pp.44-55
    • /
    • 2022
  • In order to automatically provide information on fruits in agricultural related broadcasting contents, instance image segmentation of target fruits is required. In addition, the information on the 3D pose of the corresponding fruit may be meaningfully used. This paper represents research that provides information about tomatoes in video content. A large amount of data is required to learn the instance segmentation, but it is difficult to obtain sufficient training data. Therefore, the training data is generated through a data augmentation technique based on a small amount of real images. Compared to the result using only the real images, it is shown that the detection performance is improved as a result of learning through the synthesized image created by separating the foreground and background. As a result of learning augmented images using images created using conventional image pre-processing techniques, it was shown that higher performance was obtained than synthetic images in which foreground and background were separated. To estimate the pose from the result of object detection, a point cloud was obtained using an RGB-D camera. Then, cylinder fitting based on least square minimization is performed, and the tomato pose is estimated through the axial direction of the cylinder. We show that the results of detection, instance image segmentation, and cylinder fitting of a target object effectively through various experiments.

Automatic Extraction of Focused Video Object from Low Depth-of-Field Image Sequences (낮은 피사계 심도의 동영상에서 포커스 된 비디오 객체의 자동 검출)

  • Park, Jung-Woo;Kim, Chang-Ick
    • Journal of KIISE:Software and Applications
    • /
    • v.33 no.10
    • /
    • pp.851-861
    • /
    • 2006
  • The paper proposes a novel unsupervised video object segmentation algorithm for image sequences with low depth-of-field (DOF), which is a popular photographic technique enabling to represent the intention of photographer by giving a clear focus only on an object-of-interest (OOI). The proposed algorithm largely consists of two modules. The first module automatically extracts OOIs from the first frame by separating sharply focused OOIs from other out-of-focused foreground or background objects. The second module tracks OOIs for the rest of the video sequence, aimed at running the system in real-time, or at least, semi-real-time. The experimental results indicate that the proposed algorithm provides an effective tool, which can be a basis of applications, such as video analysis for virtual reality, immersive video system, photo-realistic video scene generation and video indexing systems.

Crab Region Extraction Method from Suncheon Bay Tidal Flat Images (순천만 갯벌 영상에서 게 영역 추출 방법)

  • Park, Sang-Hyun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.14 no.6
    • /
    • pp.1197-1206
    • /
    • 2019
  • Suncheon Bay is a very important natural resource and various efforts have been made to protect it from the environmental pollution. Although the project to monitor the environmental changes in periodically by observing the creatures in tidal flats is processing, it is being done inefficiently by people directly observing it. In this paper, we propose an object segmentation method that can be applied to the method to automatically monitor the living creatures in the tidal flats. In the proposed method, a foreground map representing the location of objects is obtained by using a temporal difference method, and a superpixel method is applied to detect the detailed boundary of an image. Finally the region of crab is extracted by combining the foreground map and the superpixel information. Experimental results show that the proposed method separates crab regions from a tidal flat image easily and accurately.

Detection of Video Scene Boundaries based on the Local and Global Context Information (지역 컨텍스트 및 전역 컨텍스트 정보를 이용한 비디오 장면 경계 검출)

  • 강행봉
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.6
    • /
    • pp.778-786
    • /
    • 2002
  • Scene boundary detection is important in the understanding of semantic structure from video data. However, it is more difficult than shot change detection because scene boundary detection needs to understand semantics in video data well. In this paper, we propose a new approach to scene segmentation using contextual information in video data. The contextual information is divided into two categories: local and global contextual information. The local contextual information refers to the foreground regions' information, background and shot activity. The global contextual information refers to the video shot's environment or its relationship with other video shots. Coherence, interaction and the tempo of video shots are computed as global contextual information. Using the proposed contextual information, we detect scene boundaries. Our proposed approach consists of three consecutive steps: linking, verification, and adjusting. We experimented the proposed approach using TV dramas and movies. The detection accuracy of correct scene boundaries is over than 80%.

An Improved Cast Shadow Removal in Object Detection (객체검출에서의 개선된 투영 그림자 제거)

  • Nguyen, Thanh Binh;Chung, Sun-Tae;Kim, Yu-Sung;Kim, Jae-Min
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2009.05a
    • /
    • pp.889-894
    • /
    • 2009
  • Accompanied by the rapid development of Computer Vision, Visual surveillance has achieved great evolution with more and more complicated processing. However there are still many problems to be resolved for robust and reliable visual surveillance, and the cast shadow occurring in motion detection process is one of them. Shadow pixels are often misclassified as object pixels so that they cause errors in localization, segmentation, tracking and classification of objects. This paper proposes a novel cast shadow removal method. As opposed to previous conventional methods, which considers pixel properties like intensity properties, color distortion, HSV color system, and etc., the proposed method utilizes observations about edge patterns in the shadow region in the current frame and the corresponding region in the background scene, and applies Laplacian edge detector to the blob regions in the current frame and the background scene. Then, the product of the outcomes of application determines whether the blob pixels in the foreground mask comes from object blob regions or shadow regions. The proposed method is simple but turns out practically very effective for Gaussian Mixture Model, which is verified through experiments.

  • PDF

A Content-Based Image Classification using Neural Network (신경망을 이용한 내용기반 영상 분류)

  • 이재원;김상균
    • Journal of Korea Multimedia Society
    • /
    • v.5 no.5
    • /
    • pp.505-514
    • /
    • 2002
  • In this Paper, we propose a method of content-based image classification using neural network. The images for classification ate object images that can be divided into foreground and background. To deal with the object images efficiently, object region is extracted with a region segmentation technique in the preprocessing step. Features for the classification are texture and shape features extracted from wavelet transformed image. The neural network classifier is constructed with the extracted features and the back-propagation learning algorithm. Among the various texture features, the diagonal moment was more effective. A test with 300 training data and 300 test data composed of 10 images from each of 30 classes shows correct classification rates of 72.3% and 67%, respectively.

  • PDF

Implementation of Surveillance System using Motion Tracking Method based on Mobile (모바일 기반의 동작 추적 기법을 이용한 감시 시스템의 구현)

  • Kim, Hyeng-Gyun;Kim, Yong-Ho;Guen, Bae-Yong
    • Journal of Advanced Navigation Technology
    • /
    • v.12 no.2
    • /
    • pp.164-169
    • /
    • 2008
  • This paper is using motion tracking by image segmentation to monitor intruders and to confirm based on mobile the relevant information. First, detect frame in animation that film fixed area, and make use of image subtraction between two frame that adjoin, segment fixed backing and target who move. Segmental foreground object to the edge detecting the location specified by the edge of the median estimate extracted by analyzing the motion of the intruders to monitor. When a motion is detected, the detected image is transmitted by using the W AP pull basis image transmission method on the mobile client data terminal.

  • PDF