• Title/Summary/Keyword: Scene Matching

Search Result 156, Processing Time 0.023 seconds

Multi-Range Approach of Stereo Vision for Mobile Robot Navigation in Uncertain Environments

  • Park, Kwang-Ho;Kim, Hyung-O;Baek, Moon-Yeol;Kee, Chang-Doo
    • Journal of Mechanical Science and Technology
    • /
    • v.17 no.10
    • /
    • pp.1411-1422
    • /
    • 2003
  • The detection of free spaces between obstacles in a scene is a prerequisite for navigation of a mobile robot. Especially for stereo vision-based navigation, the problem of correspondence between two images is well known to be of crucial importance. This paper describes multi-range approach of area-based stereo matching for grid mapping and visual navigation in uncertain environment. Camera calibration parameters are optimized by evolutionary algorithm for successful stereo matching. To obtain reliable disparity information from both images, stereo images are to be decomposed into three pairs of images with different resolution based on measurement of disparities. The advantage of multi-range approach is that we can get more reliable disparity in each defined range because disparities from high resolution image are used for farther object a while disparities from low resolution images are used for close objects. The reliable disparity map is combined through post-processing for rejecting incorrect disparity information from each disparity map. The real distance from a disparity image is converted into an occupancy grid representation of a mobile robot. We have investigated the possibility of multi-range approach for the detection of obstacles and visual mapping through various experiments.

Time Analysis of Structural Element and Theme Association of Television News Imagery (텔레비전 뉴스 영상의 구조적 요소와 주제연관성 시계열 분석)

  • Park, Dug-Chun
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.7
    • /
    • pp.100-109
    • /
    • 2011
  • This thesis is a content analysis on whether the proportion of structural element and theme association of television news imagery is different, depending on the historical background, and on what it means, which can be the index of scene-based and realistic report. Most researches of television news are horizontal studies of the same period, making light of vertical studies reflecting the change of age. Therefore, This study analyzed 729 items composed of 11,945 shots extracted from MBC Newsdesk from 1987, to 2007, the samples of which were extracted by systematic random sampling with five years' interval. This content analysis found out that there was high proportion of scene-based and realistic report such as 'sound-bite', 'event footage', 'direct matching' in the year 1987, 2007, and high proportion of 'corroboration shot', 'file footage', 'indirect reference', 'literal matching only' in the year 1997, which revealed the fact that reality-based report had not been faithfully accomplished in 1997.

3D image mosaicking technique using multiple planes for urban visualization (복수 투영면을 사용한 도심지 가시화용 3 차원 모자이크 기술)

  • CHON Jaechoon;KIM Hyongsuk
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.3 s.303
    • /
    • pp.41-50
    • /
    • 2005
  • A novel image mosaicking technique suitable for 3D urban visualization is proposed. It is not effective to apply 2D image mosaicking techniques for urban visualization when, for example, one is filming a sequence of images from a side-looking video camera along a road in an urban area. The proposed method presents the roadside scene captured by a side-looking video camera as a continuous set of textured planar faces, which are termed 'multiple planes' in this paper. The exterior parameters of each frame are first calculated through automatically selected matching feature points. The matching feature points are also used to estimate a plane approximation of the scene geometry for each frame. These planes are concatenated to create an approximate model on which images are back-projected as textures. Here, we demonstrate algorithm that creates efficient image mosaics in 3D space from a sequence of real images.

Efficient video matching method for illegal video detection (불법 동영상 검출을 위한 효율적인 동영상 정합 방법)

  • Choi, Minseok
    • Journal of Digital Convergence
    • /
    • v.20 no.1
    • /
    • pp.179-184
    • /
    • 2022
  • With the development of information and communication technology, the production and distribution of digital contents is rapidly increasing, and the distribution of illegally copied contents also increases, causing various problems. In order to prevent illegal distribution of contents, a DRM (Digital Rights Management)-based approach can be used, but in a situation where the contents are already copied and distributed, a method of searching and detecting the duplicated contents is required. In this paper, a duplication detection method based on the contents of video content is proposed. The proposed method divides the video into scene units using the visual rhythm extracted from the video, and hierarchically applies the playback time and color feature values of each divided scene to quickly and efficiently detect duplicate videos in a large database. Through experiments, it was shown that the proposed method can reliably detect various replication modifications.

3D feature point extraction technique using a mobile device (모바일 디바이스를 이용한 3차원 특징점 추출 기법)

  • Kim, Jin-Kyum;Seo, Young-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.256-257
    • /
    • 2022
  • In this paper, we introduce a method of extracting three-dimensional feature points through the movement of a single mobile device. Using a monocular camera, a 2D image is acquired according to the camera movement and a baseline is estimated. Perform stereo matching based on feature points. A feature point and a descriptor are acquired, and the feature point is matched. Using the matched feature points, the disparity is calculated and a depth value is generated. The 3D feature point is updated according to the camera movement. Finally, the feature point is reset at the time of scene change by using scene change detection. Through the above process, an average of 73.5% of additional storage space can be secured in the key point database. By applying the algorithm proposed to the depth ground truth value of the TUM Dataset and the RGB image, it was confirmed that the\re was an average distance difference of 26.88mm compared with the 3D feature point result.

  • PDF

Detecting near-duplication Video Using Motion and Image Pattern Descriptor (움직임과 영상 패턴 서술자를 이용한 중복 동영상 검출)

  • Jin, Ju-Kyong;Na, Sang-Il;Jenong, Dong-Seok
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.4
    • /
    • pp.107-115
    • /
    • 2011
  • In this paper, we proposed fast and efficient algorithm for detecting near-duplication based on content based retrieval in large scale video database. For handling large amounts of video easily, we split the video into small segment using scene change detection. In case of video services and copyright related business models, it is need to technology that detect near-duplicates, that longer matched video than to search video containing short part or a frame of original. To detect near-duplicate video, we proposed motion distribution and frame descriptor in a video segment. The motion distribution descriptor is constructed by obtaining motion vector from macro blocks during the video decoding process. When matching between descriptors, we use the motion distribution descriptor as filtering to improving matching speed. However, motion distribution has low discriminability. To improve discrimination, we decide to identification using frame descriptor extracted from selected representative frames within a scene segmentation. The proposed algorithm shows high success rate and low false alarm rate. In addition, the matching speed of this descriptor is very fast, we confirm this algorithm can be useful to practical application.

A Method for Reconstructing Original Images for Captions Areas in Videos Using Block Matching Algorithm (블록 정합을 이용한 비디오 자막 영역의 원 영상 복원 방법)

  • 전병태;이재연;배영래
    • Journal of Broadcast Engineering
    • /
    • v.5 no.1
    • /
    • pp.113-122
    • /
    • 2000
  • It is sometimes necessary to remove the captions and recover original images from video images already broadcast, When the number of images requiring such recovery is small, manual processing is possible, but as the number grows it would be very difficult to do it manually. Therefore, a method for recovering original image for the caption areas in needed. Traditional research on image restoration has focused on restoring blurred images to sharp images using frequency filtering or video coding for transferring video images. This paper proposes a method for automatically recovering original image using BMA(Block Matching Algorithm). We extract information on caption regions and scene change that is used as a prior-knowledge for recovering original image. From the result of caption information detection, we know the start and end frames of captions in video and the character areas in the caption regions. The direction for the recovery is decided using information on the scene change and caption region(the start and end frame for captions). According to the direction, we recover the original image by performing block matching for character components in extracted caption region. Experimental results show that the case of stationary images with little camera or object motion is well recovered. We see that the case of images with motion in complex background is also recovered.

  • PDF

A 3-D Vision Sensor Implementation on Multiple DSPs TMS320C31 (다중 TMS320C31 DSP를 사용한 3-D 비젼센서 Implementation)

  • Oksenhendler, V.;Bensrhair, Abdelaziz;Miche, Pierre;Lee, Sang-Goog
    • Journal of Sensor Science and Technology
    • /
    • v.7 no.2
    • /
    • pp.124-130
    • /
    • 1998
  • High-speed 3D vision systems are essential for autonomous robot or vehicle control applications. In our study, a stereo vision process has been developed. It consists of three steps : extraction of edges in right and left images, matching corresponding edges and calculation of the 3D map. This process is implemented in a VME 150/40 Imaging Technology vision system. It is a modular system composed by a display, an acquisition, a four Mbytes image frame memory, and three computational cards. Programmable accelerator computational modules are running at 40 MHz and are based on TMS320C31 DSP with a $64{\times}32$ bit instruction cache and two $1024{\times}32$ bit internal RAMs. Each is equipped with 512 Kbytes static RAM, 4 Mbytes image memory, 1 Mbytes flash EEPROM and a serial port. Data transfers and communications between modules are provided by three 8 bit global video bus, and three local configurable pipeline 8 bit video bus. The VME bus is dedicated to system management. Tasks between DSPs are distributed as follows: two DSPs are used to edges detection, one for the right image and the other for the left one. The last processor computes the matching process and the 3D calculation. With $512{\times}512$ pixels images, this sensor generates dense 3D maps at a rate of about 1 Hz depending of the scene complexity. Results can surely be improved by using a special suited multiprocessors cards.

  • PDF

Content based Video Copy Detection Using Spatio-Temporal Ordinal Measure (시공간 순차 정보를 이용한 내용기반 복사 동영상 검출)

  • Jeong, Jae-Hyup;Kim, Tae-Wang;Yang, Hun-Jun;Jin, Ju-Kyong;Jeong, Dong-Seok
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.2
    • /
    • pp.113-121
    • /
    • 2012
  • In this paper, we proposed fast and efficient algorithm for detecting near-duplication based on content based retrieval in large scale video database. For handling large amounts of video easily, we split the video into small segment using scene change detection. In case of video services and copyright related business models, it is need to technology that detect near-duplicates, that longer matched video than to search video containing short part or a frame of original. To detect near-duplicate video, we proposed motion distribution and frame descriptor in a video segment. The motion distribution descriptor is constructed by obtaining motion vector from macro blocks during the video decoding process. When matching between descriptors, we use the motion distribution descriptor as filtering to improving matching speed. However, motion distribution has low discriminability. To improve discrimination, we decide to identification using frame descriptor extracted from selected representative frames within a scene segmentation. The proposed algorithm shows high success rate and low false alarm rate. In addition, the matching speed of this descriptor is very fast, we confirm this algorithm can be useful to practical application.

A Scene Change Detection using Motion Estimation in Animation Sequence (움직임 추정을 이용한 애니메이션 영상의 장면전환 검출)

  • Kwak, Sung-Keun
    • Journal of the Korea Computer Industry Society
    • /
    • v.9 no.4
    • /
    • pp.149-156
    • /
    • 2008
  • There is the temporal correlation of a animation sequence between the motion vector of current block and the motion vector of previous block. In this paper, we propose the scene change detection algorithm for block matching using the temporal correlation of the animation sequence and the center-biased property of motion vectors. The proposed algorithm determines the location of a better starting point for the search of an exact motion vector using the point of the smallest SAD(sum of absolute difference) value by the predicted motion vector from the same block of the previous frame and the predictor candidate point on each search region. Simulation results show that the proposed algorithm has better detection performance, such as recall rate, then the existing method. The algorithm has the advantage of speed, simplicity and accuracy. In addition, it requires less amount of storage.

  • PDF