• Title/Summary/Keyword: shot segmentation

Search Result 38, Processing Time 0.028 seconds

Shot Boundary Detection Algorithm by Compensating Pixel Brightness and Object Movement (화소 밝기와 객체 이동을 이용한 비디오 샷 경계 탐지 알고리즘)

  • Lee, Joon-Goo;Han, Ki-Sun;You, Byoung-Moon;Hwang, Doo-Sung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.5
    • /
    • pp.35-42
    • /
    • 2013
  • Shot boundary detection is an essential step for efficient browsing, sorting, and classification of video data. Robust shot detection method should overcome the disturbances caused by pixel brightness and object movement between frames. In this paper, two shot boundary detection methods are presented to address these problem by using segmentation, object movement, and pixel brightness. The first method is based on the histogram that reflects object movements and the morphological dilation operation that considers pixel brightness. The second method uses the pixel brightness information of segmented and whole blocks between frames. Experiments on digitized video data of National Archive of Korea show that the proposed methods outperforms the existing pixel-based and histogram-based methods.

Content based Video Segmentation Algorithm using Comparison of Pattern Similarity (장면의 유사도 패턴 비교를 이용한 내용기반 동영상 분할 알고리즘)

  • Won, In-Su;Cho, Ju-Hee;Na, Sang-Il;Jin, Ju-Kyong;Jeong, Jae-Hyup;Jeong, Dong-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.10
    • /
    • pp.1252-1261
    • /
    • 2011
  • In this paper, we propose the comparison method of pattern similarity for video segmentation algorithm. The shot boundary type is categorized as 2 types, abrupt change and gradual change. The representative examples of gradual change are dissolve, fade-in, fade-out or wipe transition. The proposed method consider the problem to detect shot boundary as 2-class problem. We concentrated if the shot boundary event happens or not. It is essential to define similarity between frames for shot boundary detection. We proposed 2 similarity measures, within similarity and between similarity. The within similarity is defined by feature comparison between frames belong to same shot. The between similarity is defined by feature comparison between frames belong to different scene. Finally we calculated the statistical patterns comparison between the within similarity and between similarity. Because this measure is robust to flash light or object movement, our proposed algorithm make contribution towards reducing false positive rate. We employed color histogram and mean of sub-block on frame image as frame feature. We performed the experimental evaluation with video dataset including set of TREC-2001 and TREC-2002. The proposed algorithm shows the performance, 91.84% recall and 86.43% precision in experimental circumstance.

MPEG Video Segmentation using Hierarchical Frame Search (계층적 프레임 탐색을 이용한 MPEG 비디오 분할)

  • 김주민;최영우;정규식
    • Proceedings of the IEEK Conference
    • /
    • 2000.09a
    • /
    • pp.215-218
    • /
    • 2000
  • 디지털 비디오 데이터를 효율적으로 브라우징 하는데 필요한 비디오 분할에 관한 연구가 활발하게 진행되고 있다. 본 연구에서는 비디오 데이터를 Shot단위로 분할하고, Shot내부에서 카메라 동작과 객체 움직임 분석을 이용한 sub-shot으로 분할하고자 한다. 연구 방법으로는 I-frame의 DC 영상을 이용하여 픽쳐그룹을 Shot(장면이 바뀐 경우), Move(카메라 동작,객체움직임), Static(영상의 변화가 거의 없는 경우)로 세분화하고 해당 픽쳐 그룹의 P, B-frame을 검사하여 정확한 컷 발생 위치, 디졸브, 카메라동작, 객체 움직임을 검출하게 된다. 픽쳐그룹 분류에서 정확성을 높이기 위해 계층적 신경망과 다중 특징을 이용한다. 정확한 컷 발생위치 검출하기 위해서 P, B프레임의 메크로블럭 타입을 이용한 통계적 방법을 이용하고, 디졸브, 카메라 동작, 객체 움직임을 검출하기 위해서 P, B-frame의 메크로블럭 타입과 움직임 벡터를 이용한 신경망으로 검출한다. 본 연구에서는 계층적 탐색을 이용하여 시간을 단축할 수 있고, 계층적 신경망과 다중 특징을 이용하여 픽쳐 그룹을 세분화 할 수 있고, 메크로 블록 타입과 통계적 방법을 이용하여 정확한 컷 검출을 할수 있고, 신경망을 이용하여 디졸브, 카메라 동작, 객체움직임을 검출 할 수 있음을 확인한다.

  • PDF

Video Shot Boundary Detection Using Correlation of Luminance and Edge Information (명도와 에지정보의 상관계수를 이용한 비디오샷 경계검출)

  • Yu, Heon-U;Jeong, Dong-Sik;Na, Yun-Gyun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.7 no.4
    • /
    • pp.304-308
    • /
    • 2001
  • The increase of video data makes the demand of efficient retrieval, storing, and browsing technologies necessary. In this paper, a video segmentation method (scene change detection method, or shot boundary detection method) for the development of such systems is proposed. For abrupt cut detection, inter-frame similarities are computed using luminance and edge histograms and a cut is declared when the similarities are under th predetermined threshold values. A gradual scene change detection is based on the similarities between the current frame and the previous shot boundary frame. A correlation method is used to obtain universal threshold values, which are applied to various video data. Experimental results show that propose method provides 90% precision and 98% recall rates for abrupt cut, and 59% precision and 79% recall rates for gradual change.

  • PDF

Generation of Video Clips Utilizing Shot Boundary Detection (샷 경계 검출을 이용한 영상 클립 생성)

  • Kim, Hyeok-Man;Cho, Seong-Kil
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.7 no.6
    • /
    • pp.582-592
    • /
    • 2001
  • Video indexing plays an important role in the applications such as digital video libraries or web VOD which archive large volume of digital videos. Video indexing is usually based on video segmentation. In this paper, we propose a software tool called V2Web Studio which can generate video clips utilizing shot boundary detection algorithm. With the V2Web Studio, the process of clip generation consists of the following four steps: 1) Automatic detection of shot boundaries by parsing the video, 2) Elimination of errors by manually verifying the results of the detection, 3) Building a modeling structure of logical hierarchy using the verified shots, and 4) Generating multiple video clips corresponding to each logically modeled segment. The aforementioned steps are performed by shot detector, shot verifier, video modeler and clip generator in the V2Web Studio respectively.

  • PDF

Shot Motion Classification Using Partial Decoding of INTRA Picture in Compressed Video (압축비디오에서 인트라픽쳐 부분 복호화를 이용한 샷 움직임 분류)

  • Kim, Kang-Wook;Kwon, Seong-Geun
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.7
    • /
    • pp.858-865
    • /
    • 2011
  • In order to allow the user to efficiently browse, select, and retrieve a desired video part without having to deal directly with GBytes of compressed data, classification of shot motion characteristic has to be carried out as a preparation for such user interaction. The organization of video information for video database requires segmentation of a video into its constituent shots and their subsequent characterization in terms of content and camera movement in shot. In order to classify shot motion, it is a conventional way to use element of motion vector. However, there is a limit to estimate global camera motion because the way that uses motion vectors only represents local movement. For shot classification in terms of motion information, we propose a new scheme consisting of partial decoding of INTRA pictures and comparing the x, y displacement vector curve between the decoded I-frame and next P-frame in compressed video data.

A Study on Shot Segmentation and Indexing of Language Education Videos by Content-based Visual Feature Analysis (교육용 어학 영상의 내용 기반 특징 분석에 의한 샷 구분 및 색인에 대한 연구)

  • Han, Heejun
    • Journal of the Korean Society for information Management
    • /
    • v.34 no.1
    • /
    • pp.219-239
    • /
    • 2017
  • As IT technology develops rapidly and the personal dissemination of smart devices increases, video material is especially used as a medium of information transmission among audiovisual materials. Video as an information service content has become an indispensable element, and it has been used in various ways such as unidirectional delivery through TV, interactive service through the Internet, and audiovisual library borrowing. Especially, in the Internet environment, the information provider tries to reduce the effort and cost for the processing of the provided information in view of the video service through the smart device. In addition, users want to utilize only the desired parts because of the burden on excessive network usage, time and space constraints. Therefore, it is necessary to enhance the usability of the video by automatically classifying, summarizing, and indexing similar parts of the contents. In this paper, we propose a method of automatically segmenting the shots that make up videos by analyzing the contents and characteristics of language education videos and indexing the detailed contents information of the linguistic videos by combining visual features. The accuracy of the semantic based shot segmentation is high, and it can be effectively applied to the summary service of language education videos.

Modeling of Semantic Similarity for Scene Segmentation (장면 분할 기법을 위한 의미적 유사도의 모델링)

  • Jung, Eui-Son;Jeon, Seong-Jun;Cho, Dong-Hwi;Geum, Yong-Ho;Ham, Dong-gyun;Kim, Eun-Ji;Park, Seung-Bo
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.07a
    • /
    • pp.225-228
    • /
    • 2022
  • 본 논문에서는 의미적 유사도 기반의 장면 분할 방법을 제안한다. 이 방법은 의미적 접근을 통해 기존 연구에서 가졌던 한계를 극복하고 정확한 장면 분할이 가능할 것으로 기대한다. 의미적 유사도 비교를 Class 종류 비교, Class별 객체의 개수 비교, 샷 간의 Histogram비교, 객체의 관심영역(ROI) Histogram비교 총 4가지 규칙으로 정의했고 이때 도출된 4가지 유사도는 전처리를 거쳐 종합 유사도를 계산한다. 또한 의미적 접근을 통해 연속되는 Shot의 유사도를 비교하고 기준값에 따라 Shot을 묶어서 최종적으로 의미적 유사도(Semantic Similarity)에 기반한 장면의 경계(Scene Boundary) 분할 방법을 제시한다.

  • PDF

Segmentation-Based Depth Map Adjustment for Improved Grasping Pose Detection (물체 파지점 검출 향상을 위한 분할 기반 깊이 지도 조정)

  • Hyunsoo Shin;Muhammad Raheel Afzal;Sungon Lee
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.1
    • /
    • pp.16-22
    • /
    • 2024
  • Robotic grasping in unstructured environments poses a significant challenge, demanding precise estimation of gripping positions for diverse and unknown objects. Generative Grasping Convolution Neural Network (GG-CNN) can estimate the position and direction that can be gripped by a robot gripper for an unknown object based on a three-dimensional depth map. Since GG-CNN uses only a depth map as an input, the precision of the depth map is the most critical factor affecting the result. To address the challenge of depth map precision, we integrate the Segment Anything Model renowned for its robust zero-shot performance across various segmentation tasks. We adjust the components corresponding to the segmented areas in the depth map aligned through external calibration. The proposed method was validated on the Cornell dataset and SurgicalKit dataset. Quantitative analysis compared to existing methods showed a 49.8% improvement with the dataset including surgical instruments. The results highlight the practical importance of our approach, especially in scenarios involving thin and metallic objects.

Detection of Video Scene Boundaries based on the Local and Global Context Information (지역 컨텍스트 및 전역 컨텍스트 정보를 이용한 비디오 장면 경계 검출)

  • 강행봉
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.6
    • /
    • pp.778-786
    • /
    • 2002
  • Scene boundary detection is important in the understanding of semantic structure from video data. However, it is more difficult than shot change detection because scene boundary detection needs to understand semantics in video data well. In this paper, we propose a new approach to scene segmentation using contextual information in video data. The contextual information is divided into two categories: local and global contextual information. The local contextual information refers to the foreground regions' information, background and shot activity. The global contextual information refers to the video shot's environment or its relationship with other video shots. Coherence, interaction and the tempo of video shots are computed as global contextual information. Using the proposed contextual information, we detect scene boundaries. Our proposed approach consists of three consecutive steps: linking, verification, and adjusting. We experimented the proposed approach using TV dramas and movies. The detection accuracy of correct scene boundaries is over than 80%.