• Title/Summary/Keyword: Scene Number Detection

Search Result 35, Processing Time 0.026 seconds

Effective scene change detection methods using characteristics of MPEG video (MPEG 비디오의 특성 추출을 이용한 효과적인 장면 전환 검출 기법)

  • 곽영경;최윤석;고성제
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.8B
    • /
    • pp.1567-1576
    • /
    • 1999
  • In this paper, we propose new methods to detect a scene cut and a dissolve region in compressed MPEG video sequences. The scene cut detection method uses edge images obtained using DCT AC coefficients and the dissolve detection method utilizes the macroblock type information of the MPEG stream. The proposed scene cut detection method is insensitive to brightness and can detect scene changes more precisely than the methods using DC coefficients since AC edge images can express original images more exactly than DC edge images do. The proposed dissolve detection method using the number of intra macroblocks(MBs) computationally efficient since it does not require the decoding process. Experimental results show that the proposed methods perform better in detection scene changes than conventional other methods.

  • PDF

Detection of Fast Scene Changes Using a Statistical Technique (영상의 통계적 특성을 이용한 급격한 장면전화 검출 알고리즘)

  • 곽대호;박성준;이건호;최유태;송문호
    • Proceedings of the IEEK Conference
    • /
    • 2000.11d
    • /
    • pp.151-154
    • /
    • 2000
  • We propose a statically motivated scene change detection algorithm. As the difference between the neighboring frames will generate peaks at scene boundaries, the problem of detecting fast scene changes is equivalent to detecting peaks in a given sequence. In this paper, the peak detection is performed via several statistics, namely the sample means and variances. For eliminating flash lights as well as detecting fast scene changes within a small number of frames, we have opted to use a two-stage process for computing the necessary statistics. The results indicate superiority of necessary statistics. The results indicate superiority of the proposed algorithm over the previously reported algorithm.

  • PDF

MPEG-1 Video Scene Change Detection Using Horizontal and Vertical Blocks (수평과 수직 블록을 이용한 MPEG-1 비디오 장면전환 검출)

  • Lee, Min-Seop;An, Byeong-Cheol
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.2S
    • /
    • pp.629-637
    • /
    • 2000
  • The content-based information retrieval for a multimedia database uses feature information extracted from the compressed videos. This paper presents an effective method to detect scene changes from compressed videos. Scene changes are detected with DC values of DCT coefficients in MPEG-1 encoded video sequences. Instead of decoding full frames. partial macroblocks of each frame, horizontal and vertical macroblocks, are decoded to detect scene changes. This method detects abrupt scene changes by decoding minimal number of blocks and saves a lot of computation time. The performance of the proposed algorithm is analyzed based on the precision and the recall. The experimental results show the effectiveness in computation time and detection rate to detect scene changes of various MPEG-1 video streams.

  • PDF

Extraction of Smocking in Elevator Using Robust Scene Change Detection Method (강건한 장면 전환 검출 기법을 이용한 엘리베이터 내의 흡연 추출)

  • Lee, Kang-Ho;Shin, Seong-Yoon;Rhee, Yang-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.10
    • /
    • pp.89-95
    • /
    • 2013
  • Smoking in elevators is a criminal offense that is included in a misdemeanor. Because of that smoking in elevators can be very critical for our growing children and weak women. In this paper, we would like to extract criminals doing this criminal offense to smoke in elevators. Extraction method detect difference value using modified color-X2-test and it was normalized. Next, we find frames that has occurred scene change in successive frames using the four-step algorithm of scene change detection. Finally, we present the method of smoking image retrieval and extraction in stored large amount of video. In the experiment, we show process and number of scene change detection, and the number of video searched per retrieval time. The extracted smoking video is to submit as evidence for the police or court.

An adaptive motion estimation based on the temporal subband analysis (시간축 서브밴드 해석을 이용한 적응적 움직임 추정에 관한 연구)

  • 임중곤;정재호
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.21 no.6
    • /
    • pp.1361-1369
    • /
    • 1996
  • Motion estimation is one of the key components for high quality video coding. In this paper, a new motion estimation scheme for MPEG-like video coder is suggested. The proposed temporally adaptive motion estimation scheme consists of five functional blocks: Temporal subband analysis (TSBA), extraction of temporal information, scene change detection (SCD), picture type replacement (PTR), and temporally adapted block matching algorithm (TABMA). Here all the functional components are based on the temporal subband analysis. In this papre, we applied the analysis part of subband decompostion to the temporal axis of moving picture sequence, newly defined the temporal activity distribution (TAD) and average TAD, and proposed the temporally adapted block matching algorithm, the scene change detection algorithm and picture type replacement algorithm which employed the results of the temporal subband analysis. A new block matching algorithm TABMA is capable of controlling the block matching area. According to the temporal activity distribution of objects, it allocates the search areas nonuniformly. The proposed SCD and PTR can prevent unavailable motion prediction for abrupt scene changes. Computer simulation results show that the proposed motion estimation scheme improve the quality of reconstructed sequence and reduces the number of block matching trials to 40% of the numbers of trials in conventional methods. The TSBA based scene change detection algorithm can detect the abruptly changed scenes in the intentionally combined sequence of this experiment without additional computations.

  • PDF

A Study on Frame of MSE Comparison for Scene Chang Detection Retrieval (장면 전환점 검출을 위한 프레임의 평균오차 비교에 관한 연구)

  • 김단환;김형균;오무송
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2002.05a
    • /
    • pp.638-642
    • /
    • 2002
  • User in video data utilization of high-capacity can grasp whole video data at a look. Offer frame list that summarize information of video data to do so that can remake video from branch that want when need. Need index process of video data for effective video retrieval. This treatise wishes to propose effective method about scene change point detection of video that is been based on contents base index. Proposed method video data so that can grasp whole structure of video detection color value of schedule pixel for diagonal line direction in image sampling do. Data that get into sampling could grasp scene change point on one eye. Color value of pixel that detection in each frame is i frame number by i$\times$j procession to procession A, j stores to reflex height of frame. Introduce MSE and calculate mean error of each frame. If exceed mean error and schedule critical value, wish to detect the frame for scene change point.

  • PDF

Shot Group and Representative Shot Frame Detection using Similarity-based Clustering

  • Lee, Gye-Sung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.9
    • /
    • pp.37-43
    • /
    • 2016
  • This paper introduces a method for video shot group detection needed for efficient management and summary of video. The proposed method detects shots based on low-level visual properties and performs temporal and spatial clustering based on visual similarity of neighboring shots. Shot groups created from temporal clustering are further clustered into small groups with respect to visual similarity. A set of representative shot frames are selected from each cluster of the smaller groups representing a scene. Shots excluded from temporal clustering are also clustered into groups from which representative shot frames are selected. A number of video clips are collected and applied to the method for accuracy of shot group detection. We achieved 91% of accuracy of the method for shot group detection. The number of representative shot frames is reduced to 1/3 of the total shot frames. The experiment also shows the inverse relationship between accuracy and compression rate.

Adaptive Face Mask Detection System based on Scene Complexity Analysis

  • Kang, Jaeyong;Gwak, Jeonghwan
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.5
    • /
    • pp.1-8
    • /
    • 2021
  • Coronavirus disease 2019 (COVID-19) has affected the world seriously. Every person is required for wearing a mask properly in a public area to prevent spreading the virus. However, many people are not wearing a mask properly. In this paper, we propose an efficient mask detection system. In our proposed system, we first detect the faces of input images using YOLOv5 and classify them as the one of three scene complexity classes (Simple, Moderate, and Complex) based on the number of detected faces. After that, the image is fed into the Faster-RCNN with the one of three ResNet (ResNet-18, 50, and 101) as backbone network depending on the scene complexity for detecting the face area and identifying whether the person is wearing the mask properly or not. We evaluated our proposed system using public mask detection datasets. The results show that our proposed system outperforms other models.

Scen based MPEG video traffic modeling considering the correlations between frames (프레임간 상관관계를 고려한 장면기반 MPEG 비디오 트래픽 모델링)

  • 유상조;김성대;최재각
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.23 no.9A
    • /
    • pp.2289-2304
    • /
    • 1998
  • For the performance analysis and traffic control of ATM networks carrying video sequences, need an appropriate video traffic model. In this paper, we propose a new traffic model for MPEG compressed videos which are widely used for any type of video applications at th emoment. The proposed modeling scheme uses scene-based traffic characteristics and considers the correlation between frames of consecutiv GOPs. Using a simple scene detection algorithm, scene changes are modeled by state transitions and the number of GOPs of a scene state is modeled by a geometric distirbution. Frames of a scene stte are modeled by mean I, P, and B frame size. For more accurate traffic modeling, quantization errors (residual bits) that the state transition model using mean values has are compensated by autoregressive processes. We show that our model very well captures the traffic chracteristics of the original videos by performance analysis in terms of autocorrelation, histogram of frame bits genrated by the model, and cell loss rate in the ATM multiplexer with limited buffers. Our model is able to perrorm translations between levels (i.e., GOP, frame, and cell levels) and to estimate very accurately the stochastic characteristics of the original videos by each level.

  • PDF

Comparisons of Object Recognition Performance with 3D Photon Counting & Gray Scale Images

  • Lee, Chung-Ghiu;Moon, In-Kyu
    • Journal of the Optical Society of Korea
    • /
    • v.14 no.4
    • /
    • pp.388-394
    • /
    • 2010
  • In this paper the object recognition performance of a photon counting integral imaging system is quantitatively compared with that of a conventional gray scale imaging system. For 3D imaging of objects with a small number of photons, the elemental image set of a 3D scene is obtained using the integral imaging set up. We assume that the elemental image detection follows a Poisson distribution. Computational geometrical ray back propagation algorithm and parametric maximum likelihood estimator are applied to the photon counting elemental image set in order to reconstruct the original 3D scene. To evaluate the photon counting object recognition performance, the normalized correlation peaks between the reconstructed 3D scenes are calculated for the varied and fixed total number of photons in the reconstructed sectional image changing the total number of image channels in the integral imaging system. It is quantitatively illustrated that the recognition performance of the photon counting integral imaging system can be similar to that of a conventional gray scale imaging system as the number of image viewing channels in the photon counting integral imaging (PCII) system is increased up to the threshold point. Also, we present experiments to find the threshold point on the total number of image channels in the PCII system which can guarantee a comparable recognition performance with a gray scale imaging system. To the best of our knowledge, this is the first report on comparisons of object recognition performance with 3D photon counting & gray scale images.