• Title/Summary/Keyword: Video Frames

Search Result 888, Processing Time 0.024 seconds

Shot Change Detection Using Multiple Features and Binary Decision Tree (다수의 특징과 이진 분류 트리를 이용한 장면 전환 검출)

  • 홍승범;백중환
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.5C
    • /
    • pp.514-522
    • /
    • 2003
  • Contrary to the previous methods, in this paper, we propose an enhanced shot change detection method using multiple features and binary decision tree. The previous methods usually used single feature and fixed threshold between consecutive frames. However, contents such as color, shape, background, and texture change simultaneously at shot change points in a video sequence. Therefore, in this paper, we detect the shot changes effectively using multiple features, which are supplementary each other, rather than using single feature. In order to classify the shot changes, we use binary classification tree. According to this classification result, we extract important features among the multiple features and obtain threshold value for each feature. We also perform the cross-validation and droop-case to verify the performance of our method. From an experimental result, it was revealed that the EI of our method performed average of 2% better than that of the conventional shot change detection methods.

A Moving Camera Localization using Perspective Transform and Klt Tracking in Sequence Images (순차영상에서 투영변환과 KLT추적을 이용한 이동 카메라의 위치 및 방향 산출)

  • Jang, Hyo-Jong;Cha, Jeong-Hee;Kim, Gye-Young
    • The KIPS Transactions:PartB
    • /
    • v.14B no.3 s.113
    • /
    • pp.163-170
    • /
    • 2007
  • In autonomous navigation of a mobile vehicle or a mobile robot, localization calculated from recognizing its environment is most important factor. Generally, we can determine position and pose of a camera equipped mobile vehicle or mobile robot using INS and GPS but, in this case, we must use enough known ground landmark for accurate localization. hi contrast with homography method to calculate position and pose of a camera by only using the relation of two dimensional feature point between two frames, in this paper, we propose a method to calculate the position and the pose of a camera using relation between the location to predict through perspective transform of 3D feature points obtained by overlaying 3D model with previous frame using GPS and INS input and the location of corresponding feature point calculated using KLT tracking method in current frame. For the purpose of the performance evaluation, we use wireless-controlled vehicle mounted CCD camera, GPS and INS, and performed the test to calculate the location and the rotation angle of the camera with the video sequence stream obtained at 15Hz frame rate.

A shot change detection algorithm based on frame segmentation and object movement (프레임 블록화와 객체의 이동을 이용한 샷 전환 탐지 알고리즘)

  • Kim, Seung-Hyun;Hwang, Doosung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.20 no.5
    • /
    • pp.21-29
    • /
    • 2015
  • This paper proposes a shot change detection algorithm by using frame segmentation and the object changes among moving blocks. In order to detect the rapid moving changes of objects between two consecutive frames, the moving blocks on the diagonal are defined, and their histograms are calculated. When a block of the current frame is compared to the moving blocks of the next frame, the block histograms are used and the threshold of a shot change detection is automatically adjusted by Otsu's threshold method. The proposed algorithm was tested for the various types of color or gray videos such as films, dramas, animations, and video tapes in National Archives of Korea. The experimental results showed that the proposed algorithm could enhance the detection rate when compared to the studied methods that use brightness, histogram, or segmentation.

The Extraction of Camera Parameters using Projective Invariance for Virtual Studio (가상 스튜디오를 위한 카메라 파라메터의 추출)

  • Han, Seo-Won;Eom, Gyeong-Bae;Lee, Jun-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.9
    • /
    • pp.2540-2547
    • /
    • 1999
  • Chromakey method is one of key technologies for realizing virtual studio, and the blue portions of a captured image in virtual studio, are replaced with a computer generated or real image. The replaced image must be changed according to the camera parameter of studio for natural merging with the non-blue portions of a captured image. This paper proposes a novel method to extract camera parameters using the recognition of pentagonal patterns that are painted on a blue screen. We extract corresponding points between a blue screen. We extract corresponding points between a blue screen and a captured image using the projective invariant features of a pentagon. Then, calculate camera parameters using corresponding points by the modification of Tsai's method. Experimental results indicate that the proposed method is more accurate compared to conventional method and can process about twelve frames of video per a second in Pentium-MMX processor with CPU clock of 166MHz.

  • PDF

A scheme to minimize transmission delay during handoff for rt-VBR service in the wireless ATM Networks (무선 ATM에서 핸드오프 동안 실시간 VBR 서비스를 위한 전송 지연의 최소화 방안)

  • Kim, Jun-Bae;Jang, Dong-Hyeok;Lee, Seon-Suk;Lee, Jae-Hong;Kim, Seung-Hwan;Gwon, O-Seok;Park, Jin-Su
    • The KIPS Transactions:PartC
    • /
    • v.9C no.2
    • /
    • pp.277-282
    • /
    • 2002
  • In general, requirements for QoS are different according to the type of services in wire-line and wireless ATM networks, and real-time video service is more sensitive to cell transmission delay than to cell loss. Existing handoff schemes emphasizing prevention of cell loss had limitations in cell transmission delay to satisfy QoS. In this paper, a novel scheme to transmit ATM cells with low CLP(when CLP = 0) prior to others and discarding cells with high CLP(when CLF = 1) in ATM cell header among cells to be forwarded to new base station during handoffs in real-time VBR service is proposed. The proposed scheme is proven to be suitable for the satisfaction of QoS of real-time VBR service and appropriate for fast handoffs by giving high CLP value to less meaningful MPEG frames through simulations.

The Kinematic Analysis of the Last Approach Stride and Take-off Phase of BKH Athlete in the High Jump (남자 높이뛰기 BKH 선수를 중심으로 한 도움닫기 마지막 1보와 발구름 국면의 운동학적 분석)

  • Yoon, Hee-Joong;Kim, Tae-Sam;Lee, Jin-Taek
    • Korean Journal of Applied Biomechanics
    • /
    • v.15 no.3
    • /
    • pp.105-115
    • /
    • 2005
  • This study was investigated the kinematic factors of the last approach strides and. take off motion for the skill improving of BKH elite male athlete. 'The subjects chosen for the study were BKH and. KASZCZYK Emillian male athletes who were participated in 2003 Dae-Gu Universiad Games. Three high speed video cameras set in 60frames/s setting were used. for recording from the last approach strides to the apex position. After digitizing motion, the Direct Linear Transformation(DLT) technique was employed to obtain 3-D position coordinates, The kinematic factors of the distance, velocity and angle variable were calculated for Kwon3D 3.1. The following conclusions were drawn; 1. It showed longer stride length, as well as faster horizontal and lateral velocity than the success trial during the approach phase. For consistent of the approach rhythm, it appeared that the subject should a short length for obtain the breaking force by the lower COG during the approach phase. 2. The body lean angle showed a small angle by a high COG during the take-off phase. For obtain the vertical displacement of the COG and a enough space form the bar after take-off, it appeared that the subject should increase the body lean angle. 3. For obtain the vertical force during the takeoff phase, it appeared that the subject should keep straight as possible the knee joint. Therefor, the subject can be obtain a enough breaking force at the approach landing.

The Modified Block Matching Algorithm for a Hand Tracking of an HCI system (HCI 시스템의 손 추적을 위한 수정 블록 정합 알고리즘)

  • Kim Jin-Ok
    • Journal of Internet Computing and Services
    • /
    • v.4 no.4
    • /
    • pp.9-14
    • /
    • 2003
  • A GUI (graphical user interface) has been a dominant platform for HCI (human computer interaction). A GUI - based interaction has made computers simpler and easier to use. The GUI - based interaction, however, does not easily support the range of interaction necessary to meet users' needs that are natural. intuitive, and adaptive. In this paper, the modified BMA (block matching algorithm) is proposed to track a hand in a sequence of an image and to recognize it in each video frame in order to replace a mouse with a pointing device for a virtual reality. The HCI system with 30 frames per second is realized in this paper. The modified BMA is proposed to estimate a position of the hand and segmentation with an orientation of motion and a color distribution of the hand region for real - time processing. The experimental result shows that the modified BMA with the YCbCr (luminance Y, component blue, component red) color coordinate guarantees the real - time processing and the recognition rate. The hand tracking by the modified BMA can be applied to a virtual reclity or a game or an HCI system for the disable.

  • PDF

Fast Motion Estimation Algorithm Using Importance of Search Range and Adaptive Matching Criterion (탐색영역의 중요도와 적응적인 매칭기준을 이용한 고속 움직임 예측 알고리즘)

  • Choi, Hong-Seok;Kim, Jong-Nam;Jeong, Shin-Il
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.4
    • /
    • pp.129-133
    • /
    • 2015
  • In this paper, we propose a fast motion estimation algorithm which is important in the performance of video encoding. Conventional fast motion estimation algorithms have serious problems of low prediction quality in some frames and still much computation. In the paper, we propose an algorithm that reduces unnecessary computations only, while keeping prediction quality almost similar to that of the full search. The proposed algorithm uses distribution of probability of motion vectors, divides search range into several groups according to its importance, and applies adaptive block matching criteria for each group of search range. The proposed algorithm takes only 3~5% in computational amount and has decreased prediction quality about 0~0.01dB compared with the fast full search algorithm.

Compression Method for MPEG CDVA Global Feature Descriptors (MPEG CDVA 전역 특징 서술자 압축 방법)

  • Kim, Joonsoo;Jo, Won;Lim, Guentaek;Yun, Joungil;Kwak, Sangwoon;Jung, Soon-heung;Cheong, Won-Sik;Choo, Hyon-Gon;Seo, Jeongil;Choi, Yukyung
    • Journal of Broadcast Engineering
    • /
    • v.27 no.3
    • /
    • pp.295-307
    • /
    • 2022
  • In this paper, we propose a novel compression method for scalable Fisher vectors (SCFV) which is used as a global visual feature description of individual video frames in MPEG CDVA standard. CDVA standard has adopted a temporal descriptor redundancy removal technique that takes advantage of the correlation between global feature descriptors for adjacent keyframes. However, due to the variable length property of SCFV, the temporal redundancy removal scheme often results in inferior compression efficiency. It is even worse than the case when the SCFVs are not compressed at all. To enhance the compression efficiency, we propose an asymmetric SCFV difference computation method and a SCFV reconstruction method. Experiments on the FIVR dataset show that the proposed method significantly improves the compression efficiency compared to the original CDVA Experimental Model implementation.

Lip and Voice Synchronization Using Visual Attention (시각적 어텐션을 활용한 입술과 목소리의 동기화 연구)

  • Dongryun Yoon;Hyeonjoong Cho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.4
    • /
    • pp.166-173
    • /
    • 2024
  • This study explores lip-sync detection, focusing on the synchronization between lip movements and voices in videos. Typically, lip-sync detection techniques involve cropping the facial area of a given video, utilizing the lower half of the cropped box as input for the visual encoder to extract visual features. To enhance the emphasis on the articulatory region of lips for more accurate lip-sync detection, we propose utilizing a pre-trained visual attention-based encoder. The Visual Transformer Pooling (VTP) module is employed as the visual encoder, originally designed for the lip-reading task, predicting the script based solely on visual information without audio. Our experimental results demonstrate that, despite having fewer learning parameters, our proposed method outperforms the latest model, VocaList, on the LRS2 dataset, achieving a lip-sync detection accuracy of 94.5% based on five context frames. Moreover, our approach exhibits an approximately 8% superiority over VocaList in lip-sync detection accuracy, even on an untrained dataset, Acappella.