• Title/Summary/Keyword: MPEG-4 Visual

Search Result 81, Processing Time 0.026 seconds

Scene Change Detection and Key Frame Selection Using Fast Feature Extraction in the MPEG-Compressed Domain (MPEG 압축 영상에서의 고속 특징 요소 추출을 이용한 장면 전환 검출과 키 프레임 선택)

  • 송병철;김명준;나종범
    • Journal of Broadcast Engineering
    • /
    • v.4 no.2
    • /
    • pp.155-163
    • /
    • 1999
  • In this paper, we propose novel scene change detection and key frame selection techniques, which use two feature images, i.e., DC and edge images, extracted directly from MPEG compressed video. For fast edge image extraction. we suggest to utilize 5 lower AC coefficients of each DCT. Based on this scheme, we present another edge image extraction technique using AC prediction. Although the former is superior to the latter in terms of visual quality, both methods all can extract important edge features well. Simulation results indicate that scene changes such as cut. fades, and dissolves can be correctly detected by using the edge energy diagram obtained from edge images and histograms from DC images. In addition. we find that our edge images are comparable to those obtained in the spatial domain while keeping much lower computational cost. And based on HVS, a key frame of each scene can also be selected. In comparison with an existing method using optical flow. our scheme can select semantic key frames because we only use the above edge and DC images.

  • PDF

Semantic Event Detection in Golf Video Using Hidden Markov Model (은닉 마코프 모델을 이용한 골프 비디오의 시멘틱 이벤트 검출)

  • Kim Cheon Seog;Choo Jin Ho;Bae Tae Meon;Jin Sung Ho;Ro Yong Man
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.11
    • /
    • pp.1540-1549
    • /
    • 2004
  • In this paper, we propose an algorithm to detect semantic events in golf video using Hidden Markov Model. The purpose of this paper is to identify and classify the golf events to facilitate highlight-based video indexing and summarization. In this paper we first define 4 semantic events, and then design HMM model with states made up of each event. We also use 10 multiple visual features based on MPEG-7 visual descriptors to acquire parameters of HMM for each event. Experimental results showed that the proposed algorithm provided reasonable detection performance for identifying a variety of golf events.

  • PDF

A Generation Method of Spatially Encoded Video Data for Geographic Information Systems

  • Joo, In-Hak;Hwang, Tae-Hyun;Choi, Kyoung-Ho;Jang, Byung-Tae
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.801-803
    • /
    • 2003
  • In this paper, we present a method for generating and providing spatially encoded video data that can be effectively used by GIS applications. We collect the video data by a mobile mapping system called 4S-Van that is equipped by GPS, INS, CCD camera, and DVR system. The information about spatial object appearing in video, such as occupied region in each frame, attribute value, and geo-coordinate, are generated and encoded. We suggest methods that can generate such data for each frame in semi-automatic manner. We adopt standard MPEG-7 metadata format for representation of the spatially encoded video data to be generally used by GIS application. The spatial and attribute information encoded to each video frame can make visual browsing between map and video possible. The generated video data can be provided and applied to various GIS applications where location and visual data are both important.

  • PDF

Distribution of Target Bits based on Size, Motion and Distortion (크기, 움직임 및 왜곡 정보에 의한 목표비트 분배)

  • 지석상;황재정;이문호
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.4 no.5
    • /
    • pp.1085-1093
    • /
    • 2000
  • An efficient bit rate distribution technique that distributes available bits for multiple objects based on motion vector magnitude, size of object shape, and coding distortion is presented. This coding concept using the three parameters was exploited in MPEG-4 multiple object coding. But the scheme is likely to produce poor results such as allocating more bits to less important objects and degrading picture quality, due to the lack of analysis and research in view of human visual aspect. In this paper importance of each object is represented by the three parameters and visually analyzed. Target bits are distributed according to coding distortion using the pre-assigned shape and motion information.

  • PDF

A H.264 based Selective Fine Granular Scalable Coding Scheme (H.264 기반 선택적인 미세입자 스케일러블 코딩 방법)

  • 박광훈;유원혁;김규헌
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.10 no.4
    • /
    • pp.309-318
    • /
    • 2004
  • This paper proposes the H.264-based selective fine granular scalable (FGS) coding scheme that selectively uses the temporal prediction data in the enhancement layer. The base layer of the proposed scheme is basically coded by the H.264 (MPEG-4 Part 10 AVC) visual coding scheme that is the state-of-art in codig efficiency. The enhancement layer is basically coded by the same bitplane-based algorithm of the MPEG-4 (Part 2) fine granular scalable coding scheme. In this paper, we introduce a new algorithm that uses the temproal prediction mechanism inside the enhancement layer and the effective selection mechanism to decide whether the temporally-predicted data would be sent to the decoder or not. Whenever applying the temporal prediction inside the enhancement layer, the temporal redundancies may be effectively reduced, however the drift problem would be severly occurred especially at the low bitrate transmission, due to the mismatch bewteen the encoder's and decoder's reference frame images. Proposed algorithm selectively uses the temporal-prediction data inside the enhancement layer only in case those data could siginificantly reduce the temporal redundancies, to minimize the drift error and thus to improve the overall coding efficiency. Simulation results, based on several test image sequences, show that the proposed scheme has 1∼3 dB higher coding efficiency than the H.264-based FGS coding scheme, even 3∼5 dB higher coding efficiency than the MPEG-4 FGS international standard.

A Study on the constant bit-rate control for H.264/AVC encoders (H.264/AVC 부호기에서의 고정 비트율 제어 방안 연구)

  • Yun Hyeogsang;Kim Yong Han
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2003.11a
    • /
    • pp.17-20
    • /
    • 2003
  • 동등한 영상의 풀질을 유지하면서 기존의 비디오 압축 표준보다 30퍼센트 이상의 비트 절감효과를 내는 H.264/MPEG-4 Part 10 AVC(Advanced Video Coding) 표준이 최근 확정되었다. AVC는 국내 지상파 DMB(Digital Multimedia Broadcasting) 멀티미디어 서비스의 비디오 표준으로도 채택되어 많은 관심을 불러일으키고 있다. 본 논문에서는 H.264/AVC 부호기를 위한 고정 비트 율 제어 알고리듬을 구현하고 이를 실험을 통하여 검증하였다. 비디오 부호화 과정에서 각 픽처마다 발생하는 비트 량이 같지는 않지만 부호기 출력 비트 율은 일정하게 한 수 있으며 이를 고정 비트 율 제어라 한다 부호기는 입력 영상의 복잡도를 예측하고 실제 버퍼 점유도에 대한 피프백 정보를 이용하여 발생 비트 수글 조절한다. 특히 전송 대역폭 제한사항이 있는 고정 비트 율 채널 환경에서 비디오 부호기의 고정 비트 율 제어는 반드시 필요하다. JVT에서 제공하는 참조 소프트웨어인 JM(Joint Model) 소프트웨어가 아직까지 고정 비트 율 제어 기능을 지원하지 않으므로, MPEG-2 Part 2 Video 표준과 MPEG-4 Part 2 Visual 표준에서 고정 비트 율 제어에 이용되었던 TM5 고정 비트 율 제어 모델을 적용하여 H.264/AVC 부호기의 고정 비트 율 제어를 구현하였다.

  • PDF

Video Content Manipulation Using 3D Analysis for MPEG-4

  • Sull, Sanghoon
    • Journal of Broadcast Engineering
    • /
    • v.2 no.2
    • /
    • pp.125-135
    • /
    • 1997
  • This paper is concerned with realistic mainpulation of content in video sequences. Manipulation of content in video sequences is one of the content-based functionalities for MPEG-4 Visual standard. We present an approach to synthesizing video sequences by using the intermediate outputs of three-dimensional (3D) motion and depth analysis. For concreteness, we focus on video showing 3D motion of an observer relative to a scene containing planar runways (or roads). We first present a simple runway (or road) model. Then, we describe a method of identifying the runway (or road) boundary in the image using the Point of Heading Direction (PHD) which is defined as the image of, the ray along which a camera moves. The 3D motion of the camera is obtained from one of the existing 3D analysis methods. Then, a video sequence containing a runway is manipulated by (i) coloring the scene part above a vanishing line, say blue, to show sky, (ii) filling in the occluded scene parts, and (iii) overlaying the identified runway edges and placing yellow disks in them, simulating lights. Experimental results for a real video sequence are presented.

  • PDF

인터렉티브 하이브리드 미디어 응용기술 -MPEG-4 SNHC를 중심으로-

  • 김형곤
    • Broadcasting and Media Magazine
    • /
    • v.3 no.2
    • /
    • pp.44-58
    • /
    • 1998
  • 최근의 멀티미디어 기술은 정보의 디지털화와 온라인화에 따라 가전, 컴퓨터, 통신 및 방송 기술이 융화되어 가는 추세에 있으며, 대화형의 하이브리드 멀티미디어 기술을 그 특징으로 하고있다. 하이브리드 멀티미디어는 컴퓨터 그래픽 및 미디(MIDI) 기술로 인위적으로 생성한 2D/3D그래픽 및 음향을 실제의 자연적인 영상과 소리에 추가하여 합성하므로 생성된다. MPEG-4는 이렇게 인위적으로 합성되거나 자연적인 영상 혹은 음향 정보의 디지털 하이브리드 멀티미디어 부호화를 목적으로 하며, 활성화된 혼합 미디어의 내용기반 처리, 상호, 동작 및 사용자의 쉬운 접근 등을 가능하게 한다. SNHC(Synthetic-Natural Hybrid Coding)는 기존의 수동적인 미디어의 전달뿐 아니라 실시간 처리가 가능한 인터랙티브 응용 분야까지 다루고 있으며, 통합된 시공간 부호화 기법을 사용하여 시각, 청각, 2차원, 3차원 컴퓨터 그래픽스 등 다양한 형태의 표준 AV(Aural/Visual) 객체를 처리한다. 표준화는 주로mesh-segmented 비디오 부호화, 구조물 부호화, 객체간의 동기화, AV 객체 스트림의 멀티플렉싱, 혼합 미디어 형태의 시-공간 통합화 등에서 이루어지게 되는데, 이는 궁극적으로 네트워크로 연결되는 가상 환경(Virtual Environment)에서 다수의 사용자가 서로 상호작용 할 수 있는 틀을 제공하는데 있다. 이러한 틀이 제공되면, 대화형 하이브리드 멀티미디어라는 새로운 형태의 정보를 사용함으로써 기존의 미디어로는 경험하지 못하는 다양한 응용과 서비스를 경험할 수 있을 것이다.

  • PDF

A 3D Audio Broadcasting Terminal for Interactive Broadcasting Services (대화형 방송을 위한 3차원 오디오 방송단말)

  • Park Gi Yoon;Lee Taejin;Kang Kyeongok;Hong Jinwoo
    • Journal of Broadcast Engineering
    • /
    • v.10 no.1 s.26
    • /
    • pp.22-30
    • /
    • 2005
  • We implement an interactive 3D audio broadcasting terminal which synthesizes an audio scene according to the request of a user. Audio scene structure is described by the MPEG-4 AudioBIFS specifications. The user updates scene attributes and the terminal synthesizes the corresponding sound images in the 3D space. The terminal supports the MPEG-4 Audio top nodes and some visual nodes. Instead of using sensor nodes and route elements, we predefine node type-specific user interfaces to support BIFS commands for field replacement. We employ sound spatialization, directivity/shape modeling, and reverberation effects for 3D audio rendering and realistic feedback to user inputs. We also introduce a virtual concert program as an application scenario of the interactive broadcasting terminal.

Rectangular Region-based Selective Enhancement (RSE) for MPEG-4 FGS Video (MPEG-4 FGS 비디오를 위한 사각영역 기반의 선택적 향상기법)

  • 서광덕;신창호;김재균
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.6C
    • /
    • pp.634-647
    • /
    • 2003
  • In MPEG-4 FGS (fine granular scalability) video, SE (selective enhancement) function is adopted to enhance the subject quality of the region of interest (ROI). However, it has the problem of excessive bit-rate increase in the enhancement layer. We present a new rectangular region-based SE (RSE) algorithm to significantly reduce the overhead bits resulting from the standard SE. The proposed RSE is based on two new algorithms. The first is to apply the SE function to a rectangular region. By doing so, we can reduce the required bits for describing the selectively enhanced region. The second is to use constrained bit-plane scanning (CBS) to encode bit-planes of the enhancement layer. By using CBS, we can efficiently encode the ALL-ZERO symbols that are generated by applying the SE. It Is shown by simulation that the proposed RSE can provide a good visual quality for the selected rectangular region with significantly reduced overhead bits.