• Title/Summary/Keyword: video metadata

Search Result 115, Processing Time 0.026 seconds

CARA: Character Appearance Retrieval and Analysis for TV Programs

  • Jung Byunghee;Park Sungchoon;Kim Kyeongsoo
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2004.11a
    • /
    • pp.237-240
    • /
    • 2004
  • This paper describes a character retrieval system for TV programs and a set of novel algorithms for detecting and recognizing faces for the system. Our character retrieval system consists of two main components: Face Register and Face Recognizer. The Face Register detects faces in video frames and then guides users to register the detected faces of interest into the database. The Face Recognizer displays the appearance interval of each character on the timeline interface and the list of scenes with the names of characters that appear on each scene. These two components also provide a function to modify incorrect results. which is helpful to provide accurate character retrieval services. In the proposed face detection and recognition algorithms. we reduce the computation time without sacrificing the recognition accuracy by using the DCT/LDA method for face feature extraction. We also develop the character retrieval system in the form of plug-in. By plugging in our system to a cataloguing system. the metadata about the characters in a video can be automatically generated. Through this system, we can easily realize sophisticated on-demand video services which provide the search of scenes of a specific TV star.

  • PDF

A Design and Implementation of Multimedia Retrieval System based on MAF(Multimedia Application File Format) (MAF(Multimedia Application File Format) 기반 멀티미디어 검색 시스템의 설계 및 구현)

  • Gang Young-Mo;Park Joo-Hyoun;Bang Hyung-Gin;Nang Jong-Ho;Kim Hyung-Chul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.33 no.9
    • /
    • pp.574-584
    • /
    • 2006
  • Recently, ISO/IEC 23000 (also known as 'MPEG-A') has proposed a new file format called 'MAF(Multimedia Application File Format)[1]' which provides a capability of integrating/storing the widely-used compression standards for audio and video and the metadata in MPEG-7 form into a single file format. However, it is still very hard to verify the usefulness of MPEG-A in the real applications because there is still no real system that fully implements this standard. In this thesis, a design and implementation of a multimedia retrieval system based on MPEG-A standard on PC and mobile device is presented. Furthermore, an extension of MPEG-A for describing the metadata for video is also proposed. It is selected and defined as a subset of MPEG-7 MDS[4] and TV-anytime[5] for video that is useful and manageable in the mobile environments. In order to design the multimedia retrieval system based on MPEG-A, we define the system requirements in terms of portability, extensibility, compatibility, adaptability, efficiency. Based on these requirements, we design the system which composed of 3 layers: Application Layer, Middleware Layer, Platform Layer. The proposed system consists of two sub-parts, client-part and server-part. The client-part consists of MAF authoring tool, MAP player tool and MAF searching tool which allow users to create, play and search the MAF files, respectively. The server-part is composed of modules to store and manage the MAF files and metadata extracted from MAF files. We show the usefulness of the proposed system by implementing the client system both on MS-Windows platform on desk-top computer and WIPI platform on mobile phone, and validate whether it to satisfy all the system requirements. The proposed system can be used to verify the specification in the MPEG-A, and to proves the usefulness of MPEG-A in the real application.

An Experimental Study on the Effectiveness of Storyboard Surrogates in the Meanings Extraction of Digital Videos (비디오자료의 의미추출을 위한 영상초록의 효용성에 관한 실험적 연구)

  • Kim, Hyun-Hee
    • Journal of the Korean Society for information Management
    • /
    • v.24 no.4
    • /
    • pp.53-72
    • /
    • 2007
  • This study is designed to assess whether storyboard surrogates are useful enough to be utilized for indexing sources as well as for metadata elements using 12 sample videos and 14 participants. Study shows that first, the match rates of index terms and summaries are significantly different according to video types, which means storyboard surrogates are especially useful for the type of videos of conveying their meanings mainly through images. Second, participants could assign subject keywords and summaries to digital video, sacrificing a little loss of full video clips' match rates. Moreover, the match rate of index terms (0.45) is higher than that of summaries (0.40). This means storyboard surrogates could be more useful for indexing videos rather than summarizing them. The study suggests that 1)storyboard surrogates can be used as sources for indexing and abstracting digital videos; 2) using storyboard surrogates along with other metadata elements (e.g., text-based abstracts) can be more useful for users' relevance judgement; and 3)storyboard surrogates can be utilized as match sources of image-based queries. Finally, in order to improve storyboard surrogates quality, this study proposes future studies: constructing key frame extraction algorithms and designing key frame arrangement models.

Media Production Environment Using Metadata based on Advanced Authoring Format (Advanced Authoring Format기반에서 메타데이터를 활용한 영상제작환경)

  • Mok, Seon-Ah;Kim, Chul-Hyun;Paik, Joon-Ki
    • Journal of Broadcast Engineering
    • /
    • v.13 no.2
    • /
    • pp.274-282
    • /
    • 2008
  • In this paper, we propose from media production environment using meta-data based on advanced authoring format(AAF). The media production system becomes digitalized since the image degradation and data storage should be minimized. For transmitting various contents without loss of meta-data. Editing decision list(EDL)is used in the current broadcasting and cinema environment, which results in inefficient performance. We compared the proposed AAF with the existing EDL cut, and tested successful transmission of the metadata. Based on the experimental results, the proposed AAF contains more video information than EDL.

A Study on the Extraction and Integration of Learning Object Meta-data using Web Service of Databases (DBMS의 웹서비스를 이용한 학습객체 메타데이터 추출 및 통합에 관한 연구)

  • Choe, Hyun-Jong
    • Journal of The Korean Association of Information Education
    • /
    • v.7 no.2
    • /
    • pp.199-206
    • /
    • 2003
  • XML is becoming a new developing tool of web technology because of its ability of data management and flexibility in data presentation. So it's well researched that the reusability and integration with learning objects such as text, image, sound, video and plug-in programs of web contents in computer education. But the research for storing, extracting and integrating metadata about learning object was needed prior to implementing online learning system to integrate and manage it. Therefore this study propose a new method of using web service of DBMS for extracting learning object's metadata in database server which located in 3-tier system. To evaluate the efficiency of proposed method, The test server and two DBMSs(MS SQL Server 2000 and Oracle 9i) which have 30 metadata was implemented and the response time of it was measured. The response time of it was short, but in order to using this method the additional programming with SAX/DOM was necessary.

  • PDF

Implementation of Character and Object Metadata Generation System for Media Archive Construction (미디어 아카이브 구축을 위한 등장인물, 사물 메타데이터 생성 시스템 구현)

  • Cho, Sungman;Lee, Seungju;Lee, Jaehyeon;Park, Gooman
    • Journal of Broadcast Engineering
    • /
    • v.24 no.6
    • /
    • pp.1076-1084
    • /
    • 2019
  • In this paper, we introduced a system that extracts metadata by recognizing characters and objects in media using deep learning technology. In the field of broadcasting, multimedia contents such as video, audio, image, and text have been converted to digital contents for a long time, but the unconverted resources still remain vast. Building media archives requires a lot of manual work, which is time consuming and costly. Therefore, by implementing a deep learning-based metadata generation system, it is possible to save time and cost in constructing media archives. The whole system consists of four elements: training data generation module, object recognition module, character recognition module, and API server. The deep learning network module and the face recognition module are implemented to recognize characters and objects from the media and describe them as metadata. The training data generation module was designed separately to facilitate the construction of data for training neural network, and the functions of face recognition and object recognition were configured as an API server. We trained the two neural-networks using 1500 persons and 80 kinds of object data and confirmed that the accuracy is 98% in the character test data and 42% in the object data.

Stereo Audio Matched with 3D Video (3D영상에 정합되는 스테레오 오디오)

  • Park, Sung-Wook;Chung, Tae-Yun
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.2
    • /
    • pp.153-158
    • /
    • 2011
  • This paper presents subjective experimental results to understand how audio should be changed when a video clip is watched in 3D than 2D. This paper divided auditory perceptual information into two categories; distance and azimuth that a sound source contributes mostly, and spaciousness that scene or environment contribute mostly. According to the experiment for distance and azimuth, i.e. sound localization, we found that distance and azimuth of sound sources were magnified when heard with 3D than 2D video. This lead us to conclude 3D sound for localization should be designed to have more distance and azimuth than 2D sound. Also we found 3D sound are preferred to be played with not only 3D video clip but also 2D video clip. According to the experiment for spaciousness, we found people prefer sound with more reverberation when they watch 3D video clips than 2D video clips. This can be understood that 3D video provides more spacial information than 2D video. Those subjective experimental results can help audio engineer familiar with 2D audio to create 3D audio, and be fundamental information of future research to make 2D to 3D audio conversion system. Furthermore when designing 3D broadcasting system with limited bandwidth and with 2D TV supportive, we propose to consider transmitting stereoscopic video, audio with enhanced localization, and metadata for TV sets to generate reverberation for spaciousness.

Enhancement of Pruning Order Determining for Immersive Video Test Model (이머시브 비디오 테스트 모델에서의 프루닝 기법의 개선)

  • Shin, Hong-Chang;Yun, Junyoung;Lee, Gwangsoon;Eum, Homin;Seo, Jungil
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.305-307
    • /
    • 2020
  • 이머시브 비디오 서비스를 위해 MPEG-I Visual 그룹에서는 3DoF+ 기술과 관련하여 MIV(Metadata for Immersive video)의 표준화를 진행하고 있으며, 이를 위해 다시점 영상 및 전방위 장면을 촬영한 ERP 규격의 영상들이 주어진 경우에 운동시차를 제공할 수 있는 영상 합성 레퍼런스 소프트웨어인 TMIV SW를 제공한다. TMIV는 기본적으로 송신부인 인코더와 수신부인 디코더로 구성이 되어있으며, 인코더에서 가장 중요한 기능은 다수의 입력 시점영상 간의 중복된 데이터를 찾아내서 제거하는 프루닝 과정이다. 프루닝 방법에 따라 데이터 전송량과 디코더에서의 합성 품질이 달라지기 때문에 인코더에서 핵심이라고 할 수 있다. 본 논문은 인코더의 프루닝의 효율을 높이기 위해 전체 흐름도에서 프루닝 순서 변경 과정을 추가하고 그 과정에서 시점 영상간 중첩 영역을 계산하여 이를 토대로 프루닝 순서를 결정하는 방법을 제안하였고 이를 통해 데이터 압축률이 향상됨을 확인할 수 있었고, 또한 수신부에서 영상 합성의 품질이 달라짐을 확인할 수 있었다.

  • PDF

Smart Camera Technology to Support High Speed Video Processing in Vehicular Network (차량 네트워크에서 고속 영상처리 기반 스마트 카메라 기술)

  • Son, Sanghyun;Kim, Taewook;Jeon, Yongsu;Baek, Yunju
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.1
    • /
    • pp.152-164
    • /
    • 2015
  • A rapid development of semiconductors, sensors and mobile network technologies has enable that the embedded device includes high sensitivity sensors, wireless communication modules and a video processing module for vehicular environment, and many researchers have been actively studying the smart car technology combined on the high performance embedded devices. The vehicle is increased as the development of society, and the risk of accidents is increasing gradually. Thus, the advanced driver assistance system providing the vehicular status and the surrounding environment of the vehicle to the driver using various sensor data is actively studied. In this paper, we design and implement the smart vehicular camera device providing the V2X communication and gathering environment information. And we studied the method to create the metadata from a received video data and sensor data using video analysis algorithm. In addition, we invent S-ROI, D-ROI methods that set a region of interest in a video frame to improve calculation performance. We performed the performance evaluation for two ROI methods. As the result, we confirmed the video processing speed that S-ROI is 3.0 times and D-ROI is 4.8 times better than a full frame analysis.

Video Search System using Semantic Web and MPEG-7 Metadata (시맨틱 웹과 MPEG-7 메타데이타를 이용한 비디오 검색 시스템)

  • Lee Seong-Woo;Han Jong-Hyeong;Ahn Chul-Bum;Nah Yun-Mook
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2006.05a
    • /
    • pp.165-168
    • /
    • 2006
  • 본 논문에서는 시맨틱 웹을 위한 마크업 언어인 RDF와 MPEG-7 표준을 따르는 XML 문서를 이용하여 구축된 비디오 검색 시스템에 대한 연구결과를 기술한다. 본 논문에서 제안된 시스템은 RDF를 통해 비디오 파일에 대한 정보를 웹 자원의 관점으로 정의/저장한다. 또한 비디오에 대한 내용기반 검색을 위해 비디오를 구성하는 중요 프레임들을 추출하여 그림 파일로 저장하는 한편 각 프레임에 대한 정보를 MPEG-7 표준에 따라 XML 형태로 저장한다. 이렇게 구축된 정보들을 바탕으로 비디오 검색 엔진은 사용자의 질의요청을 수행한다. 본 논문에서 제안하는 비디오 검색 시스템은 비디오에 대한 다양한 검색은 물론 비디오에 대한 내용기반 검색이 가능한 장점을 가진다.

  • PDF