• Title/Summary/Keyword: Image-to-Video

Search Result 2,715, Processing Time 0.031 seconds

Design and Implementation of the Video Query Processing Engine for Content-Based Query Processing (내용기반 질의 처리를 위한 동영상 질의 처리기의 설계 및 구현)

  • Jo, Eun-Hui;Kim, Yong-Geol;Lee, Hun-Sun;Jeong, Yeong-Eun;Jin, Seong-Il
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.3
    • /
    • pp.603-614
    • /
    • 1999
  • As multimedia application services on high-speed information network have been rapidly developed, the need for the video information management system that provides an efficient way for users to retrieve video data is growing. In this paper, we propose a video data model that integrates free annotations, image features, and spatial-temporal features for video purpose of improving content-based retrieval of video data. The proposed video data model can act as a generic video data model for multimedia applications, and support free annotations, image features, spatial-temporal features, and structure information of video data within the same framework. We also propose the video query language for efficiently providing query specification to access video clips in the video data. It can formalize various kinds of queries based on the video contents. Finally we design and implement the query processing engine for efficient video data retrieval on the proposed metadata model and the proposed video query language.

  • PDF

Face Information Conversion Mechanism to Prevent Privacy Infringement (프라이버시 침해 방지를 위한 얼굴 정보 변환 메커니즘)

  • Kim, Jinsu;Kim, Sangchoon;Park, Namje
    • The Journal of Korean Institute of Information Technology
    • /
    • v.17 no.6
    • /
    • pp.115-122
    • /
    • 2019
  • CCTV(Closed-Circuit Television) is increasingly exposed to CCTV per person as the number of installations increases every year for accident prevention and facility safety. The intelligent video surveillance system technology is attracting attention to the privacy protection of exposed subjects. The intelligent video surveillance system performs a process for the privacy protection so as to perform the action type of the subject and the judgment of the situation in the simple identification of the photographed image data, or to prevent the information, from which the information of the photographed subject is exposed. The proposed technique is applied to the video surveillance system and converts the original image information taken from the video surveillance system into similar image information so that the original image information is not leaked to the outside. In this paper, we propose an image conversion mechanism that inserts a virtual face image that approximates a preset similarity.

Effective Scene Change Detection Method for MuIUmedia Bata as Video Images using Mean Squared Error (평균오차를 이용한 멀티미디어 동영상 데이터를 위한 효율적인 장면전환 검출)

  • Jung, Chang-Ryul;Koh, Jin-Gwang;Lee, Joon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.6 no.6
    • /
    • pp.951-957
    • /
    • 2002
  • When retrieving voluminous capacity of video image data, it is necessary to provide synopsized frame lists of video image data for indexing and replaying at the exact point where the user want to retrieve. We apply Mean Squared Error method to extract certain pixel value from diagonal direction of a frame. The RGB value of a pixel extracted from each frame is saved in a matrix form, and this frame is retrievedas a scene change point if the compared value of two points met the certain condition. Also implement the algorithm and provide a way to seize entire structure of video image and the point of scene changes. finally, we analyze and prove that our method has better performance compared with the others.

A Real-Time Video Stitching Algorithm in H.264/AVC Compressed Domain (실시간 H.264/AVC 압축 영역에서의 영상 합성 알고리즘)

  • Gankhuyag, Ganzorig;Hong, Eun Gi;Kim, Giyeol;Kim, Younghwan;Choe, Yoonsik
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.39C no.6
    • /
    • pp.503-511
    • /
    • 2014
  • In this paper, a novel, real-time video stitching algorithm in an H.264/AVC compressed domain is proposed. This enables viewers to watch multiple video contents using a single device. The basic concept of this paper is that the server is asked to combine multiple streams into one bit-stream based in a compressed domain. In other words, this paper presents a new compressed domain combiner that works in boundary macroblocks of input videos with re-calculating intra prediction mode, intra prediction MVD, a re-allocation of the coefficient table, and border extension methods. The rest of the macroblocks of the input video data are achieved simply by copying them. Simulation experiments have demonstrated the possibility and effectiveness of the proposed algorithm by showing that it is able to generate more than 103 frames per second, stitching four 480p-sized images into each frame.

Lossless Inter-frame Video Coding using Extended JPEG2000

  • IMAIZUMI, Shoko;TAKAGI, Ayuko;KIYA, Hitoshi
    • Proceedings of the IEEK Conference
    • /
    • 2002.07c
    • /
    • pp.1803-1806
    • /
    • 2002
  • This paper describes an effective technique for lossless inter-frame video coding sequences based on a JPEG2000 CODEC. This technique has diminished the compression rate for lossless video coding. In this proposed method, firstly a predicted image for an in- put image is generated by motion estimation(ME), and then a difference image between the input image and the predicted image is calculated, and finally the difference image becomes an input image to a JPEG2000 encoder for lossless coding. Simulation results show the effectiveness of this method.

  • PDF

Video Content Manipulation Using 3D Analysis for MPEG-4

  • Sull, Sanghoon
    • Journal of Broadcast Engineering
    • /
    • v.2 no.2
    • /
    • pp.125-135
    • /
    • 1997
  • This paper is concerned with realistic mainpulation of content in video sequences. Manipulation of content in video sequences is one of the content-based functionalities for MPEG-4 Visual standard. We present an approach to synthesizing video sequences by using the intermediate outputs of three-dimensional (3D) motion and depth analysis. For concreteness, we focus on video showing 3D motion of an observer relative to a scene containing planar runways (or roads). We first present a simple runway (or road) model. Then, we describe a method of identifying the runway (or road) boundary in the image using the Point of Heading Direction (PHD) which is defined as the image of, the ray along which a camera moves. The 3D motion of the camera is obtained from one of the existing 3D analysis methods. Then, a video sequence containing a runway is manipulated by (i) coloring the scene part above a vanishing line, say blue, to show sky, (ii) filling in the occluded scene parts, and (iii) overlaying the identified runway edges and placing yellow disks in them, simulating lights. Experimental results for a real video sequence are presented.

  • PDF

Affine Model for Generating Stereo Mosaic Image from Video Frames (비디오 프레임 영상의 자유 입체 모자이크 영상 제작을 위한 부등각 모델 연구)

  • Noh, Myoung-Jong;Cho, Woo-Sug;Park, Jun-Ku;Koh, Jin-Woo
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.17 no.3
    • /
    • pp.49-56
    • /
    • 2009
  • Recently, a generation of high quality mosaic images from video sequences has been attempted by a variety of investigations. Among the matter of investigation, in this paper, generation on stereo mosaic utilizing airborne-video sequence images is focused upon. The stereo mosaic is made by creating left and right mosaic which are fabricated by front and rear slices having different viewing angle in consecutive video frames. For making the stereo mosaic, motion parameters which are able to define geometric relationship between consecutive video frames are determined. For determining motion parameters, affine model which is able to explain relative motion parameters is applied by this paper. The mosaicing method using relative motion parameters is called by free mosaic. The free mosaic proposed in this paper consists of 4 step processes: image registration with reference to first frame using affine model, front and rear slicing, stitching line definition and image mosaicing. As the result of experiment, the left and right mosaic image, anaglyphic image for stereo mosaic images are showed and analyzed y-parallax for checking accuracy.

  • PDF

Fast Extraction of Objects of Interest from Images with Low Depth of Field

  • Kim, Chang-Ick;Park, Jung-Woo;Lee, Jae-Ho;Hwang, Jenq-Neng
    • ETRI Journal
    • /
    • v.29 no.3
    • /
    • pp.353-362
    • /
    • 2007
  • In this paper, we propose a novel unsupervised video object extraction algorithm for individual images or image sequences with low depth of field (DOF). Low DOF is a popular photographic technique which enables the representation of the photographer's intention by giving a clear focus only on an object of interest (OOI). We first describe a fast and efficient scheme for extracting OOIs from individual low-DOF images and then extend it to deal with image sequences with low DOF in the next part. The basic algorithm unfolds into three modules. In the first module, a higher-order statistics map, which represents the spatial distribution of the high-frequency components, is obtained from an input low-DOF image. The second module locates the block-based OOI for further processing. Using the block-based OOI, the final OOI is obtained with pixel-level accuracy. We also present an algorithm to extend the extraction scheme to image sequences with low DOF. The proposed system does not require any user assistance to determine the initial OOI. This is possible due to the use of low-DOF images. The experimental results indicate that the proposed algorithm can serve as an effective tool for applications, such as 2D to 3D and photo-realistic video scene generation.

  • PDF

A Technical Analysis on Deep Learning based Image and Video Compression (딥 러닝 기반의 이미지와 비디오 압축 기술 분석)

  • Cho, Seunghyun;Kim, Younhee;Lim, Woong;Kim, Hui Yong;Choi, Jin Soo
    • Journal of Broadcast Engineering
    • /
    • v.23 no.3
    • /
    • pp.383-394
    • /
    • 2018
  • In this paper, we investigate image and video compression techniques based on deep learning which are actively studied recently. The deep learning based image compression technique inputs an image to be compressed in the deep neural network and extracts the latent vector recurrently or all at once and encodes it. In order to increase the image compression efficiency, the neural network is learned so that the encoded latent vector can be expressed with fewer bits while the quality of the reconstructed image is enhanced. These techniques can produce images of superior quality, especially at low bit rates compared to conventional image compression techniques. On the other hand, deep learning based video compression technology takes an approach to improve performance of the coding tools employed for existing video codecs rather than directly input and process the video to be compressed. The deep neural network technologies introduced in this paper replace the in-loop filter of the latest video codec or are used as an additional post-processing filter to improve the compression efficiency by improving the quality of the reconstructed image. Likewise, deep neural network techniques applied to intra prediction and encoding are used together with the existing intra prediction tool to improve the compression efficiency by increasing the prediction accuracy or adding a new intra coding process.

Soft-$\alpha$ Filter Technology for image enhancement of MPEG-2 Video (MPEG-2 비디오의 화질 향상을 위한 소프트-$\alpha$ 필터 기법)

  • 심비연;박영배
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.04b
    • /
    • pp.109-111
    • /
    • 2002
  • Visual organs play an important role in human information recognition processes. If they are expressed in a way of digital information, it makes much bigger amount of visual information among any other information. For that reason, MPEG-2 has been taken use of to represent information compressing technology in multi-media. Although the imported data would basically contain noises, when original video images are encoded into MPET-2. Accordingly, we propose soft- $\alpha$ filter to improve image quality of digital image received from the actual image and to reduce noises from them. We also propose a method combining vertical/horizontal filter and soft- $\alpha$ filter on MPEG-2 video image. We can get two kinds of effects from the advantages of this kind of combination. Firstly, it will reduce processing time ducting horizontal and vetical filtering process. It will cover time for soft- $\alpha$ filter. Secondly, it will simplify the colors in horizontal and vertical filter. Therefore we can get clearer quality without noises from soft- $\alpha$ filter.

  • PDF