• Title/Summary/Keyword: Video Frames

Search Result 888, Processing Time 0.032 seconds

Single Pixel Compressive Camera for Fast Video Acquisition using Spatial Cluster Regularization

  • Peng, Yang;Liu, Yu;Lu, Kuiyan;Zhang, Maojun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5481-5495
    • /
    • 2018
  • Single pixel imaging technology has developed for years, however the video acquisition on the single pixel camera is not a well-studied problem in computer vision. This work proposes a new scheme for single pixel camera to acquire video data and a new regularization for robust signal recovery algorithm. The method establishes a single pixel video compressive sensing scheme to reconstruct the video clips in spatial domain by recovering the difference of the consecutive frames. Different from traditional data acquisition method works in transform domain, the proposed scheme reconstructs the video frames directly in spatial domain. At the same time, a new regularization called spatial cluster is introduced to improve the performance of signal reconstruction. The regularization derives from the observation that the nonzero coefficients often tend to be clustered in the difference of the consecutive video frames. We implement an experiment platform to illustrate the effectiveness of the proposed algorithm. Numerous experiments show the well performance of video acquisition and frame reconstruction on single pixel camera.

Content-based Video Information Retrieval and Streaming System using Viewpoint Invariant Regions

  • Park, Jong-an
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.2 no.1
    • /
    • pp.43-50
    • /
    • 2009
  • This paper caters the need of acquiring the principal objects, characters, and scenes from a video in order to entertain the image based query. The movie frames are divided into frames with 2D representative images called "key frames". Various regions in a key frame are marked as key objects according to their textures and shapes. These key objects serve as a catalogue of regions to be searched and matched from rest of the movie, using viewpoint invariant regions calculation, providing the location, size, and orientation of all the objects occurring in the movie in the form of a set of structures collaborating as video profile. The profile provides information about occurrences of every single key object from every frame of the movie it exists in. This information can further ease streaming of objects over various network-based viewing qualities. Hence, the method provides an effective reduced profiling approach of automatic logging and viewing information through query by example (QBE) procedure, and deals with video streaming issues at the same time.

  • PDF

An Effective MC-BCS-SPL Algorithm and Its Performance Comparison with Respect to Prediction Structuring Method (효과적인 MC-BCS-SPL 알고리즘과 예측 구조 방식에 따른 성능 비교)

  • Ryug, Joong-seon;Kim, Jin-soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.7
    • /
    • pp.1355-1363
    • /
    • 2017
  • Recently, distributed compressed video sensing (DCVS) has been actively studied in order to achieve a low complexity video encoder by integrating both compressed sensing and distributed video coding characteristics. Conventionally, a motion compensated block compressed sensing with smoothed projected Landweber (MC-BCS-SPL) has been considered as an effective scheme of DCVS with all compressed sensing frames pursuing the simplest sampling. In this scheme, video frames are separately classified into key frames and WZ frames. However, when reconstructing WZ frame with conventional MC-BCS-SPL scheme at the decoder side, the visual qualities are poor for temporally active video sequences. In this paper, to overcome the drawbacks of the conventional scheme, an enhanced MC-BCS-SPL algorithm is proposed, which corrects the initial image with reference to the key frame using a high correlation between adjacent key frames. The proposed scheme is analyzed with respect to GOP (Group of Pictures) structuring method. Experimental results show that the proposed method performs better than conventional MC-BCS-SPL in rate-distortion.

Adaptive Video Watermarking using the Bitrate and the Motion Vector (비트율과 움직임 벡터를 이용한 적응적 동영상 워터마킹)

  • Ahn, I.Y.
    • 전자공학회논문지 IE
    • /
    • v.43 no.4
    • /
    • pp.37-42
    • /
    • 2006
  • This paper proposes a adaptive video watermarking algorithm according to bitrate and motion vector size in MPEG2 system. The watermark strength in the I-frames is adapted for quantization step size and the strength in the P-B-frames is adapted for quantization step size and motion vector of macroblock to make the watermark more robust against the accompanying degradation due to aggressively compression. A realtime watermark extraction is done directly in the DCT domain during MPEG decoding without full decoding of MPEG video. The experimental simulations show that the video quality results almost invisible difference between the watermarked frames and the original frames and the watermark is resistant to frame dropping, MPEG compression, GoP conversion and low pass filter attacks.

Video Quality Assessment based on Deep Neural Network

  • Zhiming Shi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.8
    • /
    • pp.2053-2067
    • /
    • 2023
  • This paper proposes two video quality assessment methods based on deep neural network. (i)The first method uses the IQF-CNN (convolution neural network based on image quality features) to build image quality assessment method. The LIVE image database is used to test this method, the experiment show that it is effective. Therefore, this method is extended to the video quality assessment. At first every image frame of video is predicted, next the relationship between different image frames are analyzed by the hysteresis function and different window function to improve the accuracy of video quality assessment. (ii)The second method proposes a video quality assessment method based on convolution neural network (CNN) and gated circular unit network (GRU). First, the spatial features of video frames are extracted using CNN network, next the temporal features of the video frame using GRU network. Finally the extracted temporal and spatial features are analyzed by full connection layer of CNN network to obtain the video quality assessment score. All the above proposed methods are verified on the video databases, and compared with other methods.

An Efficient Video Sequence Matching Algorithm (효율적인 비디오 시퀀스 정합 알고리즘)

  • 김상현;박래홍
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.5
    • /
    • pp.45-52
    • /
    • 2004
  • According tothe development of digital media technologies various algorithms for video sequence matching have been proposed to match the video sequences efficiently. A large number of video sequence matching methods have focused on frame-wise query, whereas a relatively few algorithms have been presented for video sequence matching or video shot matching. In this paper, we propose an efficientalgorithm to index the video sequences and to retrieve the sequences for video sequence query. To improve the accuracy and performance of video sequence matching, we employ the Cauchy function as a similarity measure between histograms of consecutive frames, which yields a high performance compared with conventional measures. The key frames extracted from segmented video shots can be used not only for video shot clustering but also for video sequence matching or browsing, where the key frame is defined by the frame that is significantly different from the previous fames. Several key frame extraction algorithms have been proposed, in which similar methods used for shot boundary detection were employed with proper similarity measures. In this paper, we propose the efficient algorithm to extract key frames using the cumulative Cauchy function measure and. compare its performance with that of conventional algorithms. Video sequence matching can be performed by evaluating the similarity between data sets of key frames. To improve the matching efficiency with the set of extracted key frames we employ the Cauchy function and the modified Hausdorff distance. Experimental results with several color video sequences show that the proposed method yields the high matching performance and accuracy with a low computational load compared with conventional algorithms.

Video Augmentation of Virtual Object by Uncalibrated 3D Reconstruction from Video Frames (비디오 영상에서의 비보정 3차원 좌표 복원을 통한 가상 객체의 비디오 합성)

  • Park Jong-Seung;Sung Mee-Young
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.4
    • /
    • pp.421-433
    • /
    • 2006
  • This paper proposes a method to insert virtual objects into a real video stream based on feature tracking and camera pose estimation from a set of single-camera video frames. To insert or modify 3D shapes to target video frames, the transformation from the 3D objects to the projection of the objects onto the video frames should be revealed. It is shown that, without a camera calibration process, the 3D reconstruction is possible using multiple images from a single camera under the fixed internal camera parameters. The proposed approach is based on the simplification of the camera matrix of intrinsic parameters and the use of projective geometry. The method is particularly useful for augmented reality applications to insert or modify models to a real video stream. The proposed method is based on a linear parameter estimation approach for the auto-calibration step and it enhances the stability and reduces the execution time. Several experimental results are presented on real-world video streams, demonstrating the usefulness of our method for the augmented reality applications.

  • PDF

Orientation Analysis between UAV Video and Photos for 3D Measurement of Bridges (교량의 3차원 측정을 위한 UAV 비디오와 사진의 표정 분석)

  • Han, Dongyeob;Park, Jae Bong;Huh, Jungwon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.36 no.6
    • /
    • pp.451-456
    • /
    • 2018
  • UAVs (Unmanned Aerial Vehicles) are widely used for maintenance and monitoring of facilities. It is necessary to acquire a high-resolution image for evaluating the appearance state of the facility in safety inspection. In addition, it is essential to acquire the video data in order to acquire data over a wide area rapidly. In general, since video data does not include position information, it is difficult to analyze the actual size of the inspection object quantitatively. In this study, we evaluated the utilization of 3D point cloud data of bridges using a matching between video frames and reference photos. The drones were used to acquire video and photographs. And exterior orientations of the video frames were generated through feature point matching with reference photos. Experimental results showed that the accuracy of the video frame data is similar to that of the reference photos. Furthermore, the point cloud data generated by using video frames represented the shape and size of bridges with usable accuracy. If the stability of the product is verified through the matching test of various conditions in the future, it is expected that the video-based facility modeling and inspection will be effectively conducted.

Correction of Rotated Frames in Video Sequences Using Modified Mojette Transform (변형된 모젯 변환을 이용한 동영상에서의 회전 프레임 보정)

  • Kim, Ji-Hong
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.1
    • /
    • pp.42-49
    • /
    • 2013
  • The camera motion is accompanied with the translation and/or the rotation of objects in frames of a video sequence. An unnecessary rotation of objects declines the quality of the moving pictures and in addition is a primary cause of the viewers' fatigue. In this paper, a novel method for correcting rotated frames in video sequences is presented, where the modified Mojette transform is applied to the motion-compensated area in each frame. The Mojette transform is one of discrete Radon transforms, and is modified for correcting the rotated frames as follows. First, the bin values in the Mojette transform are determined by using pixels on the projection line and the interpolation of pixels adjacent to the line. Second, the bin values are calculated only at some area determined by the motion estimation between current and reference frames. Finally, only one bin at each projection is computed for reducing the amount of the calculation in the Mojette transform. Through the simulation carried out on various test video sequences, it is shown that the proposed scheme has good performance for correcting the rotation of frames in moving pictures.

Development of an Perceptual Video Quality Assessment Metric Using HVS and Video Communication Parameters (인간 시각 특성과 영상통신 파라미터를 이용한 동영상 품질 메트릭 개발)

  • Lee, Won-Kyun;Jang, Seong-Hwan;Park, Heui-Cheol;Lee, Ju-Yong;Suh, Chang-Ryul;Kim, Jung-Joon
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2007.08a
    • /
    • pp.155-158
    • /
    • 2007
  • In this paper, we solved the underestimation problem of PSNR, which is caused by repeated frames, by easily synchronizing original and decoded frames using the proposed marks. Also we propose full-reference system which can be applied for measuring the quality of various kinds of video communication systems, e.g. wireless handsets, mobile phones and applications for PC. In addition, we propose a new video quality assessment metric using video communication parameters, i.e. frame rate and delay. According to the experiments, the proposed metric is not only appropriate for real-time video communication systems but also shows better correlation with the subjective video quality assessment than PSNR. The proposed measuring system and metric can be effectively used for measuring and standardizing the video quality of future communications.

  • PDF