• Title/Summary/Keyword: Encoder-decoder

Search Result 447, Processing Time 0.029 seconds

The Performance Estiamtion of Turbo Internal Interleaver Using Weight Distribution of Codewords (부호어의 무게 분포를 통한 터보 인터리버의 성능 분석)

  • 고태환;김주민;정덕진
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.3A
    • /
    • pp.173-179
    • /
    • 2002
  • In this paper, we suggest more precise performance analysis method of turbo interleavers based on two criteria; performance bounds like Union Bound and weight frequency of codewords. In order to present our new method, we employ block pseudo random, and so-called prime interleavers in compliance of 3GPP standard, respectively, We also applied this method to S-random interleavers that have different window size, S. 3GPP complied turbo encoder, decoder, and AWGN channel are implemented by using MATLAB for our performance analysis. According to our analysis, both criteria should be taken into account coincidently to predict the performance of newly designed interleavers.

Implementation of Electronic Nose System applicable to MPEG-V(ISO/IEC 23005) Standardization (MPEG-V(ISO/IEC 23005) 표준적용이 가능한 전자코 시스템 구현)

  • Lim, Hea-Jin;Choi, Jang-Sik;Jeon, Jin-Young;Byun, Hyung-Gi
    • Journal of Sensor Science and Technology
    • /
    • v.25 no.6
    • /
    • pp.388-393
    • /
    • 2016
  • MPEG-V(ISO/IEC 23005) standardizes normative sensory effects metadata and sensory devices command for adapting the sensory effects between the virtual world and the real world. MPEG-V(Virtual) standardization has been carried out by 3DG(Dimensional Graphics) ad-hoc group inside MPEG Working Group(ISO IEC JTC1/SC29/WG11). For the scent effect, one of the sensory effects within MPEG-V, we proposed an olfactory interaction model including electronic nose and scent display to the ad-hoc group. Recently, we proposed types and elements related to the electronic nose as a sensor defined in MPEG-V standard for olfactory interaction. At the 114th MPEG meeting, the types and elements were consequently reflected on MPEG-V CD(Committee Draft) 4th edition. In this paper, we implement an electronic nose system applicable to MPEG-V standard by using MPEG-V schema, encoder, and decoder in order to assess their adequacy.

Implementation of an Intelligent Video Surveillance System based on Digital Media Processor (디지털미디어프로세서 기반의 지능형 비디오 감시 시스템 구현)

  • Kim, Won-Ho
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.11 no.3
    • /
    • pp.841-846
    • /
    • 2010
  • This paper presents design and implementation of an intelligent video surveillance system. The proposed system has advantages of management efficiency and operation robustness unrelated to working condition compared to conventional CCTV based system. The system hardware is designed and implemented by using commercial chips such as digital media processor and video encoder, video decoder and the functions of software are to analyze temperature distribution of a infrared image and to detect disaster situation such as fire. The required functions are confirmed by testing of the prototype and we verified practicality of the system.

A Stabilization of MC-BCS-SPL Scheme for Distributed Compressed Video Sensing (분산 압축 비디오 센싱을 위한 MC-BCS-SPL 기법의 안정화 알고리즘)

  • Ryu, Joong-seon;Kim, Jin-soo
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.5
    • /
    • pp.731-739
    • /
    • 2017
  • Distributed compressed video sensing (DCVS) is a framework that integrates both compressed sensing and distributed video coding characteristics to achieve a low complexity video sampling. In DCVS schemes, motion estimation & motion compensation is employed at the decoder side, similarly to distributed video coding (DVC), for a low-complex encoder. However, since a simple BCS-SPL algorithm is applied to a residual arising from motion estimation and compensation in conventional MC-BCS-SPL (motion compensated block compressed sensing with smoothed projected Landweber) scheme, the reconstructed visual qualities are severly degraded in Wyner-Ziv (WZ) frames. Furthermore, the scheme takes lots of iteration to reconstruct WZ frames. In this paper, the conventional MC-BCS-SPL algorithm is improved to be operated in more effective way in WZ frames. That is, first, the proposed algorithm calculates a correlation coefficient between two reference key frames and, then, by selecting adaptively the reference frame, the residual reconstruction in pixel domain is performed to the conventional BCS-SPL scheme. Experimental results show that the proposed algorithm achieves significantly better visual qualities than conventional MC-BCS-SPL algorithm, while resulting in the significant reduction of the decoding time.

An Ultrasonic Wave Encoder and Decoder for Indoor Positioning of Mobile Marketing System

  • Kim, Young-Mo;Jang, Se-Young;Park, Byeong-Chan;Bang, Kyung-Sik;Kim, Seok-Yoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.7
    • /
    • pp.93-100
    • /
    • 2019
  • In this paper, we propose an intelligent marketing service system that can provide custom advertisements and events to both businesses and customers by identifying the location and contents using the ultrasonic signals and feature information in voice signals. We also develop the encoding and decoding algorithm of ultrasonic signals for this system and analyze the performance evaluation results. With the development of the hyper-connected society, the on-line marketing has been activated and is growing in size. Existing store marketing applications have disadvantages that customers have to find out events or promotional materials that the headquarters or stores throughusing the corresponding applications whenever they visit them. To solve these problems, there are attempts to create intelligent marketing tools using GPS technology and voice recognition technology. However, this approach has difficulties in technology development due to accuracy of location and speed of comparison and retrieval of voice recognition technology, and marketing services for customer relation are also much simplified.

Deep Learning Based Gray Image Generation from 3D LiDAR Reflection Intensity (딥러닝 기반 3차원 라이다의 반사율 세기 신호를 이용한 흑백 영상 생성 기법)

  • Kim, Hyun-Koo;Yoo, Kook-Yeol;Park, Ju H.;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.1
    • /
    • pp.1-9
    • /
    • 2019
  • In this paper, we propose a method of generating a 2D gray image from LiDAR 3D reflection intensity. The proposed method uses the Fully Convolutional Network (FCN) to generate the gray image from 2D reflection intensity which is projected from LiDAR 3D intensity. Both encoder and decoder of FCN are configured with several convolution blocks in the symmetric fashion. Each convolution block consists of a convolution layer with $3{\times}3$ filter, batch normalization layer and activation function. The performance of the proposed method architecture is empirically evaluated by varying depths of convolution blocks. The well-known KITTI data set for various scenarios is used for training and performance evaluation. The simulation results show that the proposed method produces the improvements of 8.56 dB in peak signal-to-noise ratio and 0.33 in structural similarity index measure compared with conventional interpolation methods such as inverse distance weighted and nearest neighbor. The proposed method can be possibly used as an assistance tool in the night-time driving system for autonomous vehicles.

Detection and Localization of Image Tampering using Deep Residual UNET with Stacked Dilated Convolution

  • Aminu, Ali Ahmad;Agwu, Nwojo Nnanna;Steve, Adeshina
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.9
    • /
    • pp.203-211
    • /
    • 2021
  • Image tampering detection and localization have become an active area of research in the field of digital image forensics in recent times. This is due to the widespread of malicious image tampering. This study presents a new method for image tampering detection and localization that combines the advantages of dilated convolution, residual network, and UNET Architecture. Using the UNET architecture as a backbone, we built the proposed network from two kinds of residual units, one for the encoder path and the other for the decoder path. The residual units help to speed up the training process and facilitate information propagation between the lower layers and the higher layers which are often difficult to train. To capture global image tampering artifacts and reduce the computational burden of the proposed method, we enlarge the receptive field size of the convolutional kernels by adopting dilated convolutions in the residual units used in building the proposed network. In contrast to existing deep learning methods, having a large number of layers, many network parameters, and often difficult to train, the proposed method can achieve excellent performance with a fewer number of parameters and less computational cost. To test the performance of the proposed method, we evaluate its performance in the context of four benchmark image forensics datasets. Experimental results show that the proposed method outperforms existing methods and could be potentially used to enhance image tampering detection and localization.

Context-Awareness Cat Behavior Captioning System (반려묘의 상황인지형 행동 캡셔닝 시스템)

  • Chae, Heechan;Choi, Yoona;Lee, Jonguk;Park, Daihee;Chung, Yongwha
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.1
    • /
    • pp.21-29
    • /
    • 2021
  • With the recent increase in the number of households raising pets, various engineering studies have been underway for pets. The final purpose of this study is to automatically generate situation-sensitive captions that can express implicit intentions based on the behavior and sound of cats by embedding the already mature behavioral detection technology of pets as basic element technology in the video capturing research. As a pilot project to this end, this paper proposes a high-level capturing system using optical-flow, RGB, and sound information of cat videos. That is, the proposed system uses video datasets collected in an actual breeding environment to extract feature vectors from the video and sound, then through hierarchical LSTM encoder and decoder, to identify the cat's behavior and its implicit intentions, and to perform learning to create context-sensitive captions. The performance of the proposed system was verified experimentally by utilizing video data collected in the environment where actual cats are raised.

A Study on Error-Resilient, Scalable Video Codecs Based on the Set Partitioning in Hierarchical Trees(SPIHT) Algorithm (계층적 트리의 집합 분할 알고리즘(SPIHT)에 기반한 에러에 강하고 가변적인 웨이브렛 비디오 코덱에 관한 연구)

  • Inn-Ho, Jee
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.1
    • /
    • pp.37-43
    • /
    • 2023
  • Compressed still image or video bitstreams require protection from channel errors in a wireless channel. Embedded Zerotree Coding(EZW), SPIHT could have provided unprecedented high performance in image compression with low complexity. If bit error is generated by dint of wireless channel transmission problem, the loss of synchronization on between encoder and decoder causes serious performance degradation. But wavelet zerotree coding algorithms are producing variable-length codewords, extremely sensitive to bit errors. The idea is to partition the lifting coefficients. A many partition of lifting transform coefficients distributes channel error from wireless channel to each partition. Therefore synchronization problem that caused quality deterioration in still image and video stream was improved.

Crack segmentation in high-resolution images using cascaded deep convolutional neural networks and Bayesian data fusion

  • Tang, Wen;Wu, Rih-Teng;Jahanshahi, Mohammad R.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.221-235
    • /
    • 2022
  • Manual inspection of steel box girders on long span bridges is time-consuming and labor-intensive. The quality of inspection relies on the subjective judgements of the inspectors. This study proposes an automated approach to detect and segment cracks in high-resolution images. An end-to-end cascaded framework is proposed to first detect the existence of cracks using a deep convolutional neural network (CNN) and then segment the crack using a modified U-Net encoder-decoder architecture. A Naïve Bayes data fusion scheme is proposed to reduce the false positives and false negatives effectively. To generate the binary crack mask, first, the original images are divided into 448 × 448 overlapping image patches where these image patches are classified as cracks versus non-cracks using a deep CNN. Next, a modified U-Net is trained from scratch using only the crack patches for segmentation. A customized loss function that consists of binary cross entropy loss and the Dice loss is introduced to enhance the segmentation performance. Additionally, a Naïve Bayes fusion strategy is employed to integrate the crack score maps from different overlapping crack patches and to decide whether a pixel is crack or not. Comprehensive experiments have demonstrated that the proposed approach achieves an 81.71% mean intersection over union (mIoU) score across 5 different training/test splits, which is 7.29% higher than the baseline reference implemented with the original U-Net.