• Title/Summary/Keyword: maximum a posteriori

Search Result 162, Processing Time 0.022 seconds

A Bottom-up and Top-down Based Disparity Computation

  • Kim, Jung-Gu;hong Jeong
    • Journal of Electrical Engineering and information Science
    • /
    • v.3 no.2
    • /
    • pp.211-221
    • /
    • 1998
  • It is becoming apparent that stereo matching algorithms need much information from high level cognitive processes. Otherwise, conventional algorithms based on bottom-up control alone are susceptible to local minima. We introduce a system that consists of two levels. A lower level, using a usual matching method, is based upon the local neighborhood and a second level, that can integrate the partial information, is aimed at contextual matching. Conceptually, the introduction of bottom-up and top-down feedback loop to the usual matching algorithm improves the overall performance. For this purpose, we model the image attributes using a Markov random field (MRF) and thereupon derive a maximum a posteriori (MAP) estimate. The energy equation, corresponding to the estimate, efficiently represents the natural constraints such as occlusion and the partial informations from the other levels. In addition to recognition, we derive a training method that can determine the system informations from the other levels. In addition to recognition, we derive a training method that can determine the system parameters automatically. As an experiment, we test the algorithms using random dot stereograms (RDS) as well as natural scenes. It is proven that the overall recognition error is drastically reduced by the introduction of contextual matching.

  • PDF

A 18-Mbp/s, 8-State, High-Speed Turbo Decoder

  • Jung Ji-Won;Kim Min-Hyuk;Jeong Jin-Hee
    • Journal of electromagnetic engineering and science
    • /
    • v.6 no.3
    • /
    • pp.147-154
    • /
    • 2006
  • In this paper, we propose and present implementation results of a high-speed turbo decoding algorithm. The latency caused by (de) interleaving and iterative decoding in a conventional maximum a posteriori(MAP) turbo decoder can be dramatically reduced with the proposed design. The source of the latency reduction is come from the combination of the radix-4, dual-path processing, parallel decoding, and rearly-stop algorithms. This reduced latency enables the use of the turbo decoder as a forward error correction scheme in real-time wireless communication services. The proposed scheme results in a slight degradation in bit-error rate(BER) performance for large block sizes because the effective interleaver size in a radix-4 implementation is reduced to half, relative to the conventional method. Fixed on the parameters of N=212, iteration=3, 8-states, 3 iterations, and QPSK modulation scheme, we designed the adaptive high-speed turbo decoder using the Xilinx chip (VIRTEX2P (XC2VP30-5FG676)) with the speed of 17.78 Mb/s. From the results, we confirmed that the decoding speed of the proposed decoder is faster than conventional algorithms by 8 times.

Noise Robust Speaker Verification Using Subband-Based Reliable Feature Selection (신뢰성 높은 서브밴드 특징벡터 선택을 이용한 잡음에 강인한 화자검증)

  • Kim, Sung-Tak;Ji, Mi-Kyong;Kim, Hoi-Rin
    • MALSORI
    • /
    • no.63
    • /
    • pp.125-137
    • /
    • 2007
  • Recently, many techniques have been proposed to improve the noise robustness for speaker verification. In this paper, we consider the feature recombination technique in multi-band approach. In the conventional feature recombination for speaker verification, to compute the likelihoods of speaker models or universal background model, whole feature components are used. This computation method is not effective in a view point of multi-band approach. To deal with non-effectiveness of the conventional feature recombination technique, we introduce a subband likelihood computation, and propose a modified feature recombination using subband likelihoods. In decision step of speaker verification system in noise environments, a few very low likelihood scores of a speaker model or universal background model cause speaker verification system to make wrong decision. To overcome this problem, a reliable feature selection method is proposed. The low likelihood scores of unreliable feature are substituted by likelihood scores of the adaptive noise model. In here, this adaptive noise model is estimated by maximum a posteriori adaptation technique using noise features directly obtained from noisy test speech. The proposed method using subband-based reliable feature selection obtains better performance than conventional feature recombination system. The error reduction rate is more than 31 % compared with the feature recombination-based speaker verification system.

  • PDF

A Study on the Pattern Classificatiion of the EMG Signals Using Neural Network and Probabilistic Model (신경회로망과 확률모델을 이용한 근전도신호의 패턴분류에 관한 연구)

  • 장영건;권장우;장원환;장원석;홍성홍
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.28B no.10
    • /
    • pp.831-841
    • /
    • 1991
  • A combined model of probabilistic and MLP(multi layer perceptron) model is proposed for the pattern classification of EMG( electromyogram) signals. The MLP model has a problem of not guaranteeing the global minima of error and different quality of approximations to Bayesian probabilities. The probabilistic model is, however, closely related to the estimation error of model parameters and the fidelity of assumptions. A proper combination of these will reduce the effects of the problems and be robust to input variations. Proposed model is able to get the MAP(maximum a posteriori probability) in the probabilistic model by estimating a priori probability distribution using the MLP model adaptively. This method minimize the error probability of the probabilistic model as long as the realization of the MLP model is optimal, and this is a good combination of the probabilistic model and the MLP model for the usage of MLP model reliability. Simulation results show the benefit of the proposed model compared to use the Mlp and the probabilistic model seperately and the average calculation time fro classification is about 50ms in the case of combined motion using an IBM PC 25 MHz 386model.

  • PDF

Boundary-adaptive Despeckling : Simulation Study

  • Lee, Sang-Hoon
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.3
    • /
    • pp.295-309
    • /
    • 2009
  • In this study, an iterative maximum a posteriori (MAP) approach using a Bayesian model of Markovrandom field (MRF) was proposed for despeckling images that contains speckle. Image process is assumed to combine the random fields associated with the observed intensity process and the image texture process respectively. The objective measure for determining the optimal restoration of this "double compound stochastic" image process is based on Bayes' theorem, and the MAP estimation employs the Point-Jacobian iteration to obtain the optimal solution. In the proposed algorithm, MRF is used to quantify the spatial interaction probabilistically, that is, to provide a type of prior information on the image texture and the neighbor window of any size is defined for contextual information on a local region. However, the window of a certain size would result in using wrong information for the estimation from adjacent regions with different characteristics at the pixels close to or on boundary. To overcome this problem, the new method is designed to use less information from more distant neighbors as the pixel is closer to boundary. It can reduce the possibility to involve the pixel values of adjacent region with different characteristics. The proximity to boundary is estimated using a non-uniformity measurement based on standard deviation of local region. The new scheme has been extensively evaluated using simulation data, and the experimental results show a considerable improvement in despeckling the images that contain speckle.

The Optimal Turbo Coded V-BLAST Technique in the Adaptive Modulation System corresponding to each MIMO Scheme (적응 변조 시스템에서 각 MIMO 기법에 따른 최적의 터보 부호화된 V-BLAST 기법)

  • Lee, Kyung-Hwan;Ryoo, Sang-Jin;Choi, Kwang-Wook;You, Cheol-Woo;Hong, Dae-Ki;Kim, Dae-Jin;Hwang, In-Tae;Kim, Cheol-Sung
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.44 no.6 s.360
    • /
    • pp.40-47
    • /
    • 2007
  • In this paper, we propose and analyze the Adaptive Modulation System with optimal Turbo Coded V-BLAST(Vertical-Bell-lab Layered Space-Time) technique that adopts the extrinsic information from MAP (Maximum A Posteriori) Decoder with Iterative Decoding as a priori probability in two decoding procedures of V-BLAST; the ordering and the slicing. Also, we consider and compare the Adaptive Modulation System using conventional Turbo Coded V-BLAST technique that is simply combined V-BLAST with Turbo Coding scheme and the Adaptive Modulation System using conventional Turbo Coded V-BLAST technique that is decoded by the ML (Maximum Likelihood) decoding algorithm. We observe a throughput performance and a complexity. As a result of a performance comparison of each system, it has been proved that the complexity of the proposed decoding algorithm is lower than that of the ML decoding algorithm but is higher than that of the conventional V-BLAST decoding algorithm. however, we can see that the proposed system achieves a better throughput performance than the conventional system in the whole SNR (Signal to Noise Ratio) range. And the result shows that the proposed system achieves a throughput performance close to the ML decoded system. Specifically, a simulation shows that the maximum throughput improvement in each MIMO scheme is respectively about 350 kbps, 460 kbps, and 740 kbps compared to the conventional system. It is suggested that the effect of the proposed decoding algorithm accordingly gets higher as the number of system antenna increases.

Noise-Robust Speaker Recognition Using Subband Likelihoods and Reliable-Feature Selection

  • Kim, Sung-Tak;Ji, Mi-Kyong;Kim, Hoi-Rin
    • ETRI Journal
    • /
    • v.30 no.1
    • /
    • pp.89-100
    • /
    • 2008
  • We consider the feature recombination technique in a multiband approach to speaker identification and verification. To overcome the ineffectiveness of conventional feature recombination in broadband noisy environments, we propose a new subband feature recombination which uses subband likelihoods and a subband reliable-feature selection technique with an adaptive noise model. In the decision step of speaker recognition, a few very low unreliable feature likelihood scores can cause a speaker recognition system to make an incorrect decision. To overcome this problem, reliable-feature selection adjusts the likelihood scores of an unreliable feature by comparison with those of an adaptive noise model, which is estimated by the maximum a posteriori adaptation technique using noise features directly obtained from noisy test speech. To evaluate the effectiveness of the proposed methods in noisy environments, we use the TIMIT database and the NTIMIT database, which is the corresponding telephone version of TIMIT database. The proposed subband feature recombination with subband reliable-feature selection achieves better performance than the conventional feature recombination system with reliable-feature selection.

  • PDF

Post-processing of 3D Video Extension of H.264/AVC for a Quality Enhancement of Synthesized View Sequences

  • Bang, Gun;Hur, Namho;Lee, Seong-Whan
    • ETRI Journal
    • /
    • v.36 no.2
    • /
    • pp.242-252
    • /
    • 2014
  • Since July of 2012, the 3D video extension of H.264/AVC has been under development to support the multi-view video plus depth format. In 3D video applications such as multi-view and free-view point applications, synthesized views are generated using coded texture video and coded depth video. Such synthesized views can be distorted by quantization noise and inaccuracy of 3D wrapping positions, thus it is important to improve their quality where possible. To achieve this, the relationship among the depth video, texture video, and synthesized view is investigated herein. Based on this investigation, an edge noise suppression filtering process to preserve the edges of the depth video and a method based on a total variation approach to maximum a posteriori probability estimates for reducing the quantization noise of the coded texture video. The experiment results show that the proposed methods improve the peak signal-to-noise ratio and visual quality of a synthesized view compared to a synthesized view without post processing methods.

Performance Analysis of OFDM Systems with Turbo Code in a Satellite Broadcasting Channel (위성 방송 채널에서 터보 부호화된 OFDM 시스템의 성능 분석)

  • Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.9 no.6
    • /
    • pp.175-185
    • /
    • 2009
  • In this paper, performance of OFDM systems with turbo code is analyzed and simulated in a satellite broadcasting channel. The performance is evaluated in terms of bit error probability. The satellite channel is modeled as a combination of Rayleigh fading with shadowing and Rician fading channels. As turbo decoding algorithms, MAP (maximum a posteriori), Max-Log-MAP, and SOVA (soft decision Viterbi output) algorithms are chosen and their performances are compared. From simulation results, it is demonstrated that Max-Log-MAP algorithm is promising in terms of performance and complexity. It is shown that performance is substantially improved by increasing the number of iterations and interleaver length of a turbo encoder. The results in this paper can be applied to OFDM-based satellite broadcasting systems.

  • PDF

Error Resilience in Image Transmission Using LVQ and Turbo Coding

  • Hwang, Junghyeun;Joo, Sanghyun;Kikuchi, Hisakazu;Sasaki, Shigenobu;Muramatsu, Shogo;Shin, JaeHo
    • Proceedings of the IEEK Conference
    • /
    • 2000.07a
    • /
    • pp.478-481
    • /
    • 2000
  • In this paper, we propose a joint coding system for still images using source coding and powerful error correcting code schemes. Our system comprises an LVQ (lattice vector quantization) source coding for wavelet transformed images and turbo coding for channel coding. The parameters of the image encoder and channel encoder have been optimized for an n-D (dimension) cubic lattice (D$_{n}$, Z$_{n}$), parallel concatenation fur two simple RSC (recursive systematic convolutional code) and an interleaver. For decoding the received image in the case of the AWGN (additive white gaussian noise) channel, we used an iterative joint source-channel decoding algorithm for a SISO (soft-input soft-output) MAP (maximum a posteriori) module. The performance of transmission system has been evaluated in the PSNR, BER and iteration times. A very small degradation of the PSNR and an improvement in BER were compared to a system without joint source-channel decoding at the input of the receiver.ver.

  • PDF