• Title/Summary/Keyword: computation complexity

Search Result 607, Processing Time 0.024 seconds

Fast Motion Estimation for Variable Motion Block Size in H.264 Standard (H.264 표준의 가변 움직임 블록을 위한 고속 움직임 탐색 기법)

  • 최웅일;전병우
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.6
    • /
    • pp.209-220
    • /
    • 2004
  • The main feature of H.264 standard against conventional video standards is the high coding efficiency and the network friendliness. In spite of these outstanding features, it is not easy to implement H.264 codec as a real-time system due to its high requirement of memory bandwidth and intensive computation. Although the variable block size motion compensation using multiple reference frames is one of the key coding tools to bring about its main performance gain, it demands substantial computational complexity due to SAD (Sum of Absolute Difference) calculation among all possible combinations of coding modes to find the best motion vector. For speedup of motion estimation process, therefore, this paper proposes fast algorithms for both integer-pel and fractional-pel motion search. Since many conventional fast integer-pel motion estimation algorithms are not suitable for H.264 having variable motion block sizes, we propose the motion field adaptive search using the hierarchical block structure based on the diamond search applicable to variable motion block sizes. Besides, we also propose fast fractional-pel motion search using small diamond search centered by predictive motion vector based on statistical characteristic of motion vector.

Performance Evaluation of Output Queueing ATM Switch with Finite Buffer Using Stochastic Activity Networks (SAN을 이용한 제한된 버퍼 크기를 갖는 출력큐잉 ATM 스위치 성능평가)

  • Jang, Kyung-Soo;Shin, Ho-Jin;Shin, Dong-Ryeol
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.8
    • /
    • pp.2484-2496
    • /
    • 2000
  • High speed switches have been developing to interconnect a large number of nodes. It is important to analyze the switch performance under various conditions to satisfy the requirements. Queueing analysis, in general, has the intrinsic problem of large state space dimension and complex computation. In fact, The petri net is a graphical and mathematical model. It is suitable for various applications, in particular, manufacturing systems. It can deal with parallelism, concurrence, deadlock avoidance, and asynchronism. Currently it has been applied to the performance of computer networks and protocol verifications. This paper presents a framework for modeling and analyzing ATM switch using stochastic activity networks (SANs). In this paper, we provide the ATM switch model using SANs to extend easily and an approximate analysis method to apply A TM switch models, which significantly reduce the complexity of the model solution. Cell arrival process in output-buffered Queueing A TM switch with finite buffer is modeled as Markov Modulated Poisson Process (MMPP), which is able to accurately represent real traffic and capture the characteristics of bursty traffic. We analyze the performance of the switch in terms of cell-loss ratio (CLR), mean Queue length and mean delay time. We show that the SAN model is very useful in A TM switch model in that the gates have the capability of implementing of scheduling algorithm.

  • PDF

Image Retrieval Using Multiresoluton Color and Texture Features in Wavelet Transform Domain (웨이브릿 변환 영역의 칼라 및 질감 특징을 이용한 영상검색)

  • Chun Young-Deok;Sung Joong-Ki;Kim Nam-Chul
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.1 s.307
    • /
    • pp.55-66
    • /
    • 2006
  • We propose a progressive image retrieval method based on an efficient combination of multiresolution color and torture features in wavelet transform domain. As a color feature, color autocorrelogram of the hue and saturation components is chosen. As texture features, BDIP and BVLC moments of the value component are chosen. For the selected features, we obtain multiresolution feature vectors which are extracted from all decomposition levels in wavelet domain. The multiresolution feature vectors of the color and texture features are efficiently combined by the normalization depending on their dimensions and standard deviation vector, respectively, vector components of the features are efficiently quantized in consideration of their storage space, and computational complexity in similarity computation is reduced by using progressive retrieval strategy. Experimental results show that the proposed method yields average $15\%$ better performance in precision vs. recall and average 0.2 in ANMRR than the methods using color histogram color autocorrelogram SCD, CSD, wavelet moments, EHD, BDIP and BVLC moments, and combination of color histogram and wavelet moments, respectively. Specially, the proposed method shows an excellent performance over the other methods in image DBs contained images of various resolutions.

Low-Cost Elliptic Curve Cryptography Processor Based On Multi-Segment Multiplication (멀티 세그먼트 곱셈 기반 저비용 타원곡선 암호 프로세서)

  • LEE Dong-Ho
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.42 no.8 s.338
    • /
    • pp.15-26
    • /
    • 2005
  • In this paper, we propose an efficient $GF(2^m)$ multi-segment multiplier architecture and study its application to elliptic curve cryptography processors. The multi-segment based ECC datapath has a very small combinational multiplier to compute partial products, most of its internal data buses are word-sized, and it has only a single m bit multiplexer and a single m bit register. Hence, the resource requirements of the proposed ECC datapath can be minimized as the segment number increases and word-size is decreased. Hence, as compared to the ECC processor based on digit-serial multiplication, the proposed ECC datapath is more efficient in resource usage. The resource requirement of ECC Processor implementation depends not only on the number of basic hardware components but also on the complexity of interconnection among them. To show the realistic area efficiency of proposed ECC processors, we implemented both the ECC processors based on the proposed multi-segment multiplication and digit serial multiplication and compared their FPGA resource usages. The experimental results show that the Proposed multi-segment multiplication method allows to implement ECC coprocessors, requiring about half of FPGA resources as compared to digit serial multiplication.

Integration of Condensation and Mean-shift algorithms for real-time object tracking (실시간 객체 추적을 위한 Condensation 알고리즘과 Mean-shift 알고리즘의 결합)

  • Cho Sang-Hyun;Kang Hang-Bong
    • The KIPS Transactions:PartB
    • /
    • v.12B no.3 s.99
    • /
    • pp.273-282
    • /
    • 2005
  • Real-time Object tracking is an important field in developing vision applications such as surveillance systems and vision based navigation. mean-shift algerian and Condensation algorithm are widely used in robust object tracking systems. Since the mean-shift algorithm is easy to implement and is effective in object tracking computation, it is widely used, especially in real-time tracking systems. One of the drawbacks is that it always converges to a local maximum which may not be a global maximum. Therefore, in a cluttered environment, the Mean-shift algorithm does not perform well. On the other hand, since it uses multiple hypotheses, the Condensation algorithm is useful in tracking in a cluttered background. Since it requires a complex object model and many hypotheses, it contains a high computational complexity. Therefore, it is not easy to apply a Condensation algorithm in real-time systems. In this paper, by combining the merits of the Condensation algorithm and the mean-shift algorithm we propose a new model which is suitable for real-time tracking. Although it uses only a few hypotheses, the proposed method use a high-likelihood hypotheses using mean-shift algorithm. As a result, we can obtain a better result than either the result produced by the Condensation algorithm or the result produced by the mean-shift algorithm.

Moving Object Contour Detection Using Spatio-Temporal Edge with a Fixed Camera (고정 카메라에서의 시공간적 경계 정보를 이용한 이동 객체 윤곽선 검출 방법)

  • Kwak, Jae-Ho;Kim, Whoi-Yul
    • Journal of Broadcast Engineering
    • /
    • v.15 no.4
    • /
    • pp.474-486
    • /
    • 2010
  • In this paper, we propose a new method for detection moving object contour using spatial and temporal edge. In general, contour pixels of the moving object are likely present around pixels with high gradient value along the time axis and the spatial axis. Therefore, we can detect the contour of the moving objects by finding pixels which have high gradient value in the time axis and spatial axis. In this paper, we introduce a new computation method, termed as temporal edge, to compute an gradient value along the time axis for any pixel on an image. The temporal edge can be computed using two input gray images at time t and t-2 using the Sobel operator. Temporal edge is utilized to detect a candidate region of the moving object contour and then the detected candidate region is used to extract spatial edge information. The final contour of the moving object is detected using the combination of these two edge information, which are temporal edge and spatial edge, and then the post processing such as a morphological operation and a background edge removing procedure are applied to remove noise regions. The complexity of the proposed method is very low because it dose not use any background scene and high complex operation, therefore it can be applied to real-time applications. Experimental results show that the proposed method outperforms the conventional contour extraction methods in term of processing effort and a ghost effect which is occurred in the case of entropy method.

Adaptive Discrete Wavelet Transform Based on Block Energy for JPEG2000 Still Images (JPEG2000 정지영상을 위한 블록 에너지 기반 적응적 이산 웨이블릿 변환)

  • Kim, Dae-Won
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.8 no.1
    • /
    • pp.22-31
    • /
    • 2007
  • The proposed algorithm in this paper is based on the wavelet decomposition and the energy computation of composed blocks so the amount of calculation and complexity is minimized by adaptively replacing the DWT coefficients and managing the resources effectively. We are now living in the world of a lot. of multimedia applications for many digital electric appliances and mobile devices. Among so many multimedia applications, the digital image compression is very important technology for digital cameras to store and transmit digital images to other sites and JPEG2000 is one of the cutting edge technology to compress still images efficiently. The digital cm technology is mainly using the digital image compression features so that those images could be efficiently saved locally and transferred to other sites without any losses. JPEG2000 standard is applicable for processing the digital images usefully to keep, send and receive through wired and/or wireless networks. The discrete wavelet transform (DWT) is one of the main differences to the previous digital image compression standard such as JPEG, performing the DWT to the entire image rather than splitting into many blocks. Several digital images m tested with this method and restored to compare to the results of conventional DWT which shows that the proposed algorithm get the better result without any significant degradation in terms of MSE & PSNR and the number of zero coefficients when the energy based adaptive DWT is applied.

  • PDF

Improvement of Address Pointer Assignment in DSP Code Generation (DSP용 코드 생성에서 주소 포인터 할당 성능 향상 기법)

  • Lee, Hee-Jin;Lee, Jong-Yeol
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.1
    • /
    • pp.37-47
    • /
    • 2008
  • Exploitation of address generation units which are typically provided in DSPs plays an important role in DSP code generation since that perform fast address computation in parallel to the central data path. Offset assignment is optimization of memory layout for program variables by taking advantage of the capabilities of address generation units, consists of memory layout generation and address pointer assignment steps. In this paper, we propose an effective address pointer assignment method to minimize the number of address calculation instructions in DSP code generation. The proposed approach reduces the time complexity of a conventional address pointer assignment algorithm with fixed memory layouts by using minimum cost-nodes breaking. In order to contract memory size and processing time, we employ a powerful pruning technique. Moreover our proposed approach improves the initial solution iteratively by changing the memory layout for each iteration because the memory layout affects the result of the address pointer assignment algorithm. We applied the proposed approach to about 3,000 sequences of the OffsetStone benchmarks to demonstrate the effectiveness of the our approach. Experimental results with benchmarks show an average improvement of 25.9% in the address codes over previous works.

A Comparison Study with the Vatiation of Isocenter and Collimator in Stereotactic Radiosurgery (방사선 수술시 Isocenter, 콜리메이터 변수에 따른 선량 분포 비교연구)

  • 오승종;박정훈;곽철은;이형구;최보영;이태규;김문찬;서태석
    • Progress in Medical Physics
    • /
    • v.13 no.3
    • /
    • pp.129-134
    • /
    • 2002
  • The radiosurgery is planned that prescribed dose was irradiated to tumor for obtaining expected remedial value in stereotactic radiosurgery. The planning for many irregular tumor shape requires long computation time and skilled planners. Due to the rapid development in computer power recently, many optimization methods using computer has been proposed, although the practical method is still trial and error type of plan. In this study, many beam variables were considered and many tumor shapes were assumed cylinderical ideal models. Then, beam variables that covered the target within 50% isodose curve were searched, the result was compared and analysed. The beam variables considered were isocenter separation distance, number of isocenters and collimator size. Dose distributions obtained with these variables were analysed by dose volume histogram(DVH) and dose profile at orthogonal plane. According to the results compared, the use of more isocenters than specified isocenter dosen't improve DVH and dose profile but only increases complexity of plan. The best result of DVH and dose profile are obtainedwhen isocenter separation was 1.0-1.2 in using same number of isocenter.

  • PDF

A Sequential Estimation Algorithm for TDOA/FDOA Extraction for VHF Communication Signals (VHF 대역 통신 신호에서 TDOA/FDOA 정보 추출을 위한 순차 추정 알고리즘)

  • Kim, Dong-Gyu;Kim, Yong-Hee;Park, Jin-Oh;Lee, Moon Seok;Park, Young-Mi;Kim, Hyoung-Nam
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.7
    • /
    • pp.60-68
    • /
    • 2014
  • In modern electronic warfare systems, a demand on the more accurate estimation method based on TDOA and FDOA has been increased. TDOA/FDOA localization consists of two-stage procedures; the extraction of information from signals, and the estimation of emitter location. CAF(complex ambiguity function) is known as a basic method in the extraction stage. However, when we extract TDOA and FDOA information from VHF(very high frequency) communication signals, conventional CAF algorithms may not work within a permitted time because of much computation. Therefore, in this paper, an improved sequential estimation algorithm based on CAF is proposed for effective calculation of extracting TDOA and FDOA estimates in terms of computational complexity. The proposed method is compared with the conventional CAF-based algorithms through simulation. In addition, we derive the optimal performance based on the CRLB(Cramer-Lao lower bound) to check the extraction performance of the proposed method.