• Title/Summary/Keyword: complexity of algorithms

Search Result 976, Processing Time 0.028 seconds

Low-Complexity Network Coding Algorithms for Energy Efficient Information Exchange

  • Wang, Yu;Henning, Ian D.
    • Journal of Communications and Networks
    • /
    • v.10 no.4
    • /
    • pp.396-402
    • /
    • 2008
  • The use of network coding in wireless networks has been proposed in the literature for energy efficient broadcast. However, the decoding complexity of existing algorithms is too high for low-complexity devices. In this work we formalize the all-to-all information exchange problem and shows how to optimize the transmission scheme in terms of energy efficiency. Furthermore, we prove by construction that there exists O(1) -complexity network coding algorithms for grid networks which can achieve such optimality. We also present low-complexity heuristics for random. topology networks. Simulation results show that network coding algorithms outperforms forwarding algorithms in most cases.

Adaptive Frame Rate Up-Conversion Algorithms using Block Complexity Information

  • Lee, Kangjun
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.8
    • /
    • pp.813-820
    • /
    • 2018
  • This paper proposes new frame rate up-conversion algorithms. Adaptive motion estimation based on block complexity information are used to obtain more accurate motion vectors. Because the information on block complexity is extracted from the motion estimation prediction size from the original frame, additional computational complexity is not imparted. In experimental results, the proposed algorithms provide robust frame interpolation performance for whole test sequences. Also, the computational complexity of the proposed algorithm is reduced to a benchmark algorithm.

DISTRIBUTED ALGORITHMS SOLVING THE UPDATING PROBLEMS

  • Park, Jung-Ho;Park, Yoon-Young;Choi, Sung-Hee
    • Journal of applied mathematics & informatics
    • /
    • v.9 no.2
    • /
    • pp.607-620
    • /
    • 2002
  • In this paper, we consider the updating problems to reconstruct the biconnected-components and to reconstruct the weighted shortest path in response to the topology change of the network. We propose two distributed algorithms. The first algorithm solves the updating problem that reconstructs the biconnected-components after the several processors and links are added and deleted. Its bit complexity is O((n'+a+d)log n'), its message complexity is O(n'+a+d), the ideal time complexity is O(n'), and the space complexity is O(e long n+e' log n'). The second algorithm solves the updating problem that reconstructs the weighted shortest path. Its message complexity and ideal-time complexity are $O(u^2+a+n')$ respectively.

Low-Complexity Sub-Pixel Motion Estimation Utilizing Shifting Matrix in Transform Domain

  • Ryu, Chul;Shin, Jae-Young;Park, Eun-Chan
    • Journal of Electrical Engineering and Technology
    • /
    • v.11 no.4
    • /
    • pp.1020-1026
    • /
    • 2016
  • Motion estimation (ME) algorithms supporting quarter-pixel accuracy have been recently introduced to retain detailed motion information for high quality of video in the state-of-the-art video compression standard of H.264/AVC. Conventional sub-pixel ME algorithms in the spatial domain are faced with a common problem of computational complexity because of embedded interpolation schemes. This paper proposes a low-complexity sub-pixel motion estimation algorithm in the transform domain utilizing shifting matrix. Simulations are performed to compare the performances of spatial-domain ME algorithms and transform-domain ME algorithms in terms of peak signal-to-noise ratio (PSNR) and the number of bits per frame. Simulation results confirm that the transform-domain approach not only improves the video quality and the compression efficiency, but also remarkably alleviates the computational complexity, compared to the spatial-domain approach.

Nonlinear Quality Indices Based on a Novel Lempel-Ziv Complexity for Assessing Quality of Multi-Lead ECGs Collected in Real Time

  • Zhang, Yatao;Ma, Zhenguo;Dong, Wentao
    • Journal of Information Processing Systems
    • /
    • v.16 no.2
    • /
    • pp.508-521
    • /
    • 2020
  • We compared a novel encoding Lempel-Ziv complexity (ELZC) with three common complexity algorithms i.e., approximate entropy (ApEn), sample entropy (SampEn), and classic Lempel-Ziv complexity (CLZC) so as to determine a satisfied complexity and its corresponding quality indices for assessing quality of multi-lead electrocardiogram (ECG). First, we calculated the aforementioned algorithms on six artificial time series in order to compare their performance in terms of discerning randomness and the inherent irregularity within time series. Then, for analyzing sensitivity of the algorithms to content level of different noises within the ECG, we investigated their change trend in five artificial synthetic noisy ECGs containing different noises at several signal noise ratios. Finally, three quality indices based on the ELZC of the multi-lead ECG were proposed to assess the quality of 862 real 12-lead ECGs from the MIT databases. The results showed the ELZC could discern randomness and the inherent irregularity within six artificial time series, and also reflect content level of different noises within five artificial synthetic ECGs. The results indicated the AUCs of three quality indices of the ELZC had statistical significance (>0.500). The ELZC and its corresponding three indices were more suitable for multi-lead ECG quality assessment than the other three algorithms.

WHAT CAN WE SAY ABOUT THE TIME COMPLEXITY OF ALGORITHMS \ulcorner

  • Park, Chin-Hong
    • Journal of applied mathematics & informatics
    • /
    • v.8 no.3
    • /
    • pp.959-973
    • /
    • 2001
  • We shall discuss one of some techniques needed to analyze algorithms. It is called a big-O function technique. The measures of efficiency of an algorithm have two cases. One is the time used by a computer to solve the problem using this algorithm when the input values are of a specified size. The other one is the amount of computer memory required to implement the algorithm when the input values are of a specified size. Mainly, we will restrict our attention to time complexity. To figure out the Time Complexity in nonlinear problems of Numerical Analysis seems to be almost impossible.

Low Complexity Super Resolution Algorithm for FOD FMCW Radar Systems (이물질 탐지용 FMCW 레이더를 위한 저복잡도 초고해상도 알고리즘)

  • Kim, Bong-seok;Kim, Sangdong;Lee, Jonghun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.13 no.1
    • /
    • pp.1-8
    • /
    • 2018
  • This paper proposes a low complexity super resolution algorithm for frequency modulated continuous wave (FMCW) radar systems for foreign object debris (FOD) detection. FOD radar has a requirement to detect foreign object in small units in a large area. However, The fast Fourier transform (FFT) method, which is most widely used in FMCW radar, has a disadvantage in that it can not distinguish between adjacent targets. Super resolution algorithms have a significantly higher resolution compared with the detection algorithm based on FFT. However, in the case of the large number of samples, the computational complexity of the super resolution algorithms is drastically high and thus super resolution algorithms are difficult to apply to real time systems. In order to overcome this disadvantage of super resolution algorithm, first, the proposed algorithm coarsely obtains the frequency of the beat signal by employing FFT. Instead of using all the samples of the beat signal, the number of samples is adjusted according to the frequency of the beat signal. By doing so, the proposed algorithm significantly reduces the computational complexity of multiple signal classifier (MUSIC) algorithm. Simulation results show that the proposed method achieves accurate location even though it has considerably lower complexity than the conventional super resolution algorithms.

A Study on Optimization of Hardware Complexity of a FFT Processor for IEEE 802.11n WLAN (IEEE 802.11n WLAN을 위한 FFT 프로세서의 하드웨어 복잡도 최적화에 대한 연구)

  • Choi, Rakhun;Park, Jungjun;Lim, Taemin;Lee, Jinyong;Kim, Younglok
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.6 no.4
    • /
    • pp.243-248
    • /
    • 2011
  • A FFT/IFFT processor is the key component for orthogonal frequency division multiplexing (OFDM) systems based IEEE 802.11n wireless local area network (WLAN). There exists many radix algorithms according to the structure of butterfly as FFT sub-module, each has the pros and cons on hardware complexity. Here, mixed radix algorithms for 64 and 128 FFT/IFFT processors are proposed, which reduce hardware complexity by using mixture of radix-23 and radix-4 algorithms. The proposed algorithm finish calculation within 3.2${\mu}s$ in order to meet IEEE 802.11n standard requirements and it has less hardware complexity compared with conventional algorithms.

Adaptive De-interlacing Algorithm using Method Selection based on Degree of Local Complexity (지역 복잡도 기반 방법 선택을 이용한 적응적 디인터레이싱 알고리듬)

  • Hong, Sung-Min;Park, Sang-Jun;Jeong, Je-Chang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.4C
    • /
    • pp.217-225
    • /
    • 2011
  • In this paper, we propose an adaptive de-interlacing algorithm that is based on the degree of local complexity. The conventional intra field de-interlacing algorithms show the different performance according to the ways which find the edge direction. Furthermore, FDD (Fine Directional De-interlacing) algorithm has the better performance than other algorithms but the computational complexity of FDD algorithm is too high. In order to alleviate these problems, the proposed algorithm selects the most efficient de-interacing algorithm among LA (Line Average), MELA (Modified Edge-based Line Average), and LCID (Low-Complexity Interpolation Method for De-interlacing) algorithms which have low complexity and good performance. The proposed algorithm is trained by the DoLC (Degree of Local Complexity) for selection of the algorithms mentioned above. Simulation results show that the proposed algorithm not only has the low complexity but also performs better objective and subjective image quality performances compared with the conventional intra-field methods.

General Algorithms for Construction of Broadcast and Multicast Trees with Applications to Wireless Networks

  • Nguyen Gam D.
    • Journal of Communications and Networks
    • /
    • v.7 no.3
    • /
    • pp.263-277
    • /
    • 2005
  • In this paper, we introduce algorithms for constructing broadcasting and multicasting trees. These algorithms are general because they may be used for tree cost functions that are of arbitrary form. Thus, essentially the same algorithmic procedures are used for different tree cost functions. We evaluate the effectiveness of the general algorithms by applying them to different cost functions that are often used to model wired and wireless net­works. Besides providing a unifying framework for dealing with many present and future tree-construction applications, these algorithms typically outperform some existing algorithms that are specifically designed for energy-aware wireless networks. These general algorithms perform well at the expense of higher computational complexity. They are centralized algorithms, requiring the full network information for tree construction. Thus, we also present variations of these general algorithms to yield other algorithms that have lower complexity and distributed implementation.