• 제목/요약/키워드: complexity of algorithms

검색결과 976건 처리시간 0.024초

Low-Complexity Network Coding Algorithms for Energy Efficient Information Exchange

  • Wang, Yu;Henning, Ian D.
    • Journal of Communications and Networks
    • /
    • 제10권4호
    • /
    • pp.396-402
    • /
    • 2008
  • The use of network coding in wireless networks has been proposed in the literature for energy efficient broadcast. However, the decoding complexity of existing algorithms is too high for low-complexity devices. In this work we formalize the all-to-all information exchange problem and shows how to optimize the transmission scheme in terms of energy efficiency. Furthermore, we prove by construction that there exists O(1) -complexity network coding algorithms for grid networks which can achieve such optimality. We also present low-complexity heuristics for random. topology networks. Simulation results show that network coding algorithms outperforms forwarding algorithms in most cases.

Adaptive Frame Rate Up-Conversion Algorithms using Block Complexity Information

  • Lee, Kangjun
    • 한국멀티미디어학회논문지
    • /
    • 제21권8호
    • /
    • pp.813-820
    • /
    • 2018
  • This paper proposes new frame rate up-conversion algorithms. Adaptive motion estimation based on block complexity information are used to obtain more accurate motion vectors. Because the information on block complexity is extracted from the motion estimation prediction size from the original frame, additional computational complexity is not imparted. In experimental results, the proposed algorithms provide robust frame interpolation performance for whole test sequences. Also, the computational complexity of the proposed algorithm is reduced to a benchmark algorithm.

DISTRIBUTED ALGORITHMS SOLVING THE UPDATING PROBLEMS

  • Park, Jung-Ho;Park, Yoon-Young;Choi, Sung-Hee
    • Journal of applied mathematics & informatics
    • /
    • 제9권2호
    • /
    • pp.607-620
    • /
    • 2002
  • In this paper, we consider the updating problems to reconstruct the biconnected-components and to reconstruct the weighted shortest path in response to the topology change of the network. We propose two distributed algorithms. The first algorithm solves the updating problem that reconstructs the biconnected-components after the several processors and links are added and deleted. Its bit complexity is O((n'+a+d)log n'), its message complexity is O(n'+a+d), the ideal time complexity is O(n'), and the space complexity is O(e long n+e' log n'). The second algorithm solves the updating problem that reconstructs the weighted shortest path. Its message complexity and ideal-time complexity are $O(u^2+a+n')$ respectively.

Low-Complexity Sub-Pixel Motion Estimation Utilizing Shifting Matrix in Transform Domain

  • Ryu, Chul;Shin, Jae-Young;Park, Eun-Chan
    • Journal of Electrical Engineering and Technology
    • /
    • 제11권4호
    • /
    • pp.1020-1026
    • /
    • 2016
  • Motion estimation (ME) algorithms supporting quarter-pixel accuracy have been recently introduced to retain detailed motion information for high quality of video in the state-of-the-art video compression standard of H.264/AVC. Conventional sub-pixel ME algorithms in the spatial domain are faced with a common problem of computational complexity because of embedded interpolation schemes. This paper proposes a low-complexity sub-pixel motion estimation algorithm in the transform domain utilizing shifting matrix. Simulations are performed to compare the performances of spatial-domain ME algorithms and transform-domain ME algorithms in terms of peak signal-to-noise ratio (PSNR) and the number of bits per frame. Simulation results confirm that the transform-domain approach not only improves the video quality and the compression efficiency, but also remarkably alleviates the computational complexity, compared to the spatial-domain approach.

Nonlinear Quality Indices Based on a Novel Lempel-Ziv Complexity for Assessing Quality of Multi-Lead ECGs Collected in Real Time

  • Zhang, Yatao;Ma, Zhenguo;Dong, Wentao
    • Journal of Information Processing Systems
    • /
    • 제16권2호
    • /
    • pp.508-521
    • /
    • 2020
  • We compared a novel encoding Lempel-Ziv complexity (ELZC) with three common complexity algorithms i.e., approximate entropy (ApEn), sample entropy (SampEn), and classic Lempel-Ziv complexity (CLZC) so as to determine a satisfied complexity and its corresponding quality indices for assessing quality of multi-lead electrocardiogram (ECG). First, we calculated the aforementioned algorithms on six artificial time series in order to compare their performance in terms of discerning randomness and the inherent irregularity within time series. Then, for analyzing sensitivity of the algorithms to content level of different noises within the ECG, we investigated their change trend in five artificial synthetic noisy ECGs containing different noises at several signal noise ratios. Finally, three quality indices based on the ELZC of the multi-lead ECG were proposed to assess the quality of 862 real 12-lead ECGs from the MIT databases. The results showed the ELZC could discern randomness and the inherent irregularity within six artificial time series, and also reflect content level of different noises within five artificial synthetic ECGs. The results indicated the AUCs of three quality indices of the ELZC had statistical significance (>0.500). The ELZC and its corresponding three indices were more suitable for multi-lead ECG quality assessment than the other three algorithms.

WHAT CAN WE SAY ABOUT THE TIME COMPLEXITY OF ALGORITHMS \ulcorner

  • Park, Chin-Hong
    • Journal of applied mathematics & informatics
    • /
    • 제8권3호
    • /
    • pp.959-973
    • /
    • 2001
  • We shall discuss one of some techniques needed to analyze algorithms. It is called a big-O function technique. The measures of efficiency of an algorithm have two cases. One is the time used by a computer to solve the problem using this algorithm when the input values are of a specified size. The other one is the amount of computer memory required to implement the algorithm when the input values are of a specified size. Mainly, we will restrict our attention to time complexity. To figure out the Time Complexity in nonlinear problems of Numerical Analysis seems to be almost impossible.

이물질 탐지용 FMCW 레이더를 위한 저복잡도 초고해상도 알고리즘 (Low Complexity Super Resolution Algorithm for FOD FMCW Radar Systems)

  • 김봉석;김상동;이종훈
    • 대한임베디드공학회논문지
    • /
    • 제13권1호
    • /
    • pp.1-8
    • /
    • 2018
  • This paper proposes a low complexity super resolution algorithm for frequency modulated continuous wave (FMCW) radar systems for foreign object debris (FOD) detection. FOD radar has a requirement to detect foreign object in small units in a large area. However, The fast Fourier transform (FFT) method, which is most widely used in FMCW radar, has a disadvantage in that it can not distinguish between adjacent targets. Super resolution algorithms have a significantly higher resolution compared with the detection algorithm based on FFT. However, in the case of the large number of samples, the computational complexity of the super resolution algorithms is drastically high and thus super resolution algorithms are difficult to apply to real time systems. In order to overcome this disadvantage of super resolution algorithm, first, the proposed algorithm coarsely obtains the frequency of the beat signal by employing FFT. Instead of using all the samples of the beat signal, the number of samples is adjusted according to the frequency of the beat signal. By doing so, the proposed algorithm significantly reduces the computational complexity of multiple signal classifier (MUSIC) algorithm. Simulation results show that the proposed method achieves accurate location even though it has considerably lower complexity than the conventional super resolution algorithms.

IEEE 802.11n WLAN을 위한 FFT 프로세서의 하드웨어 복잡도 최적화에 대한 연구 (A Study on Optimization of Hardware Complexity of a FFT Processor for IEEE 802.11n WLAN)

  • 최락훈;박정준;임태민;이진용;김영록
    • 대한임베디드공학회논문지
    • /
    • 제6권4호
    • /
    • pp.243-248
    • /
    • 2011
  • A FFT/IFFT processor is the key component for orthogonal frequency division multiplexing (OFDM) systems based IEEE 802.11n wireless local area network (WLAN). There exists many radix algorithms according to the structure of butterfly as FFT sub-module, each has the pros and cons on hardware complexity. Here, mixed radix algorithms for 64 and 128 FFT/IFFT processors are proposed, which reduce hardware complexity by using mixture of radix-23 and radix-4 algorithms. The proposed algorithm finish calculation within 3.2${\mu}s$ in order to meet IEEE 802.11n standard requirements and it has less hardware complexity compared with conventional algorithms.

지역 복잡도 기반 방법 선택을 이용한 적응적 디인터레이싱 알고리듬 (Adaptive De-interlacing Algorithm using Method Selection based on Degree of Local Complexity)

  • 홍성민;박상준;정제창
    • 한국통신학회논문지
    • /
    • 제36권4C호
    • /
    • pp.217-225
    • /
    • 2011
  • 본 논문에서는 영상의 지역 특성별로 보간 방법을 적응적으로 선택하여 적용하는 효과적인 디인터레이싱 알고리듬을 제안한다. 기존의 알고리듬들의 경우 각기 다른 방법으로 방향성을 구하기 때문에 영상의 지역 특성별로 성능이 다르게 나오는 경우가 있다. 또한, FDD(Fine Directional De-interlacing) 알고리듬의 경우 PSNR(Peak Signal-to-Noise Ratio)은 다른 알고리듬들에 비해 높게 나오지만 계산량이 많다는 단점이 있다. 이를 보안하기 위해 본 논문에서는 여러 영상들에서 계산량은 적으면서 화질 성능은 뛰어난 LA(Line Average), MELA(Modified Edge-based Line Average), LCID(Low-Complexity Interpolation Method for De-interlacing) 알고리듬들 중 지역복잡도 (DoLC, Degree of Local Complexity)별로 효과적인 알고리듬을 학습하여 이를 이용하여 보간을 수행하는 디인터레이싱 방법을 제안한다. 실험 결과 제안하는 방법은 좋은 성능에 비해 계산량이 적은 LCID 알고리듬과 비슷한 계산량을 보이면서 객관적 화질이 우수한 FDD, MELA 알고리듬보다 PSNR로 대표되는 객관적 화질과 주관적 화질 측면에서 우수한 결과를 나타내는 것을 알 수 있다.

General Algorithms for Construction of Broadcast and Multicast Trees with Applications to Wireless Networks

  • Nguyen Gam D.
    • Journal of Communications and Networks
    • /
    • 제7권3호
    • /
    • pp.263-277
    • /
    • 2005
  • In this paper, we introduce algorithms for constructing broadcasting and multicasting trees. These algorithms are general because they may be used for tree cost functions that are of arbitrary form. Thus, essentially the same algorithmic procedures are used for different tree cost functions. We evaluate the effectiveness of the general algorithms by applying them to different cost functions that are often used to model wired and wireless net­works. Besides providing a unifying framework for dealing with many present and future tree-construction applications, these algorithms typically outperform some existing algorithms that are specifically designed for energy-aware wireless networks. These general algorithms perform well at the expense of higher computational complexity. They are centralized algorithms, requiring the full network information for tree construction. Thus, we also present variations of these general algorithms to yield other algorithms that have lower complexity and distributed implementation.