• Title/Summary/Keyword: data complexity

Search Result 2,379, Processing Time 0.033 seconds

Complexity Reduction of MPEG-4 ER-BSAC Decoder Using Significance Tree Structure (중요도 트리 구조를 이용한 MPEG-4 ER-BSAC 디코더의 복잡도 개선)

  • Ahn, Young-Uk;Jung, Gyu-Heok;Kim, Gyu-Jin;Lee, In-Sung
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.355-356
    • /
    • 2006
  • MPEG-4 ER-BSAC decoder employes a full search method for maximum significance search and arithmetic decoding position search in spectral data decoding procedure. Then the search procedure have the most complexity. This paper proposes the new search method, the maximum significance tree structure, for the optimized implementation of BSAC decoder.

  • PDF

A Performance Improvement Study On Hierarchical Clustering (Centroid Linkage) Using A Priority Queue (Priority Queue 를 이용한 Hierarchical Clustering (Centroid Linkage) 성능 개선)

  • Jeon, Yongkweon;Yoon, Sungroh
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2010.11a
    • /
    • pp.1837-1838
    • /
    • 2010
  • 기존 hierarchical clustering 은 Time complexity 와 space complexity 가 Large data set 을 clustering 하기에는 적당하지 못하며 이것을 일반 PC 의 메모리 내에서 해결하는데 어려움이 있다. 따라서 본 연구에서는 이러한 어려움을 극복하기 위해 기존 Hierarchical clustering 중 Centroid Linkage 에 새로운 Algorithm 을 제안하여 보다 적은 메모리를 사용하고 빠르게 처리하는 방법을 제안하고자 한다.

Self-organized Learning in Complexity Growing of Radial Basis Function Networks

  • Arisariyawong, Somwang;Charoenseang, Siam
    • Proceedings of the IEEK Conference
    • /
    • 2002.07a
    • /
    • pp.30-33
    • /
    • 2002
  • To obtain good performance of radial basis function (RBF) neural networks, it needs very careful consideration in design. The selection of several parameters such as the number of centers and widths of the radial basis functions must be considered carefully since they critically affect the network's performance. We propose a learning algorithm for growing of complexity of RBF neural networks which is adapted automatically according to the complexity of tasks. The algorithm generates a new basis function based on the errors of network, the percentage of decreasing rate of errors and the nearest distance from input data to the center of hidden unit. The RBF's center is located at the point where the maximum of absolute interference error occurs in the input space. The width is calculated based on the standard deviation of distance between the center and inputs data. The steepest descent method is also applied for adjusting the weights, centers, and widths. To demonstrate the performance of the proposed algorithm, general problem of function estimation is evaluated. The results obtained from the simulation show that the proposed algorithm for RBF neural networks yields good performance in terms of convergence and accuracy compared with those obtained by conventional multilayer feedforward networks.

  • PDF

A Software Size Estimation Using Weighted FFP (가중치를 적용한 FFP 소프트웨어 규모 측정)

  • Park Juseok
    • Journal of Internet Computing and Services
    • /
    • v.6 no.2
    • /
    • pp.37-47
    • /
    • 2005
  • Most of the methods of estimating the size of software are based on the functions provided to costumers and in the process of granting the score to each function we consider the complexity during the process. The FFP technique has advantages applied to vast areas like data management. real-time system, algorithmic software, etc, but on the other hand, has disadvantage on estimating sizes for weights for necessary function elements. This paper proposes the estimating method for software size by considering the complexity of each function elements in full function point calculation method applied to a new developed project and maintenance projects. For this, based on function point by using surveyed data proved the validity of proposed method. The valid result. was that the function elements, the attributes used in size estimation of software, est mated better estimated sizes than in the case of other weights being applied.

  • PDF

Possibility of Chaotic Motion in the R&D Activities in Korea

  • Loh, Jeunghwee
    • Journal of Information Technology Applications and Management
    • /
    • v.21 no.3
    • /
    • pp.1-17
    • /
    • 2014
  • In this study, various characteristics of R&D related economic variables were studied to analyze complexity of science and technology activities in Korea, as reliance of R&D activities of the private sector is growing by the day. In comparison to other countries, this means that it is likely to be fluctuated by economic conditions. This complexity characteristic signifies that the result of science and technology activities can be greatly different from the anticipated results - depending on the influences from economic conditions and the results of science and technology activities which may be unpredictable. After reviewing the results of 17 variables related to science and technology characteristics of complex systems intended for time-series data - in the total R&D expenditure, and private R&D expenditure, numbers of SCI papers, the existence of chaotic characteristics were. using Lyapunov Exponent, Hurst Exponent, BDS test. This result reveals science and technology activity of the three most important components in Korea which are; heavy dependence on initial condition, the long term memory of time series, and non-linear structure. As stable R&D investment and result are needed in order to maintain steady development of Korea economy, the R&D structure should be less influenced by business cycles and more effective technology development policy for improving human resource development must be set in motion. And to minimize the risk of new technology, the construction of sophisticated technology forecasting system should take into account, for development of R&D system.

Design of an Expandable VLSI Rebound Sorter (확장형 VLSI 리바운드 정렬기의 설계)

  • Yun, Ji-Heon;Ahn, Byoung-Chul
    • The Transactions of the Korea Information Processing Society
    • /
    • v.2 no.3
    • /
    • pp.433-442
    • /
    • 1995
  • This paper presents an improved VLSI implementation of a parallel sorter to achieve O(Ν) time complexity. Many fast VLSI sort algorithms have been proposed for sorting N elements in O(log Ν) time. However, most such algorithms proposed have complex network structure without considering data input and output time. They are also very difficult to expand or to use in real applications. After analyzing the chip area and time complexity of several parallel sort algorithms with overlapping data input and output time, the most effective algorithm, the rebound sort algorithm, is implemented in VLSI with some improvements. To achieve O(Ν) time complexity, an improved rebound sorter is able to sort 8 16-bits records on a chip. And it is possible to sort more than 8 records by connecting chips in a chain vertically.

  • PDF

New low-complexity segmentation scheme for the partial transmit sequence technique for reducing the high PAPR value in OFDM systems

  • Jawhar, Yasir Amer;Ramli, Khairun Nidzam;Taher, Montadar Abas;Shah, Nor Shahida Mohd;Audah, Lukman;Ahmed, Mustafa Sami;Abbas, Thamer
    • ETRI Journal
    • /
    • v.40 no.6
    • /
    • pp.699-713
    • /
    • 2018
  • Orthogonal frequency division multiplexing (OFDM) has been the overwhelmingly prevalent choice for high-data-rate systems due to its superior advantages compared with other modulation techniques. In contrast, a high peak-to-average-power ratio (PAPR) is considered the fundamental obstacle in OFDM systems since it drives the system to suffer from in-band distortion and out-of-band radiation. The partial transmit sequence (PTS) technique is viewed as one of several strategies that have been suggested to diminish the high PAPR trend. The PTS relies upon dividing an input data sequence into a number of subblocks. Hence, three common types of the subblock segmentation methods have been adopted - interleaving (IL-PTS), adjacent (Ad-PTS), and pseudorandom (PR-PTS). In this study, a new type of subblock division scheme is proposed to improve the PAPR reduction capacity with a low computational complexity. The results indicate that the proposed scheme can enhance the PAPR reduction performance better than the IL-PTS and Ad-PTS schemes. Additionally, the computational complexity of the proposed scheme is lower than that of the PR-PTS and Ad-PTS schemes.

Optimal Path Planning Algorithm for Visiting Multiple Mission Points in Dynamic Environments (동적 변화 환경에서 다중 임무점 방문을 위한 최적 경로 계획 알고리즘)

  • Lee, Hohyeong;Chang, Woohyuk;Jang, Hwanchol
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.47 no.5
    • /
    • pp.379-387
    • /
    • 2019
  • The complexity of path planning for visiting multiple mission points is even larger than that of single pair path planning. Deciding a path for visiting n mission points requires conducting $n^2+n$ times of single pair path planning. We propose Multiple Mission $D^*$ Lite($MMD^*L$) which is an optimal path planning algorithm for visiting multiple mission points in dynamic environments. $MMD^*L$ reduces the complexity by reusing the computational data of preceding single pair path planning. Simulation results show that the complexity reduction is significant while its path optimality is not compromised.

IMI-Heap: An Implicit Double-Ended Priority Queue with Constant Insertion Amortized Time Complexity (IMI-힙: 상수 삽입 전이 시간 복잡도를 가진 묵시 양단 우선순위 큐)

  • Jung, Haejae
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.8 no.2
    • /
    • pp.29-34
    • /
    • 2019
  • Priority queues, one of the fundamental data structures, have been studied for a long time by computer scientists. This paper proposes an implicit double-ended priority queue, called IMI-heap, in which insert operation takes constant amortized time and each of removal operation of the minimum key or the maximum key takes O(logn) time. To the author's knowledge, all implicit double-ended priority queues that have been published, perform insert, removeMin and removeMax operations in O(logn) time each. So, the proposed IMI-heap is superior than the published heaps in terms of insertion time complexity.The abstract should concisely state what was done, how it was done, principal results, and their significance.

Data Dissemination in Wireless Sensor Networks with Instantly Decodable Network Coding

  • Gou, Liang;Zhang, Gengxin;Bian, Dongming;Zhang, Wei;Xie, Zhidong
    • Journal of Communications and Networks
    • /
    • v.18 no.5
    • /
    • pp.846-856
    • /
    • 2016
  • Wireless sensor networks (WSNs) are widely applied in monitoring and control of environment parameters. It is sometimes necessary to disseminate data through wireless links after they are deployed in order to adjust configuration parameters of sensors or distribute management commands and queries to sensors. Several approaches have been proposed recently for data dissemination in WSNs. However, none of these approaches achieves both high efficiency and low complexity simultaneously. To address this problem, cluster-tree based network architecture, which divides a WSN into hierarchies and clusters is proposed. Upon this architecture, data is delivered from base station to all sensors in clusters hierarchy by hierarchy. In each cluster, father broadcasts data to all his children with instantly decodable network coding (IDNC), and a novel scheme targeting to maximize total transmission gain (MTTG) is proposed. This scheme employs a new packet scheduling algorithm to select IDNC packets, which uses weight status feedback matrix (WSFM) directly. Analysis and simulation results indicate that the transmission efficiency approximate to the best existing approach maximum weight clique, but with much lower computational overhead. Hence, the energy efficiency achieves both in data transmission and processing.