• Title/Summary/Keyword: computation-intensive

Search Result 108, Processing Time 0.022 seconds

An Empirical Analysis of Worldwide Cyberinfrastructure

  • Cho, Manhyung
    • Asian Journal of Innovation and Policy
    • /
    • v.4 no.3
    • /
    • pp.381-396
    • /
    • 2015
  • Cyberinfrastructure is a research infrastructure that provides an environment in which research communities can get access to distributed resources and collaborate at unprecedented levels of computation, storage, and network capacity. The Worldwide LHC Computing Grid (WLCG) is a global collaborative project of computing or data centers that enables access to scientific data generated by the Large Hadron Collider (LHC) experiments at CERN. This case study analyzes the WLCG as a model of cyberinfrastructure in research collaboration. WLCG provides a useful case of how cyberinfrastructure can work in providing an infrastructure for collaborative researches under data-intensive paradigm. Cyberinfrastructure plays the critical role of facilitating collaboration of diverse and widely separated communities of researchers. Data-intensive science requires new strategies for research support and significant development of cyberinfrastructure. The sustainability of WLCG depends on the resources of partner organizations and virtual organizations at international levels, essential for research collaboration.

A design of dual AC-3 and MPEG-2 audio decoder (AC-3와 MPEG-2 오디오 공용 복호화기의 설계)

  • Ko, Woo-Suk;Yoo, Sun-Kook;Park, Sung-Wook;Jung, Nam-Hoon;Kim, Joon-Seok;Lee, Keun-Sup;Youn, Dae-Hee
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.23 no.6
    • /
    • pp.1433-1442
    • /
    • 1998
  • The thesis presents a dual audio decoder which can decode both AC-3 and MPEG-2 bitstream. The MPEG-2 synthesis processi s optimized via FFT to establish the common data path with AC-'3s. A dual audio decoder consists of a DSP core which performs the control-intensive part of each algorithm and a common synthesis filter which perfomrs the computation-intensive part. All the components of the dual audio decoder have been described in VHDL and simulated with a SYNOPSYS tool. The software modeling of the DSP core was used for functional validation. After being synthesized using 0.6 .mu.m-3ML technology standard cell, the dual audio decoder was simulated at gate-level with a COMPASS tool for hardware validation.

  • PDF

PhysioCover: Recovering the Missing Values in Physiological Data of Intensive Care Units

  • Kim, Sun-Hee;Yang, Hyung-Jeong;Kim, Soo-Hyung;Lee, Guee-Sang
    • International Journal of Contents
    • /
    • v.10 no.2
    • /
    • pp.47-58
    • /
    • 2014
  • Physiological signals provide important clues in the diagnosis and prediction of disease. Analyzing these signals is important in health and medicine. In particular, data preprocessing for physiological signal analysis is a vital issue because missing values, noise, and outliers may degrade the analysis performance. In this paper, we propose PhysioCover, a system that can recover missing values of physiological signals that were monitored in real time. PhysioCover integrates a gradual method and EM-based Principle Component Analysis (PCA). This approach can (1) more readily recover long- and short-term missing data than existing methods, such as traditional EM-based PCA, linear interpolation, 5-average and Missing Value Singular Value Decomposition (MSVD), (2) more effectively detect hidden variables than PCA and Independent component analysis (ICA), and (3) offer fast computation time through real-time processing. Experimental results with the physiological data of an intensive care unit show that the proposed method assigns more accurate missing values than previous methods.

Pose Estimation of a Cylindrical Object Using Genetic Algorithm (유전자 알고리즘을 이용한 원기둥형 물체의 자세 추정 방법)

  • Jeong Kyuwon
    • Transactions of the Korean Society of Machine Tool Engineers
    • /
    • v.14 no.3
    • /
    • pp.54-59
    • /
    • 2005
  • The cylindrical object are widely used as mechanical parts in the manufacturing process. In order to handling those objects using a robot or an automated machine automatically, the pose of the object must be known. The pose can be described by two rotation angles; one angle about the x axis and the other about the y axis. In the many previous researches these angles were obtained by the computationally intensive algorithm, that is, fitting the data as a polynomial and doing pseudo inverse. So that, this method required high performance microprocessor. In this paper in order to avoid complex computation, a new method based on a genetic algorithm is proposed and analyzed through a series of simulations. This algorithm utilized the geometry of the cylindrical shape. The simulation results show that this method find the pose angles very well In most cases, but the computation time is randomly changed because the genetic algorithm is basically one of the random search method.

Data-Compression-Based Resource Management in Cloud Computing for Biology and Medicine

  • Zhu, Changming
    • Journal of Computing Science and Engineering
    • /
    • v.10 no.1
    • /
    • pp.21-31
    • /
    • 2016
  • With the application and development of biomedical techniques such as next-generation sequencing, mass spectrometry, and medical imaging, the amount of biomedical data have been growing explosively. In terms of processing such data, we face the problems surrounding big data, highly intensive computation, and high dimensionality data. Fortunately, cloud computing represents significant advantages of resource allocation, data storage, computation, and sharing and offers a solution to solve big data problems of biomedical research. In order to improve the efficiency of resource management in cloud computing, this paper proposes a clustering method and adopts Radial Basis Function in order to compress comprehensive data sets found in biology and medicine in high quality, and stores these data with resource management in cloud computing. Experiments have validated that with such a data-compression-based resource management in cloud computing, one can store large data sets from biology and medicine in fewer capacities. Furthermore, with reverse operation of the Radial Basis Function, these compressed data can be reconstructed with high accuracy.

Marginal Likelihoods for Bayesian Poisson Regression Models

  • Kim, Hyun-Joong;Balgobin Nandram;Kim, Seong-Jun;Choi, Il-Su;Ahn, Yun-Kee;Kim, Chul-Eung
    • Communications for Statistical Applications and Methods
    • /
    • v.11 no.2
    • /
    • pp.381-397
    • /
    • 2004
  • The marginal likelihood has become an important tool for model selection in Bayesian analysis because it can be used to rank the models. We discuss the marginal likelihood for Poisson regression models that are potentially useful in small area estimation. Computation in these models is intensive and it requires an implementation of Markov chain Monte Carlo (MCMC) methods. Using importance sampling and multivariate density estimation, we demonstrate a computation of the marginal likelihood through an output analysis from an MCMC sampler.

A Throughput Computation Method for Throughput Driven Floorplan (처리량 기반 평면계획을 위한 처리량 계산 방법)

  • Kang, Min-Sung;Rim, Chong-Suck
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.44 no.12
    • /
    • pp.18-24
    • /
    • 2007
  • As VLSI technology scales to nano-meter order, relatively increasing global wire-delay has added complexity to system design. Global wire-delay could be reduced by inserting pipeline-elements onto wire but it should be coupled with LIP(Latency Intensive Protocol) to have correct system timing. This combination however, drops the throughput although it ensures system functionality. In this paper, we propose a computation method useful for minimizing throughput deterioration when pipeline-elements are inserted to reduce global wire-delay. We apply this method while placing blocks in the floorplanning stage. When the necessary for this computation is reflected on the floorplanning cost function, the throughput increases by 16.97% on the average when compared with the floorplanning that uses the conventional heuristic throughput-evaluation-method.

Energy-Efficient Resource Allocation for Application Including Dependent Tasks in Mobile Edge Computing

  • Li, Yang;Xu, Gaochao;Ge, Jiaqi;Liu, Peng;Fu, Xiaodong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.6
    • /
    • pp.2422-2443
    • /
    • 2020
  • This paper studies a single-user Mobile Edge Computing (MEC) system where mobile device (MD) includes an application consisting of multiple computation components or tasks with dependencies. MD can offload part of each computation-intensive latency-sensitive task to the AP integrated with MEC server. In order to accomplish the application faultlessly, we calculate out the optimal task offloading strategy in a time-division manner for a predetermined execution order under the constraints of limited computation and communication resources. The problem is formulated as an optimization problem that can minimize the energy consumption of mobile device while satisfying the constraints of computation tasks and mobile device resources. The optimization problem is equivalently transformed into solving a nonlinear equation with a linear inequality constraint by leveraging the Lagrange Multiplier method. And the proposed dual Bi-Section Search algorithm Bi-JOTD can efficiently solve the nonlinear equation. In the outer Bi-Section Search, the proposed algorithm searches for the optimal Lagrangian multiplier variable between the lower and upper boundaries. The inner Bi-Section Search achieves the Lagrangian multiplier vector corresponding to a given variable receiving from the outer layer. Numerical results demonstrate that the proposed algorithm has significant performance improvement than other baselines. The novel scheme not only reduces the difficulty of problem solving, but also obtains less energy consumption and better performance.

Modified 3-step Search Motion Estimation Algorithm for Effective Early Termination (효과적인 조기 중단 기법을 위한 변형된 3단계 탐색 움직임 추정 알고리즘)

  • Yang, Hyeon-Cheol;Lee, Seong-Soo
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.47 no.7
    • /
    • pp.70-77
    • /
    • 2010
  • Motion estimation occupies most of the required computation in video compression, and many fast search algorithms were propsoed to reduce huge computation. SAD (sum-of-absolute difference) calculation is the most computation-intensive process in the motion estimation. Early termination is widely used in SAD calculation, where SAD calculation is terminated and it proceeds to next search position if partial SAD during SAD calculation exceeds current minimum SAD. In this paper, we proposed a modified 3-step search algorithm for effective early termination where only search order of search positions are adaptive rearranged. Simulation results show that the proposed motion estimation algorithm reduces computation by 17~30% over conventional 3-step search algorithm without extra computation, while maintaining same performance.

Efficient Parallel TLD on CPU-GPU Platform for Real-Time Tracking

  • Chen, Zhaoyun;Huang, Dafei;Luo, Lei;Wen, Mei;Zhang, Chunyuan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.1
    • /
    • pp.201-220
    • /
    • 2020
  • Trackers, especially long-term (LT) trackers, now have a more complex structure and more intensive computation for nowadays' endless pursuit of high accuracy and robustness. However, computing efficiency of LT trackers cannot meet the real-time requirement in various real application scenarios. Considering heterogeneous CPU-GPU platforms have been more popular than ever, it is a challenge to exploit the computing capacity of heterogeneous platform to improve the efficiency of LT trackers for real-time requirement. This paper focuses on TLD, which is the first LT tracking framework, and proposes an efficient parallel implementation based on OpenCL. In this paper, we firstly make an analysis of the TLD tracker and then optimize the computing intensive kernels, including Fern Feature Extraction, Fern Classification, NCC Calculation, Overlaps Calculation, Positive and Negative Samples Extraction. Experimental results demonstrate that our efficient parallel TLD tracker outperforms the original TLD, achieving the 3.92 speedup on CPU and GPU. Moreover, the parallel TLD tracker can run 52.9 frames per second and meet the real-time requirement.