• 제목/요약/키워드: running average

검색결과 279건 처리시간 0.027초

A Study on Filtering Techniques for Dynamic Analysis of Data Races in Multi-threaded Programs

  • Ha, Ok-Kyoon;Yoo, Hongseok
    • 한국컴퓨터정보학회논문지
    • /
    • 제22권11호
    • /
    • pp.1-7
    • /
    • 2017
  • In this paper, we introduce three monitoring filtering techniques which reduce the overheads of dynamic data race detection. It is well known that detecting data races dynamically in multi-threaded programs is quite hard and troublesome task, because the dynamic detection techniques need to monitor all execution of a multi-threaded program and to analyse every conflicting memory and thread operations in the program. Thus, the main drawback of the dynamic analysis for detecting data races is the heavy additional time and space overheads for running the program. For the practicality, we also empirically compare the efficiency of three monitoring filtering techniques. The results using OpenMP benchmarks show that the filtering techniques are practical for dynamic data race detection, since they reduce the average runtime overhead to under 10% of that of the pure detection.

CUDA를 이용한 Particle Swarm Optimization 구현 (Implementation of Particle Swarm Optimization Method Using CUDA)

  • 김조환;김은수;김종욱
    • 전기학회논문지
    • /
    • 제58권5호
    • /
    • pp.1019-1024
    • /
    • 2009
  • In this paper, particle swarm optimization(PSO) is newly implemented by CUDA(Compute Unified Device Architecture) and is applied to function optimization with several benchmark functions. CUDA is not CPU but GPU(Graphic Processing Unit) that resolves complex computing problems using parallel processing capacities. In addition, CUDA helps one to develop GPU softwares conveniently. Compared with the optimization result of PSO executed on a general CPU, CUDA saves about 38% of PSO running time as average, which implies that CUDA is a promising frame for real-time optimization and control.

기존선을 주행하는 무궁화호 열차의 소음원 모델링과 음향강도 평가에 관한 연구 (Study on the Noise Source Modeling and the Source Strength Estimation of Mugungwha Trains Running on the Conventional Railway)

  • 장승호;장은혜
    • 한국소음진동공학회논문집
    • /
    • 제23권11호
    • /
    • pp.1020-1026
    • /
    • 2013
  • An accurate railway environmental noise prediction model is required to make the proper solution of the railway noise problems. In this paper, an engineering model for predicting the noise of conventional passenger cars is presented considering the acoustic source strength in octave-band frequencies and the propagation over grounds with varying surface properties. Since the formation of a train can be variable, the source strength of each locomotive and passenger car was estimated by measuring the pass-by noise and analysing the wheel-rail rolling noise. Some validation cases show on the average small differences between the predictions of the present model and the measurement results.

실내온도조절을 위한 인버터 열펌프의 주파수 제어에 관한 연구 (A study of frequency control of an inverter heat pump for indoor air temperature adjustment)

  • 박윤철;민만기
    • 대한기계학회논문집B
    • /
    • 제21권10호
    • /
    • pp.1262-1272
    • /
    • 1997
  • An experimental study on the frequency control of an inverter heat pump to get the desired indoor room temperature has been conducted for the performance characteristics during the steady, 4, 8, and 16 step frequency operations. The heat pump model used in this study was operated to meet the experimental conditions of ASHRAE standard. The performance of the system was tested by measuring the temperature and pressure of the refrigerant, and cooling capacity, power consumption, etc. of the system. As the controlling frequency steps increased, the running time of the compressor increased as well as the electric consumption of the system and the cooling energy due to the wall heating load. However, the average cooling COP was improved.

Bandwidth-aware Memory Placement on Hybrid Memories targeting High Performance Computing Systems

  • Lee, Jongmin
    • 한국컴퓨터정보학회논문지
    • /
    • 제24권8호
    • /
    • pp.1-8
    • /
    • 2019
  • Modern computers provide tremendous computing capability and a large memory system. Hybrid memories consist of next generation memory devices and are adopted in high performance systems. However, the increased complexity of the microprocessor makes it difficult to operate the system effectively. In this paper, we propose a simple data migration method called Bandwidth-aware Data Migration (BDM) to efficiently use memory systems for high performance processors with hybrid memory. BDM monitors the status of applications running on the system using hardware performance monitoring tools and migrates the appropriate pages of selected applications to High Bandwidth Memory (HBM). BDM selects applications whose bandwidth usages are high and also evenly distributed among the threads. Experimental results show that BDM improves execution time by an average of 20% over baseline execution.

Hybrid in-memory storage for cloud infrastructure

  • Kim, Dae Won;Kim, Sun Wook;Oh, Soo Cheol
    • 인터넷정보학회논문지
    • /
    • 제22권5호
    • /
    • pp.57-67
    • /
    • 2021
  • Modern cloud computing is rapidly changing from traditional hypervisor-based virtual machines to container-based cloud-native environments. Due to limitations in I/O performance required for both virtual machines and containers, the use of high-speed storage (SSD, NVMe, etc.) is increasing, and in-memory computing using main memory is also emerging. Running a virtual environment on main memory gives better performance compared to other storage arrays. However, RAM used as main memory is expensive and due to its volatile characteristics, data is lost when the system goes down. Therefore, additional work is required to run the virtual environment in main memory. In this paper, we propose a hybrid in-memory storage that combines a block storage such as a high-speed SSD with main memory to safely operate virtual machines and containers on main memory. In addition, the proposed storage showed 6 times faster write speed and 42 times faster read operation compared to regular disks for virtual machines, and showed the average 12% improvement of container's performance tests.

스피드 러닝 프레임워크 분석 : 일반 게임플레이어와 스피드 러너 간의 비교를 중심으로 (Framework analysis of Speed-running : Towards a Comparison Between the Normal Gameplayer and Speed Runners)

  • Thiago, Araujo Silva;Song, Seung-Keun
    • 한국정보통신학회논문지
    • /
    • 제24권2호
    • /
    • pp.337-339
    • /
    • 2020
  • Speed-run is a gameplay style in what the player tries to finish the game as quickly as possible. The purpose of this study is to observe and analyze the difference between a normal player and a speed runner in a speed run to better understand their interaction with the game and to extrapolate the applications of the analyzed frameworks to scenarios that extent normal gameplay. For this purpose, MDA (Mechanic, Dynamic, Aesthetics) and DPE (Experience, Design, Play) were used. As a result, the average player was found to focus on aesthetics or affects while the speed runner focused on mechanics and gameplay.

KAWS: Coordinate Kernel-Aware Warp Scheduling and Warp Sharing Mechanism for Advanced GPUs

  • Vo, Viet Tan;Kim, Cheol Hong
    • Journal of Information Processing Systems
    • /
    • 제17권6호
    • /
    • pp.1157-1169
    • /
    • 2021
  • Modern graphics processor unit (GPU) architectures offer significant hardware resource enhancements for parallel computing. However, without software optimization, GPUs continuously exhibit hardware resource underutilization. In this paper, we indicate the need to alter different warp scheduler schemes during different kernel execution periods to improve resource utilization. Existing warp schedulers cannot be aware of the kernel progress to provide an effective scheduling policy. In addition, we identified the potential for improving resource utilization for multiple-warp-scheduler GPUs by sharing stalling warps with selected warp schedulers. To address the efficiency issue of the present GPU, we coordinated the kernel-aware warp scheduler and warp sharing mechanism (KAWS). The proposed warp scheduler acknowledges the execution progress of the running kernel to adapt to a more effective scheduling policy when the kernel progress attains a point of resource underutilization. Meanwhile, the warp-sharing mechanism distributes stalling warps to different warp schedulers wherein the execution pipeline unit is ready. Our design achieves performance that is on an average higher than that of the traditional warp scheduler by 7.97% and employs marginal additional hardware overhead.

주축 진동특성을 이용한 정밀가공 성능평가 (Evaluation of Precision Cutting Performance by Bending Vibration Made Shapes of Main Spindle)

  • 박보용;김종관
    • 한국정밀공학회지
    • /
    • 제10권3호
    • /
    • pp.191-197
    • /
    • 1993
  • In this paper, experimental studies are mainly carried out for the evaluation of precision cutting performance of a machine tool spincle running at high speed with the low load, in consideration of the bending vibration characteristics. As a result a process in presented for the practical application in the machine tools industry to evaluate the cutting performance in design stage of spindles.

  • PDF

Optimal dwelling time prediction for package tour using K-nearest neighbor classification algorithm

  • Aria Bisma Wahyutama;Mintae Hwang
    • ETRI Journal
    • /
    • 제46권3호
    • /
    • pp.473-484
    • /
    • 2024
  • We introduce a machine learning-based web application to help travel agents plan a package tour schedule. K-nearest neighbor (KNN) classification predicts the optimal tourists' dwelling time based on a variety of information to automatically generate a convenient tour schedule. A database collected in collaboration with an established travel agency is fed into the KNN algorithm implemented in the Python language, and the predicted dwelling times are sent to the web application via a RESTful application programming interface provided by the Flask framework. The web application displays a page in which the agents can configure the initial data and predict the optimal dwelling time and automatically update the tour schedule. After conducting a performance evaluation by simulating a scenario on a computer running the Windows operating system, the average response time was 1.762 s, and the prediction consistency was 100% over 100 iterations.