• Title/Summary/Keyword: Performance benchmark

Search Result 853, Processing Time 0.014 seconds

Development of New Benchmark and Benchmark-observation Method for Effective Performence Rating Training of Assembling and Machining Operations (조립작업과 기계가공작업의 수행도평가훈련을 위한 기본표준과 기본표준관측법의 개발)

  • 박성학;장영기
    • Journal of the Korean Professional Engineers Association
    • /
    • v.22 no.3
    • /
    • pp.5-13
    • /
    • 1989
  • A major problem of stopwatch time study is how to do for the accurate and consistent performance rating, which is one of the critical variables to determine the accuracy of work measurement and should be still dependent upon time observer's judgement. Therefore the time observer's ability for the performance rating is very important, and must be improved by correct training method and procedure. This paper developed a new benchmark and benchmark-observation method for the effective performance rating training of assembling and machining operations. The trainees' ability in the accuracy and consistency of the performance rating ,improved significantly after being trained by subject method. The percentage improvement in rating accuracy and consistency values was 34.7% and 49% respectively. In addition, benchmark-practice method for the performance rating training is not significant, so it is proofed that the skill of a certain operation is not important for the improvement of the rating ability.

  • PDF

A New Benchmark for Object-relational DBMSs (객체-관계형 데이터베이스 시스템을 위한 새로운 성능 평가 방법론)

  • Kim, Seong-Jin;Lee, Sang-Ho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.7
    • /
    • pp.1997-2007
    • /
    • 2000
  • This paper presents a new benchmark for object-relational database systems, which are regarded as the next-generation database system. This benchmark has been developed to evaluate system performance peculiar to object-relational database systems. The design philosophy, test databases, an test queries of the benchmark are presented. This benchmark features scaleability, use of a synthesized database only, and a query-oriented evaluation. We have implemented his benchmark with two commerical object-relational database systems and the experimental results are also reported.

  • PDF

The EBORD Benchmark for Database Systems (데이터베이스 시스템을 위한 EBORD 성능 평가 방법론)

  • Jeong, Hoe-Jin;Lee, Sang-Ho
    • Journal of Information Technology Applications and Management
    • /
    • v.12 no.2
    • /
    • pp.29-43
    • /
    • 2005
  • The paper presents the EBORD (Extended Benchmark for Object-Relational Databases) benchmark, which is an extension of the BORD benchmark for object-relational databases. The EBORD benchmark is developed to evaluate the database common functions that should be supported in modern database systems. Besides the 36 test queries already defined in the BORD benchmark, totally 22 test queries in five categories are newly defined in order to measure the index-relevant performance issues and database import capabilities. The EBORD benchmark also features scalability, use of a synthesized database, and a query-oriented evaluation. In order to show the feasibility of the proposed benchmark, we implement it with two commercial database systems. The experimental results and analyses are also reported.

  • PDF

Development of the Performance Benchmark Tool for Data Stream Management Systems Combined with DBMS (DBMS와 결합된 데이터스트림관리시스템을 위한 성능 평가 도구 개발)

  • Kim, Gyoung-Bae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.8
    • /
    • pp.1-11
    • /
    • 2010
  • Many applications of DSMS(Data Stream Management System) require not only to process real-time stream data efficiently but also to provide high quality services such as data mining and data warehouse combining with DBMS(Database Management System) to users. In this paper we execute the performance benchmark of the combined system of DSMS and DBMS that is developed for high quality services. We use the stream data of network monitoring application system and combine the traditional representative DSMSs and DBMSs in a single system for the performance testing. We develop the total performance benchmark tool implementing JAVA language for the our testing. For our performance testing, we combine DSMS such as STREAM and Coral8 and DBMS such MySQL and Oracle10g respectively.

A study for Verification Procedures on Open-source Software Via Benchmark Testing (벤치마크 테스트를 통한 공개소프트웨어 검증 절차에 관한 연구)

  • Kim, Doo-Yeon;Rhew, Sung-Yul
    • Journal of Information Technology Services
    • /
    • v.5 no.3
    • /
    • pp.99-108
    • /
    • 2006
  • Public institutions are considering adoption of open-source software in the process of information projects. However, there doesn't exist reliable information about an adoption process for open-source software. Performance and stability problems of this software also persist, as a result, current open-source software is not widely used. As a software market and industry grows, Benchmark test has been performed more often than before in order to help customers understand and select the most appropriate product among myriad similar ones. It is certain that more objective and trustful data evidence should be obtained by way of utilizing the procedures and methods of Benchmark Test in decision making process for selecting an open-source software. For this research, Benchmark test was applied as a way of demonstrating performance verification of an open-source software in the public institutions. It is certain that more objective and trustful data evidence should be obtained by way of utilizing the procedures and methods of Benchmark Test in decision making process for selecting an open-source software. It also introduces a case study of a information system, which selected and implemented open-source software, in order to confirm the validity of this research. This research will serve as a guideline to adopt open-source software in governments as well as public institutions.

Analyses of the Effect of System Environment on Filebench Benchmark (시스템 환경이 Filebench 벤치마크에 미치는 영향 분석)

  • Song, Yongju;Kim, Junghoon;Kang, Dong Hyun;Lee, Minho;Eom, Young Ik
    • Journal of KIISE
    • /
    • v.43 no.4
    • /
    • pp.411-418
    • /
    • 2016
  • In recent times, NAND flash memory has become widely used as secondary storage for computing devices. Accordingly, to take advantage of NAND flash memory, new file systems have been actively studied and proposed. The performance of these file systems is generally measured with benchmark tools. However, since benchmark tools are executed by software simulation methods, many researchers get non-uniform benchmark results depending on the system environments. In this paper, we use Filebench, one of the most popular and representative benchmark tools, to analyze benchmark results and study the reasons why the benchmark result variations occur. Our experimental results show the differences in benchmark results depending on the system environments. In addition, this study substantiates the fact that system performance is affected mainly by background I/O requests and fsync operations.

Design and Implementation of I/O Performance Benchmarking Framework for Linux Container

  • Oh, Gijun;Son, Suho;Yang, Junseok;Ahn, Sungyong
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.1
    • /
    • pp.180-186
    • /
    • 2021
  • In cloud computing service it is important to share the system resource among multiple instances according to user requirements. In particular, the issue of efficiently distributing I/O resources across multiple instances is paid attention due to the rise of emerging data-centric technologies such as big data and deep learning. However, it is difficult to evaluate the I/O resource distribution of a Linux container, which is one of the core technologies of cloud computing, since conventional I/O benchmarks does not support features related to container management. In this paper, we propose a new I/O performance benchmarking framework that can easily evaluate the resource distribution of Linux containers using existing I/O benchmarks by supporting container-related features and integrated user interface. According to the performance evaluation result with trace-replay benchmark, the proposed benchmark framework has induced negligible performance overhead while providing convenience in evaluating the I/O performance of multiple Linux containers.

Performance Scalability of SPEC CPU2000 Benchmark over CPU Clock Speed (CPU 주파수 속도에 대한 SPEC CPU2000 성능 변화)

  • Yi, Jong-Su;Kim, Jun-Seong
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.42 no.5
    • /
    • pp.1-8
    • /
    • 2005
  • SPEC CPU2000 is an widely used benchmark program, both in industry and in academy, for measuring compute-intensive performance of computer systems with various architectures. However, there has been little effort to investigate its characteristics with respect to hardware components. This paper presents the performance scalability of SPEC CPU2000 benchmark over CPU clock speed. For an Intel x86-based system running at various clock speed, we measure the performance of SPEC CPU2000 benchmark, and analyze the characteristic of SPEC CPU2000 in a system aspect. In the experiment, we found that the overall performance of SPEC CPU2000 increases monotonically and linearly as the CPU clock speed increases and that the scale efficiencies of SPEC CPU2000 component benchmarks are quite evenly distributed.

Performance Analysis of Processors for Next Generation Satellites (차세대 위성 프로세서 선정을 위한 성능 분석)

  • Yoo, Bum-Soo;Choi, Jong-Wook;Jeong, Jae-Yeop;Kim, Sun-Wook
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.1
    • /
    • pp.51-61
    • /
    • 2019
  • There are strict evaluation processes before using new processors to satellites. Engineers evaluate processors from various viewpoints including specification, development environment, and cost. From a viewpoint of computation power, manufacturers provide benchmark results with processors, and engineers decide which processors are adequate to their satellites by comparing the benchmark results with requirements of their satellites. However, the benchmark results depends on a test environment of manufacturers, and it is quite difficult to achieve similar performance in a target environment. Therefore, it is necessary to evaluate the processors in the target environment. This paper compares performance of a processor, AT697F/LEON2, in software testbed (STB) with three development boards of XC2V/LEON3, GR712RC/LEON3, and GR740/LEON4. Seven benchmark functions of Dhrystone, Stanford, Coremark, Whetstone, Flops, NBench, and MiBench are selected. Results are analyzed with hardware and software properties: hardware properties of core architecture, number of cores, cache, and memory; and software properties of build options and compilers. Based on the analysis, this paper describes a guideline for choosing processors for next generation satellites.

Performance Comparison of Symbolic Manipulation Programs using a Validation Method for Numerical Solution (수치해 검증방법을 이용한 기호 연산 프로그램 성능 비교)

  • Yang, Sung-Wook;Lee, Sang-Chul
    • Journal of the Korean Society for Aviation and Aeronautics
    • /
    • v.23 no.2
    • /
    • pp.69-74
    • /
    • 2015
  • We propose a rigorous and practical methodology to evaluate the performance of symbolic manipulation program such as Mathematica, Maple, and Maxima. First, we demonstrate an inverse method to construct the benchmark problems of an initial value problems. The benchmark problems associated with the discrete version of the Chebyshev polynomials provide a rigorous and objective measure to evaluate the performance of symbolic manipulation programs. We compare three symbolic manipulation programs, which are Mathematica, Maple and Maxima, using this methodology. The computation time, the used memory and the perturbation terms are chosen for comparison parameters.