• Title/Summary/Keyword: I/O stack

Search Result 48, Processing Time 0.034 seconds

Optimizing I/O Stack for Fast Storage Devices (고속 저장 장치를 위한 입출력 스택 최적화)

  • Han, Hyuck
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.5
    • /
    • pp.251-258
    • /
    • 2016
  • Recently, the demand for fast storage devices is rapidly increasing in cloud platforms, social network services, etc. Despite the development of fast storage devices, the traditional Linux I/O stack is not able to exploit the full extent of the performance improvement since it has been optimized for disk-based storage devices. In this paper, we propose an optimized I/O stack which can fully utilize the I/O bandwidth and latency of fast storage devices. To this end, we design a new I/O interface to replace the current block I/O interface and optimize our I/O interface. Our optimized I/O interface bypasses operations/layers in block I/O subsystems of the current Linux I/O stack to fully exploit fast storage devices. We also optimize the Linux file systems such as ext2 and ext4 to run on our I/O interface. We evaluate our I/O stack with multiple benchmarks and the experimental results show that our I/O stack achieves 1.7 times better throughput compared to traditional Linux I/O stack.

Implementation of Light-weight I/O Stack for NVMe-over-Fabrics

  • Ahn, Sungyong
    • International journal of advanced smart convergence
    • /
    • v.9 no.3
    • /
    • pp.253-259
    • /
    • 2020
  • Most of today's large-scale cloud systems and enterprise data centers are distributing resources to improve scalability and resource utilization. NVMe-over-Fabric protocol allows submitting NVMe commands to a remote NVMe SSD through RDMA (Remote Direct Memory Access) network. It is attracting attention recently because it is possible to construct a disaggregation storage system with low latency through the protocol. However, the current I/O stack of NVMe-over-Fabric has an inefficient structure for maintaining compatibility with the traditional I/O stack. Therefore, in this paper, we propose a new mechanism to reduce I/O latency and CPU overhead by modifying I/O path of NVMe-over-Fabric to pass through legacy block layer. According to the performance evaluation results, the proposed mechanism is able to reduce the I/O latency and CPU overhead by up to 22% and 24% compared to the existing NVMe-over-Fabrics protocol, respectively.

Performance Analysis of NVMe SSDs and Design of Direct Access Engine on Virtualized Environment (가상화 환경에서 NVMe SSD 성능 분석 및 직접 접근 엔진 개발)

  • Kim, Sewoog;Choi, Jongmoo
    • KIISE Transactions on Computing Practices
    • /
    • v.24 no.3
    • /
    • pp.129-137
    • /
    • 2018
  • NVMe(Non-Volatile Memory Express) SSD(Solid State Drive) is a high-performance storage that makes use of flash memory as a storage cell, PCIe as an interface and NVMe as a protocol on the interface. It supports multiple I/O queues which makes it feasible to process parallel-I/Os on multi-core environments and to provide higher bandwidth than SATA SSDs. Hence, NVMe SSD is considered as a next generation-storage for data-center and cloud computing system. However, in the virtualization system, the performance of NVMe SSD is not fully utilized due to the bottleneck of the software I/O stack. Especially, when it uses I/O stack of the hypervisor or the host operating system like Xen and KVM, I/O performance degrades seriously due to doubled-I/O stack between host and virtual machine. In this paper, we propose a new I/O engine, called Direct-AIO (Direct-Asynchronous I/O) engine, that can access NVMe SSD directly for I/O performance improvements on QEMU emulator. We develop our proposed I/O engine and analyze I/O performance differences between the existed I/O engine and Direct-AIO engine.

An Empirical Study on Linux I/O stack for the Lifetime of SSD Perspective (SSD 수명 관점에서 리눅스 I/O 스택에 대한 실험적 분석)

  • Jeong, Nam Ki;Han, Tae Hee
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.9
    • /
    • pp.54-62
    • /
    • 2015
  • Although NAND flash-based SSD (Solid-State Drive) provides superior performance in comparison to HDD (Hard Disk Drive), it has a major drawback in write endurance. As a result, the lifetime of SSD is determined by the workload and thus it becomes a big challenge in current technology trend of such as the shifting from SLC (Single Level Cell) to MLC (Multi Level cell) and even TLC (Triple Level Cell). Most previous studies have dealt with wear-leveling or improving SSD lifetime regarding hardware architecture. In this paper, we propose the optimal configuration of host I/O stack focusing on file system, I/O scheduler, and link power management using JEDEC enterprise workloads in terms of WAF (Write Amplification Factor) which represents the efficiency perspective of SSD life time especially for host write processing into flash memory. Experimental analysis shows that the optimum configuration of I/O stack for the perspective of SSD lifetime is MinPower-Dead-XFS which prolongs the lifetime of SSD approximately 2.6 times in comparison with MaxPower-Cfq-Ext4, the best performance combination. Though the performance was reduced by 13%, this contributions demonstrates a considerable aspect of SSD lifetime in relation to I/O stack optimization.

Numerical Simulation on Self-heating for Interlayer Tunneling Spectroscopy in $Bi_2Sr_2CaCu_2O_{8+x}$

  • Park, Jae-Hyun;Lee, Hu-Jong
    • Progress in Superconductivity
    • /
    • v.9 no.1
    • /
    • pp.18-22
    • /
    • 2007
  • For interlayer tunneling spectroscopy using a small stack of $Bi_2Sr_2CaCu_2O_{8+x}$ (Bi-2212) intrinsic junctions in a high-bias range, large self-heating takes place due to the poor thermal conductivity of Bi-2212. In this study, we numerically estimate the self-heating around a Bi-2212 sample stack for I-V or dI/dV-V measurements. Our results show that the temperature discrepancy between the Bi-2212 sample stack and top Au electrodes due to bias-induced self-heating is small enough along the c-axis direction of Bi-2212. On the other hand, the lateral temperature discrepancy between the sample stack and the Bi-2212 on-chip thermometer stack can be as large as ${\sim}20\;K$ for the highest bias required to observe the pseudogap hump structure. We thus suggest a new in-situ ac thermometry, employing the Au current-bias electrode itself deposited on top of the sample stack as the resistive thermometer layer, which is supposed to allow safe temperature measurements for the interlayer tunneling spectroscopy.

  • PDF

Cycle Detection of Discrete Logarithm using an Array (배열을 이용한 이산대수의 사이클 검출)

  • Sang-Un Lee
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.5
    • /
    • pp.15-20
    • /
    • 2023
  • Until now, Pollard's Rho algorithm has been known as the most efficient way for discrete algebraic problems to decrypt symmetric keys. However, the algorithm is being studied on how to further reduce the complexity of O(${\sqrt{p}}$) performance, along with the disadvantage of having to store the giant stride m=⌈${\sqrt{p}}$⌉ data. This paper proposes an array method for cycle detection in discrete logarithms. The proposed method reduces the number of updates of stack memory by at least 73%. This is done by only updating the array when (xi<0.5xi-1)∩(xi<0.5(p-1)). The proposed array method undergoes the same number of modular calculation as stack method, but significantly reduces the number of updates and the execution time for array through the use of a binary search method.

Development of Outdoor 50W-LED Module using Heat-pipe and Stack-fin (히트파이프 및 스택핀을 이용한 50W급 옥외용 LED 모듈개발)

  • Hong, Seok-Gi;Jeong, Hee-Suk;Ryeom, Jeong-duk
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.29 no.12
    • /
    • pp.15-21
    • /
    • 2015
  • We proposed 50W-LED modules of using Heat-pipe and Stack-fin and produced LED modules was evaluated heat dissipation characteristics with comparison of the conventional die-casting type. It verified the application of products by applying it to 150W-LED road luminaires through simulation. The LED module was measures aimed design temperature of the Stack-fin and showed 26% upward heat dissipation effect than a conventional die-casting type. The luminous efficacy of 150W-LED road luminaires using this LED module reached over 112lm/W, and the simulation results showed average of horizontal luminance, overall luminance uniformity($U_O$) and lane luminance uniformity($U_I$) that is suitable for five-lane road with the KS standards.

The research on embody in mobile system efficiently using Embedded linux kernel (Embedded linux kernel을 이용한 효율적인 모바일 단말 구현에 관한 연구)

  • 이용훈;윤원동;김영근
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1601-1604
    • /
    • 2003
  • 본 논문에서는 PDA. Hand PC(HPC)등과 같은 모바일 단말에 운영체제로써 Embedded Linux를 채택하였을 경우 고려해야 할 Hardware 사양, I/O interrupt latency에 따른 성능. 스케줄링 정책에 따른 성능에 대하여 논한다. 대상 타겟으로 사용한 HPC 의 하드웨어 사양에 설명하고, Embedded Linux와의 연동에 있어서 문제점을 살펴본다. 또한 각종 I/O device들의 Interrupt latency에 따른 성능저하와 스케줄링 정책에 의한 성능저하에 대하여 분석하고. 해결 방안에 대하여 논한다. 마지막으로 실제 예로서 Mobile IPv6 S/W Stack을 이용한 실제 검증을 수행하고 성능 향상 방안을 제시한다.

  • PDF

An Adaptive Polling Selection Technique for Ultra-Low Latency Storage Systems (초저지연 저장장치를 위한 적응형 폴링 선택 기법)

  • Chun, Myoungjun;Kim, Yoona;Kim, Jihong
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.2
    • /
    • pp.63-69
    • /
    • 2019
  • Recently, ultra-low latency flash storage devices such as Z-SSD and Optane SSD were introduced with the significant technological improvement in the storage devices which provide much faster response time than today's other NVMe SSDs. With such ultra-low latency, $10{\mu}s$, storage devices the cost of context switch could be an overhead during interrupt-driven I/O completion process. As an interrupt-driven I/O completion process could bring an interrupt handling overhead, polling or hybrid-polling for the I/O completion is known to perform better. In this paper, we analyze tail latency problem in a polling process caused by process scheduling in data center environment where multiple applications run simultaneously under one system and we introduce our adaptive polling selection technique which dynamically selects efficient processing method between two techniques according to the system's conditions.