• Title/Summary/Keyword: I/O latency

Search Result 75, Processing Time 0.035 seconds

Event Routing Scheme to Improve I/O Latency of SMP VM (SMP 가상 머신의 I/O 지연 시간 감소를 위한 이벤트 라우팅 기법)

  • Shin, Jungsub;Kim, Hagyoung
    • Journal of KIISE
    • /
    • v.42 no.11
    • /
    • pp.1322-1331
    • /
    • 2015
  • According to the hypervisor scheduler, the vCPU (virtual CPU) operates under two states: the running state and the stop state. When the vCPU is in the stop state, incoming events are delayed until that vCPU's state changes to the running state. The latency in handling such events that are sent to the vCPU is regarded as the I/O latency. Since a SMP (symmetric multiprocessing) VM (virtual machine) incorporates multiple vCPUs, the event latency on a SMP VM can vary according to specific vCPU that receives the event. In this paper, we propose a new scheme named event routing that sends events according to the operation state of each vCPU to reduce the event latency on an SMP VM. We implemented the proposed event routing scheme in Xen ARM hypervisor and confirmed the reduction of I/O latency from measuring the network RTT (round trip time) and the TCP bandwidth under a variety of testing conditions. The network RTT decreases by up to 94% and the TCP bandwidth increases up to 35% when compare to native Xen ARM.

Implementation of Light-weight I/O Stack for NVMe-over-Fabrics

  • Ahn, Sungyong
    • International journal of advanced smart convergence
    • /
    • v.9 no.3
    • /
    • pp.253-259
    • /
    • 2020
  • Most of today's large-scale cloud systems and enterprise data centers are distributing resources to improve scalability and resource utilization. NVMe-over-Fabric protocol allows submitting NVMe commands to a remote NVMe SSD through RDMA (Remote Direct Memory Access) network. It is attracting attention recently because it is possible to construct a disaggregation storage system with low latency through the protocol. However, the current I/O stack of NVMe-over-Fabric has an inefficient structure for maintaining compatibility with the traditional I/O stack. Therefore, in this paper, we propose a new mechanism to reduce I/O latency and CPU overhead by modifying I/O path of NVMe-over-Fabric to pass through legacy block layer. According to the performance evaluation results, the proposed mechanism is able to reduce the I/O latency and CPU overhead by up to 22% and 24% compared to the existing NVMe-over-Fabrics protocol, respectively.

An Adaptive Polling Selection Technique for Ultra-Low Latency Storage Systems (초저지연 저장장치를 위한 적응형 폴링 선택 기법)

  • Chun, Myoungjun;Kim, Yoona;Kim, Jihong
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.14 no.2
    • /
    • pp.63-69
    • /
    • 2019
  • Recently, ultra-low latency flash storage devices such as Z-SSD and Optane SSD were introduced with the significant technological improvement in the storage devices which provide much faster response time than today's other NVMe SSDs. With such ultra-low latency, $10{\mu}s$, storage devices the cost of context switch could be an overhead during interrupt-driven I/O completion process. As an interrupt-driven I/O completion process could bring an interrupt handling overhead, polling or hybrid-polling for the I/O completion is known to perform better. In this paper, we analyze tail latency problem in a polling process caused by process scheduling in data center environment where multiple applications run simultaneously under one system and we introduce our adaptive polling selection technique which dynamically selects efficient processing method between two techniques according to the system's conditions.

Analyzing the Overhead of the Memory Mapped File I/O for In-Memory File Systems (메모리 파일시스템에서 메모리 매핑을 이용한 파일 입출력의 오버헤드 분석)

  • Choi, Jungsik;Han, Hwansoo
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.10
    • /
    • pp.497-503
    • /
    • 2016
  • Emerging next-generation storage technologies such as non-volatile memory will help eliminate almost all of the storage latency that has plagued previous storage devices. In conventional storage systems, the latency of slow storage devices dominates access latency; hence, software efficiency is not critical. With low-latency storage, software costs can quickly dominate memory latency. Hence, researchers have proposed the memory mapped file I/O to avoid the software overhead. Mapping a file into the user memory space enables users to access the file directly. Therefore, it is possible to avoid the complicated I/O stack. This minimizes the number of user/kernel mode switchings. In addition, there is no data copy between kernel and user areas. Despite of the benefits in the memory mapped file I/O, its overhead still needs to be addressed, as the existing mechanism for the memory mapped file I/O is designed for slow block devices. In this paper, we identify the overheads of the memory mapped file I/O via experiments.

iSCSI Intermediate Target System for Reducing I/O Latency between Mobile Device and Mass Storage (모바일 기기와 대용량 스토리지간의 I/O latency를 줄이기 위한 iSCSI 중개 서버 시스템)

  • Yang Yuan;Choi Wonil;Park Myong-Sun
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11a
    • /
    • pp.892-894
    • /
    • 2005
  • 모바일 기기가 널리 보급되면서 모바일 기기에서 멀티미디어 데이터나 데이터베이스 같은 대용량 데이터 사용에 대한 요구가 커지고 있다. 따라서 모바일 기기의 저장 공간의 한계를 극복하기 위한 노력의 하나로 iSCSI 를 이용해 모바일 기기에 원격 스토리지 서비스를 제공하는 연구가 진행되었다. 그러나 모바일 기기에 iSCSI 를 적용했을 때, 지연 시간에 민강한 iSCSI의 구조적인 특성 때문에 iSCSI 클라이언트가 서버에서 멀리 떨어진 위치로 이동하게 되면 iSCSI I/O 성능이 급격히 떨어지는 문제가 발생한다. 본 논문에서는 모바일 기기가 스토리지 서버로부터 거리가 멀어졌을 때 전송 지연시간이 증가함에 따라 iSCSI 성능이 급격하게 떨어지는 단점을 개선하기 위해 중개서버(Intermediate Target)를 이용해 iSCSI Target을 지역화하여 성능을 높이는 방안을 제안한다.

  • PDF

Delayed-constrained scheduling for a game server virtual machine (게임서버 가상머신을 위한 지연 제약 스케줄링)

  • kang, Kiwook;Ko, Cherl-hong;Hong, Jiman;Paik, Doowon
    • Journal of Korea Game Society
    • /
    • v.16 no.4
    • /
    • pp.57-68
    • /
    • 2016
  • As the size of the data used in the game application increase gradually, the physical resources of game server grow. Accordingly, it is necessary to/it is required to introduce I/O virtualization in game server to improve I/O performance of the server. But it is difficult to ensure high responsiveness in game server where I/O delay change from time to time. To maximize the benefit of I/O virtualization, guaranteeing I/O response time is very important and it is necessary to have I/O scheduling techniques to manage the I/O latency in the order of priority of virtual machines. In this paper, we propose an efficient delay-constrained scheduling algorithm in a virtualization environment to ensure maximum I/O latency. In addition, a reduced amount of loss of the packet was found to increase the fairness of scheduling in the experiments with the proposed scheme.

The research on embody in mobile system efficiently using Embedded linux kernel (Embedded linux kernel을 이용한 효율적인 모바일 단말 구현에 관한 연구)

  • 이용훈;윤원동;김영근
    • Proceedings of the IEEK Conference
    • /
    • 2003.07d
    • /
    • pp.1601-1604
    • /
    • 2003
  • 본 논문에서는 PDA. Hand PC(HPC)등과 같은 모바일 단말에 운영체제로써 Embedded Linux를 채택하였을 경우 고려해야 할 Hardware 사양, I/O interrupt latency에 따른 성능. 스케줄링 정책에 따른 성능에 대하여 논한다. 대상 타겟으로 사용한 HPC 의 하드웨어 사양에 설명하고, Embedded Linux와의 연동에 있어서 문제점을 살펴본다. 또한 각종 I/O device들의 Interrupt latency에 따른 성능저하와 스케줄링 정책에 의한 성능저하에 대하여 분석하고. 해결 방안에 대하여 논한다. 마지막으로 실제 예로서 Mobile IPv6 S/W Stack을 이용한 실제 검증을 수행하고 성능 향상 방안을 제시한다.

  • PDF

Para-virtualized Library for Bare-metal Network Performance in Virtualized Environment (가상화 환경의 고성능 I/O를 위한 반가상화 라이브러리)

  • Lee, Dongwoo;Cho, Youngjoong;Eom, Young Ik
    • Journal of KIISE
    • /
    • v.41 no.9
    • /
    • pp.605-610
    • /
    • 2014
  • Now, virtualization is no more emerging research area, and we can easily find its application in our circumstance. Nevertheless, I/O workloads are reluctant to be applied in virtual environment since they still suffer from unacceptable performance degradation due to virtualization latency. Many previous papers identified that virtual I/O overhead is mainly caused by exits and redundant I/O stack, and proposed several techniques to reduce them. However, they still have some limitations. In this paper, we introduce a novel I/O virtualization framework which improves I/O performance by exploiting multicore architecture. We applied our framework to the virtual network, and it improves TCP throughput up to 169%, and decreases UDP latency up to 38% on the network with the 10Gbps NIC.

An Analysis on I/O latency of Guest OS based on Xen-Arm Virtualization (Xen-ARM에서 게스트 운영체제의 I/O 지연 시간 분석)

  • Kwak, Keun-Hwan;Jo, Jae-Hyun;Yoo, See-Hwan;Yoo, Chuck
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2011.06a
    • /
    • pp.531-534
    • /
    • 2011
  • 최근에는 가상화 기술이 임베디드 시스템 분야에도 적용 되면서, 다양한 종류의 임베디드 시스템이 하나의 물리 머신에서 동시에 실행되어야 하는 경우가 늘어나고 있다. 임베디드 시스템은 다양한 특성을 가지는데 그 중 하나가 바로 I/O 처리에 되한 지연시간의 보장이다. 본 논문에서는 임베디드 가상화 환경에서 I/O 지연 시간이 발생하는 원인을 분석하고 이를 해결 하기 위한 방법을 제안한다. 그리고 실험을 통해서 I/O 지연시간이 개선 될 수 있음을 보인다.

Implementation of Storage Service Protocol on Infiniband based Network (인피니밴드 네트웍에서 RDMA 기반의 저장장치 서비스 프로토콜개발)

  • Joen Ki-Man;Park Chang-Won;Kim Young-Hwan
    • 한국정보통신설비학회:학술대회논문집
    • /
    • 2006.08a
    • /
    • pp.77-81
    • /
    • 2006
  • Because of the rapid increasing of network user, there are some problems to tolerate the network overhead. Recently, the research and technology of the user-level for high performance and low latency than TCP/IP which relied upon the kernel for processing the messages. For example, there is an Infiniband technology. The Infiniband Trade Association (IBTA) has been proposed as an industry standard for both communication between processing node and I/O devices and for inter-processor communication. It replaces the traditional bus-based interconnect with a switch-based network for connecting processing node and I/O devices. Also Infiniband uses RDMA (Remote DMA) for low latency of CPU and OS to communicate between Remote nodes. In this paper, we develop the SRP (SCSI RDMA Protocol) which is Storage Access Protocol on Infiniband network. And will compare to FC (Fibre Channle) based I-SCSI (Internet SCSI) that it is used to access storage on Etherent Fabric.

  • PDF