• 제목/요약/키워드: latency reducing

검색결과 165건 처리시간 0.029초

Locally weighted linear regression prefetching method for hybrid memory system (하이브리드 메모리 시스템의 지역 가중 선형회귀 프리페치 방법)

  • Tang, Qian;Kim, Jeong-Geun;Kim, Shin-Dug
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 한국정보처리학회 2020년도 추계학술발표대회
    • /
    • pp.12-15
    • /
    • 2020
  • Data access characteristics can directly affect the efficiency of the system execution. This research is to design an accurate predictor by using historical memory access information, where highly accessible data can be migrated from low-speed storage (SSD/HHD) to high-speed memory (Memory/CPU Cache) in advance, thereby reducing data access latency and further improving overall performance. For this goal, we design a locally weighted linear regression prefetch scheme to cope with irregular access patterns in large graph processing applications for a DARM-PCM hybrid memory structure. By analyzing the testing result, the appropriate structural parameters can be selected, which greatly improves the cache prefetching performance, resulting in overall performance improvement.

Deep Reinforcement Learning-Based Edge Caching in Heterogeneous Networks

  • Yoonjeong, Choi; Yujin, Lim
    • Journal of Information Processing Systems
    • /
    • 제18권6호
    • /
    • pp.803-812
    • /
    • 2022
  • With the increasing number of mobile device users worldwide, utilizing mobile edge computing (MEC) devices close to users for content caching can reduce transmission latency than receiving content from a server or cloud. However, because MEC has limited storage capacity, it is necessary to determine the content types and sizes to be cached. In this study, we investigate a caching strategy that increases the hit ratio from small base stations (SBSs) for mobile users in a heterogeneous network consisting of one macro base station (MBS) and multiple SBSs. If there are several SBSs that users can access, the hit ratio can be improved by reducing duplicate content and increasing the diversity of content in SBSs. We propose a Deep Q-Network (DQN)-based caching strategy that considers time-varying content popularity and content redundancy in multiple SBSs. Content is stored in the SBS in a divided form using maximum distance separable (MDS) codes to enhance the diversity of the content. Experiments in various environments show that the proposed caching strategy outperforms the other methods in terms of hit ratio.

Suggestion of a method for reducing latency in the vehicle network (차량 네트워크에서의 지연 시간 절감 방법 제안)

  • Kim, Minkyu;Kim, Jonghun;Do, Young-Soo;Jeon, Jae-Wook
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 한국정보처리학회 2022년도 추계학술발표대회
    • /
    • pp.162-164
    • /
    • 2022
  • 본 논문에서는 차량 ECU(Electronic Control Unit)에서 제한된 메모리 공간의 효율적인 사용과 추가적인 프로세싱 시간(메모리 할당과 해제)을 감소시키는 방안을 제안한다. 메모리 할당을 하고 사용이 된 이후 바로 해제하는 기존의 방식 대신, 버퍼에 넣어두고 이후에 메모리 할당하는 상황이 생기면 버퍼에 넣어둔 할당된 공간을 재사용하는 방법을 사용하여 네트워크 전환 과정에서의 데이터 처리시간 차이를 비교 및 분석하고자 한다.

A Study on Design and Cache Replacement Policy for Cascaded Cache Based on Non-Volatile Memories (비휘발성 메모리 시스템을 위한 저전력 연쇄 캐시 구조 및 최적화된 캐시 교체 정책에 대한 연구)

  • Juhee Choi
    • Journal of the Semiconductor & Display Technology
    • /
    • 제22권3호
    • /
    • pp.106-111
    • /
    • 2023
  • The importance of load-to-use latency has been highlighted as state-of-the-art computing cores adopt deep pipelines and high clock frequencies. The cascaded cache was recently proposed to reduce the access cycle of the L1 cache by utilizing differences in latencies among banks of the cache structure. However, this study assumes the cache is comprised of SRAM, making it unsuitable for direct application to non-volatile memory-based systems. This paper proposes a novel mechanism and structure for lowering dynamic energy consumption. It inserts monitoring logic to keep track of swap operations and write counts. If the ratio of swap operations to total write counts surpasses a set threshold, the cache controller skips the swap of cache blocks, which leads to reducing write operations. To validate this approach, experiments are conducted on the non-volatile memory-based cascaded cache. The results show a reduction in write operations by an average of 16.7% with a negligible increase in latencies.

  • PDF

Artificial neural network for safety information dissemination in vehicle-to-internet networks

  • Ramesh B. Koti;Mahabaleshwar S. Kakkasageri;Rajani S. Pujar
    • ETRI Journal
    • /
    • 제45권6호
    • /
    • pp.1065-1078
    • /
    • 2023
  • In vehicular networks, diverse safety information can be shared among vehicles through internet connections. In vehicle-to-internet communications, vehicles on the road are wirelessly connected to different cloud networks, thereby accelerating safety information exchange. Onboard sensors acquire traffic-related information, and reliable intermediate nodes and network services, such as navigational facilities, allow to transmit safety information to distant target vehicles and stations. Using vehicle-to-network communications, we minimize delays and achieve high accuracy through consistent connectivity links. Our proposed approach uses intermediate nodes with two-hop separation to forward information. Target vehicle detection and routing of safety information are performed using machine learning algorithms. Compared with existing vehicle-to-internet solutions, our approach provides substantial improvements by reducing latency, packet drop, and overhead.

A reinforcement learning-based network path planning scheme for SDN in multi-access edge computing

  • MinJung Kim;Ducsun Lim
    • International journal of advanced smart convergence
    • /
    • 제13권2호
    • /
    • pp.16-24
    • /
    • 2024
  • With an increase in the relevance of next-generation integrated networking environments, the need to effectively utilize advanced networking techniques also increases. Specifically, integrating Software-Defined Networking (SDN) with Multi-access Edge Computing (MEC) is critical for enhancing network flexibility and addressing challenges such as security vulnerabilities and complex network management. SDN enhances operational flexibility by separating the control and data planes, introducing management complexities. This paper proposes a reinforcement learning-based network path optimization strategy within SDN environments to maximize performance, minimize latency, and optimize resource usage in MEC settings. The proposed Enhanced Proximal Policy Optimization (PPO)-based scheme effectively selects optimal routing paths in dynamic conditions, reducing average delay times to about 60 ms and lowering energy consumption. As the proposed method outperforms conventional schemes, it poses significant practical applications.

Technology Standard Trends in Distributed and Edge Cloud Computing (분산 및 에지 클라우드 기술 표준 동향)

  • M.K. In;K.C. Lee;S.Y. Lee
    • Electronics and Telecommunications Trends
    • /
    • 제39권3호
    • /
    • pp.69-78
    • /
    • 2024
  • Cloud computing technology based on centralized high-performance computing has brought about major changes across the information technology industry and led to new paradigms. However, with the rapid development of the industry and increasing need for mass generation and real-time processing of data across various fields, centralized cloud computing is lagging behind the demand. This is particularly critical in emerging technologies such as autonomous driving, the metaverse, and augmented/virtual reality that require the provision of services with ultralow latency for real-time performance. To address existing limitations, distributed and edge cloud computing technologies have recently gained attention. These technologies allow for data to be processed and analyzed closer to their point of generation, substantially reducing the response times and optimizing the network bandwidth usage. We describe distributed and edge cloud computing technologies and explore the latest trends in their standardization.

Page Replacement Algorithm for Improving Performance of Hybrid Main Memory (하이브리드 메인 메모리의 성능 향상을 위한 페이지 교체 기법)

  • Lee, Minhoe;Kang, Dong Hyun;Kim, Junghoon;Eom, Young Ik
    • KIISE Transactions on Computing Practices
    • /
    • 제21권1호
    • /
    • pp.88-93
    • /
    • 2015
  • In modern computer systems, DRAM is commonly used as main memory due to its low read/write latency and high endurance. However, DRAM is volatile memory that requires periodic power supply (i.e., memory refresh) to sustain the data stored in it. On the other hand, PCM is a promising candidate for replacement of DRAM because it is non-volatile memory, which could sustain the stored data without memory refresh. PCM is also available for byte-addressable access and in-place update. However, PCM is unsuitable for using main memory of a computer system because it has two limitations: high read/write latency and low endurance. To take the advantage of both DRAM and PCM, a hybrid main memory, which consists of DRAM and PCM, has been suggested and actively studied. In this paper, we propose a novel page replacement algorithm for hybrid main memory. To cope with the weaknesses of PCM, our scheme focuses on reducing the number of PCM writes in the hybrid main memory. Experimental results shows that our proposed page replacement algorithm reduces the number of PCM writes by up to 80.5% compared with the other page replacement algorithms.

Data De-duplication and Recycling Technique in SSD-based Storage System for Increasing De-duplication Rate and I/O Performance (SSD 기반 스토리지 시스템에서 중복률과 입출력 성능 향상을 위한 데이터 중복제거 및 재활용 기법)

  • Kim, Ju-Kyeong;Lee, Seung-Kyu;Kim, Deok-Hwan
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • 제49권12호
    • /
    • pp.149-155
    • /
    • 2012
  • SSD is a storage device of having high-performance controller and cache buffer and consists of many NAND flash memories. Because NAND flash memory does not support in-place update, valid pages are invalidated when update and erase operations are issued in file system and then invalid pages are completely deleted via garbage collection. However, garbage collection performs many erase operations of long latency and then it reduces I/O performance and increases wear leveling in SSD. In this paper, we propose a new method of de-duplicating valid data and recycling invalid data. The method de-duplicates valid data and then recycles invalid data so that it improves de-duplication ratio. Due to reducing number of writes and garbage collection, the method could increase I/O performance and decrease wear leveling in SSD. Experimental result shows that it can reduce maximum 20% number of garbage collections and 9% I/O latency than those of general case.

Design and Implementation of Kernel Binder Cache for Accelerating Android IPC (안드로이드 IPC 가속화를 위한 커널 바인더 캐쉬의 설계 및 구현)

  • Yeon, Jeseong;Koh, Kern;Lee, Eunji
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • 제16권5호
    • /
    • pp.33-38
    • /
    • 2016
  • In Android platform, as applications invoke various service functions through IPC (Inter-Process Communication), IPC performance is critical to the responsiveness in Android. However, Android offers long IPC latency of hundreds of micro-seconds due to complicated software stacks between the kernel Binder and the user-level process Context Manager. This separation provides modularity and flexibility, but degrades the responsiveness of services owing to additional context switching and inefficient request handling. In this paper, we anatomize Android IPC mechanisms and observe that 55% of IPC latency comes from the communication overhead between Binder and Context Manager. Based on this observation, this paper proposes a kernel Binder cache that retains a popular subset of service function mappings, thereby reducing the requests transferred to the user-level daemon. The proposed Binder cache is implemented in Android 5.0 and experimental results with various benchmarks show that the proposed cache architecture improves performance by 52.9% on average.