• Title/Summary/Keyword: cache storage

Search Result 138, Processing Time 0.024 seconds

A Technique of Replacing XML Semantic Cache (XML 시맨틱 캐쉬의 교체 기법)

  • Hong, Jung-Woo;Kang, Hyun-Chul
    • The Journal of Society for e-Business Studies
    • /
    • v.12 no.3
    • /
    • pp.211-234
    • /
    • 2007
  • In e-business, XML is a major format of data and it is essential to efficiently process queries against XML data. XML query caching has received much attention for query performance improvement. In employing XML query caching, some efficient technique of cache replacement is required. The previous techniques considered as a replacement unit either the whole query result or the path in the query result. The former is simple to employ but it is not efficient whereas the latter is more efficient and yet the size difference among the potential victims is large, and thus, efficiency of caching would be limited. In this paper, we propose a new technique where the element in the query result is are placement unit to overcome the limitations of the previous techniques. The proposed technique could enhance the cache efficiency to a great extent because it would not pick a victim whose size is too large to store a new cached item, the variance in the size of victims would be small, and the unused space of the cache storage would be small. A technique of XML semantic cache replacement is presented which is based on the replacement function that takes into account cache hit ratio, last access time, fetch time, size of XML semantic region, size of element in XML semantic region, etc. We implemented a prototype XML semantic cache system that employs the proposed technique, and conducted a detailed set of experiments over a LAN environment. The experimental results showed that our proposed technique outperformed the previous ones.

  • PDF

Embedded Node Cache Management for Hybrid Storage Systems (하이브리드 저장 시스템을 위한 내장형 노드 캐시 관리)

  • Byun, Si-Woo;Hur, Moon-Haeng;Roh, Chang-Bae
    • Proceedings of the KIEE Conference
    • /
    • 2007.04a
    • /
    • pp.157-159
    • /
    • 2007
  • The conventional hard disk has been the dominant database storage system for over 25 years. Recently, hybrid systems which incorporate the advantages of flash memory into the conventional hard disks are considered to be the next dominant storage systems to support databases for desktops and server computers. Their features are satisfying the requirements like enhanced data I/O, energy consumption and reduced boot time, and they are sufficient to hybrid storage systems as major database storages. However, we need to improve traditional index node management schemes based on B-Tree due to the relatively slow characteristics of hard disk operations, as compared to flash memory. In order to achieve this goal, we propose a new index node management scheme called FNC-Tree. FNC-Tree-based index node management enhanced search and update performance by caching data objects in unused free area of flash leaf nodes to reduce slow hard disk I/Os in index access processes.

  • PDF

An performance analysis on SSD caching mechanism in Linux (리눅스 SSD caching mechanism 의 성능 비교 및 분석)

  • Heo, Sang-Bok;Park, Jinhee;Jo, Heeseung
    • Smart Media Journal
    • /
    • v.4 no.2
    • /
    • pp.62-67
    • /
    • 2015
  • During several decades, hard disk drive(HDD) has been used in most computer systems as secondary storage and, however, the performance enhancement of HDD is limited by its mechanical properties. On the other hand, although the flash memory based solid state drive (SSD) has more advantages over HDD such as high performance and low noise, SSD is still too expensive for common usage and expected to take several years to replace HDD completely. Therefore, SSD caching mechanism using the SSD as a cache of high capacity HDD has been highlighted lately. The representatives of SSD caching mechanisms are typically bcache, dm-cache, Flashcache, and EnhanceIO. Each of them has its own internal mechanism and implementation, and this makes them to show their own pros. and cons. In this paper, we analyze the characteristics of each SSD caching mechanisms and compare the performance of them under various workloads. We expect that our contribution will be useful to enhance the performance of SSD caching mechanisms.

Game Theoretic Cache Allocation Scheme in Wireless Networks (게임이론 기반 무선 통신에서의 캐시 할당 기법)

  • Le, Tra Huong Thi;Kim, Do Hyeon;Hong, Choong Seon
    • Journal of KIISE
    • /
    • v.44 no.8
    • /
    • pp.854-859
    • /
    • 2017
  • Caching popular videos in the storage of base stations is an efficient method to reduce the transmission latency. This paper proposes an incentive proactive cache mechanism in the wireless network to motivate the content providers (CPs) to participate in the caching procedure. The system consists of one/many Infrastructure Provider (InP) and many CPs. The InP aims to define the price it charges the CPs to maximize its revenue while the CPs compete to determine the number of files they cache at the InP's base stations (BSs). We conceive this system within the framework of Stackelberg game where InP is considered as the leader and CPs are the followers. By using backward induction, we show closed form of the amount of cache space that each CP renting on each base station and then solve the optimization problem to calculate the price that InP leases each CP. This is different from the existing works in that we consider the non-uniform pricing scheme. The numerical results show that InP's profit in the proposed scheme is higher than in the uniform pricing.

Design and Implementation of File System Using Local Buffer Cache for Digital Convergence Devices (디지털 컨버전스 기기를 위한 지역 버퍼 캐쉬 파일 시스템 설계 및 구현)

  • Jeong, Geun-Jae;Cho, Moon-Haeng;Lee, Cheol-Hoon
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.8
    • /
    • pp.21-30
    • /
    • 2007
  • Due to the growth of embedded systems and the development of semi-conductor and storage devices, digital convergence devises is ever growing. Digital convergence devices are equipments into which various functions such as communication, playing movies and wave files and electronic dictionarys are integrated. Example are portable multimedia players(PMPs), personal digital assistants(PDAs), and smart phones. Therefore, these devices need an efficient file system which manages and controls various types of files. In designing such file systems, the size constraint for small embedded systems as well as performance and compatibility should be taken into account. In this paper, we suggest the partial buffer cache technique. Contrary to the traditional buffer cache, the partial buffer cache is used for only the FAT meta data and write-only data. Simulation results show that we could enhance the write performance more than 30% when the file size is larger than about 100 KBytes.

Hybrid Main Memory based Buffer Cache Scheme by Using Characteristics of Mobile Applications (모바일 애플리케이션의 특성을 이용한 하이브리드 메모리 기반 버퍼 캐시 정책)

  • Oh, Chansoo;Kang, Dong Hyun;Lee, Minho;Eom, Young Ik
    • Journal of KIISE
    • /
    • v.42 no.11
    • /
    • pp.1314-1321
    • /
    • 2015
  • Mobile devices employ buffer cache mechanisms, just as in computer systems such as desktops or servers, to mitigate the performance gap between main memory and secondary storage. However, DRAM has a problem in that it accelerates battery consumption by performing refresh operations periodically to maintain the stored data. In this paper, we propose a novel buffer cache scheme to increase the battery lifecycle in mobile devices based on a hybrid main memory architecture consisting of DRAM and non-volatile PCM. We also suggest a new buffer cache policy that allocates buffers based on process states to optimize the performance and endurance of PCM. In particular, our algorithm allocates each page to the appropriate position corresponding to the state of the application that owns the page, and tries to ensure a rapid response of foreground applications even with a small amount of DRAM memory. The experimental results indicate that the proposed scheme reduces the elapsed time of foreground applications by 58% on average and power consumption by 23% on average without negatively impacting the performance of background applications.

Design and Implementation of the Spatial Data Cache Based on Agents for Providing Mobile Map Services (모바일 지도 서비스를 위한 에이전트 기반의 공간 데이터 캐쉬의 설계 및 구현)

  • Lim, Duk-Sung;Lee, Jai-Ho;Hong, Bong-Hee
    • The KIPS Transactions:PartD
    • /
    • v.10D no.2
    • /
    • pp.175-186
    • /
    • 2003
  • Mobile clients like a PDA need a cache and a spatial index to search and access map data efficiently. When a server transmits spatial objects to a mobile client which has a low storage capacity, some of them can be duplicated in a cache of the mobile client. Moreover, the cost for strong added data in the cache and reconfiguring spatial index is very high in the mobile client with low computing power. The scheme for processing duplicated objects and disturbing tasks of the mobile client which has low computing power is needed. In this paper, we classfy the method for strorng duplicated objects and present the scheme for the both caching objects and reconfiguring a spatial index of cached objects using the clipping technique. We propose the caching system based on an agent in order to distribute the overhead of a mobile client as well as to provide efficiently map services. We design and implement it, and evaluate the performance.

Directory Cache Coherence Scheme using the Number-Balanced Binary Tree (수 평형 이진트리를 이용한 디렉토리 캐쉬 일관성 유지 기법)

  • Seo, Dae-Wha
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.3
    • /
    • pp.821-830
    • /
    • 1997
  • The directory-based cache coherence scheme is an attractive approach to solve the caceh coherence problem in a large-scale shared-memory multiprocessor.However, the exsting directory-based schemes have some problens such as the enormous storage overhead for a directory, the long invalidation latency, the heavy network condes-tion, and the low scalability.For resolving these problems, we propose a new directroy- based caceh coherence scheme which is suitable for building scalable, shred-memory multiprocessors.In this scheme, each directory en-try ofr a given memory block is a number-balanced binaty tree(NBBT) stucture.The NBBT has several proper-ties to effciently maintain the directory for the cache consistency such that the shape is unique, the maximum depth is [log$_2$n], and the tree has the minimum number of leaf nodes among the binarry tree with n nodes.Therefore, this scheme can reduce the storage overhead, the network traffic, and the inbalidation latency and can ensutr the high- scalability the large-scale shared-memory multiprocessors.

  • PDF

Acceleration Method of RAID Level 5 for DDR-SSD (DDR-SSD를 위한 RAID 레벨 5의 고속화 방법)

  • Gu, Bon-Gen;Kwak, Yun-Sik;Jeong, Seung-Kook;Hwang, Jung-Yeon
    • Journal of Advanced Navigation Technology
    • /
    • v.13 no.5
    • /
    • pp.684-690
    • /
    • 2009
  • In this paper, we propose the acceleration method of the DDR-SSD RAID level 5. The DDR-SSD is the storage device of the Next Generation Storage(NGS) system. The DDR-SSD has different characteristics with HDD and Flash SSD. That's why the DDR-SSD RAID level 5 does not provide the best performance when the normal acceleration method is used. In this paper, to accelerate the DDR-SSD RAID level 5 operation, we propose the parity cache and the architecture of the parity cell. The parity cache stores only parity blocks. This acceleration method proposed in this paper reduce the number of the disk access and the overhead of parity operations.

  • PDF

Design of Global Buffer Manager in SAN-based Cluster File Systems (SAN 환경의 대용량 클러스터 파일 시스템을 위한 광역 버퍼 관리기의 설계)

  • Lee, Kyu-Woong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.11
    • /
    • pp.2404-2410
    • /
    • 2011
  • This paper describes the design overview of cluster file system $SANique^{TM}$ based on SAN(Storage Area Network) environment. The design issues and problems of the conventional global buffer manager are also illustrated under a large set of clustered computing hosts. We propose the efficient global buffer management method that provides the more scalability and availability. In our proposed global buffer management method, we reuse the maintained list of lock information from our cluster lock manager. The global buffer manger can easily find and determine the location of requested data block cache based on that lock information. We present the pseudo code of the global buffer manager and illustration of global cache operation in cluster environment.