• Title/Summary/Keyword: File Caching

Search Result 52, Processing Time 0.029 seconds

Performance Evaluation of Flash Memory-Based File Storages: NAND vs. NOR (플래시 메모리 기반의 파일 저장 장치에 대한 성능분석)

  • Sung, Min-Young
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.9 no.3
    • /
    • pp.710-716
    • /
    • 2008
  • This paper covers the performance evaluation of two flash memory-based file storages, NAND and NOR, which are the major flash types. To evaluate their performances, we set up separate file storages for the two types of flash memories on a PocketPC-based experimental platform. Using the platform, we measured and compared the I/O throughputs in terms of buffer size, amount of used space, and kernel-level write caching. According to the results from our experiments, the overall performance of the NAND-based storage is higher than that of NOR by up to 4.8 and 5.7 times in write and read throughputs, respectively. The experimental results show the relative strengths and weaknesses of the two schemes and provide insights which we believe assist in the design of flash memory-based file storages.

A Mapping Table Caching Scheme for NAND Flash-based Mobile Storage Devices (NAND 플래시 기반 모바일 저장장치를 위한 사상 테이블 캐싱 기법)

  • Yang, Soo-Hyeon;Ryu, Yeon-Seung
    • The Journal of Society for e-Business Studies
    • /
    • v.15 no.4
    • /
    • pp.21-31
    • /
    • 2010
  • Recently e-business such as online financial trade and online shopping using mobile computes are widely spread. Most of mobile computers use NAND flash memory-based storage devices for storing data. Flash memory storage devices use a software called flash translation layer to translate logical address from a file system to physical address of flash memory by using mapping tables. The legacy FTLs have a problem that they must maintain very large mapping tables in the RAM. In order to address this issues, in this paper, we proposed a new caching scheme of mapping tables. We showed through the trace-driven simulations that the proposed caching scheme reduces the space overhead dramatically but does not increase the time overhead. In the case of online transaction workload in e-business environment, in particular, the proposed scheme manifests better performance in reducing the space overhead.

Distributed File Placement and Coverage Expansion Techniques for Network Throughput Enhancement in Small-cell Network (소형셀 네트워크 전송용량 향상을 위한 분산 파일저장 및 커버리지 확장 기법)

  • Hong, Jun-Pyo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.1
    • /
    • pp.183-189
    • /
    • 2018
  • This paper proposes distributed file placement and coverage expansion techniques for mitigating the traffic bottleneck in backhaul for small-cell networks. In order to minimize the backhaul load with limited memory space, the proposed scheme controls the coverage and file placement of base station according to file popularity distribution and memory space of base stations. In other words, since the cache hit ratio is low when there is small memory capacity or widespread file popularity distribution, the base stations expand its coverage and cache different set of files for the user located in overlapped area to exploit multiple cached file sets of base stations. Our simulation results show that the proposed scheme outperforms the conventional cache strategy in terms of network throughput when there is small memory capacity or widespread file popularity distribution.

An Empirical Study on the Construction Strategy of Web-caching Network (효과적인 웹-캐싱 네트웍 구축전략에 관한 실증 연구)

  • 이주헌;조병룡
    • The Journal of Information Technology and Database
    • /
    • v.8 no.2
    • /
    • pp.41-60
    • /
    • 2001
  • Despite the growth in Internet users, demand for multi-medial, large data files and resulting explosive growth in data traffic, there has been lack of investment in Middle-Mile, interconnection of various networks, resulting in bottleneck effect, which is acerbating. One strategy to overcome such network bottleneck is Content Delivery Network (CDN). CDN does not achieve efficient delivery of large file data through physical improvement/increase in network capacity, but by delivering large file contents, the cause of bottlenecks, from distributed servers. Since it is impracticable to physically improve networks capacity to accommodate the growth in internet traffic, CON, by strong CPs contents at cache servers deployed at major ISPs networks, is able to deliver requested contents to the requesting Web clients without the loss of data and long latency.

  • PDF

A Performance Improvement technique for Flash-based File System Using File-Unit Caching (파일수준의 캐시기능을 통한 플래시 파일 시스템의 성능 향상 기법)

  • Lee, Jun-Hee;Seo, Min-Yeol;Maeng, Ji-Chan;Ryu, Min-Soo
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2007.06b
    • /
    • pp.298-303
    • /
    • 2007
  • 비휘발성 메모리의 하나인 플래시 메모리는 저전력 및 저비용 등의 장점으로 인해 임베디드 시스템에 필수적인 요소로 사용되고 있다. 이러한 장점에 반해, DRAM과 같은 휘발성 반도체 메모리와 비교하여 데이터 쓰기는 느리고, 데이터 업데이트를 위한 블록 소거 (erase) 동작의 오버헤드라는 단점이 있다. 특히 블록 소거 동작은 횟수가 제한되어 있으며, 이는 플래시 메모리의 수명을 결정하는데 중요한 요소이다. 본 논문에서는 플래시 메모리 기반 파일시스템에서 DRAM과 같은 동적 메모리를 사용하여 블록 소거의 횟수를 줄이고 입출력 속도를 향상시키는 기법을 제안한다.

  • PDF

Performance Evaluation of Disk I/O for Web Proxy Servers (웹 프락시 서버의 디스크 I/O 성능 평가)

  • Shim Jong-Ik
    • The KIPS Transactions:PartC
    • /
    • v.12C no.4 s.100
    • /
    • pp.603-608
    • /
    • 2005
  • Disk I/O is a major performance bottleneck of web proxy server. Today's most web proxy sowers are design to run on top of a general purpose file system. But general purpose file system can not efficiently handle web cache workload, small files, leading to the performance degradation of entire web proxy servers. In this paper we evaluate the performance potential of raw disk to reduce disk I/O overhead of web proxy servers. To show the performance potential of raw disk, we design a storage management system called Block-structured Storage Management System (BSMS). And we also actually implement web proxy server that incorporate BSMS in Squid. Comprehensive experimental evaluations show that raw disk can be a good solution to improve disk I/O performance significantly for web proxy servers.

A Study on the Block Lookup and Replacement in Global Memory (전역적 메모리에서의 블록 룩업과 재배치에 관한 연구)

  • 이영섭;김은경;정병수
    • Proceedings of the IEEK Conference
    • /
    • 2000.11c
    • /
    • pp.51-54
    • /
    • 2000
  • Due to the emerging of high-speed network, lots of interests of access to remote data have increased. Those interests motivate using of Cooperative Caching that uses remote cache like local cache by sharing other clients' cache. The conventional algorithm like GMS(Global Memory Service) has some disadvantages that occurred bottleneck and decreasing performance because of exchanges of many messages to server or manager. On the other hand, Hint-based algorithm resolves a GMS's server bottleneck as each client has hint information of all blocks. But Hint-based algorithm also causes some problems such as inaccurate information in it, if it has too old hint information. In this paper, we offer the policy that supplement bottleneck and inaccuracy; by using file identifier that can search for the lookup table and by exchanging oldest block information between each client periodically.

  • PDF

Caching and Prefetching Policies Using Program Page Reference Patterns on a File System Layer for NAND Flash Memory (NAND 플래시 메모리용 파일 시스템 계층에서 프로그램의 페이지 참조 패턴을 고려한 캐싱 및 선반입 정책)

  • Kim, Gyeong-San;Kim, Seong-Jo
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.777-778
    • /
    • 2006
  • In this thesis, we design and implement a Flash Cache Core Module (FCCM) which operates on the YAFFS NAND flash memory. The FCCM applies memory replacement policy and prefetching policy based on the page reference pattern of applications. Also, implement the Clean-First memory replacement technique considering the characteristics of flash memory. In this method the decision is made according to page hit to apply prefetched waiting area. The FCCM decrease I/O hit frequency up to 37%, Compared with the linux cache and prefetching policy. Also, it operated using less memory for prefetching(maximum 24% and average 16%) compared with the linux kernel.

  • PDF

A Caching Scheme of Meta-Information for the Linux Cluster File System (리눅스 클러스터 파일 시스템을 위한 메타정보 캐쉬 기법)

  • 홍재연;김형식
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.10c
    • /
    • pp.316-318
    • /
    • 2002
  • 클러스터 파일 시스템은 고가용성(high availability) 및 결함내성(fault tolerance)을 만족하고 확장성이 뛰어나기 때문에 멀티미디어 서비스 등으로 활용범위를 넓혀왔다. 클러스터 시스템에서 일반적으로 제공되는 단일 시스템 이미지 (single system image) 기술은 저장된 위치에 관계없이 디렉토리나 파일에 접근할 수 있는 장점을 제공하지만 실제 저장된 위치에 따라 접근 시간의 편차가 발생된다. 본 논문에서는 파일이나 디렉토리에 관한 메타정보를 캐쉬에 저장함으로써 클러스터 파일 시스템에서의 접근 시간을 단축하기 위한 방법을 제안한다. 파일과 디렉토리의 접근 형태에 적합한 캐쉬 배치(placement) 기법과 재배치(replacement) 기법을 제시하고 캐쉬 일관성 유지를 위한 알고리즘을 보인다. 제안된 방법은 멀티미디어 서비스 등의 응용에서 효과적으로 접근시간을 단축할 수 있을 것으로 예상된다.

  • PDF

An Efficient Cache Management Scheme for Load Balancing in Distributed Environments with Different Memory Sizes (상이한 메모리 크기를 가지는 분산 환경에서 부하 분산을 위한 캐시 관리 기법)

  • Choi, Kitae;Yoon, Sangwon;Park, Jaeyeol;Lim, Jongtae;Lee, Seokhee;Bok, Kyoungsoo;Yoo, Jaesoo
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.8
    • /
    • pp.543-548
    • /
    • 2015
  • Recently, volume of data has been growing dramatically along with the growth of social media and digital devices. However, the existing disk-based distributed file systems have limits to their performance of data processing or data access, due to I/O processing costs and bottlenecks. To solve this problem, the caching technique is being used to manage data in the memory. In this paper, we propose a cache management scheme to handle load balancing in a distributed memory environment. The proposed scheme distributes the data according to the memory size, n distributed environments with different memory sizes. If overloaded nodes occur, it redistributes the the access time of the caching data. In order to show the superiority of the proposed scheme, we compare it with an existing distributed cache management scheme through performance evaluation.