• Title/Summary/Keyword: Cache server

Search Result 220, Processing Time 0.026 seconds

Cache Policies for WWW Sewers to Reduce Workload (웹 서버 작업부하 감소를 위한 캐시 정책)

  • Lim Jaehyun
    • Journal of Internet Computing and Services
    • /
    • v.5 no.6
    • /
    • pp.103-110
    • /
    • 2004
  • In this paper we study and analyze the influence of caching strategies on the performance of WWW servers, We propose a new strategy called file type based caching that aims to obtain a well-balanced mixture between large and small files in the cache, and moreover, it provides good performance for both small and large file as expected, By using the type based caching good results are obtained for both the hit rate and the byte hit rate.

  • PDF

Efficient Content-based Load Distribution for Web Server Clusters (웹 서버 클러스터를 위한 효율적인 내용 기반의 부하 분배)

  • Chung Ji Yung;Kim Sungsoo
    • Journal of KIISE:Information Networking
    • /
    • v.32 no.1
    • /
    • pp.60-67
    • /
    • 2005
  • A cluster consists of a collection of interconnected stand-alone computers working together and provides a high-availability solution in application area such as web services or information systems. Content-based load distribution for web server clusters uses the detailed data found in the application layer to intelligently route user requests among web servers. In this paper, we propose a content-based load distribution algorithm that considers cache hit and load information of the web servers under the web server clusters. In addition, we expand this algorithm in order to manage user requests for dynamic file. Specially, our algorithm does not keep track of any frequency of access information or try to model the contents of the caches of the web servers.

Mapping Cache for High-Performance Memory Mapped File I/O in Memory File Systems (메모리 파일 시스템 기반 고성능 메모리 맵 파일 입출력을 위한 매핑 캐시)

  • Kim, Jiwon;Choi, Jungsik;Han, Hwansoo
    • Journal of KIISE
    • /
    • v.43 no.5
    • /
    • pp.524-530
    • /
    • 2016
  • The desire to access data faster and the growth of next-generation memories such as non-volatile memories, contribute to the development of research on memory file systems. It is recommended that memory mapped file I/O, which has less overhead than read-write I/O, is utilized in a high-performance memory file system. Memory mapped file I/O, however, brings a page table overhead, which becomes one of the big overheads that needs to be resolved in the entire file I/O performance. We find that same overheads occur unnecessarily, because a page table of a file is removed whenever a file is opened after being closed. To remove the duplicated overhead, we propose the mapping cache, a technique that does not delete a page table of a file but saves the page table to be reused when the mapping of the file is released. We demonstrate that mapping cache improves the performance of traditional file I/O by 2.8x and web server performance by 12%.

A User-Level File System for Streaming Media Caching (스트리밍 미디어 캐슁을 위한 사용자 수준 화일 시스템)

  • Oh, Jae-Hak;Cha, Ho-Jeong
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.8
    • /
    • pp.472-483
    • /
    • 2002
  • This paper presents the design and implementation of a cache file system, umcFS, which is specifically designed to provide an efficient caching and transmission of streaming media. The proposed file system is based on the concept of file disk and implemented as an application level on top of a general-purpose file system. The file disk favors the continuity of cached media and provides an efficient I/O mechanism for cache server. umcFS statically allocates control blocks as well as media cache blocks. These blocks are referenced by the single-level indirect management structure. As the file system is designed as an application level, it is easy to develop and port to other systems. The performance of the implemented system shows that umcFS performs about 13% better than the native file system in randomly accessing the cache blocks of 1024KB.

General Web Cache Implementation Using NIO (NIO를 이용한 범용 웹 캐시 구현)

  • Lee, Chul-Hui;Shin, Yong-Hyeon
    • Journal of Advanced Navigation Technology
    • /
    • v.20 no.1
    • /
    • pp.79-85
    • /
    • 2016
  • Network traffic is increased rapidly, due to mobile and social network, such as smartphones and facebook, in recent web environment. In this paper, we improved web response time of existing system using direct buffer of NIO and DMA. This solved the disadvantage of JAVA, such as CPU performance reduction due to the blocking of I/O, garbage collection of buffer. Key values circulated many data due to priority change put on a hash map operated easily and apply a priority modification algorithm. Large response data is separated and stored at a fast direct buffer and improved performance. This paper showed that the proposed method using NIO was much improved performance, in many test situations of cache hit and cache miss.

Automatic Detection of Memory Subsystem Parameters for Embedded Systems (임베디드 시스템을 위한 메모리 서브시스템 파라미터의 자동 검출)

  • Ha, Tae-Jun;Seo, Sang-Min;Chun, Po-Sung;Lee, Jae-Jin
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.5
    • /
    • pp.350-354
    • /
    • 2009
  • To optimize the performance of software programs, it is important to know certain hardware parameters such as the CPU speed, the cache size, the number of TLB entries, and the parameters of the memory subsystem. There exist several ways to obtain the values of various hardware parameters. Firstly. the values can be taken from the hardware manual. Secondly, the parameters can be obtained by calling functions provided by the operating systems. Finally, hardware detection programs can find the desired values. Such programs are usually executed on PC or server systems and report the CPU speed, the cache size, the number of TLB entries, and so on. However, they do not sufficiently detect the parameters of one of the most important parts of the computer concerning performance, namely the memory bank layout in the memory subsystem. In this paper, we present an algorithm to detect the memory bank parameters. We run an implementation of our algorithm on various embedded systems and compare the detected values with the real hardware parameters. The results show that the presented algorithm detects the cache size, the number of TLB entries, and the memory bank layout with high accuracy.

An Address Translation Technique Large NAND Flash Memory using Page Level Mapping (페이지 단위 매핑 기반 대용량 NAND플래시를 위한 주소변환기법)

  • Seo, Hyun-Min;Kwon, Oh-Hoon;Park, Jun-Seok;Koh, Kern
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.3
    • /
    • pp.371-375
    • /
    • 2010
  • SSD is a storage medium based on NAND Flash memory. Because of its short latency, low power consumption, and resistance to shock, it's not only used in PC but also in server computers. Most SSDs use FTL to overcome the erase-before-overwrite characteristic of NAND flash. There are several types of FTL, but page mapped FTL shows better performance than others. But its usefulness is limited because of its large memory footprint for the mapping table. For example, 64MB memory space is required only for the mapping table for a 64GB MLC SSD. In this paper, we propose a novel caching scheme for the mapping table. By using the mapping-table-meta-data we construct a fully associative cache, and translate the address within O(1) time. The simulation results show more than 80 hit ratio with 32KB cache and 90% with 512KB cache. The overall memory footprint was only 1.9% of 64MB. The time overhead of cache miss was measured lower than 2% for most workload.

Improving Search Performance of Tries Data Structures for Network Filtering by Using Cache (네트워크 필터링에서 캐시를 적용한 트라이 구조의 탐색 성능 개선)

  • Kim, Hoyeon;Chung, Kyusik
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.3 no.6
    • /
    • pp.179-188
    • /
    • 2014
  • Due to the tremendous amount and its rapid increase of network traffic, the performance of network equipments are becoming an important issue. Network filtering is one of primary functions affecting the performance of the network equipment such as a firewall or a load balancer to process the packet. In this paper, we propose a cache based tri method to improve the performance of the existing tri method of searching for network filtering. When several packets are exchanged at a time between a server and a client, the tri method repeats the same search procedure for network filtering. However, the proposed method can avoid unnecessary repetition of search procedure by exploiting cache so that the performance of network filtering can be improved. We performed network filtering experiments for the existing method and the proposed method. Experimental results showed that the proposed method could process more packets up to 790,000 per second than the existing method. When the size of cache list is 11, the proposed method showed the most outstanding performance improvement (18.08%) with respect to memory usage increase (7.75%).

Improvement of Partial Update for the Web Map Tile Service (실시간 타일 지도 서비스를 위한 타일이미지 갱신 향상 기법)

  • Cho, Sunghwan;Ga, Chillo;Yu, Kiyun
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.5
    • /
    • pp.365-373
    • /
    • 2013
  • Tile caching technology is a commonly used method that optimizes the delivery of map imagery across the internet in modern WebGIS systems. However the poor performance of the map tile cache update is one of the major causes that hamper the wider use of this technique for datasets with frequent updates. In this paper, we introduce a new algorithm, namely, Partial Area Cache Update (PACU) that significantly minimizes redundant update of map tiles where the update frequency of source map data is very large. The performance of our algorithm is verified with the cadastral map data of Pyeongtaek of Gyeonggi Province, where approximately 3,100 changes occur in a day among the 331,594 parcels. The experiment results show that the performance of the PACU algorithm is 6.6 times faster than the ESRI ArcGIS SERVER$^{(r)}$. This algorithm significantly contributes in solving the frequent update problem and enable Web Map Tile Services for data that requires frequent update.

A Shared Cache Directory based Wireless Internet Proxy Server Cluster (공유 캐시 디렉토리 기반의 무선 인터넷 프록시 서버 클러스터)

  • Kwak Hu-Keun;Chung Kyu-Sik
    • The KIPS Transactions:PartA
    • /
    • v.13A no.4 s.101
    • /
    • pp.343-350
    • /
    • 2006
  • In this paper, wireless internet proxy server clusters are used for the wireless internet because their caching, distillation, and clustering functions are helpful to overcome the limitations and needs of the wireless internet. A wireless Internet proxy server cluster needs a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. In our former research, we proposed the CD-A structure which can be scalable in a systematic way and has a simple communication structure but it has no cooperative caching. A hash based load balancing can be used to solve the problem, but it can not deal with Hot Spot request problem. In this paper, we proposed a shared storage based wireless internet proxy server cluster which has a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. The proposed method shares one cache directory and it has advantages: advantages of the existing CD-A structure, cooperative caching, and serving Hot Spot requests. We performed experiments using 16 PCs and experimental results show high performance improvement of the proposed system compared to the existing systems in Hot Spot requests.