• Title/Summary/Keyword: Network cache

Search Result 273, Processing Time 0.026 seconds

A Study on the Performance of Prefetching Web Cache Proxy (Prefetch하는 웹 캐쉬 프록시의 성능에 대한 연구)

  • 백윤철
    • Journal of the Korea Computer Industry Society
    • /
    • v.2 no.11
    • /
    • pp.1453-1464
    • /
    • 2001
  • Explosive growth of Internet populations results performance degradations of web service. Popular web sites cannot provide proper level of services to many requests, and such poor services cannot give user a satisfaction. Web caching, the remedy of this problem, reduces the amount of network traffic and gives fast response to user. In this paper, we analyze the characteristics of web cache traffics using traces of NLANR(National Laboratory for Applied Network Research) root caches and Education network cache in Seoul National University. Based on this analysis, we suggest a method of prefetching and we discuss the gains of our prefetching. As a result, we find proposed prefetching enhances hit rate up to 3%, response time up to 5%.

  • PDF

A Study on Caching Management Technique in Mobile Ad-hoc Network (Mobile Ad-hoc Network에서 캐싱 관리 기법에 관한 연구)

  • Yang, Hwan Seok;Yoo, Seung Jae
    • Convergence Security Journal
    • /
    • v.12 no.4
    • /
    • pp.91-96
    • /
    • 2012
  • MANET is developed technique fairly among many field of wireless network. Nodes which consist of MANET transmit data using multi-hop wireless connection. Caching scheme is technique which can improve data access capacity and availability of nodes. Previous studies were achieved about dynamic routing protocol to improve multi-hop connection of moving nodes. But management and maintenance of effective cache information because of movement of nodes is not easy. In this study, we proposed cluster-based caching scheme to manage connection by decreasing overhead and moving of nodes as moving node discovers cache of wish information. And HLP was used to maintain effective cache table in each cluster head. Efficiency of proposed technique in this study was confirmed by experiment.

Document Replacement Policy by Site Popularity in Web Cache (웹 캐시에서 사이트의 인기도에 의한 도큐먼트 교체정책)

  • Yoo, Hang-Suk;Jang, Tea-Mu
    • Journal of Korea Game Society
    • /
    • v.3 no.1
    • /
    • pp.67-73
    • /
    • 2003
  • Most web caches save documents temporarily into themselves on the basis of those documents. And when a corresponding document exists within the cache on wei s request, web cache sends the document to corresponding user. On the contrary, when there is not any document within the cache, web cache requests a new document to the related server to copy the document into the cache and then rum it back to user. Here, web cache uses a replacement policy to change existing document into a new one due to exceeded capacity of cache. Typical replacement policy includes document-based LRU or LFU technique and other various replacement policies are used to replace the documents within cache effectively. However, these replacement policies function only with regard to the time and frequency of document request, not considering the popularity of each web site. Based on replacement policies with regard to documents on frequent requests and the popularity of each web site, this paper aims to present the document replacement policies with regard to the popularity of each web site, which are suitable for latest network environments to enhance the hit-ratio of cache and efficiently manage the contents of cache by effectively replacing documents on intermittent requests by new ones.

  • PDF

A Chinese Restaurant Game for Distributed Cooperative Caching in Small Cell Networks

  • Chen, Junliang;Wang, Gang;Wang, Fuxiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.1
    • /
    • pp.222-236
    • /
    • 2019
  • Wireless content caching in small cell networks has recently been considered as a promising way to alleviate the congestion of the backhaul in emerging heterogenous cellular network. However, how to select files which are cached in SBSs and how to make SBSs work together is an important issue for cooperative cache research for the propose of reducing file download time. In this paper, a Cooperative-Greedy strategy (CGS) among cache-enabled small base stations (SBSs) in small cell network is proposed, in order to minimize the download time of files. This problem is formulated as a Chinese restaurant game.Using this game model, we can configure file caching schemes based on file popularity and the spectrum resources allocated to several adjacent SBSs. Both the existence and uniquencess of a Nash equilibrium are proved. In the theoretical analysis section, SBSs cooperate with each other in order to cache popular files as many as possible near UEs. Simulation results show that the CGS scheme outperforms other schemes in terms of the file-download time.

Contactless Smart Card as a Cache for Geocaching

  • Burda, Karel
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.7
    • /
    • pp.205-210
    • /
    • 2021
  • In this paper, the possibility of using a contactless smart card as a cache for geocaching is analyzed. Geocaching is an outdoor game in which players search for hidden boxes, or caches based on geographical coordinates. The problems with this game are the possibility of players cheating and the need to maintain the caches. And then there is the problem of the ignorant public accidentally discovering a cache and considering it an explosive device. This paper proposes a concept for a possible solution to the above problems by replacing the boxes with conventional contactless smart cards. Also, this concept makes geocaching more attractive by using various games. This paper proposes a system architecture as well as the cryptographic protocol required for secure communication between the player's smartphone and the card.

BIM Geometry Cache Structure for Data Streaming with Large Volume (대용량 BIM 형상 데이터 스트리밍을 위한 캐쉬 구조)

  • Kang, Tae-Wook
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.18 no.9
    • /
    • pp.1-8
    • /
    • 2017
  • The purpose of this study is to propose a cache structure for processing large-volume building information modeling (BIM) geometry data,whereit is difficult to allocate physical memory. As the number of BIM orders has increased in the public sector, it is becoming more common to visualize and calculate large-volume BIM geometry data. Design and review collaboration can require a lot of time to download large-volume BIM data through the network. If the BIM data exceeds the physical free-memory limit, visualization and geometry computation cannot be possible. In order to utilize large amounts of BIM data on insufficient physical memory or a low-bandwidth network, it is advantageous to cache only the data necessary for BIM geometry rendering and calculation time. Thisstudy proposes acache structure for efficiently rendering and calculating large-volume BIM geometry data where it is difficult to allocate enough physical memory.

Wireless Sensor Networks have Applied the Routing History Cache Routing Algorithm (무선센서 네트워크에서 Routing History Cache를 이용한 라우팅 알고리즘)

  • Lee, Doo-Wan;Jang, Kyung-Sik
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2009.10a
    • /
    • pp.1018-1021
    • /
    • 2009
  • Wireless Sensor Network collects a data from the specific area and the control is composed of small sensor nodes. Like this sensors to after that is established at the beginning are operated with the battery, the operational duration until several years must be continued from several months and will be able to apply the resources which is restricted in efficiently there must be. In this paper RHC (rounting history cache) applies in Directed Diffusion which apply a data central concept a reliability and an efficiency in data transfer course set. RHC algorithms which proposes each sensor node updated RHC of oneself with periodic and because storing the optimization course the course and, every event occurrence hour they reset the energy is wasted the fact that a reliability with minimization of duplication message improved.

  • PDF

Joint Optimization Algorithm Based on DCA for Three-tier Caching in Heterogeneous Cellular Networks

  • Zhang, Jun;Zhu, Qi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2650-2667
    • /
    • 2021
  • In this paper, we derive the expression of the cache hitting probability with random caching policy and propose the joint optimization algorithm based on difference of convex algorithm (DCA) in the three-tier caching heterogeneous cellular network assisted by macro base stations, helpers and users. Under the constraint of the caching capacity of caching devices, we establish the optimization problem to maximize the cache hitting probability of the network. In order to solve this problem, a convex function is introduced to convert the nonconvex problem to a difference of convex (DC) problem and then we utilize DCA to obtain the optimal caching probability of macro base stations, helpers and users for each content respectively. Simulation results show that when the density of caching devices is relatively low, popular contents should be cached to achieve a good performance. However, when the density of caching devices is relatively high, each content ought to be cached evenly. The algorithm proposed in this paper can achieve the higher cache hitting probability with the same density.

Document Replacement Policy by Web Site Popularity (웹 사이트의 인기도에 의한 도큐먼트 교체정책)

  • Yoo, Hang-Suk;Chang, Tae-Mu
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.1
    • /
    • pp.227-232
    • /
    • 2008
  • General web caches save documents temporarily into themselves on the basis of those documents. And when a corresponding document exists within the cache on user's request. web cache sends the document to corresponding user. On the contrary. when there is not any document within the cache, web cache requests a new document to the related server to copy the document into the cache and then turn it back to user. Here, web cache uses a replacement policy to change existing document into a new one due to exceeded capacity of cache. Typical replacement policy includes document-based LRU or LFU technique and other various replacement policies are used to replace the documents within cache effectively. However. these replacement policies function only with regard to the time and frequency of document request. not considering the popularity of each web site. Based on replacement policies with regard to documents on frequent requests and the popularity of each web site, this paper aims to present the document replacement policies with regard to the popularity of each web site, which are suitable for latest network environments to enhance the hit-ratio of cache and efficiently manage the contents of cache by effectively replacing documents on intermittent requests by new ones.

  • PDF

Keeping-ownership Cache Replacement Policies for Remote Access Caches of NUMA System (NUMA 시스템에서 소유권에 근거한 원격 캐시 교체 정책)

  • 신숭현;곽종욱;장성태;전주식
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.8
    • /
    • pp.473-486
    • /
    • 2004
  • NUMA systems have remote access caches(RAC) in each local node to reduce the overhead for repeated remote memory accesses. By this RAC, memory latency and network traffic can be reduced and the performance of the multiprocessor system can be improved. Until now, several cache replacement policies have been proposed in recent years, and there also is cache replacement policy for multiprocessor systems. In this paper, we propose a cache replacement policy which is based on cache line coherence information. In this policy, the cache line that does not have an ownership is replaced first with respect to cache line that has an ownership. Like this way, the overhead to transfer ownership is avoided and the memory latency can be decreased. We also propose “Keeping-Ownership replacement policy with MRU (KOM)” and “Keeping-Ownership replacement policy with Reference Bit(KORB)” to reduce the frequent replacement penalty of the ownership-lacking cache line. We compare and analyze these with LRU and Pseudo LRU(PLRU). The simulation shows that KOM outperforms the PLRU by 25%, and KORB outperforms the PLRU by 13%. Although the hardware cost of KOM is very small, the performance of KOM is nearly equal to that of the LRU.