• Title/Summary/Keyword: Distributed Cache

Search Result 94, Processing Time 0.027 seconds

A Study on Location-based Routing Technique for Improving the Performance of P2P in MANET (MANET에서 P2P 성능 향상을 위한 위치기반 라우팅 기법에 관한 연구)

  • Yang, Hwanseok
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.11 no.2
    • /
    • pp.37-45
    • /
    • 2015
  • The technology development of MANET and dissemination of P2P services has been made very widely. In particular, the development of many application services for the integration of P2P services in MANET has been made actively. P2P networks are commonly used because of the advantages of efficient use of network bandwidth and rapid information exchange. In P2P network, the infrastructure managing each node in the middle does not exist and each node is a structure playing a role as the sender and receiver. Such a structure is very similar to the structure of the MANET. However, it is difficult to provide reliable P2P service due to the high mobility of mobile nodes. In this paper, we propose location-based routing technique in order to provide efficient file sharing and management between nodes. GMN managing the group is elected after network is configured to the area of a certain size. Each node is assigned an identifier of 12 bit dynamically to provide routing which uses location information to the identifier. ZGT is managed in the GMN in order to provide management of group nodes and distributed cache information. The distributed cache technique is applied to provide a rapid retrieval of the sharing files in the each node. The excellent performance of the proposed technique was confirmed through experiments.

A Caching Mechanism for Remote Queries in Distributed Directory Systems (분산 디렉토리 시스템에서의 원격 질의에 대한 캐싱 기법)

  • Lee, Kang-Woo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.1
    • /
    • pp.50-56
    • /
    • 2000
  • In this paper, for improving the speed of query processing on distributed directory system, we proposed a caching mechanism which is store the queries and their on the remote site objects in the cache of local site. For this, first, cached information which is stored in distributed directory systems is classified as application data and system data. And cache system architecture is designed according to classified information. Second least-TTL replacement mechanism which uses the weighted value of geographical information and access frequency for replacements are developed for each cache. Finally, performance evaluations are performed by comparing the proposed caching mechanism and other mechanisms (LRU, LFU replacement). Our least-TTL mechanism shows a performance improvement of 25% over the LRU and that of 30% over LFU.

  • PDF

The Need of Cache Partitioning on Shared Cache of Integrated Graphics Processor between CPU and GPU (내장형 GPU 환경에서 CPU-GPU 간의 공유 캐시에서의 캐시 분할 방식의 필요성)

  • Sung, Hanul;Eom, Hyeonsang;Yeom, HeonYoung
    • KIISE Transactions on Computing Practices
    • /
    • v.20 no.9
    • /
    • pp.507-512
    • /
    • 2014
  • Recently, Distributed computing processing begins using both CPU(Central processing unit) and GPU(Graphic processing unit) to improve the performance to overcome darksilicon problem which cannot use all of the transistors because of the electric power limitation. There is an integrated graphics processor that CPU and GPU share memory and Last level cache(LLC). But, There is no LLC access rules between CPU and GPU, so if GPU and CPU processes run together at the same time, performance of both processes gets worse because of the contention on the LLC. This Paper gives evidence to prove the need of the Cache Partitioning and is mentioned about the cache partitioning design using page coloring to allocate the L3 Cache space only for the GPU process to guarantee GPU process performance.

Distributed Cache for High-Performance in real time cloud (실시간 클라우드 환경에서 HDFS의 고 성능을 위한 분산캐시)

  • Choi, Ji Hyeon;Youn, Hee Yong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2014.07a
    • /
    • pp.351-352
    • /
    • 2014
  • 분산 파일시스템은 서로 분산된 여러 서버들을 가지고 파일 시스템을 구성함으로써 높은 확장성과 고가용성을 지원한다. HDFS는 대용량 데이터 저장장치로 처리되고 있지만 실시간 파일 접근에 관한 고려는 부족하다. 파일을 읽을 때 네임노드와 데이터 노드는 상호 작용을 하지만 엄청난 대용량의 데이터 그리고 동시작업량이 많을 때 접근수행속가 급격하게 감소하게 된다. 따라서 실시간 클라우드 서비스 환경에서 HDFS 파일 접근 수행속도를 향상시키기 위한 연구가 이슈이다. 본 논문에서는 HDFS의 위에 분산 캐시를 둔 새로운 캐시시스템을 제안한다.

  • PDF

Remote Cache Replacement Policy using Processor Locality in Multi-Processor System (다중 프로세서 시스템에서 프로세서 지역성을 이용한 원격 캐쉬 교체 정책)

  • Han Sang Yoon;Kwak Jong Wook;Jhang Seong Tae;Jhon Chu Shik
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.32 no.11_12
    • /
    • pp.541-556
    • /
    • 2005
  • The memory access latency of the system has been a primary factor of performance degradation in single-processor system and multi-processor system. The remote memory access latency takes a lot of overhead over the local memory access latency especially in the distributed shared-memory system. To resolve this problem, the multi-level cache architecture that contains a remote cache in the multi-processor system has been proposed. In this paper, we propose a new cache replacement policy that improves the performance of the multi-processor system with the remote cache. If the multi-level cache keeps the multi-level inclusion(MLI) property and uses the LRU(Least Recently Used) cache replacement policy, the LRU information of the higher-level cache(a processor cache) would be different with that of the lower-level cache(a remote cache). In this situation, the replacement of a remote cache line can induce the exchange of a processor cache line that is used by the processor. It is a main factor of performance degradation in a whole system. To alleviate this disadvantage of the LRU replacement polity, the new policy analyses tht processor's remote memory access pattern of each node and uses this information to reduce the number of invalidations of the useful cache line in the higher-level cache. The new replacement policy of the remote cache can improve the performance by $3.5\%$ in maximum and $2.5\%$ in average on SPLASH-2 benchmarks, compared to the general LRU cache replacement policy.

A Bankruptcy Game for Optimize Caching Resource Allocation in Small Cell Networks

  • Zhang, Liying;Wang, Gang;Wang, Fuxiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.5
    • /
    • pp.2319-2337
    • /
    • 2019
  • In this paper, we study the distributed cooperative caching for Internet content providers in a small cell of heterogeneous network (HetNet). A general framework based on bankruptcy game model is put forth for finding the optimal caching policy. In this framework, the small cell and different content providers are modeled as bankrupt company and players, respectively. By introducing strategic decisions into the bankruptcy game, we propose a caching value assessment algorithm based on analytic hierarchy process in the framework of bankruptcy game theory to optimize the caching strategy and increase cache hit ratio. Our analysis shows that resource utilization can be improved through cooperative sharing while considering content providers' satisfaction. When the cache value is measured by multiple factors, not just popularity, the cache hit rate for user access is also increased. Simulation results show that our approach can improve the cache hit rate while ensuring the fairness of the distribution.

Reducing Outgoing Traffic of Proxy Cache by Using Client-Cluster

  • Kim Kyung-Baek;Park Dae-Yeon
    • Journal of Communications and Networks
    • /
    • v.8 no.3
    • /
    • pp.330-338
    • /
    • 2006
  • Many web cache systems and policies concerning them have been proposed. These studies, however, consider large objects less useful than small objects in terms of performance, and evict them as soon as possible. Even if this approach increases the hit rate, the byte hit rate decreases and the connections occurring over congested links to outside networks waste more bandwidth in obtaining large objects. This paper puts forth a client-cluster approach for improving the web cache system. The client-cluster is composed of the residual resources of clients and utilizes them as exclusive storage for large objects. This proposed system achieves not only a high hit rate but also a high byte hit rate, while reducing outgoing traffic. The distributed hash table (DHT) based peer-to-peer lookup protocol is utilized to manage the client-cluster. With the natural characteristics of this protocol, the proposed system with the client-cluster is self-organizing, fault-tolerant, well-balanced, and scalable. Additionally, the large objects are managed with an index based allocation method, which balances the loads of all clients well. The performance of the cache system is examined via a trace driven simulation and an effective enhancement of the proxy cache performance is demonstrated.

CACHE:Context-aware Clustering Hierarchy and Energy efficient for MANET (CACHE:상황인식 기반의 계층적 클러스터링 알고리즘에 관한 연구)

  • Mun, Chang-min;Lee, Kang-Hwan
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2009.10a
    • /
    • pp.571-573
    • /
    • 2009
  • Mobile Ad-hoc Network(MANET) needs efficient node management because the wireless network has energy constraints. Mobility of MANET would require the topology change frequently compared with a static network. To improve the routing protocol in MANET, energy efficient routing protocol would be required as well as considering the mobility would be needed. Previously proposed a hybrid routing CACH prolong the network lifetime and decrease latency. However the algorithm has a problem when node density is increase. In this paper, we propose a new method that the CACHE(Context-aware Clustering Hierarchy and Energy efficient) algorithm. The proposed analysis could not only help in defining the optimum depth of hierarchy architecture CACH utilize, but also improve the problem about node density.

  • PDF

PECAN: Peer Cache Adaptation for Peer-to-Peer Video-on-Demand Streaming

  • Kim, Jong-Tack;Bahk, Sae-Woong
    • Journal of Communications and Networks
    • /
    • v.14 no.3
    • /
    • pp.286-295
    • /
    • 2012
  • To meet the increased demand of video-on-demand (VoD) services, peer-to-peer (P2P) mesh-based multiple video approaches have been recently proposed, where each peer is able to find a video segment interested without resort to the video server. However, they have not considered the constraint of the server's upload bandwidth and the fairness between upload and download amounts at each peer. In this paper, we propose a novel P2P VoD streaming system, named peer cache adaptation (PECAN) where each peer adjusts its cache capacity adaptively to meet the server's upload bandwidth constraint and achieve the fairness. For doing so, we first propose a new cache replacement algorithm that designs the number of caches for a segment to be proportional to its popularity. Second, we mathematically prove that if the cache capacity of a peer is proportional to its segment request rate, the fairness between upload and download amounts at each peer can be achieved. Third, we propose a method that determines each peer's cache capacity adaptively according to the constraint of the server's upload bandwidth. Against the proposed design objective, some selfish peers may not follow our protocol to increase their payoff. To detect such peers, we design a simple distributed reputation and monitoring system. Through simulations, we show that PECAN meets the server upload bandwidth constraint, and achieves the fairness well at each peer. We finally verify that the control overhead in PECAN caused by the search, reputation, and monitoring systems is very small, which is an important factor for real deployment.

Cache Coherence Protocols in NUMA Multiprocessors (NUMA 다중 프로세서에서의 캐쉬 일관성 프로토콜)

  • Moh, Sang-Man;Hahn, Woo-Jong;Yoon, Suk-Han
    • Electronics and Telecommunications Trends
    • /
    • v.13 no.5 s.53
    • /
    • pp.11-22
    • /
    • 1998
  • Recently, scalable multiprocessor systems are actively developed for general-purpose computing, which are based on distributed shared memory (DSM) architecture to boost up both programmability and scalability. In this paper, we survey and analyze cache coherence protocols in non-uniform memory access (NUMA) multiprocessor systems. In particular, it has been easily inferred that specialized hardware suitable for NUMA multiprocessor systems with commodity symmetric multiprocessors (SMPs) is highly required. The cache coherence protocol combined with specialized hardware can significantly improve the performance and scalability of NUMA multiprocessor systems, providing better programmability.