• Title/Summary/Keyword: Local Cache

Search Result 72, Processing Time 0.023 seconds

Doughnut: An improved P2P Pastry Overlay Network with Efficient Locality and Caching (Doughnut: 효율적인 지역성 및 캐슁을 사용하는 향상된 P2P Pastry 오버레이 네트워크)

  • Kim, Myung-Won;Kwak, Hu-Keun;Chung, Kyu-Sik
    • The KIPS Transactions:PartC
    • /
    • v.16C no.2
    • /
    • pp.245-256
    • /
    • 2009
  • Pastry overlay network is one of structured P2Ps using DHT(Distributed Hash Table). To reduce the number of messages among nodes, Rosary and LAR have been proposed by exploiting spatial locality and caching, respectively, in the Pastry. Rosary consists of Inter-Pastry and Intra-Pastry. A root node is assigned as a representative in each Intra-Pastry and it has the responsibility of Inter-Pastry and Intra-Pastry routing. Therefore, Rosary has several disadvantages; 1) low fault tolerance in case of root node failure 2) routing hop count increases because of the use of root nodes compared to the existing structured P2Ps, and 3) the communication load is concentrated in some specific areas. LAR has inefficient problems in that caching is not distributed among nodes in Intra-Pastry and caching is used by only nodes in the Intra-Pastry. In this paper, we propose an improved Pastry called Doughnut to overcome the above problems of Rosary and LAR. By dividing nodes with the local characteristics, the Doughnut consists of Inter-Pastry and Intra-Pastry, and all nodes have the responsibility of Inter-Pastry and Intra-Pastry routing. This results in that all nodes perform the role of the existing root node. This solves the problems of the reducing of fault-tolerance, the increasing of routing hop count, and the not-distributed communication load. Also Doughnut can use cache effectively because it guarantees the even cache distribution in local(Intra-Pastry) and the cache contents in local can be used in the other local. The proposed algorithm is implemented using simulator and the experimental results show the effectiveness of the proposed method compared to the existing method.

A Study on the Route Optimization with Cache Information on PMIPv6 with Inter-LMA (Inter-LMA 환경을 고려한 PMIPv6 환경에서의 캐쉬 정보를 활용한 경로 최적화 기법에 관한 연구)

  • Choi, Young-Hyun;Kim, Ho-Yeon;Chung, Tai-Myoung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2011.11a
    • /
    • pp.617-620
    • /
    • 2011
  • 본 논문은 Inter-Local Mobility Anchor 환경에서의 경로 최적화 기법을 Proxy Mobile IPv6 환경에서 캐쉬 정보를 활용하여 삼각 라우팅 문제를 해결하기 위해 경로 최적화 수행 시 빠른 처리 속도와 안정성을 가지는 방법에 대해 연구하였다. 기존의 경로 최적화 방법 중 inter-Local Mobility Anchor 환경에서 Liebsch의 방법은 모든 경우를 확인하기에 어떤 상황에서도 동일한 시그널링 코스트가 사용되지만 경로 최적화를 위한 프록시 바인딩의 시그널링 코스트가 많고, Dutta의 방법은 경로 최적화를 위한 프록시 바인딩의 시그널링 코스트는 효과적으로 감소시켰으나 Mobile Access Gateway 사이의 프록시 바인딩 과정이 없음으로 인하여 발생할 수 있는 문제점을 간과하였다. 본 논문에서는 Inter-Local Mobility Anchor 환경에서도 시그널링 비용은 증가시키지 않으면서 Local Mobility Anchor와 Mobile Access Gateway로부터 메시지를 받을 경우 응답 메시지를 전송하여 신뢰성을 향상시킨 경로 최적화 방법을 캐쉬 정보를 활용하는 방안에 관하여 연구하였다.

Asynchronous Cache Invalidation Strategy to Support Read-Only Transaction in Mobile Environments (이동 컴퓨팅 환경에서 읽기-전용 트랜잭션을 지원하기 위한 비동기적 캐쉬 무효화 기법)

  • Kim, Il-Do;Nam, Sung-Hun
    • The KIPS Transactions:PartC
    • /
    • v.10C no.3
    • /
    • pp.325-334
    • /
    • 2003
  • In stateless server, if an asynchronous cache invalidation scheme attempts to support local processing of read-only transaction in mobile client/sever database systems, a critical problem may occur ; the asynchronous invalidation reports provide no guarantees of waiting time for mobile transactions requesting commit. To solve this problem, the server in our algorithm broadcasts two kind of messages, asynchronous invalidation report to reduce transaction latency and periodic guide message to avoid the uncertainty of waiting time for the next invalidation report. The asynchronous invalidation report has its own sequence number and the periodic guide message has the sequence number of the most recently broadcast asynchronous invalidation report. A mobile client checks its cache validity by using the sequence numbers of these messages.

A Local Buffer Allocation Scheme for Multimedia Data on Linux (리눅스 상에서 멀티미디어 데이타를 고려한 지역 버퍼 할당 기법)

  • 신동재;박성용;양지훈
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.9 no.4
    • /
    • pp.410-419
    • /
    • 2003
  • The buffer cache of general operating systems such as Linux manages file data by using global block replacement policy and read ahead. As a result, multimedia data with a low locality of reference and various consumption rate have low cache hit ratio and consume additional buffers because of read ahead. In this paper we have designed and implemented a new buffer allocation algorithm for multimedia data on Linux. Our approach keeps one read-ahead cache per every opened multimedia file and dynamically changes the read-ahead group size based on the buffer consumption rate of the file. This distributes resources fairly and optimizes the buffer consumption. This paper compares the system performance with that of Linux 2.4.17 in terms of buffer consumption and buffer hit ratio.

A Cache Consistency Control for B-Tree Indices in a Database Sharing System (데이타베이스 공유 시스템에서 B-트리 인덱스를 위한 캐쉬 일관성 제어)

  • On, Gyeong-O;Jo, Haeng-Rae
    • The KIPS Transactions:PartD
    • /
    • v.8D no.5
    • /
    • pp.593-604
    • /
    • 2001
  • A database sharing system (DSS) refers to a system for high performance transaction processing. In the DSS, the processing nodes are coupled via a high speed network and share a common database at the disk level. Each node has a local memory and a separate copy of operating system. To reduce the number of disk accesses, the node caches data pages and index pages in its memory buffer. In general, B-tree index pages are accessed more often and thus cached at more processing nodes, than their corresponding data pages. There are also complicated operations in the B-tree such as Fetch, Fetch Next, Insertion and Deletion. Therefore, an efficient cache consistency scheme supporting high level concurrency is required. In this paper, we propose cache consistency schemes using identifiers of index pages and page_LSN of leaf page. The propose schemes can improve the system throughput by reducing the required message traffic between nodes and index re-traversal.

  • PDF

An Enhanced Route Optimization Scheme for Multiple LMAs in PMIPv6 Domain (다중 LMA 환경을 고려한 Proxy Mobile IP 기반의 향상된 경로 최적화 방안)

  • Jang, Jong-Min;Seo, Won-Kyeong;Choi, Jea-In;Cho, You-Ze
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.1A
    • /
    • pp.82-89
    • /
    • 2011
  • The Proxy Mobile IPv6 is a network-based localized mobility management protocol. In the PMIPv6, Mobile Nodes are topologically anchored at a Local Mobility Anchor, which forwards all data packets for registered Mobile Nodes. Since all data packets destined for the Mobile Nodes a1ways traverse the Mobile Nodes's Local Mobility Anchor, the LMA might be bottleneck and the end-to-end de1ay are increased. Therefore, in this paper, we proposed an enhanced Route Optimization scheme in Multiple Local Mobility Anchors environment. In order to rapid1y detect Route Optimization, we designed Domain Information Table in Mobility Access Gateway. Furthermore, we use Correspondent Binding Cache in Local Mobility Anchor to maintain Route Optimization information during Mobile Nodes's handover. To solve packet loss and reodering problems during handover, we propose a new buffering and forwarding scheme.

Content Delivery Network Based on MST Algorithm (MST 알고리즘 기반 콘텐츠 전송 네트워크에 관한 연구)

  • Lee, Hyung-ok;Kang, Mi-young;Nam, Ji-seung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.2
    • /
    • pp.178-188
    • /
    • 2016
  • The traffic in the wired and wireless networks has increased exponentially because of increase of smart phone and improvement of PC performance. Multimedia services and file transmission such as Facebook, Youtube occupy a large part of the traffic. CDN is a technique that duplicates the contents on a remote web server of content provider to local CDN servers near clients and chooses the optimal CDN server for providing the content to the client in the event of a content request. In this paper, the content request message between CDN servers and the client used the SCRP algorithm utilizing the MST algorithm and the traffic throughput was optimized. The average response time for the content request is reduced by employing HC_LRU cache algorithm that improves the cache hit ratio. The proposed SCRP and HC_LRU algorithm may build a scalable content delivery network system that efficiently utilizes network resources, achieves traffic localization and prevents bottlenecks.

A Remote Cache Replacement Policy for the Chordal Ring Based CC-NUMA System (코달링 구조의 CC-NUMA 시스템을 위한 원격 캐쉬 교체 정책)

  • Kim Soo-Han;Kim In-Suk;Kim Bong-Joon;Jhang Seong-Tae
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.11
    • /
    • pp.643-657
    • /
    • 2004
  • The chordal ring based CC-NUMA system contains many links to transmit transactions between a local node and a remote node because of its structural characteristics. However, the inclination that the transactions concentrate on the ring link increases both the traffic of the ring link and the response time, which degrades the overall performance of the chordal ring based CC-NUMA system. In this paper we suggest a new remote cache replacement policy that considers both the number of total links and the number of ring links to traverse for the transactions. Our proposed replacement policy can balance data between the ring link and the chordal link properly because it reflects the characteristics of chordal ring based CC-NUMA system well.

Design of a Parallel Rendering Processor Architecture with Effective Memory System (효과적인 메모리 구조를 갖는 병렬 렌더링 프로세서 설계)

  • Park Woo-Chan;Yoon Duk-Ki;Kim Kyoung-Su
    • The KIPS Transactions:PartA
    • /
    • v.13A no.4 s.101
    • /
    • pp.305-316
    • /
    • 2006
  • Current rendering processors are organized mainly to process a triangle as fast as possible and recently parallel 3D rendering processors, which can process multiple triangles in parallel with multiple rasterizers, begin to appear. For high performance in processing triangles, it is desirable for each rasterizer have its own local pixel cache. However, the consistency problem may occur in accessing the data at the same address simultaneously by more than one rasterizer. In this paper, we propose a parallel rendering processor architecture resolving such consistency problem effectively. Moreover, the proposed architecture reduces the latency due to a pixel cache miss significantly. For the above two goals, effective memory organizations including a new pixel cache architecture are presented. The experimental results show that the proposed architecture achieves almost linear speedup at best case even in sixteen rasterizers.

A Cache-based Reconfigurable Accelerator in Die-stacked DRAM (3차원 구조 DRAM의 캐시 기반 재구성형 가속기)

  • Kim, Yongjoo
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.4 no.2
    • /
    • pp.41-46
    • /
    • 2015
  • The demand on low power and high performance system is soaring due to the extending of mobile and small electronic device market. The 3D die-stacking technology is widely studying for next generation integration technology due to its high density and low access time. We proposed the 3D die-stacked DRAM including a reconfigurable accelerator in a logic layer of DRAM. Also we discuss and suggest a cache-based local memory for a reconfigurable accelerator in a logic layer. The reconfigurable accelerator in logic layer of 3D die-stacked DRAM reduces the overhead of data management and transfer due to the characteristics of its location, so that can increase the performance highly. The proposed system archives 24.8 speedup in maximum.