• Title/Summary/Keyword: Caching System

Search Result 184, Processing Time 0.029 seconds

A Novel Service Migration Method Based on Content Caching and Network Condition Awareness in Ultra-Dense Networks

  • Zhou, Chenjun;Zhu, Xiaorong;Zhu, Hongbo;Zhao, Su
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.6
    • /
    • pp.2680-2696
    • /
    • 2018
  • The collaborative content caching system is an effective solution developed in recent years to reduce transmission delay and network traffic. In order to decrease the service end-to-end transmission delay for future 5G ultra-dense networks (UDN), this paper proposes a novel service migration method that can guarantee the continuity of service and simultaneously reduce the traffic flow in the network. In this paper, we propose a service migration optimization model that minimizes the cumulative transmission delay within the constraints of quality of service (QoS) guarantee and network condition. Subsequently, we propose an improved firefly algorithm to solve this optimization problem. Simulation results show that compared to traditional collaborative content caching schemes, the proposed algorithm can significantly decrease transmission delay and network traffic flow.

Caching Algorithm for Core Network Offloading in Smallcell Environment (소형셀 환경에서 코어망 오프로딩을 위한 캐시 알고리즘)

  • Jung, So-Yi;Kim, Jae-Hyun
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.3
    • /
    • pp.32-38
    • /
    • 2015
  • In this paper, we propose a smallcell local caching algorithm under user's context in smallcell environment. The proposed system reduces traffic to core network and the network cost, but increases it's performance. The proposed algorithm precache suitable files using smallcell's regional characteristics and target's personality. It can adjusts a storage allocation to make effective usage of our limited cache storage capacity. In order to evaluate the performance of the proposed cache algorithm, we define the cache efficiency, the decrement of core network traffic. The simulation results show that the proposed algorithm can improve performance by about 200% compared to existing web cache scheme.

Community Model for Smart TV over the Top Services

  • Pandey, Suman;Won, Young Joon;Choi, Mi-Jung;Gil, Joon-Min
    • Journal of Information Processing Systems
    • /
    • v.12 no.4
    • /
    • pp.577-590
    • /
    • 2016
  • We studied the current state-of-the-art of Smart TV, the challenges and the drawbacks. Mainly we discussed the lack of end-to-end solution. We then illustrated the differences between Smart TV and IPTV from network service provider point of view. Unlike IPTV, viewer of Smart TV's over-the-top (OTT) services could be global, such as foreign nationals in a country or viewers having special viewing preferences. Those viewers are sparsely distributed. The existing TV service deployment models over Internet are not suitable for such viewers as they are based on content popularity, hence we propose a community based service deployment methodology with proactive content caching on rendezvous points (RPs). In our proposal, RPs are intermediate nodes responsible for caching routing and decision making. The viewer's community formation is based on geographical locations and similarity of their interests. The idea of using context information to do proactive caching is itself not new, but we combined this with "in network caching" mechanism of content centric network (CCN) architecture. We gauge the performance improvement achieved by a community model. The result shows that when the total numbers of requests are same; our model can have significantly better performance, especially for sparsely distributed communities.

A Caching Strategy Considering Data Popularity in Pull-Based Data Broadcast Systems (풀 기반 데이타 방송 시스템에서의 데이타 인기도를 고려한 캐싱 전략)

  • Shin Dong-Cheon
    • Journal of KIISE:Information Networking
    • /
    • v.33 no.4
    • /
    • pp.324-332
    • /
    • 2006
  • A caching is a useful technique to alleviate performance degradation due to the inherent narrow bandwidth by reducing contention of broadcast requests. In this paper, we propose a caching strategy for pull-based data broadcast system which considers data popularity among clients. In addition, the proposed strategy also reflects recentness of data access based on data broadcast version. Then, we evaluate the performance of proposed strategy through a simulation approach. According to the results, the strategy considering both hit ratio and miss cost shows better performance than the traditional LRU. In addition, the strategy considering data popularity among clients shows better performance in some cases.

Semantic schema data processing using cache mechanism (캐쉬메카니즘을 이용한 시맨틱 스키마 데이터 처리)

  • Kim, Byung-Gon;Oh, Sung-Kyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.3
    • /
    • pp.89-97
    • /
    • 2011
  • In semantic web information system like ontology that access distributed information from network, efficient query processing requires an advanced caching mechanism to reduce the query response time. P2P network system have become an important infra structure in web environment. In P2P network system, when the query is initiated, reducing the demand of data transformation to source peer is important aspect of efficient query processing. Caching of query and query result takes a particular advantage by adding or removing a query term. Many of the answers may already be cached and can be delivered to the user right away. In web environment, semantic caching method has been proposed which manages the cache as a collection of semantic regions. In this paper, we propose the semantic caching technique in cluster environment of peers. Especially, using schema data filtering technique and schema similarity cache replacement method, we enhanced the query processing efficiency.

Caching and Prefetching Policies Using Program Page Reference Patterns on a File System Layer for NAND Flash Memory (NAND 플래시 메모리용 파일 시스템 계층에서 프로그램의 페이지 참조 패턴을 고려한 캐싱 및 선반입 정책)

  • Park, Sang-Oh;Kim, Kyung-San;Kim, Sung-Jo
    • The KIPS Transactions:PartA
    • /
    • v.14A no.4
    • /
    • pp.235-244
    • /
    • 2007
  • Caching and prefetching policies have been used in most of computer systems to compensate speed differences between primary memory and secondary storage devices. In this paper, we design and implement a Flash Cache Core Module(FCCM) on the YAFFS which operates on a file system layer for NAND flash memory. The FCCM is independent of the underlying kernel in order to support its stability and compatibility. Also, we implement the Dirty-Last memory replacement technique considering the characteristics of flash memory, and the waiting queue for pages to be prefetched according to page hit. The FCCM reduced the number of I/Os and the amount of prefetched pages by maximum 55%(20% on average) and maximum 55%(24% on average), respectively, comparing with caching and prefetching policies of Linux.

Broadcasting and Caching Schemes for Location-dependent Queries in Urban Areas (도심환경에서 위치의존 질의를 위한 방송과 캐싱 기법)

  • Jung Il-dong;Yu Young-ho;Lee Jong-hwan;Kim Kyongsok
    • Journal of KIISE:Databases
    • /
    • v.32 no.1
    • /
    • pp.56-70
    • /
    • 2005
  • The results of location-dependent queries(LDQ) generally depend on the current locations of query issuers. Many mechanisms, e.g. broadcast scheme, hoarding, or racking policy, have been developed to improve the system peformance and provide better services, which are specialized for LDQs. Considering geographical adjacency of data and characteristics oi target area, caching policy and broadcast scheme affect the overall performance in LDQ. For this reason, we propose both the caching policy and broadcast scheme, which these features are reflected in. Based on the adjacency of data in LDQ, our broadcast scheme use Hilbert curve to cluster data. Moreover, in order to develop the caching policy suitable for LDQ on urban area, we apply the moving distance of a MH(Mobile Host) to our caching policy We evaluate the performance of the caching policy measuring the workload of MHs and the correctness of LDQ results and the performance of the broadcast scheme measuring the average setup-time of MHs in our experiments. Finally, we expect that our caching policy Provides more correct answers when executing LDQ in focal cache and leads significant improvement of the performance of MHs. It also seems quite probable that our broadcast scheme leads improvement of battery life of the MH.

A Shared Cache Directory based Wireless Internet Proxy Server Cluster (공유 캐시 디렉토리 기반의 무선 인터넷 프록시 서버 클러스터)

  • Kwak Hu-Keun;Chung Kyu-Sik
    • The KIPS Transactions:PartA
    • /
    • v.13A no.4 s.101
    • /
    • pp.343-350
    • /
    • 2006
  • In this paper, wireless internet proxy server clusters are used for the wireless internet because their caching, distillation, and clustering functions are helpful to overcome the limitations and needs of the wireless internet. A wireless Internet proxy server cluster needs a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. In our former research, we proposed the CD-A structure which can be scalable in a systematic way and has a simple communication structure but it has no cooperative caching. A hash based load balancing can be used to solve the problem, but it can not deal with Hot Spot request problem. In this paper, we proposed a shared storage based wireless internet proxy server cluster which has a systematic scalability, simple communication structure, cooperative caching, and serving Hot Spot requests. The proposed method shares one cache directory and it has advantages: advantages of the existing CD-A structure, cooperative caching, and serving Hot Spot requests. We performed experiments using 16 PCs and experimental results show high performance improvement of the proposed system compared to the existing systems in Hot Spot requests.

Caching Strategies of Content-Centric Networking for the Performance of TICN Backbone (TICN 백본망 성능향상을 위한 CCN 적용 캐싱전략)

  • Park, Heungsoon;Kwon, Taewook
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.17 no.5
    • /
    • pp.637-642
    • /
    • 2014
  • Military communications have limited resources such as poor bandwidth and insufficient link capacity. Numerous hosts in conventional IP networks, including the Tactical Information Communication Network (TICN), the tactical communication system of the Army, focus heavy traffic on a single server via end-to-end connections and in which data delivery latency is caused by exchanging routing information. Content-Centric Networking (CCN), one framework of Information-Centric Networking (ICN), is a novel Internet paradigm in which data can be accessed by the name of the content instead of the location information. This paper provides content caching strategies for applying CCN to military tactical IP networks based on TICN environment.

5G Network Communication, Caching, and Computing Algorithms Based on the Two-Tier Game Model

  • Kim, Sungwook
    • ETRI Journal
    • /
    • v.40 no.1
    • /
    • pp.61-71
    • /
    • 2018
  • In this study, we developed hybrid control algorithms in smart base stations (SBSs) along with devised communication, caching, and computing techniques. In the proposed scheme, SBSs are equipped with computing power and data storage to collectively offload the computation from mobile user equipment and to cache the data from clouds. To combine in a refined manner the communication, caching, and computing algorithms, game theory is adopted to characterize competitive and cooperative interactions. The main contribution of our proposed scheme is to illuminate the ultimate synergy behind a fully integrated approach, while providing excellent adaptability and flexibility to satisfy the different performance requirements. Simulation results demonstrate that the proposed approach can outperform existing schemes by approximately 5% to 15% in terms of bandwidth utilization, access delay, and system throughput.