• Title/Summary/Keyword: Pre-caching Technology

Search Result 5, Processing Time 0.019 seconds

A Bankruptcy Game for Optimize Caching Resource Allocation in Small Cell Networks

  • Zhang, Liying;Wang, Gang;Wang, Fuxiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.5
    • /
    • pp.2319-2337
    • /
    • 2019
  • In this paper, we study the distributed cooperative caching for Internet content providers in a small cell of heterogeneous network (HetNet). A general framework based on bankruptcy game model is put forth for finding the optimal caching policy. In this framework, the small cell and different content providers are modeled as bankrupt company and players, respectively. By introducing strategic decisions into the bankruptcy game, we propose a caching value assessment algorithm based on analytic hierarchy process in the framework of bankruptcy game theory to optimize the caching strategy and increase cache hit ratio. Our analysis shows that resource utilization can be improved through cooperative sharing while considering content providers' satisfaction. When the cache value is measured by multiple factors, not just popularity, the cache hit rate for user access is also increased. Simulation results show that our approach can improve the cache hit rate while ensuring the fairness of the distribution.

Novel Pre-pushing & Downloading Model in Mobile Peer-assisted Streaming Network

  • Lin, Fuhong;Zhou, Xianewi;Chen, Changjia
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.12
    • /
    • pp.3135-3148
    • /
    • 2013
  • The popularization of streaming networks and mobile devices brings a new kind of network called mobile peer-assisted streaming network. In this network, service provider not only offers downloading services, but also pre-pushes resources to users for caching. Users can download their demanded resources while working as servers for uploading the cached data. Also the mobile characteristic makes high package losing probability in this network. So we study how the service provider pre-pushes or the user downloads resources efficiently and effectively while considering the package losing in this paper. We build utility models for service provider and user, and solve the models using Bellman's theory to achieve Nash Equilibrium which shows the service provider's optimal pre-pushing speed and user's optimal downloading speed. The numerical simulation demonstrates the efficiency and effectiveness of our proposed pre-pushing and downloading scheme by comparing to the traditional scheme.

Content Distribution for 5G Systems Based on Distributed Cloud Service Network Architecture

  • Jiang, Lirong;Feng, Gang;Qin, Shuang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.11
    • /
    • pp.4268-4290
    • /
    • 2015
  • Future mobile communications face enormous challenges as traditional voice services are replaced with increasing mobile multimedia and data services. To address the vast data traffic volume and the requirement of user Quality of Experience (QoE) in the next generation mobile networks, it is imperative to develop efficient content distribution technique, aiming at significantly reducing redundant data transmissions and improving content delivery performance. On the other hand, in recent years cloud computing as a promising new content-centric paradigm is exploited to fulfil the multimedia requirements by provisioning data and computing resources on demand. In this paper, we propose a cooperative caching framework which implements State based Content Distribution (SCD) algorithm for future mobile networks. In our proposed framework, cloud service providers deploy a plurality of cloudlets in the network forming a Distributed Cloud Service Network (DCSN), and pre-allocate content services in local cloudlets to avoid redundant content transmissions. We use content popularity and content state which is determined by content requests, editorial updates and new arrivals to formulate a content distribution optimization model. Data contents are deployed in local cloudlets according to the optimal solution to achieve the lowest average content delivery latency. We use simulation experiments to validate the effectiveness of our proposed framework. Numerical results show that the proposed framework can significantly improve content cache hit rate, reduce content delivery latency and outbound traffic volume in comparison with known existing caching strategies.

Pre-Processing of Query Logs in Web Usage Mining

  • Abdullah, Norhaiza Ya;Husin, Husna Sarirah;Ramadhani, Herny;Nadarajan, Shanmuga Vivekanada
    • Industrial Engineering and Management Systems
    • /
    • v.11 no.1
    • /
    • pp.82-86
    • /
    • 2012
  • In For the past few years, query log data has been collected to find user's behavior in using the site. Many researches have studied on the usage of query logs to extract user's preference, recommend personalization, improve caching and pre-fetching of Web objects, build better adaptive user interfaces, and also to improve Web search for a search engine application. A query log contain data such as the client's IP address, time and date of request, the resources or page requested, status of request HTTP method used and the type of browser and operating system. A query log can offer valuable insight into web site usage. A proper compilation and interpretation of query log can provide a baseline of statistics that indicate the usage levels of website and can be used as tool to assist decision making in management activities. In this paper we want to discuss on the tasks performed of query logs in pre-processing of web usage mining. We will use query logs from an online newspaper company. The query logs will undergo pre-processing stage, in which the clickstream data is cleaned and partitioned into a set of user interactions which will represent the activities of each user during their visits to the site. The query logs will undergo essential task in pre-processing which are data cleaning and user identification.

Research on Web Cache Infection Methods and Countermeasures (웹 캐시 감염 방법 및 대응책 연구)

  • Hong, Sunghyuck;Han, Kun-Hee
    • Journal of Convergence for Information Technology
    • /
    • v.9 no.2
    • /
    • pp.17-22
    • /
    • 2019
  • Cache is a technique that improves the client's response time, thereby reducing the bandwidth and showing an effective side. However, there are vulnerabilities in the cache technique as well as in some techniques. Web caching is convenient, but it can be exploited by hacking and cause problems. Web cache problems are mainly caused by cache misses and excessive cache line fetch. If the cache miss is high and excessive, the cache will become a vulnerability, causing errors such as transforming the secure data and causing problems for both the client and the system of the user. If the user is aware of the cache infection and the countermeasure against the error, the user will no longer feel the cache error or the problem of the infection occurrence. Therefore, this study proposed countermeasures against four kinds of cache infections and errors, and suggested countermeasures against web cache infections.