• Title/Summary/Keyword: Caching Optimization

검색결과 24건 처리시간 0.018초

모바일 애드 혹 스트리밍 서비스를 위한 프록시 기반 캐싱 최적화 (Proxy-based Caching Optimization for Mobile Ad Hoc Streaming Services)

  • 이종득
    • 디지털융복합연구
    • /
    • 제10권4호
    • /
    • pp.207-215
    • /
    • 2012
  • 본 논문에서는 무선 모바일 애드 혹 네트워크상에서 스트리밍 미디어 서비스 향상을 위한 프록시 기반 캐싱 최적화 기법을 제안한다. 제안된 기법은 WLAN상에서 미디어 서버와 노드들 간의 통신을 위해 프록시를 사용하며, 프록시는 무선 엑세스 포인터 근처에 설치한다. 그리고 캐싱 최적화가 수행되도록 NFCO(non-full cache optimization)과 CFO(cache full optimization)기법을 제안한다. NFCO와 CFO는 프록시에서 스트리밍이 수행될 때 캐시의 성능을 최적화하기 위해 사용된다. 제안된 기법의 성능을 알아보기 위하여 본 논문에서는 제안된 기법과 서버-중심 스트리밍 기법, 그리고 전송율 왜곡 스트리밍 기법과의 성능을 비교 분석하였다. 그 결과 제안된 기법이 서버-중심 스트리밍 기법과 전송율 왜곡 스트리밍 기법에 비해서 최적화 성능이 우수함을 알게 되었다.

Joint Optimization Algorithm Based on DCA for Three-tier Caching in Heterogeneous Cellular Networks

  • Zhang, Jun;Zhu, Qi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권7호
    • /
    • pp.2650-2667
    • /
    • 2021
  • In this paper, we derive the expression of the cache hitting probability with random caching policy and propose the joint optimization algorithm based on difference of convex algorithm (DCA) in the three-tier caching heterogeneous cellular network assisted by macro base stations, helpers and users. Under the constraint of the caching capacity of caching devices, we establish the optimization problem to maximize the cache hitting probability of the network. In order to solve this problem, a convex function is introduced to convert the nonconvex problem to a difference of convex (DC) problem and then we utilize DCA to obtain the optimal caching probability of macro base stations, helpers and users for each content respectively. Simulation results show that when the density of caching devices is relatively low, popular contents should be cached to achieve a good performance. However, when the density of caching devices is relatively high, each content ought to be cached evenly. The algorithm proposed in this paper can achieve the higher cache hitting probability with the same density.

Many-objective joint optimization for dependency-aware task offloading and service caching in mobile edge computing

  • Xiangyu Shi;Zhixia Zhang;Zhihua Cui;Xingjuan Cai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제18권5호
    • /
    • pp.1238-1259
    • /
    • 2024
  • Previous studies on joint optimization of computation offloading and service caching policies in Mobile Edge Computing (MEC) have often neglected the impact of dependency-aware subtasks, edge server resource constraints, and multiple users on policy formulation. To remedy this deficiency, this paper proposes a many-objective joint optimization dependency-aware task offloading and service caching model (MaJDTOSC). MaJDTOSC considers the impact of dependencies between subtasks on the joint optimization problem of task offloading and service caching in multi-user, resource-constrained MEC scenarios, and takes the task completion time, energy consumption, subtask hit rate, load variability, and storage resource utilization as optimization objectives. Meanwhile, in order to better solve MaJDTOSC, a many-objective evolutionary algorithm TSMSNSGAIII based on a three-stage mating selection strategy is proposed. Simulation results show that TSMSNSGAIII exhibits an excellent and stable performance in solving MaJDTOSC with different number of users setting and can converge faster. Therefore, it is believed that TSMSNSGAIII can provide appropriate sub-task offloading and service caching strategies in multi-user and resource-constrained MEC scenarios, which can greatly improve the system offloading efficiency and enhance the user experience.

Mobile Device-to-Device (D2D) Content Delivery Networking: A Design and Optimization Framework

  • Kang, Hye Joong;Kang, Chung Gu
    • Journal of Communications and Networks
    • /
    • 제16권5호
    • /
    • pp.568-577
    • /
    • 2014
  • We consider a mobile content delivery network (mCDN) in which special mobile devices designated as caching servers (caching-server device: CSD) can provide mobile stations with popular contents on demand via device-to-device (D2D) communication links. On the assumption that mobile CSD's are randomly distributed by a Poisson point process (PPP), an optimization problem is formulated to determine the probability of storing the individual content in each server in a manner that minimizes the average caching failure rate. Further, we present a low-complexity search algorithm, optimum dual-solution searching algorithm (ODSA), for solving this optimization problem. We demonstrate that the proposed ODSA takes fewer iterations, on the order of O(log N) searches, for caching N contents in the system to find the optimal solution, as compared to the number of iterations in the conventional subgradient method, with an acceptable accuracy in practice. Furthermore, we identify the important characteristics of the optimal caching policies in the mobile environment that would serve as a useful aid in designing the mCDN.

Device Caching Strategy Maximizing Expected Content Quality

  • Choi, Minseok
    • 한국컴퓨터정보학회논문지
    • /
    • 제26권1호
    • /
    • pp.111-118
    • /
    • 2021
  • 본 논문에서는 디바이스 캐싱 네트워크에서 다양한 퀄리티의 콘텐츠를 캐싱하는 기술을 제안한다. 하나의 파일을 온전히 캐싱하는 기존 기술들과 다르게, 저자는 콘텐츠의 일부 조각을 캐싱하는 것을 허용하였고, 사용자가 스스로 캐시 히트를 달성할 수 있는 경우를 고려하였다. 캐싱하는 콘텐츠의 퀄리티와 캐시 히트율 간의 트레이드오프를 분석하고, 사용자가 소비하는 콘텐츠의 기대 퀄리티를 최대화하는 디바이스 캐싱 기법을 제안한다. 퀄리티와 파일 크기의 관계 파라미터에 따라 볼록 최적화 문제와 DC programming 문제 두 가지 방식으로 나누어서 캐싱 문제를 풀어냈다. 퀄리티 증가 폭에 비해 파일 크기가 더 빠르게 증가하면, 인기도에 따라 캐싱할 콘텐츠의 부분 조각이 점차 증가하는 반면, 파일 크기가 더 느리게 증가하면, 일부 인기도가 높은 콘텐츠는 전체를 캐싱하고 그렇지 않은 콘텐츠는 아예 캐싱하지 않는 결과를 낸다.

MDS Coded Caching for Device-to-Device Content Sharing Against Eavesdropping

  • Shi, Xin;Wu, Dan;Wang, Meng;Yang, Lianxin;Wu, Yan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권9호
    • /
    • pp.4484-4501
    • /
    • 2019
  • In this paper, we put forward a delay-aware secure maximum distance separable (MDS) coded caching scheme to resist the eavesdropping attacks for device-to-device (D2D) content sharing by combining MDS coding with distributed caching. In particular, we define the average system delay to show the potential coupling of delay-content awareness, and learn the secure constraints to ensure that randomly distributed eavesdroppers cannot obtain enough encoded packets to recover their desired contents. Accordingly, we model such a caching problem as an optimization problem to minimize the average system delay with secure constraints and simplify it to its convex relaxation. Then we develop a delay-aware secure MDS coded caching algorithm to obtain the optimal caching policy. Extensive numerical results are provided to demonstrate the excellent performance of our proposed algorithm. Compared with the random coded caching scheme, uniform coded caching scheme and popularity based coded caching scheme, our proposed scheme has 3.7%, 3.3% and 0.7% performance gains, respectively.

클라우드 스토리지 최적화를 위한 고속 캐싱 및 대용량 파일 전송 기법 (A Scheme on High-Performance Caching and High-Capacity File Transmission for Cloud Storage Optimization)

  • 김태훈;김정한;엄영익
    • 한국통신학회논문지
    • /
    • 제37권8C호
    • /
    • pp.670-679
    • /
    • 2012
  • 최근 클라우드 컴퓨팅 환경의 보급과 함께 스토리지의 데이터양이 급증함에 따라 그에 따른 스토리지 저장 비용이 빠르게 증가하고 있다. 더불어, 사용자들의 다양한 서비스 및 데이터 요청으로 클라우드 스토리지의 부하 또한 급증하고 있다. 이러한 문제를 해결하기 위해 분산 파일 시스템을 통한 저비용 고성능 스토리지 환경을 제공하고자 하는 기존의 연구가 있었으나, 이에는 데이터 병렬처리, 임의위치 접근처리, 빈번한 작은 워크로드 접근처리 등의 취약점이 존재한다. 최근에는 캐싱 기술을 이용하여 이를 개선하려는 연구가 주목받고 있다. 본 논문에서는 분산 파일 시스템 환경에서 병렬 캐싱, 분산 캐싱과 공유 자원을 고려한 데이터 병렬 전송방법을 제공하는 CHPC(Cloud storage High-Performance Caching) 구조를 제안하며, 또한 이를 기존의 방법들과 비교 평가하여 스토리지 부하를 최적화하는 방법을 제시한다. 더불어, 제안 기법이 기존 클라우드 시스템에 비하여 스토리지 서버의 디스크 입출력 감소, 서버로 데이터의 요청이 집중되어 발생하는 병목현상 방지, 각 클라이언트의 중복되는 페이지 캐시 제거, 데이터 전송률 향상의 장점을 가짐을 보인다.

다중스레드 모델의 성능 향상을 위한 가용 레지스터 기반 캐슁 기법 (A Register-Based Caching Technique for the Advanced Performance of Multithreaded Models)

  • 고훈준;권영필;유원희
    • 정보처리학회논문지A
    • /
    • 제8A권2호
    • /
    • pp.107-116
    • /
    • 2001
  • A multithreaded model is a hybrid one which combines locality of execution of the von Neumann model with asynchronous data availability and implicit parallelism of the dataflow model. Much researches that have been made toward the advanced performance of multithreaded models are about the cache memory which have been proved to be efficient in the von Neumann model. To use an instruction cache or operand cache, the multithreaded models must have cache memories. If cache memories are added to the multithreaded model, they may have the disadvantage of high implementation cost in the mode. To solve these problems, we did not add cache memory but applied the method of executing the caching by using available registers of the multithreaded models. The available register-based caching method is one that use the registers which are not used on the execution of threads. It may accomplish the same effect as the cache memory. The multithreaded models can compute the number of available registers to be used during the process of the register optimization, and therefore this method can be easily applied on the models. By applying this method, we can also remove the access conflict and the bottleneck of frame memories. When we applied the proposed available register-based caching method, we found that there was an improved performance of the multithreaded model. Also, when the available-register-based caching method is compared with the cache based caching method, we found that there was the almost same execution overhead.

  • PDF

A Novel Service Migration Method Based on Content Caching and Network Condition Awareness in Ultra-Dense Networks

  • Zhou, Chenjun;Zhu, Xiaorong;Zhu, Hongbo;Zhao, Su
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권6호
    • /
    • pp.2680-2696
    • /
    • 2018
  • The collaborative content caching system is an effective solution developed in recent years to reduce transmission delay and network traffic. In order to decrease the service end-to-end transmission delay for future 5G ultra-dense networks (UDN), this paper proposes a novel service migration method that can guarantee the continuity of service and simultaneously reduce the traffic flow in the network. In this paper, we propose a service migration optimization model that minimizes the cumulative transmission delay within the constraints of quality of service (QoS) guarantee and network condition. Subsequently, we propose an improved firefly algorithm to solve this optimization problem. Simulation results show that compared to traditional collaborative content caching schemes, the proposed algorithm can significantly decrease transmission delay and network traffic flow.

Cache Optimization on Hot-Point Proxy Caching Using Weighted-Rank Cache Replacement Policy

  • Ponnusamy, S.P.;Karthikeyan, E.
    • ETRI Journal
    • /
    • 제35권4호
    • /
    • pp.687-696
    • /
    • 2013
  • The development of proxy caching is essential in the area of video-on-demand (VoD) to meet users' expectations. VoD requires high bandwidth and creates high traffic due to the nature of media. Many researchers have developed proxy caching models to reduce bandwidth consumption and traffic. Proxy caching keeps part of a media object to meet the viewing expectations of users without delay and provides interactive playback. If the caching is done continuously, the entire cache space will be exhausted at one stage. Hence, the proxy server must apply cache replacement policies to replace existing objects and allocate the cache space for the incoming objects. Researchers have developed many cache replacement policies by considering several parameters, such as recency, access frequency, cost of retrieval, and size of the object. In this paper, the Weighted-Rank Cache replacement Policy (WRCP) is proposed. This policy uses such parameters as access frequency, aging, and mean access gap ratio and such functions as size and cost of retrieval. The WRCP applies our previously developed proxy caching model, Hot-Point Proxy, at four levels of replacement, depending on the cache requirement. Simulation results show that the WRCP outperforms our earlier model, the Dual Cache Replacement Policy.