• Title/Summary/Keyword: Cloud offloading

Search Result 51, Processing Time 0.018 seconds

Mobile Energy Efficiency Study using Cloud Computing in LTE (LTE에서 클라우드 컴퓨팅을 이용한 모바일 에너지 효율 연구)

  • Jo, Bokyun;Suh, Doug Young
    • Journal of Broadcast Engineering
    • /
    • v.19 no.1
    • /
    • pp.24-30
    • /
    • 2014
  • This study investigates computing offloading effect of cloud in real-time video personal broadcast service, whose server is mobile device. Mobile device does not have enough computing resource for encoding video. The computing burden is offloaded to cloud, which has abundant resources in terms of computing, power, and storage compared to mobile device. By reducing computing burden, computation energy can be saved while transmission data amount increases because of decreasing compression efficiency. This study shows that the optimal operation point can be found adaptively to time-varying LTE communication condition result of tradeoff analysis between offloaded computation burden and increase in amount of transmitted data.

A Context-aware Task Offloading Scheme in Collaborative Vehicular Edge Computing Systems

  • Jin, Zilong;Zhang, Chengbo;Zhao, Guanzhe;Jin, Yuanfeng;Zhang, Lejun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.2
    • /
    • pp.383-403
    • /
    • 2021
  • With the development of mobile edge computing (MEC), some late-model application technologies, such as self-driving, augmented reality (AR) and traffic perception, emerge as the times require. Nevertheless, the high-latency and low-reliability of the traditional cloud computing solutions are difficult to meet the requirement of growing smart cars (SCs) with computing-intensive applications. Hence, this paper studies an efficient offloading decision and resource allocation scheme in collaborative vehicular edge computing networks with multiple SCs and multiple MEC servers to reduce latency. To solve this problem with effect, we propose a context-aware offloading strategy based on differential evolution algorithm (DE) by considering vehicle mobility, roadside units (RSUs) coverage, vehicle priority. On this basis, an autoregressive integrated moving average (ARIMA) model is employed to predict idle computing resources according to the base station traffic in different periods. Simulation results demonstrate that the practical performance of the context-aware vehicular task offloading (CAVTO) optimization scheme could reduce the system delay significantly.

An Overview of Mobile Edge Computing: Architecture, Technology and Direction

  • Rasheed, Arslan;Chong, Peter Han Joo;Ho, Ivan Wang-Hei;Li, Xue Jun;Liu, William
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.10
    • /
    • pp.4849-4864
    • /
    • 2019
  • Modern applications such as augmented reality, connected vehicles, video streaming and gaming have stringent requirements on latency, bandwidth and computation resources. The explosion in data generation by mobile devices has further exacerbated the situation. Mobile Edge Computing (MEC) is a recent addition to the edge computing paradigm that amalgamates the cloud computing capabilities with cellular communications. The concept of MEC is to relocate the cloud capabilities to the edge of the network for yielding ultra-low latency, high computation, high bandwidth, low burden on the core network, enhanced quality of experience (QoE), and efficient resource utilization. In this paper, we provide a comprehensive overview on different traits of MEC including its use cases, architecture, computation offloading, security, economic aspects, research challenges, and potential future directions.

A Dynamic Task Distribution approach using Clustering of Data Centers and Virtual Machine Migration in Mobile Cloud Computing (모바일 클라우드 컴퓨팅에서 데이터센터 클러스터링과 가상기계 이주를 이용한 동적 태스크 분배방법)

  • Mateo, John Cristopher A.;Lee, Jaewan
    • Journal of Internet Computing and Services
    • /
    • v.17 no.6
    • /
    • pp.103-111
    • /
    • 2016
  • Offloading tasks from mobile devices to available cloud servers were improved since the introduction of the cloudlet. With the implementation of dynamic offloading algorithms, mobile devices can choose the appropriate server for the set of tasks. However, current task distribution approaches do not consider the number of VM, which can be a critical factor in the decision making. This paper proposes a dynamic task distribution on clustered data centers. A proportional VM migration approach is also proposed, where it migrates virtual machines to the cloud servers proportionally according to their allocated CPU, in order to prevent overloading of resources in servers. Moreover, we included the resource capacity of each data center in terms of the maximum CPU in order to improve the migration approach in cloud servers. Simulation results show that the proposed mechanism for task distribution greatly improves the overall performance of the system.

A Novel Architecture for Mobile Crowd and Cloud computing for Health care

  • kumar, Rethina;Ganapathy, Gopinath;Kang, Jeong-Jin
    • International Journal of Advanced Culture Technology
    • /
    • v.6 no.4
    • /
    • pp.226-232
    • /
    • 2018
  • The rapid pace of growth in internet usage and rich mobile applications and with the advantage of incredible usage of internet enabled mobile devices the Green Mobile Crowd Computing will be the suitable area to research combining with cloud services architecture. Our proposed Framework will deploy the eHealth among various health care sectors and pave a way to create a Green Mobile Application to provide a better and secured way to access the Products/ Information/ Knowledge, eHealth services, experts / doctors globally. This green mobile crowd computing and cloud architecture for healthcare information systems are expected to lower costs, improve efficiency and reduce error by also providing better consumer care and service with great transparency to the patient universally in the field of medical health information technology. Here we introduced novel architecture to use of cloud services with crowd sourcing.

Validation of Cloud Robotics System in 5G MEC for Remote Execution of Robot Engines (5G MEC 기반 로봇 엔진 원격 구동을 위한 클라우드 로보틱스 시스템 구성 및 실증)

  • Gu, Sewan;Kang, Sungkyu;Jeong, Wonhong;Moon, Hyungil;Yang, Hyunseok;Kim, Youngjae
    • The Journal of Korea Robotics Society
    • /
    • v.17 no.2
    • /
    • pp.118-123
    • /
    • 2022
  • We implemented a real-time cloud robotics application by offloading robot navigation engine over to 5G Mobile Edge Computing (MEC) sever. We also ran a fleet management system (FMS) in the server and controlled the movements of multiple robots at the same time. The mobile robots under the test were connected to the server through 5G SA network. Public 5G network, which is already commercialized, has been temporarily modified to support this validation by the network operator. Robot engines are containerized based on micro-service architecture and have been deployed using Kubernetes - a container orchestration tool. We successfully demonstrated that mobile robots are able to avoid obstacles in real-time when the engines are remotely running in 5G MEC server. Test results are compared with 5G Public Cloud and 4G (LTE) Public Cloud as well.

A Study on MEC Network Application Functions for Autonomous Driving (자율주행을 위한 MEC 적용 기능의 연구)

  • Kang-Hyun Nam
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.3
    • /
    • pp.427-432
    • /
    • 2023
  • In this study, MEC (: Multi-access Edge Computing) proposes a cloud service network configuration for various tests of autonomous vehicles to which V2X (: Vehicle to Everything) is applied in Wave, LTE, and 5G networks and MEC App (: Application) applied V2X service function test verification of two domains (operator (KT, SKT, LG U+), network type (Wave, LTE (including 3G), 5G)) in a specific region. In 4G networks of domestic operators (SKT, KT, LG U+ and Wave), MEC summarized the improvement effects through V2X function blocks and traffic offloading for the purpose of bringing independent network functions. And with a high level of QoS value in the V2X VNF of the 5G network, the traffic steering function scenario was demonstrated on the destination-specific traffic path.

Collaborative Inference for Deep Neural Networks in Edge Environments

  • Meizhao Liu;Yingcheng Gu;Sen Dong;Liu Wei;Kai Liu;Yuting Yan;Yu Song;Huanyu Cheng;Lei Tang;Sheng Zhang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.7
    • /
    • pp.1749-1773
    • /
    • 2024
  • Recent advances in deep neural networks (DNNs) have greatly improved the accuracy and universality of various intelligent applications, at the expense of increasing model size and computational demand. Since the resources of end devices are often too limited to deploy a complete DNN model, offloading DNN inference tasks to cloud servers is a common approach to meet this gap. However, due to the limited bandwidth of WAN and the long distance between end devices and cloud servers, this approach may lead to significant data transmission latency. Therefore, device-edge collaborative inference has emerged as a promising paradigm to accelerate the execution of DNN inference tasks where DNN models are partitioned to be sequentially executed in both end devices and edge servers. Nevertheless, collaborative inference in heterogeneous edge environments with multiple edge servers, end devices and DNN tasks has been overlooked in previous research. To fill this gap, we investigate the optimization problem of collaborative inference in a heterogeneous system and propose a scheme CIS, i.e., collaborative inference scheme, which jointly combines DNN partition, task offloading and scheduling to reduce the average weighted inference latency. CIS decomposes the problem into three parts to achieve the optimal average weighted inference latency. In addition, we build a prototype that implements CIS and conducts extensive experiments to demonstrate the scheme's effectiveness and efficiency. Experiments show that CIS reduces 29% to 71% on the average weighted inference latency compared to the other four existing schemes.

CTaG: An Innovative Approach for Optimizing Recovery Time in Cloud Environment

  • Hung, Pham Phuoc;Aazam, Mohammad;Huh, Eui-Nam
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.4
    • /
    • pp.1282-1301
    • /
    • 2015
  • Traditional infrastructure has been superseded by cloud computing, due to its cost-effective and ubiquitous computing model. Cloud computing not only brings multitude of opportunities, but it also bears some challenges. One of the key challenges it faces is recovery of computing nodes, when an Information Technology (IT) failure occurs. Since cloud computing mainly depends upon its nodes, physical servers, that makes it very crucial to recover a failed node in time and seamlessly, so that the customer gets an expected level of service. Work has already been done in this regard, but it has still proved to be trivial. In this study, we present a Cost-Time aware Genetic scheduling algorithm, referred to as CTaG, not only to globally optimize the performance of the cloud system, but also perform recovery of failed nodes efficiently. While modeling our work, we have particularly taken into account the factors of network bandwidth and customer's monetary cost. We have implemented our algorithm and justify it through extensive simulations and comparison with similar existing studies. The results show performance gain of our work over the others, in some particular scenarios.

Service Image Placement Mechanism Based on the Logical Fog Network (논리적 포그 네트워크 기반의 서비스 이미지 배치 기법)

  • Choi, Jonghwa;Ahn, Sanghyun
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.11
    • /
    • pp.250-255
    • /
    • 2020
  • For the resolution of the latency problem of the cloud center-based cloud computing, fog computing was proposed that allows end devices to offload computations to nearby fog nodes. In the fog computing, virtualized service images are placed on fog nodes and, if service images are placed close to end devices, the duplicate service image placement problem may occur. Therefore, in this paper, we propose a service image placement mechanism based on the logical fog network that reduces duplicate service images by considering the pattern of collected service requests. For the performance evaluation of the proposed mechanism, through simulations, we compare ours with the on-demand mechanism placing a service image upon the receipt of a service request. We consider the performance factors like the number of service images, the number of non-accommodated service requests, and the network cost.