• Title/Summary/Keyword: Edge Computing Model

Search Result 116, Processing Time 0.023 seconds

Resource Allocation and Offloading Decisions of D2D Collaborative UAV-assisted MEC Systems

  • Jie Lu;Wenjiang Feng;Dan Pu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.1
    • /
    • pp.211-232
    • /
    • 2024
  • In this paper, we consider the resource allocation and offloading decisions of device-to-device (D2D) cooperative UAV-assisted mobile edge computing (MEC) system, where the device with task request is served by unmanned aerial vehicle (UAV) equipped with MEC server and D2D device with idle resources. On the one hand, to ensure the fairness of time-delay sensitive devices, when UAV computing resources are relatively sufficient, an optimization model is established to minimize the maximum delay of device computing tasks. The original non-convex objective problem is decomposed into two subproblems, and the suboptimal solution of the optimization problem is obtained by alternate iteration of two subproblems. On the other hand, when the device only needs to complete the task within a tolerable delay, we consider the offloading priorities of task to minimize UAV computing resources. Then we build the model of joint offloading decision and power allocation optimization. Through theoretical analysis based on KKT conditions, we elicit the relationship between the amount of computing task data and the optimal resource allocation. The simulation results show that the D2D cooperation scheme proposed in this paper is effective in reducing the completion delay of computing tasks and saving UAV computing resources.

Emotion-aware Task Scheduling for Autonomous Vehicles in Software-defined Edge Networks

  • Sun, Mengmeng;Zhang, Lianming;Mei, Jing;Dong, Pingping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.11
    • /
    • pp.3523-3543
    • /
    • 2022
  • Autonomous vehicles are gradually being regarded as the mainstream trend of future development of the automobile industry. Autonomous driving networks generate many intensive and delay-sensitive computing tasks. The storage space, computing power, and battery capacity of autonomous vehicle terminals cannot meet the resource requirements of the tasks. In this paper, we focus on the task scheduling problem of autonomous driving in software-defined edge networks. By analyzing the intensive and delay-sensitive computing tasks of autonomous vehicles, we propose an emotion model that is related to task urgency and changes with execution time and propose an optimal base station (BS) task scheduling (OBSTS) algorithm. Task sentiment is an important factor that changes with the length of time that computing tasks with different urgency levels remain in the queue. The algorithm uses task sentiment as a performance indicator to measure task scheduling. Experimental results show that the OBSTS algorithm can more effectively meet the intensive and delay-sensitive requirements of vehicle terminals for network resources and improve user service experience.

Resource Allocation Strategy of Internet of Vehicles Using Reinforcement Learning

  • Xi, Hongqi;Sun, Huijuan
    • Journal of Information Processing Systems
    • /
    • v.18 no.3
    • /
    • pp.443-456
    • /
    • 2022
  • An efficient and reasonable resource allocation strategy can greatly improve the service quality of Internet of Vehicles (IoV). However, most of the current allocation methods have overestimation problem, and it is difficult to provide high-performance IoV network services. To solve this problem, this paper proposes a network resource allocation strategy based on deep learning network model DDQN. Firstly, the method implements the refined modeling of IoV model, including communication model, user layer computing model, edge layer offloading model, mobile model, etc., similar to the actual complex IoV application scenario. Then, the DDQN network model is used to calculate and solve the mathematical model of resource allocation. By decoupling the selection of target Q value action and the calculation of target Q value, the phenomenon of overestimation is avoided. It can provide higher-quality network services and ensure superior computing and processing performance in actual complex scenarios. Finally, simulation results show that the proposed method can maintain the network delay within 65 ms and show excellent network performance in high concurrency and complex scenes with task data volume of 500 kbits.

Fundamental Function Design of Real-Time Unmanned Monitoring System Applying YOLOv5s on NVIDIA TX2TM AI Edge Computing Platform

  • LEE, SI HYUN
    • International journal of advanced smart convergence
    • /
    • v.11 no.2
    • /
    • pp.22-29
    • /
    • 2022
  • In this paper, for the purpose of designing an real-time unmanned monitoring system, the YOLOv5s (small) object detection model was applied on the NVIDIA TX2TM AI (Artificial Intelligence) edge computing platform in order to design the fundamental function of an unmanned monitoring system that can detect objects in real time. YOLOv5s was applied to the our real-time unmanned monitoring system based on the performance evaluation of object detection algorithms (for example, R-CNN, SSD, RetinaNet, and YOLOv5). In addition, the performance of the four YOLOv5 models (small, medium, large, and xlarge) was compared and evaluated. Furthermore, based on these results, the YOLOv5s model suitable for the design purpose of this paper was ported to the NVIDIA TX2TM AI edge computing system and it was confirmed that it operates normally. The real-time unmanned monitoring system designed as a result of the research can be applied to various application fields such as an security or monitoring system. Future research is to apply NMS (Non-Maximum Suppression) modification, model reconstruction, and parallel processing programming techniques using CUDA (Compute Unified Device Architecture) for the improvement of object detection speed and performance.

A Monitoring Scheme Based on Artificial Intelligence in Mobile Edge Cloud Computing Environments (모바일 엣지 클라우드 환경에서 인공지능 기반 모니터링 기법)

  • Lim, JongBeom;Choi, HeeSeok;Yu, HeonChang
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.2
    • /
    • pp.27-32
    • /
    • 2018
  • One of the crucial issues in mobile edge cloud computing environments is to monitor mobile devices. Due to the inherit properties of mobile devices, they are prone to unstable behavior that leads to failures. In order to satisfy the service level agreement (SLA), the mobile edge cloud administrators should take appropriate measures through a monitoring scheme. In this paper, we propose a monitoring scheme of mobile devices based on artificial intelligence in mobile edge cloud computing environments. The proposed monitoring scheme is able to measure faults of mobile devices based on previous and current monitoring information. To this end, we adapt the hidden markov chain model, one of the artificial intelligence technologies, to monitor mobile devices. We validate our monitoring scheme based on the hidden markov chain model. The proposed monitoring scheme can also be used in general cloud computing environments to monitor virtual machines.

Task offloading under deterministic demand for vehicular edge computing

  • Haotian Li ;Xujie Li ;Fei Shen
    • ETRI Journal
    • /
    • v.45 no.4
    • /
    • pp.627-635
    • /
    • 2023
  • In vehicular edge computing (VEC) networks, the rapid expansion of intelligent transportation and the corresponding enormous numbers of tasks bring stringent requirements on timely task offloading. However, many tasks typically appear within a short period rather than arriving simultaneously, which makes it difficult to realize effective and efficient resource scheduling. In addition, some key information about tasks could be learned due to the regular data collection and uploading processes of sensors, which may contribute to developing effective offloading strategies. Thus, in this paper, we propose a model that considers the deterministic demand of multiple tasks. It is possible to generate effective resource reservations or early preparation decisions in offloading strategies if some feature information of the deterministic demand can be obtained in advance. We formulate our scenario as a 0-1 programming problem to minimize the average delay of tasks and transform it into a convex form. Finally, we proposed an efficient optimal offloading algorithm that uses the interior point method. Simulation results demonstrate that the proposed algorithm has great advantages in optimizing offloading utility.

Efficient Task Offloading Decision Based on Task Size Prediction Model and Genetic Algorithm

  • Quan T. Ngo;Dat Van Anh Duong;Seokhoon Yoon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.3
    • /
    • pp.16-26
    • /
    • 2024
  • Mobile edge computing (MEC) plays a crucial role in improving the performance of resource-constrained mobile devices by offloading computation-intensive tasks to nearby edge servers. However, existing methods often neglect the critical consideration of future task requirements when making offloading decisions. In this paper, we propose an innovative approach that addresses this limitation. Our method leverages recurrent neural networks (RNNs) to predict task sizes for future time slots. Incorporating this predictive capability enables more informed offloading decisions that account for upcoming computational demands. We employ genetic algorithms (GAs) to fine-tune fitness functions for current and future time slots to optimize offloading decisions. Our objective is twofold: minimizing total processing time and reducing energy consumption. By considering future task requirements, our approach achieves more efficient resource utilization. We validate our method using a real-world dataset from Google-cluster. Experimental results demonstrate that our proposed approach outperforms baseline methods, highlighting its effectiveness in MEC systems.

Stability-based On-demand Multi-path Distance Vector Protocol for Edge Internet of Things

  • Dongzhi Cao;Peng Liang;Tongjuan Wu;Shiqiang Zhang;Zhenhu Ning
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.10
    • /
    • pp.2658-2681
    • /
    • 2023
  • In edge computing scenarios, IoT end devices play a crucial role in relaying and forwarding data to significantly improve IoT network performance. However, traditional routing mechanisms are not applicable to this scenario due to differences in network size and environment. Therefore, it becomes crucial to establish an effective and reliable data transmission path to ensure secure communication between devices. In this paper, we propose a trusted path selection strategy that comprehensively considers multiple attributes, such as link stability and edge cooperation, and selects a stable and secure data transmission path based on the link life cycle, energy level, trust level, and authentication status. In addition, we propose the Stability-based On-demand Multipath Distance Vector (STAOMDV) protocol based on the Ad hoc AOMDV protocol. The STAOMDV protocol implements the collection and updating of link stability attributes during the route discovery and maintenance process. By integrating the STAOMDV protocol with the proposed path selection strategy, a dependable and efficient routing mechanism is established for IoT networks in edge computing scenarios. Simulation results validate that the proposed STAOMDV model achieves a balance in network energy consumption and extends the overall network lifespan.

Design of A new Algorithm by Using Standard Deviation Techniques in Multi Edge Computing with IoT Application

  • HASNAIN A. ALMASHHADANI;XIAOHENG DENG;OSAMAH R. AL-HWAIDI;SARMAD T. ABDUL-SAMAD;MOHAMMED M. IBRAHM;SUHAIB N. ABDUL LATIF
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.4
    • /
    • pp.1147-1161
    • /
    • 2023
  • The Internet of Things (IoT) requires a new processing model that will allow scalability in cloud computing while reducing time delay caused by data transmission within a network. Such a model can be achieved by using resources that are closer to the user, i.e., by relying on edge computing (EC). The amount of IoT data also grows with an increase in the number of IoT devices. However, building such a flexible model within a heterogeneous environment is difficult in terms of resources. Moreover, the increasing demand for IoT services necessitates shortening time delay and response time by achieving effective load balancing. IoT devices are expected to generate huge amounts of data within a short amount of time. They will be dynamically deployed, and IoT services will be provided to EC devices or cloud servers to minimize resource costs while meeting the latency and quality of service (QoS) constraints of IoT applications when IoT devices are at the endpoint. EC is an emerging solution to the data processing problem in IoT. In this study, we improve the load balancing process and distribute resources fairly to tasks, which, in turn, will improve QoS in cloud and reduce processing time, and consequently, response time.

Stacked Sparse Autoencoder-DeepCNN Model Trained on CICIDS2017 Dataset for Network Intrusion Detection (네트워크 침입 탐지를 위해 CICIDS2017 데이터셋으로 학습한 Stacked Sparse Autoencoder-DeepCNN 모델)

  • Lee, Jong-Hwa;Kim, Jong-Wouk;Choi, Mi-Jung
    • KNOM Review
    • /
    • v.24 no.2
    • /
    • pp.24-34
    • /
    • 2021
  • Service providers using edge computing provide a high level of service. As a result, devices store important information in inner storage and have become a target of the latest cyberattacks, which are more difficult to detect. Although experts use a security system such as intrusion detection systems, the existing intrusion systems have low detection accuracy. Therefore, in this paper, we proposed a machine learning model for more accurate intrusion detections of devices in edge computing. The proposed model is a hybrid model that combines a stacked sparse autoencoder (SSAE) and a convolutional neural network (CNN) to extract important feature vectors from the input data using sparsity constraints. To find the optimal model, we compared and analyzed the performance as adjusting the sparsity coefficient of SSAE. As a result, the model showed the highest accuracy as a 96.9% using the sparsity constraints. Therefore, the model showed the highest performance when model trains only important features.