• Title/Summary/Keyword: Edge Computing Model

Search Result 116, Processing Time 0.02 seconds

Analysis on Lightweight Methods of On-Device AI Vision Model for Intelligent Edge Computing Devices (지능형 엣지 컴퓨팅 기기를 위한 온디바이스 AI 비전 모델의 경량화 방식 분석)

  • Hye-Hyeon Ju;Namhi Kang
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.24 no.1
    • /
    • pp.1-8
    • /
    • 2024
  • On-device AI technology, which can operate AI models at the edge devices to support real-time processing and privacy enhancement, is attracting attention. As intelligent IoT is applied to various industries, services utilizing the on-device AI technology are increasing significantly. However, general deep learning models require a lot of computational resources for inference and learning. Therefore, various lightweighting methods such as quantization and pruning have been suggested to operate deep learning models in embedded edge devices. Among the lightweighting methods, we analyze how to lightweight and apply deep learning models to edge computing devices, focusing on pruning technology in this paper. In particular, we utilize dynamic and static pruning techniques to evaluate the inference speed, accuracy, and memory usage of a lightweight AI vision model. The content analyzed in this paper can be used for intelligent video control systems or video security systems in autonomous vehicles, where real-time processing are highly required. In addition, it is expected that the content can be used more effectively in various IoT services and industries.

Survey on the Performance Enhancement in Serverless Computing: Current and Future Directions (성능 향상을 위한 서버리스 컴퓨팅 동향과 발전 방향)

  • Eunyoung Lee
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.2
    • /
    • pp.60-75
    • /
    • 2024
  • The demand of users, who want to focus on the core functionality of their applications without having to manage complex virtual environments in the cloud environment, has created a new computing model called serverless computing. Within the serverless paradigm, resource provisioning and server administration tasks are delegated to cloud services, facilitating application development exclusively focused on program logic. Serverless computing has upgraded the utilization of cloud computing by reducing the burden on cloud service users, and it is expected to become the basic model of cloud computing in the future. A serverless platform is responsible for managing the cloud virtual environment on behalf of users, and it is also responsible for executing serverless functions that compose applications in the cloud environment. Considering the characteristics of serverless computing in which users are billed in proportion to the resources used, the efficiency of the serverless platform is a very important factor for both users and service providers. This paper aims to identify various factors that affect the performance of serverless computing and analyze the latest research trends related to it. Drawing upon the analysis, the future directions for serverless computing that address key challenges and opportunities in serverless computing are proposed.

A Privacy-preserving and Energy-efficient Offloading Algorithm based on Lyapunov Optimization

  • Chen, Lu;Tang, Hongbo;Zhao, Yu;You, Wei;Wang, Kai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.8
    • /
    • pp.2490-2506
    • /
    • 2022
  • In Mobile Edge Computing (MEC), attackers can speculate and mine sensitive user information by eavesdropping wireless channel status and offloading usage pattern, leading to user privacy leakage. To solve this problem, this paper proposes a Privacy-preserving and Energy-efficient Offloading Algorithm (PEOA) based on Lyapunov optimization. In this method, a continuous Markov process offloading model with a buffer queue strategy is built first. Then the amount of privacy of offloading usage pattern in wireless channel is defined. Finally, by introducing the Lyapunov optimization, the problem of minimum average energy consumption in continuous state transition process with privacy constraints in the infinite time domain is transformed into the minimum value problem of each timeslot, which reduces the complexity of algorithms and helps obtain the optimal solution while maintaining low energy consumption. The experimental results show that, compared with other methods, PEOA can maintain the amount of privacy accumulation in the system near zero, while sustaining low average energy consumption costs. This makes it difficult for attackers to infer sensitive user information through offloading usage patterns, thus effectively protecting user privacy and safety.

Design and Evaluation of an Edge-Fog Cloud-based Hierarchical Data Delivery Scheme for IoT Applications (사물인터넷 응용을 위한 에지-포그 클라우드 기반 계층적 데이터 전달 방법의 설계 및 평가)

  • Bae, Ihn-Han
    • Journal of Internet Computing and Services
    • /
    • v.19 no.1
    • /
    • pp.37-47
    • /
    • 2018
  • The number of capabilities of Internet of Things (IoT) devices will exponentially grow over the next years. These devices may generate a vast amount of time-constrained data. In the context of IoT, data management should act as a layer between the objects and devices generating the data and the applications accessing the data for analysis purposes and services. In addition, most of IoT services will be content-centric rather than host centric to increase the data availability and the efficiency of data delivery. IoT will enable all the communication devices to be interconnected and make the data generated by or associated with devices or objects globally accessible. Also, fog computing keeps data and computation close to end users at the edge of network, and thus provides a new breed of applications and services to end users with low latency, high bandwidth, and geographically distributed. In this paper, we propose Edge-Fog cloud-based Hierarchical Data Delivery ($EFcHD^2$) method that effectively and reliably delivers IoT data to associated with IoT applications with ensuring time sensitivity. The proposed $EFcHD^2$ method stands on basis of fully decentralized hybrid of Edge and Fog compute cloud model, Edge-Fog cloud, and uses information-centric networking and bloom filters. In addition, it stores the replica of IoT data or the pre-processed feature data by edge node in the appropriate locations of Edge-Fog cloud considering the characteristic of IoT data: locality, size, time sensitivity and popularity. Then, the performance of $EFcHD^2$ method is evaluated through an analytical model, and is compared to fog server-based and Content-Centric Networking (CCN)-based data delivery methods.

Enhancing cloud computing security: A hybrid machine learning approach for detecting malicious nano-structures behavior

  • Xu Guo;T.T. Murmy
    • Advances in nano research
    • /
    • v.15 no.6
    • /
    • pp.513-520
    • /
    • 2023
  • The exponential proliferation of cutting-edge computing technologies has spurred organizations to outsource their data and computational needs. In the realm of cloud-based computing environments, ensuring robust security, encompassing principles such as confidentiality, availability, and integrity, stands as an overarching imperative. Elevating security measures beyond conventional strategies hinges on a profound comprehension of malware's multifaceted behavioral landscape. This paper presents an innovative paradigm aimed at empowering cloud service providers to adeptly model user behaviors. Our approach harnesses the power of a Particle Swarm Optimization-based Probabilistic Neural Network (PSO-PNN) for detection and recognition processes. Within the initial recognition module, user behaviors are translated into a comprehensible format, and the identification of malicious nano-structures behaviors is orchestrated through a multi-layer neural network. Leveraging the UNSW-NB15 dataset, we meticulously validate our approach, effectively characterizing diverse manifestations of malicious nano-structures behaviors exhibited by users. The experimental results unequivocally underscore the promise of our method in fortifying security monitoring and the discernment of malicious nano-structures behaviors.

Communication Resource Allocation Strategy of Internet of Vehicles Based on MEC

  • Ma, Zhiqiang
    • Journal of Information Processing Systems
    • /
    • v.18 no.3
    • /
    • pp.389-401
    • /
    • 2022
  • The business of Internet of Vehicles (IoV) is growing rapidly, and the large amount of data exchange has caused problems of large mobile network communication delay and large energy loss. A strategy for resource allocation of IoV communication based on mobile edge computing (MEC) is thus proposed. First, a model of the cloud-side collaborative cache and resource allocation system for the IoV is designed. Vehicles can offload tasks to MEC servers or neighboring vehicles for communication. Then, the communication model and the calculation model of IoV system are comprehensively analyzed. The optimization objective of minimizing delay and energy consumption is constructed. Finally, the on-board computing task is coded, and the optimization problem is transformed into a knapsack problem. The optimal resource allocation strategy is obtained through genetic algorithm. The simulation results based on the MATLAB platform show that: The proposed strategy offloads tasks to the MEC server or neighboring vehicles, making full use of system resources. In different situations, the energy consumption does not exceed 300 J and 180 J, with an average delay of 210 ms, effectively reducing system overhead and improving response speed.

A Joint Allocation Algorithm of Computing and Communication Resources Based on Reinforcement Learning in MEC System

  • Liu, Qinghua;Li, Qingping
    • Journal of Information Processing Systems
    • /
    • v.17 no.4
    • /
    • pp.721-736
    • /
    • 2021
  • For the mobile edge computing (MEC) system supporting dense network, a joint allocation algorithm of computing and communication resources based on reinforcement learning is proposed. The energy consumption of task execution is defined as the maximum energy consumption of each user's task execution in the system. Considering the constraints of task unloading, power allocation, transmission rate and calculation resource allocation, the problem of joint task unloading and resource allocation is modeled as a problem of maximum task execution energy consumption minimization. As a mixed integer nonlinear programming problem, it is difficult to be directly solve by traditional optimization methods. This paper uses reinforcement learning algorithm to solve this problem. Then, the Markov decision-making process and the theoretical basis of reinforcement learning are introduced to provide a theoretical basis for the algorithm simulation experiment. Based on the algorithm of reinforcement learning and joint allocation of communication resources, the joint optimization of data task unloading and power control strategy is carried out for each terminal device, and the local computing model and task unloading model are built. The simulation results show that the total task computation cost of the proposed algorithm is 5%-10% less than that of the two comparison algorithms under the same task input. At the same time, the total task computation cost of the proposed algorithm is more than 5% less than that of the two new comparison algorithms.

Fat Client-Based Abstraction Model of Unstructured Data for Context-Aware Service in Edge Computing Environment (에지 컴퓨팅 환경에서의 상황인지 서비스를 위한 팻 클라이언트 기반 비정형 데이터 추상화 방법)

  • Kim, Do Hyung;Mun, Jong Hyeok;Park, Yoo Sang;Choi, Jong Sun;Choi, Jae Young
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.3
    • /
    • pp.59-70
    • /
    • 2021
  • With the recent advancements in the Internet of Things, context-aware system that provides customized services become important to consider. The existing context-aware systems analyze data generated around the user and abstract the context information that expresses the state of situations. However, these datasets is mostly unstructured and have difficulty in processing with simple approaches. Therefore, providing context-aware services using the datasets should be managed in simplified method. One of examples that should be considered as the unstructured datasets is a deep learning application. Processes in deep learning applications have a strong coupling in a way of abstracting dataset from the acquisition to analysis phases, it has less flexible when the target analysis model or applications are modified in functional scalability. Therefore, an abstraction model that separates the phases and process the unstructured dataset for analysis is proposed. The proposed abstraction utilizes a description name Analysis Model Description Language(AMDL) to deploy the analysis phases by each fat client is a specifically designed instance for resource-oriented tasks in edge computing environments how to handle different analysis applications and its factors using the AMDL and Fat client profiles. The experiment shows functional scalability through examples of AMDL and Fat client profiles targeting a vehicle image recognition model for vehicle access control notification service, and conducts process-by-process monitoring for collection-preprocessing-analysis of unstructured data.

Development of an intelligent edge computing device equipped with on-device AI vision model (온디바이스 AI 비전 모델이 탑재된 지능형 엣지 컴퓨팅 기기 개발)

  • Kang, Namhi
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.5
    • /
    • pp.17-22
    • /
    • 2022
  • In this paper, we design a lightweight embedded device that can support intelligent edge computing, and show that the device quickly detects an object in an image input from a camera device in real time. The proposed system can be applied to environments without pre-installed infrastructure, such as an intelligent video control system for industrial sites or military areas, or video security systems mounted on autonomous vehicles such as drones. The On-Device AI(Artificial intelligence) technology is increasingly required for the widespread application of intelligent vision recognition systems. Computing offloading from an image data acquisition device to a nearby edge device enables fast service with less network and system resources than AI services performed in the cloud. In addition, it is expected to be safely applied to various industries as it can reduce the attack surface vulnerable to various hacking attacks and minimize the disclosure of sensitive data.

TinyML Gamma Radiation Classifier

  • Moez Altayeb;Marco Zennaro;Ermanno Pietrosemoli
    • Nuclear Engineering and Technology
    • /
    • v.55 no.2
    • /
    • pp.443-451
    • /
    • 2023
  • Machine Learning has introduced many solutions in data science, but its application in IoT faces significant challenges, due to the limitations in memory size and processing capability of constrained devices. In this paper we design an automatic gamma radiation detection and identification embedded system that exploits the power of TinyML in a SiPM micro radiation sensor leveraging the Edge Impulse platform. The model is trained using real gamma source data enhanced by software augmentation algorithms. Tests show high accuracy in real time processing. This design has promising applications in general-purpose radiation detection and identification, nuclear safety, medical diagnosis and it is also amenable for deployment in small satellites.