• Title/Summary/Keyword: Server selection algorithm

Search Result 28, Processing Time 0.021 seconds

Path selection algorithm for multi-path system based on deep Q learning (Deep Q 학습 기반의 다중경로 시스템 경로 선택 알고리즘)

  • Chung, Byung Chang;Park, Heasook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.1
    • /
    • pp.50-55
    • /
    • 2021
  • Multi-path system is a system in which utilizes various networks simultaneously. It is expected that multi-path system can enhance communication speed, reliability, security of network. In this paper, we focus on path selection in multi-path system. To select optimal path, we propose deep reinforcement learning algorithm which is rewarded by the round-trip-time (RTT) of each networks. Unlike multi-armed bandit model, deep Q learning is applied to consider rapidly changing situations. Due to the delay of RTT data, we also suggest compensation algorithm of the delayed reward. Moreover, we implement testbed learning server to evaluate the performance of proposed algorithm. The learning server contains distributed database and tensorflow module to efficiently operate deep learning algorithm. By means of simulation, we showed that the proposed algorithm has better performance than lowest RTT about 20%.

The Remote Access Algorithm by Object Replication (객체 복제 기법에 의한 원격 접근 알고리즘)

  • Yun, Dong-Sik;Lee, Byeong-Gwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.3
    • /
    • pp.799-807
    • /
    • 2000
  • In this paper, object replication Client/server under distributed computing system is design and implementation. Today many end-users have a computer communication by using internet in the distributed system of client/server. If many users request services to a specific remote server, the server should have got a overhead for hat service processing, delayed the speed for replay, and bring a bottleneck in communication network. Therefore object replication method was proposed to solve this problems. The growth of internet works and distributed applications has increased the need for large scale replicated systems. However, existing replication protocols do not address scale and autonomy issues adequately. Further, current application protocol require consistency of different levels, and therefore should be the selection function of consistency in them, in order to have particular semantics of each level. In this paper, server overhead and bottleneck happening in remote procedure call be using server object replication. Therefore access transparency can be improved by sharing object duplicately. So it will Keep up with the consistency within the replicated objects.

  • PDF

A DASH System Using the A3C-based Deep Reinforcement Learning (A3C 기반의 강화학습을 사용한 DASH 시스템)

  • Choi, Minje;Lim, Kyungshik
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.17 no.5
    • /
    • pp.297-307
    • /
    • 2022
  • The simple procedural segment selection algorithm commonly used in Dynamic Adaptive Streaming over HTTP (DASH) reveals severe weakness to provide high-quality streaming services in the integrated mobile networks of various wired and wireless links. A major issue could be how to properly cope with dynamically changing underlying network conditions. The key to meet it should be to make the segment selection algorithm much more adaptive to fluctuation of network traffics. This paper presents a system architecture that replaces the existing procedural segment selection algorithm with a deep reinforcement learning algorithm based on the Asynchronous Advantage Actor-Critic (A3C). The distributed A3C-based deep learning server is designed and implemented to allow multiple clients in different network conditions to stream videos simultaneously, collect learning data quickly, and learn asynchronously, resulting in greatly improved learning speed as the number of video clients increases. The performance analysis shows that the proposed algorithm outperforms both the conventional DASH algorithm and the Deep Q-Network algorithm in terms of the user's quality of experience and the speed of deep learning.

Distributing Network Loads in Tree-based Content Distribution System

  • Han, Seung Chul;Chung, Sungwook;Lee, Kwang-Sik;Park, Hyunmin;Shin, Minho
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.1
    • /
    • pp.22-37
    • /
    • 2013
  • Content distribution to a large number of concurrent clients stresses both server and network. While the server limitation can be circumvented by deploying server clusters, the network limitation is far less easy to cope with, due to the difficulty in measuring and balancing network load. In this paper, we use two useful network load metrics, the worst link stress (WLS) and the degree of interference (DOI), and formulate the problem as partitioning the clients into disjoint subsets subject to the server capacity constraint so that the WLS and the DOI are reduced for each session and also well balanced across the sessions. We present a network load-aware partition algorithm, which is practicable and effective in achieving the design goals. Through experiments on PlanetLab, we show that the proposed scheme has the remarkable advantages over existing schemes in reducing and balancing the network load. We expect the algorithm and performance metrics can be easily applied to various Internet applications, such as media streaming, multicast group member selection.

Band Selection Algorithm based on Expected Value for Pixel Classification (픽셀 분류를 위한 기댓값 기반 밴드 선택 알고리즘)

  • Chang, Duhyeuk;Jung, Byeonghyeon;Heo, Junyoung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.6
    • /
    • pp.107-112
    • /
    • 2022
  • In an embedded system such as a drone, it is difficult to store, transfer and analyze the entire hyper-spectral image to a server in real time because it takes a lot of power and time. Therefore, the hyper-spectral image data is transmitted to the server through dimension reduction or compression pre-processing. Feature selection method are used to send only the bands for analysis purpose, and these algorithms usually take a lot of processing time depending on the size of the image, even though the efficiency is high. In this paper, by improving the temporal disadvantage of the band selection algorithm, the time taken 24 hours was reduced to around 60-180 seconds based on the 40000*682 image resolution of 8GB data, and the use of 7.6GB RAM was significantly reduced to 2.3GB using 45 out of 150 bands. However, in terms of pixel classification performance, more than 98% of analysis results were derived similarly to the previous one.

Enhanced and applicable algorithm for Big-Data by Combining Sparse Auto-Encoder and Load-Balancing, ProGReGA-KF

  • Kim, Hyunah;Kim, Chayoung
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.1
    • /
    • pp.218-223
    • /
    • 2021
  • Pervasive enhancement and required enforcement of the Internet of Things (IoTs) in a distributed massively multiplayer online architecture have effected in massive growth of Big-Data in terms of server over-load. There have been some previous works to overcome the overloading of server works. However, there are lack of considered methods, which is commonly applicable. Therefore, we propose a combing Sparse Auto-Encoder and Load-Balancing, which is ProGReGA for Big-Data of server loads. In the process of Sparse Auto-Encoder, when it comes to selection of the feature-pattern, the less relevant feature-pattern could be eliminated from Big-Data. In relation to Load-Balancing, the alleviated degradation of ProGReGA can take advantage of the less redundant feature-pattern. That means the most relevant of Big-Data representation can work. In the performance evaluation, we can find that the proposed method have become more approachable and stable.

Evaluating and Distributing Algorithms based on Capacities of Duplicated Servers for Traffic Management (트래픽 관리를 위한 부본서버 성능평가 및 분배 알고리즘)

  • Han, Jeoong-Hye;Lee, Kyung-Hee
    • The KIPS Transactions:PartC
    • /
    • v.10C no.1
    • /
    • pp.33-38
    • /
    • 2003
  • Most of the existing algorithms try to disseminate the multimedia contents of internet service provider (ISP), without taking into account characteristics and capacities of duplicated servers. However, they are less reliable without prior information on capacities of duplicated servers. In this paper we propose two algorithms, performance rating algorithm of hardware and capacity algorithm, inspired by the need of improving QoS of delivering multimedia contents without incurring long access delays when the capacities of duplicated servers are significantly different and clients locate in a fixed geographical domain Our simulation results show that they are better than HTTP response time algorithm when the multimedia contents are large and quite different from performances between duplicated servers.

A Study for Path Tracking of Vehicle Robot Using Ultrasonic Positioning System (초음파 위치 센서를 이용한 차량 로봇의 경로 추종에 관한 연구)

  • Yoon, Suk-Min;Yeu, Tae-Kyeong;Park, Soung-Jea;Hong, Sup;Kim, Sang-Bong
    • Proceedings of the KSME Conference
    • /
    • 2008.11a
    • /
    • pp.795-800
    • /
    • 2008
  • The paper presents research for the established experiment environment of multi vehicle robot, localization algorithm that is based on vehicle control, and path tracking. The established experiment environment consists of ultrasonic positioning system, vehicle robot, server and wireless module. Ultrasonic positioning system measures positioning for using ultrasonic sensor and generates many errors because of the influence of environment such as a reflection of wall. For a solution of this fact, localization algorithm is proposed to determine a location using vehicle kinematics and selection of a reliable location data. And path tracking algorithm is proposed to apply localization algorithm and LOS, finally, that algorithms are verified via simulation and experimental

  • PDF

Article Data Prefetching Policy using User Access Patterns in News-On-demand System (주문형 전자신문 시스템에서 사용자 접근패턴을 이용한 기사 프리패칭 기법)

  • Kim, Yeong-Ju;Choe, Tae-Uk
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.5
    • /
    • pp.1189-1202
    • /
    • 1999
  • As compared with VOD data, NOD article data has the following characteristics: it is created at any time, has a short life cycle, is selected as not one article but several articles by a user, and has high access locality in time. Because of these intrinsic features, user access patterns of NOD article data are different from those of VOD. Thus, building NOD system using the existing techniques of VOD system leads to poor performance. In this paper, we analysis the log file of a currently running electronic newspaper, show that the popularity distribution of NOD articles is different from Zipf distribution of VOD data, and suggest a new popularity model of NOD article data MS-Zipf(Multi-Selection Zipf) distribution and its approximate solution. Also we present a life cycle model of NOD article data, which shows changes of popularity over time. Using this life cycle model, we develop LLBF (Largest Life-cycle Based Frequency) prefetching algorithm and analysis he performance by simulation. The developed LLBF algorithm supports the similar level in hit-ratio to the other prefetching algorithms such as LRU(Least Recently Used) etc, while decreasing the number of data replacement in article prefetching and reducing the overhead of the prefetching in system performance. Using the accurate user access patterns of NOD article data, we could analysis correctly the performance of NOD server system and develop the efficient policies in the implementation of NOD server system.

  • PDF

A Cloud-Edge Collaborative Computing Task Scheduling and Resource Allocation Algorithm for Energy Internet Environment

  • Song, Xin;Wang, Yue;Xie, Zhigang;Xia, Lin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.6
    • /
    • pp.2282-2303
    • /
    • 2021
  • To solve the problems of heavy computing load and system transmission pressure in energy internet (EI), we establish a three-tier cloud-edge integrated EI network based on a cloud-edge collaborative computing to achieve the tradeoff between energy consumption and the system delay. A joint optimization problem for resource allocation and task offloading in the threetier cloud-edge integrated EI network is formulated to minimize the total system cost under the constraints of the task scheduling binary variables of each sensor node, the maximum uplink transmit power of each sensor node, the limited computation capability of the sensor node and the maximum computation resource of each edge server, which is a Mixed Integer Non-linear Programming (MINLP) problem. To solve the problem, we propose a joint task offloading and resource allocation algorithm (JTOARA), which is decomposed into three subproblems including the uplink transmission power allocation sub-problem, the computation resource allocation sub-problem, and the offloading scheme selection subproblem. Then, the power allocation of each sensor node is achieved by bisection search algorithm, which has a fast convergence. While the computation resource allocation is derived by line optimization method and convex optimization theory. Finally, to achieve the optimal task offloading, we propose a cloud-edge collaborative computation offloading schemes based on game theory and prove the existence of Nash Equilibrium. The simulation results demonstrate that our proposed algorithm can improve output performance as comparing with the conventional algorithms, and its performance is close to the that of the enumerative algorithm.