• Title/Summary/Keyword: Path of Reinforcement

Search Result 134, Processing Time 0.029 seconds

A reinforcement learning-based network path planning scheme for SDN in multi-access edge computing

  • MinJung Kim;Ducsun Lim
    • International journal of advanced smart convergence
    • /
    • v.13 no.2
    • /
    • pp.16-24
    • /
    • 2024
  • With an increase in the relevance of next-generation integrated networking environments, the need to effectively utilize advanced networking techniques also increases. Specifically, integrating Software-Defined Networking (SDN) with Multi-access Edge Computing (MEC) is critical for enhancing network flexibility and addressing challenges such as security vulnerabilities and complex network management. SDN enhances operational flexibility by separating the control and data planes, introducing management complexities. This paper proposes a reinforcement learning-based network path optimization strategy within SDN environments to maximize performance, minimize latency, and optimize resource usage in MEC settings. The proposed Enhanced Proximal Policy Optimization (PPO)-based scheme effectively selects optimal routing paths in dynamic conditions, reducing average delay times to about 60 ms and lowering energy consumption. As the proposed method outperforms conventional schemes, it poses significant practical applications.

Map-Based Obstacle Avoidance Algorithm for Mobile Robot Using Deep Reinforcement Learning (심층 강화학습을 이용한 모바일 로봇의 맵 기반 장애물 회피 알고리즘)

  • Sunwoo, Yung-Min;Lee, Won-Chang
    • Journal of IKEEE
    • /
    • v.25 no.2
    • /
    • pp.337-343
    • /
    • 2021
  • Deep reinforcement learning is an artificial intelligence algorithm that enables learners to select optimal behavior based on raw and, high-dimensional input data. A lot of research using this is being conducted to create an optimal movement path of a mobile robot in an environment in which obstacles exist. In this paper, we selected the Dueling Double DQN (D3QN) algorithm that uses the prioritized experience replay to create the moving path of mobile robot from the image of the complex surrounding environment. The virtual environment is implemented using Webots, a robot simulator, and through simulation, it is confirmed that the mobile robot grasped the position of the obstacle in real time and avoided it to reach the destination.

Search of Optimal Path and Implementation using Network based Reinforcement Learning Algorithm and sharing of System Information (네트워크기반의 강화학습 알고리즘과 시스템의 정보공유화를 이용한 최단경로의 검색 및 구현)

  • Min, Seong-Joon;Oh, Kyung-Seok;Ahn, June-Young;Heo, Hoon
    • Proceedings of the KIEE Conference
    • /
    • 2005.10b
    • /
    • pp.174-176
    • /
    • 2005
  • This treatise studies composing process that renew information mastered by interactive experience between environment and system via network among individuals. In the previous study map information regarding free space is learned by using of reinforced learning algorithm, which enable each individual to construct optimal action policy. Based on those action policy each individuals can obtain optimal path. Moreover decision process to distinguish best optimal path by comparing those in the network composed of each individuals is added. Also information about the finally chosen path is being updated. A self renewing method of each system information by sharing the each individual data via network is proposed Data enrichment by shilling the information of many maps not in the single map is tried Numerical simulation is conducted to confirm the propose concept. In order to prove its suitability experiment using micro-mouse by integrating and comparing the information between individuals is carried out in various types of map to reveal successful result.

  • PDF

Leveraging Visibility-Based Rewards in DRL-based Worker Travel Path Simulation for Improving the Learning Performance

  • Kim, Minguk;Kim, Tae Wan
    • Korean Journal of Construction Engineering and Management
    • /
    • v.24 no.5
    • /
    • pp.73-82
    • /
    • 2023
  • Optimization of Construction Site Layout Planning (CSLP) heavily relies on workers' travel paths. However, traditional path generation approaches predominantly focus on the shortest path, often neglecting critical variables such as individual wayfinding tendencies, the spatial arrangement of site objects, and potential hazards. These oversights can lead to compromised path simulations, resulting in less reliable site layout plans. While Deep Reinforcement Learning (DRL) has been proposed as a potential alternative to address these issues, it has shown limitations. Despite presenting more realistic travel paths by considering these variables, DRL often struggles with efficiency in complex environments, leading to extended learning times and potential failures. To overcome these challenges, this study introduces a refined model that enhances spatial navigation capabilities and learning performance by integrating workers' visibility into the reward functions. The proposed model demonstrated a 12.47% increase in the pathfinding success rate and notable improvements in the other two performance measures compared to the existing DRL framework. The adoption of this model could greatly enhance the reliability of the results, ultimately improving site operational efficiency and safety management such as by reducing site congestion and accidents. Future research could expand this study by simulating travel paths in dynamic, multi-agent environments that represent different stages of construction.

Design for earthquake-resistant short RC structural walls

  • Zygouris, Nick St.;Kotsovos, Gerasimos M.;Kotsovos, Michael D.
    • Earthquakes and Structures
    • /
    • v.8 no.3
    • /
    • pp.713-732
    • /
    • 2015
  • The application of the compressive force path method for the design of earthquake-resistant reinforced concrete structural walls with a shear span-to-depth ratio larger than 2.5 has been shown by experiment to lead to a significant reduction of the code specified transverse reinforcement within the critical lengths without compromising the code requirements for structural performance. The present work complements these findings with experimental results obtained from tests on structural walls with a shear span-to-depth ratio smaller than 2.5. The results show that the compressive force path method is capable of safeguarding the code performance requirements without the need of transverse reinforcement confining concrete within the critical lengths. Moreover, it is shown that ductility can be considerably increased by improving the strength of the two bottom edges of the walls through the use of structural steel elements extending to a small distance of the order of 100 mm from the wall base.

Design of track path-finding simulation using Unity ML Agents

  • In-Chul Han;Jin-Woong Kim;Soo Kyun Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.2
    • /
    • pp.61-66
    • /
    • 2024
  • This paper aims to design a simulation for path-finding of objects in a simulation or game environment using reinforcement learning techniques. The main feature of this study is that the objects in the simulation are trained to avoid obstacles at random locations generated on a given track and to automatically explore path to get items. To implement the simulation, ML Agents provided by Unity Game Engine were used, and a learning policy based on PPO (Proximal Policy Optimization) was established to form a reinforcement learning environment. Through the reinforcement learning-based simulation designed in this study, we were able to confirm that the object moves on the track by avoiding obstacles and exploring path to acquire items as it learns, by analyzing the simulation results and learning result graph.

Path selection algorithm for multi-path system based on deep Q learning (Deep Q 학습 기반의 다중경로 시스템 경로 선택 알고리즘)

  • Chung, Byung Chang;Park, Heasook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.1
    • /
    • pp.50-55
    • /
    • 2021
  • Multi-path system is a system in which utilizes various networks simultaneously. It is expected that multi-path system can enhance communication speed, reliability, security of network. In this paper, we focus on path selection in multi-path system. To select optimal path, we propose deep reinforcement learning algorithm which is rewarded by the round-trip-time (RTT) of each networks. Unlike multi-armed bandit model, deep Q learning is applied to consider rapidly changing situations. Due to the delay of RTT data, we also suggest compensation algorithm of the delayed reward. Moreover, we implement testbed learning server to evaluate the performance of proposed algorithm. The learning server contains distributed database and tensorflow module to efficiently operate deep learning algorithm. By means of simulation, we showed that the proposed algorithm has better performance than lowest RTT about 20%.

Designing a Reinforcement Learning-Based 3D Object Reconstruction Data Acquisition Simulation (강화학습 기반 3D 객체복원 데이터 획득 시뮬레이션 설계)

  • Young-Hoon Jin
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.6
    • /
    • pp.11-16
    • /
    • 2023
  • The technology of 3D reconstruction, primarily relying on point cloud data, is essential for digitizing objects or spaces. This paper aims to utilize reinforcement learning to achieve the acquisition of point clouds in a given environment. To accomplish this, a simulation environment is constructed using Unity, and reinforcement learning is implemented using the Unity package known as ML-Agents. The process of point cloud acquisition involves initially setting a goal and calculating a traversable path around the goal. The traversal path is segmented at regular intervals, with rewards assigned at each step. To prevent the agent from deviating from the path, rewards are increased. Additionally, rewards are granted each time the agent fixates on the goal during traversal, facilitating the learning of optimal points for point cloud acquisition at each traversal step. Experimental results demonstrate that despite the variability in traversal paths, the approach enables the acquisition of relatively accurate point clouds.

Grasping a Target Object in Clutter with an Anthropomorphic Robot Hand via RGB-D Vision Intelligence, Target Path Planning and Deep Reinforcement Learning (RGB-D 환경인식 시각 지능, 목표 사물 경로 탐색 및 심층 강화학습에 기반한 사람형 로봇손의 목표 사물 파지)

  • Ryu, Ga Hyeon;Oh, Ji-Heon;Jeong, Jin Gyun;Jung, Hwanseok;Lee, Jin Hyuk;Lopez, Patricio Rivera;Kim, Tae-Seong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.9
    • /
    • pp.363-370
    • /
    • 2022
  • Grasping a target object among clutter objects without collision requires machine intelligence. Machine intelligence includes environment recognition, target & obstacle recognition, collision-free path planning, and object grasping intelligence of robot hands. In this work, we implement such system in simulation and hardware to grasp a target object without collision. We use a RGB-D image sensor to recognize the environment and objects. Various path-finding algorithms been implemented and tested to find collision-free paths. Finally for an anthropomorphic robot hand, object grasping intelligence is learned through deep reinforcement learning. In our simulation environment, grasping a target out of five clutter objects, showed an average success rate of 78.8%and a collision rate of 34% without path planning. Whereas our system combined with path planning showed an average success rate of 94% and an average collision rate of 20%. In our hardware environment grasping a target out of three clutter objects showed an average success rate of 30% and a collision rate of 97% without path planning whereas our system combined with path planning showed an average success rate of 90% and an average collision rate of 23%. Our results show that grasping a target object in clutter is feasible with vision intelligence, path planning, and deep RL.

Improvement of hysteretic constitutive model for reinforcements considering buckling

  • Weng Weipeng;Xie Xu;Wang Tianjia;Li Shuailing
    • Earthquakes and Structures
    • /
    • v.25 no.1
    • /
    • pp.57-67
    • /
    • 2023
  • The buckling of longitudinal reinforcements under seismic loading accelerates the degradation of the bearing capacity of reinforced concrete columns. The traditional hysteretic constitutive model of reinforcement, which does not consider buckling, usually overestimates the seismic performance of pier columns. Subsequent researchers have also proposed many models including the buckling effects. However, the accuracy of these hysteretic constitutive models proposed for simulating the buckling behavior is inadequate. In this study, based on their works, the influence of historical events on buckling is considered, the path of the re-tensioning phase is corrected by adjusting the boundary lines, and the positions of the onset buckling point and compressive buckling path during each buckling deformation are corrected by introducing correction parameters and a boundary line. A modified hysteretic constitutive model is obtained, that can more accurately reflect the buckling behavior of reinforcements. Finally, a series of hysteresis tests of reinforcements with different slenderness ratios were then conducted. The experimental results verify the effectiveness of the proposed modified model. Indicating that the modified model can more accurately simulate the equivalent stress-strain relationship of the buckling reinforcement segment.