• Title/Summary/Keyword: Q learning

Search Result 426, Processing Time 0.028 seconds

LoRa Network based Parking Dispatching System : Queuing Theory and Q-learning Approach (LoRa 망 기반의 주차 지명 시스템 : 큐잉 이론과 큐러닝 접근)

  • Cho, Youngho;Seo, Yeong Geon;Jeong, Dae-Yul
    • Journal of Digital Contents Society
    • /
    • v.18 no.7
    • /
    • pp.1443-1450
    • /
    • 2017
  • The purpose of this study is to develop an intelligent parking dispatching system based on LoRa network technology. During the local festival, many tourists come into the festival site simultaneously after sunset. To handle the traffic jam and parking dispatching, many traffic management staffs are engaged in the main road to guide the cars to available parking lots. Nevertheless, the traffic problems are more serious at the peak time of festival. Such parking dispatching problems are complex and real-time traffic information dependent. We used Queuing theory to predict inbound traffics and to measure parking service performance. Q-learning algorithm is used to find fastest routes and dispatch the vehicles efficiently to the available parking lots.

A Development of Nurse Scheduling Model Based on Q-Learning Algorithm

  • JUNG, In-Chul;KIM, Yeun-Su;IM, Sae-Ran;IHM, Chun-Hwa
    • Korean Journal of Artificial Intelligence
    • /
    • v.9 no.1
    • /
    • pp.1-7
    • /
    • 2021
  • In this paper, We focused the issue of creating a socially problematic nurse schedule. The nurse schedule should be prepared in consideration of three shifts, appropriate placement of experienced workers, the fairness of work assignment, and legal work standards. Because of the complex structure of the nurse schedule, which must reflect various requirements, in most hospitals, the nurse in charge writes it by hand with a lot of time and effort. This study attempted to automatically create an optimized nurse schedule based on legal labor standards and fairness. We developed an I/O Q-Learning algorithm-based model based on Python and Web Application for automatic nurse schedule. The model was trained to converge to 100 by creating an Fairness Indicator Score(FIS) that considers Labor Standards Act, Work equity, Work preference. Manual nurse schedules and this model are compared with FIS. This model showed a higher work equity index of 13.31 points, work preference index of 1.52 points, and FIS of 16.38 points. This study was able to automatically generate nurse schedule based on reinforcement Learning. In addition, as a result of creating the nurse schedule of E hospital using this model, it was possible to reduce the time required from 88 hours to 3 hours. If additional supplementation of FIS and reinforcement Learning techniques such as DQN, CNN, Monte Carlo Simulation and AlphaZero additionally utilize a more an optimized model can be developed.

Retained Message Delivery Scheme utilizing Reinforcement Learning in MQTT-based IoT Networks (MQTT 기반 IoT 네트워크에서 강화학습을 활용한 Retained 메시지 전송 방법)

  • Yeunwoong Kyung;Tae-Kook Kim;Youngjun Kim
    • Journal of Internet of Things and Convergence
    • /
    • v.10 no.2
    • /
    • pp.131-135
    • /
    • 2024
  • In the MQTT protocol, if the retained flag of a message published by a publisher is set, the message is stored in the broker as a retained message. When a new subscriber performs a subscribe, the broker immediately sends the retained message. This allows the new subscriber to perform updates on the current state via the retained message without waiting for new messages from the publisher. However, sending retained messages can become a traffic overhead if new messages are frequently published by the publisher. This situation could be considered an overhead when new subscribers frequently subscribe. Therefore, in this paper, we propose a retained message delivery scheme by considering the characteristics of the published messages. We model the delivery and waiting actions to new subscribers from the perspective of the broker using reinforcement learning, and determine the optimal policy through Q learning algorithm. Through performance analysis, we confirm that the proposed method shows improved performance compared to existing methods.

VC-DIMENSION AND DISTANCE CHAINS IN 𝔽dq

  • ;Ruben Ascoli;Livia Betti;Justin Cheigh;Alex Iosevich;Ryan Jeong;Xuyan Liu;Brian McDonald;Wyatt Milgrim;Steven J. Miller;Francisco Romero Acosta;Santiago Velazquez Iannuzzelli
    • Korean Journal of Mathematics
    • /
    • v.32 no.1
    • /
    • pp.43-57
    • /
    • 2024
  • Given a domain X and a collection H of functions h : X → {0, 1}, the Vapnik-Chervonenkis (VC) dimension of H measures its complexity in an appropriate sense. In particular, the fundamental theorem of statistical learning says that a hypothesis class with finite VC-dimension is PAC learnable. Recent work by Fitzpatrick, Wyman, the fourth and seventh named authors studied the VC-dimension of a natural family of functions ℋ'2t(E) : 𝔽2q → {0, 1}, corresponding to indicator functions of circles centered at points in a subset E ⊆ 𝔽2q. They showed that when |E| is large enough, the VC-dimension of ℋ'2t(E) is the same as in the case that E = 𝔽2q. We study a related hypothesis class, ℋdt(E), corresponding to intersections of spheres in 𝔽dq, and ask how large E ⊆ 𝔽dq needs to be to ensure the maximum possible VC-dimension. We resolve this problem in all dimensions, proving that whenever |E| ≥ Cdqd-1/(d-1) for d ≥ 3, the VC-dimension of ℋdt(E) is as large as possible. We get a slightly stronger result if d = 3: this result holds as long as |E| ≥ C3q7/3. Furthermore, when d = 2 the result holds when |E| ≥ C2q7/4.

Affective Outcome According to KCUE-Q1(Korean College and University Education Questionnaire) in Nursing Students (간호대학생의 KCUE-Q1(Korean College & University Education Questionnaire)에 따른 비인지적 학습성과)

  • Kim, Ok-Hyun;Choi, Eun-Ju
    • The Journal of the Korea Contents Association
    • /
    • v.14 no.10
    • /
    • pp.862-872
    • /
    • 2014
  • The purpose of this study was to examine the affective outcome according to KCUE-Q1 in nursing students. The survey was performed with 424 nursing students of 3-year or 4-year that scheduled to graduate from college. Collected data was analyzed using descriptive statistics, t-test, one-way ANOVA and Pearson's correlation coefficient. The level of affective outcome was 2.69. The affective outcome was significantly different according to gender, grade, nursing education accreditation and club activity. Learning passion showed a positive correlation with learning experience and educational outcome. Learning experience showed a positive correlation with leaning passion and educational outcome. The findings of this study indicates a need to develop outcome-based nursing curriculum for nursing students. It is also necessary to evaluate affective outcome in undergraduate nursing students.

Nursing students' and instructors' perception of simulation-based learning

  • Lee, Ji Young;Park, Sunah
    • International Journal of Advanced Culture Technology
    • /
    • v.8 no.1
    • /
    • pp.44-55
    • /
    • 2020
  • The degree of mutual understanding between nursing students and instructors regarding simulation-based education remains unknown. The purpose of this study was to identify the subjectivity of nursing students and instructors about simulation-based learning, and was intended to expand the mutual understand by employing the co-orientation model. Q-methodology was used to identify the perspectives of 46 nursing students and 38 instructors. Perception types found among students in relation to simulation-based learning were developmental training seekers, instructor-dependent seekers, and learning achievement seekers. The instructors estimated the student perception types as passive and dependent, positive commitment, demanding role as facilitators, and psychological burden. Perception types found among instructors included nursing capacity enhancement seekers, self-reflection seekers, and reality seekers. The students classified the instructors' perception types as nursing competency seekers, learning reinforcement seekers, and debriefing-oriented seekers. As a result of the analysis of these relations in the co-orientation model, instructors identified psychological burden and passive and dependent cognitive frameworks among students; however, these were not reported in the students' perspectives. Likewise, the reality seekers type found among the perception types of instructors was not identified by the students. These findings can help develop and implement simulation-based curricula aimed at maximizing the learning effect of nursing students.

Optimal Design of Semi-Active Mid-Story Isolation System using Supervised Learning and Reinforcement Learning (지도학습과 강화학습을 이용한 준능동 중간층면진시스템의 최적설계)

  • Kang, Joo-Won;Kim, Hyun-Su
    • Journal of Korean Association for Spatial Structures
    • /
    • v.21 no.4
    • /
    • pp.73-80
    • /
    • 2021
  • A mid-story isolation system was proposed for seismic response reduction of high-rise buildings and presented good control performance. Control performance of a mid-story isolation system was enhanced by introducing semi-active control devices into isolation systems. Seismic response reduction capacity of a semi-active mid-story isolation system mainly depends on effect of control algorithm. AI(Artificial Intelligence)-based control algorithm was developed for control of a semi-active mid-story isolation system in this study. For this research, an practical structure of Shiodome Sumitomo building in Japan which has a mid-story isolation system was used as an example structure. An MR (magnetorheological) damper was used to make a semi-active mid-story isolation system in example model. In numerical simulation, seismic response prediction model was generated by one of supervised learning model, i.e. an RNN (Recurrent Neural Network). Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm The numerical simulation results presented that the DQN algorithm can effectively control a semi-active mid-story isolation system resulting in successful reduction of seismic responses.

Reward Design of Reinforcement Learning for Development of Smart Control Algorithm (스마트 제어알고리즘 개발을 위한 강화학습 리워드 설계)

  • Kim, Hyun-Su;Yoon, Ki-Yong
    • Journal of Korean Association for Spatial Structures
    • /
    • v.22 no.2
    • /
    • pp.39-46
    • /
    • 2022
  • Recently, machine learning is widely used to solve optimization problems in various engineering fields. In this study, machine learning is applied to development of a control algorithm for a smart control device for reduction of seismic responses. For this purpose, Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm. A single degree of freedom (SDOF) structure with a smart tuned mass damper (TMD) was used as an example structure. A smart TMD system was composed of MR (magnetorheological) damper instead of passive damper. Reward design of reinforcement learning mainly affects the control performance of the smart TMD. Various hyper-parameters were investigated to optimize the control performance of DQN-based control algorithm. Usually, decrease of the time step for numerical simulation is desirable to increase the accuracy of simulation results. However, the numerical simulation results presented that decrease of the time step for reward calculation might decrease the control performance of DQN-based control algorithm. Therefore, a proper time step for reward calculation should be selected in a DQN training process.

Generation of ship's passage plan based on deep reinforcement learning (심층 강화학습 기반의 선박 항로계획 수립)

  • Hyeong-Tak Lee;Hyun Yang;Ik-Soon Cho
    • Proceedings of the Korean Institute of Navigation and Port Research Conference
    • /
    • 2023.11a
    • /
    • pp.230-231
    • /
    • 2023
  • This study proposes a deep reinforcement learning-based algorithm to automatically generate a ship's passage plan. First, Busan Port and Gwangyang Port were selected as target areas, and a container ship with a draft of 16m was designated as the target vessel. The experimental results showed that the ship's passage plan generated using deep reinforcement learning was more efficient than the Q-learning-based algorithm used in previous research. This algorithm presents a method to generate a ship's passage plan automatically and can contribute to improving maritime safety and efficiency.

  • PDF

Practical Implementation and Stability Analysis of ALOHA-Q for Wireless Sensor Networks

  • Kosunalp, Selahattin;Mitchell, Paul Daniel;Grace, David;Clarke, Tim
    • ETRI Journal
    • /
    • v.38 no.5
    • /
    • pp.911-921
    • /
    • 2016
  • This paper presents the description, practical implementation, and stability analysis of a recently proposed, energy-efficient, medium access control protocol for wireless sensor networks, ALOHA-Q, which employs a reinforcement-learning framework as an intelligent transmission strategy. The channel performance is evaluated through a simulation and experiments conducted using a real-world test-bed. The stability of the system against possible changes in the environment and changing channel conditions is studied with a discussion on the resilience level of the system. A Markov model is derived to represent the system behavior and estimate the time in which the system loses its operation. A novel scheme is also proposed to protect the lifetime of the system when the environment and channel conditions do not sufficiently maintain the system operation.