• 제목/요약/키워드: Dynamic Learning

검색결과 1,157건 처리시간 0.032초

Torque Ripple Minimization of PMSM Using Parameter Optimization Based Iterative Learning Control

  • Xia, Changliang;Deng, Weitao;Shi, Tingna;Yan, Yan
    • Journal of Electrical Engineering and Technology
    • /
    • 제11권2호
    • /
    • pp.425-436
    • /
    • 2016
  • In this paper, a parameter optimization based iterative learning control strategy is presented for permanent magnet synchronous motor control. This paper analyzes the mechanism of iterative learning control suppressing PMSM torque ripple and discusses the impact of controller parameters on steady-state and dynamic performance of the system. Based on the analysis, an optimization problem is constructed, and the expression of the optimal controller parameter is obtained to adjust the controller parameter online. Experimental research is carried out on a 5.2kW PMSM. The results show that the parameter optimization based iterative learning control proposed in this paper achieves lower torque ripple during steady-state operation and short regulating time of dynamic response, thus satisfying the demands for both steady state and dynamic performance of the speed regulating system.

데이터 스트림 기술에 기반 한 개인화된 교육 시스템 개발 (Develop of a Personalized Learning System based on Data Stream Technology)

  • 조성호
    • 컴퓨터교육학회논문지
    • /
    • 제8권4호
    • /
    • pp.49-56
    • /
    • 2005
  • 가상교육 시스템이 동적 콘텐츠 전달 기술을 가지고 있지 않기 때문에 같은 수업에 속한 모든 학생들은 똑같은 콘텐츠를 접하게 된다. 본 논문은 데이터 스트림 기술을 이용하여 설계되고 구현된 개인화된 학습 시스템을 소개한다. 제안하는 시스템은 학습자의 학습수준과 능력에 따라 강의 콘텐츠를 변경할 수 있는 기술과 인터페이스를 가지고 있다. 제안하는 시스템은 동적 콘텐츠 전달기술과 학습자 수준 테스트 시스템으로 구성된다. 본 논문에서는 개인화된 학습 시스템의 설계와 구현 시 고려 사항에 대하여 설명한다.

  • PDF

기계학습을 활용한 항공표적 긴급표적처리 발전방안 연구 (A Study on Methodology for Air Target Dynamic Targeting Applying Machine Learning)

  • 강정현;임동순;최봉완
    • 한국군사과학기술학회지
    • /
    • 제22권4호
    • /
    • pp.555-566
    • /
    • 2019
  • In order to prepare for the future warfare environment, which requires a faster operational tempo, it is necessary to utilize the fourth industrial revolution technology in the field of military operations. This study propose a methodology, 'machine learning based dynamic targeting', which can contribute to reduce required man-hour for dynamic targeting. Specifically, a decision tree algorithm is considered to apply to dynamic targeting process. The algorithm learns target prioritization patterns from JIPTL(Joint Integrated Prioritized Target List) which is the result of the deliberate targeting, and then learned algorithm rapidly(almost real-time) determines priorities for new targets that occur during ATO(Air Tasking Order) execution. An experiment is performed with artificially generated data to demonstrate the applicability of the methodology.

학습조직 구현방안: 공공조직의 조직학습 및 폐기학습, 지식관리를 중심으로 한 시스템 다이내믹 접근 (Implementational Architecture of Learning Organizations: System Dynamic Approach to Organizational Learning, Unlearning, and Knowledge Management in Public Sector Organizations)

  • 홍민기
    • 한국시스템다이내믹스연구
    • /
    • 제17권3호
    • /
    • pp.51-90
    • /
    • 2016
  • Learning is naturally embedded in organizational ongoing-processes and routines. Recent many research models of organizational failure ignore how failing masks breakdowns and recoveries of organization-embedded learning as a naturally occurring process. Organizational learning is the platform in tandem with base-modules of organization in this point. Organizations learn and unlearn while they acquire, discard, and forget organizational experiences or knowledges. These processes in public sector organizations are different from learning behaviors in private sector. This study expects to explore architectural components of learning organization in public sector, focusing on distinct characteristics of public organizations, and to implement learning model based on system thinking(system dynamic) approach.

감독 지식을 융합하는 강화 학습 기법을 사용하는 셀룰러 네트워크에서 동적 채널 할당 기법 (A Dynamic Channel Assignment Method in Cellular Networks Using Reinforcement learning Method that Combines Supervised Knowledge)

  • 김성완;장형수
    • 한국정보과학회논문지:컴퓨팅의 실제 및 레터
    • /
    • 제14권5호
    • /
    • pp.502-506
    • /
    • 2008
  • 최근에 제안된 강화 학습 기법인 "potential-based" reinforcement learning(RL) 기법은 다수 학습들과 expert advice들을 감독 지식으로 강화 학습 알고리즘에 융합하는 것을 가능하게 했고 그 효용성은 최적 정책으로의 이론적 수렴성 보장으로 증명되었다. 본 논문에서는 potential-based RL 기법을 셀룰러 네트워크에서의 채널 할당 문제에 적용한다. Potential-based RL 기반의 동적 채널 할당 기법이 기존의 fixed channel assignment, Maxavail, Q-learning-based dynamic channel assignment 채널 할당 기법들보다 효율적으로 채널을 할당한다. 또한, potential-based RL 기법이 기존의 강화 학습 알고리즘인 Q-learning, SARSA(0)에 비하여 최적 정책에 더 빠르게 수렴함을 실험적으로 보인다.

Whole learning algorithm of the neural network for modeling nonlinear and dynamic behavior of RC members

  • Satoh, Kayo;Yoshikawa, Nobuhiro;Nakano, Yoshiaki;Yang, Won-Jik
    • Structural Engineering and Mechanics
    • /
    • 제12권5호
    • /
    • pp.527-540
    • /
    • 2001
  • A new sort of learning algorithm named whole learning algorithm is proposed to simulate the nonlinear and dynamic behavior of RC members for the estimation of structural integrity. A mathematical technique to solve the multi-objective optimization problem is applied for the learning of the feedforward neural network, which is formulated so as to minimize the Euclidean norm of the error vector defined as the difference between the outputs and the target values for all the learning data sets. The change of the outputs is approximated in the first-order with respect to the amount of weight modification of the network. The governing equation for weight modification to make the error vector null is constituted with the consideration of the approximated outputs for all the learning data sets. The solution is neatly determined by means of the Moore-Penrose generalized inverse after summarization of the governing equation into the linear simultaneous equations with a rectangular matrix of coefficients. The learning efficiency of the proposed algorithm from the viewpoint of computational cost is verified in three types of problems to learn the truth table for exclusive or, the stress-strain relationship described by the Ramberg-Osgood model and the nonlinear and dynamic behavior of RC members observed under an earthquake.

C-COMA: 동적 다중 에이전트 환경을 위한 지속적인 강화 학습 모델 (C-COMA: A Continual Reinforcement Learning Model for Dynamic Multiagent Environments)

  • 정규열;김인철
    • 정보처리학회논문지:소프트웨어 및 데이터공학
    • /
    • 제10권4호
    • /
    • pp.143-152
    • /
    • 2021
  • 다양한 실세계 응용 분야들에서 공동의 목표를 위해 여러 에이전트들이 상호 유기적으로 협력할 수 있는 행동 정책을 배우는 것은 매우 중요하다. 이러한 다중 에이전트 강화 학습(MARL) 환경에서 기존의 연구들은 대부분 중앙-집중형 훈련과 분산형 실행(CTDE) 방식을 사실상 표준 프레임워크로 채택해왔다. 하지만 이러한 다중 에이전트 강화 학습 방식은 훈련 시간 동안에는 경험하지 못한 새로운 환경 변화가 실전 상황에서 끊임없이 발생할 수 있는 동적 환경에서는 효과적으로 대처하기 어렵다. 이러한 동적 환경에 효과적으로 대응하기 위해, 본 논문에서는 새로운 다중 에이전트 강화 학습 체계인 C-COMA를 제안한다. C-COMA는 에이전트들의 훈련 시간과 실행 시간을 따로 나누지 않고, 처음부터 실전 상황을 가정하고 지속적으로 에이전트들의 협력적 행동 정책을 학습해나가는 지속 학습 모델이다. 본 논문에서는 대표적인 실시간 전략게임인 StarcraftII를 토대로 동적 미니게임을 구현하고 이 환경을 이용한 다양한 실험들을 수행함으로써, 제안 모델인 C-COMA의 효과와 우수성을 입증한다.

새로운 패션 의류 이미지 분류 (New Fashion Clothing Image Classification)

  • 신성윤;이현창;신광성;김형진;이재완
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2021년도 추계학술대회
    • /
    • pp.555-556
    • /
    • 2021
  • 우리는 패션 의류 이미지의 빠르고 정확한 분류를 달성하기 위해 최적화된 동적 붕괴 학습률과 개선된 모델 구조를 가진 딥 러닝 모델을 기반으로 하는 새로운 방법을 제안한다.

  • PDF

A Dynamic Channel Switching Policy Through P-learning for Wireless Mesh Networks

  • Hossain, Md. Kamal;Tan, Chee Keong;Lee, Ching Kwang;Yeoh, Chun Yeow
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제10권2호
    • /
    • pp.608-627
    • /
    • 2016
  • Wireless mesh networks (WMNs) based on IEEE 802.11s have emerged as one of the prominent technologies in multi-hop communications. However, the deployment of WMNs suffers from serious interference problem which severely limits the system capacity. Using multiple radios for each mesh router over multiple channels, the interference can be reduced and improve system capacity. Nevertheless, interference cannot be completely eliminated due to the limited number of available channels. An effective approach to mitigate interference is to apply dynamic channel switching (DCS) in WMNs. Conventional DCS schemes trigger channel switching if interference is detected or exceeds a predefined threshold which might cause unnecessary channel switching and long protocol overheads. In this paper, a P-learning based dynamic switching algorithm known as learning automaton (LA)-based DCS algorithm is proposed. Initially, an optimal channel for communicating node pairs is determined through the learning process. Then, a novel switching metric is introduced in our LA-based DCS algorithm to avoid unnecessary initialization of channel switching. Hence, the proposed LA-based DCS algorithm enables each pair of communicating mesh nodes to communicate over the least loaded channels and consequently improve network performance.

Optimizing Energy Efficiency in Mobile Ad Hoc Networks: An Intelligent Multi-Objective Routing Approach

  • Sun Beibei
    • 대한임베디드공학회논문지
    • /
    • 제19권2호
    • /
    • pp.107-114
    • /
    • 2024
  • Mobile ad hoc networks represent self-configuring networks of mobile devices that communicate without relying on a fixed infrastructure. However, traditional routing protocols in such networks encounter challenges in selecting efficient and reliable routes due to dynamic nature of these networks caused by unpredictable mobility of nodes. This often results in a failure to meet the low-delay and low-energy consumption requirements crucial for such networks. In order to overcome such challenges, our paper introduces a novel multi-objective and adaptive routing scheme based on the Q-learning reinforcement learning algorithm. The proposed routing scheme dynamically adjusts itself based on measured network states, such as traffic congestion and mobility. The proposed approach utilizes Q-learning to select routes in a decentralized manner, considering factors like energy consumption, load balancing, and the selection of stable links. We present a formulation of the multi-objective optimization problem and discuss adaptive adjustments of the Q-learning parameters to handle the dynamic nature of the network. To speed up the learning process, our scheme incorporates informative shaped rewards, providing additional guidance to the learning agents for better solutions. Implemented on the widely-used AODV routing protocol, our proposed approaches demonstrate better performance in terms of energy efficiency and improved message delivery delay, even in highly dynamic network environments, when compared to the traditional AODV. These findings show the potential of leveraging reinforcement learning for efficient routing in ad hoc networks, making the way for future advancements in the field of mobile ad hoc networking.