• 제목/요약/키워드: a Learning Gain

검색결과 311건 처리시간 0.022초

A Neurofuzzy Algorithm-Based Advanced Bilateral Controller for Telerobot Systems

  • Cha, Dong-hyuk;Cho, Hyung-Suck
    • Transactions on Control, Automation and Systems Engineering
    • /
    • 제4권1호
    • /
    • pp.100-107
    • /
    • 2002
  • The advanced bilateral control algorithm, which can enlarge a reflected force by combining force reflection and compliance control, greatly enhances workability in teleoperation. In this scheme the maximum boundaries of a compliance controller and a force reflection gain guaranteeing stability and good task performance greatly depend upon characteristics of a slave arm, a master arm, and an environment. These characteristics, however, are generally unknown in teleoperation. It is, therefore, very difficult to determine such maximum boundary of the gain. The paper presented a novel method for design of an advanced bilateral controller. The factors affecting task performance and stability in the advanced bilateral controller were analyzed and a design guideline was presented. The neurofuzzy compliance model (NFCM)-based bilateral control proposed herein is an algorithm designed to automatically determine the suitable compliance for a given task or environment. The NFCM, composed of a fuzzy logic controller (FLC) and a rule-learning mechanism, is used as a compliance controller. The FLC generates compliant motions according to contact forces. The rule-learning mechanism, which is based upon the reinforcement learning algorithm, trains the rule-base of the FLC until the given task is done successfully. Since the scheme allows the use of large force reflection gain, it can assure good task performance. Moreover, the scheme does not require any priori knowledge on a slave arm dynamics, a slave arm controller and an environment, and thus, it can be easily applied to the control of any telerobot systems. Through a series of experiments effectiveness of the proposed algorithm has been verified.

반복 학습 제어의 수렴 특성에 관한 연구 (A Study on Convergence Property of Iterative Learning Control)

  • 박광현;변증남
    • 전자공학회논문지SC
    • /
    • 제38권4호
    • /
    • pp.11-19
    • /
    • 2001
  • 본 논문에서는 반복 학습 제어의 수렴 특성에 대해 다룬다. 우선, 기존의 ${\lambda}$-노옴을 사용하여 반복 학습 법칙의 수렴성을 증명한 것과는 달리 상한노옴(sup-norm)을 사용한 수렴성 증명방법을 보인다. 또한, 구간화된 학습 방법을 사용한 반복 학습 법칙을 제안하고, 임의의 시간구간에 대해 상한노옴 관점에서 출력 오차의 단조감소적 수렴 특성을 얻을 수 있음을 보인다. 마지막으로, 제안한 구간화된 학습 방법에서의 나누어진 시간 구간이 학습 이득값에 의해 영향을 받는다는 것을 보이고, 적절한 학습 이득값을 선택함에 따라 학습 속도가 증가함을 보인다. 제안한 반복 학습 법칙의 유효성을 보이기 위하여 두 가지 수치 예를 보인다.

  • PDF

기계학습 알고리즘을 이용한 UAS 제어계수 실시간 자동 조정 시스템 (UAS Automatic Control Parameter Tuning System using Machine Learning Module)

  • 문미선;송강;송동호
    • 한국항행학회논문지
    • /
    • 제14권6호
    • /
    • pp.874-881
    • /
    • 2010
  • 무인기의 자동 비행 제어 시스템은 기체의 형태, 크기, 무게 등의 정적 및 동적 변화에 따라 스스로 비행계수를 조정하여 목표 비행궤적을 정확히 따라가도록 제어할 필요가 있다. 본 논문에서는 PID 제어 기법을 이용하는 비행제어시스템에 기계학습모듈(MLM)을 추가하여 기체의 특성 변화에 따라 제어계수를 비행중 실시간 자동으로 조정하는 시스템을 제안한다. MLM은 선형회귀분석과 보정학습을 이용하여 설계되었으며 MLM을 통해 학습된 제어계수의 적합성을 평가하는 평가모듈(EvM)을 함께 모델링 하였다. 이 시스템은 FDC 비버 시뮬레이터를 기반으로 실험하였으며 그 결과를 분석 제시하였다.

PID 학습제어기를 이용한 가변부하 직류서보전동기의 실시간 제어 (Real-Time Control of DC Sevo Motor with Variable Load Using PID-Learning Controller)

  • 김상훈;정인석;강영호;남문현;김낙교
    • 대한전기학회논문지:시스템및제어부문D
    • /
    • 제50권3호
    • /
    • pp.107-113
    • /
    • 2001
  • This paper deals with speed control of DC servo motor using a PID controller with a gain tuning based on a Back-Propagation(BP) Learning Algorithm. Conventionally a PID controller has been used in the industrial control. But a PID controller should produce suitable parameters for each system. Also, variables of the PID controller should be changed according to environments, disturbances and loads. In this paper described by a experiment that contained a method using a PID controller with a gain tuning based on a Back-Propagation(BP) Learning Algorithm, we developed speed characteristics of a DC servo motor on variable loads. The parameters of the controller are determined by neural network performed on on-line system after training the neural network on off-line system.

  • PDF

심층 강화학습을 이용한 시변 비례 항법 유도 기법 (Time-varying Proportional Navigation Guidance using Deep Reinforcement Learning)

  • 채혁주;이단일;박수정;최한림;박한솔;안경수
    • 한국군사과학기술학회지
    • /
    • 제23권4호
    • /
    • pp.399-406
    • /
    • 2020
  • In this paper, we propose a time-varying proportional navigation guidance law that determines the proportional navigation gain in real-time according to the operating situation. When intercepting a target, an unidentified evasion strategy causes a loss of optimality. To compensate for this problem, proper proportional navigation gain is derived at every time step by solving an optimal control problem with the inferred evader's strategy. Recently, deep reinforcement learning algorithms are introduced to deal with complex optimal control problem efficiently. We adapt the actor-critic method to build a proportional navigation gain network and the network is trained by the Proximal Policy Optimization(PPO) algorithm to learn an evasion strategy of the target. Numerical experiments show the effectiveness and optimality of the proposed method.

Solving Survival Gridworld Problem Using Hybrid Policy Modified Q-Based Reinforcement

  • Montero, Vince Jebryl;Jung, Woo-Young;Jeong, Yong-Jin
    • 전기전자학회논문지
    • /
    • 제23권4호
    • /
    • pp.1150-1156
    • /
    • 2019
  • This paper explores a model-free value-based approach for solving survival gridworld problem. Survival gridworld problem opens up a challenge involving taking risks to gain better rewards. Classic value-based approach in model-free reinforcement learning assumes minimal risk decisions. The proposed method involves a hybrid on-policy and off-policy updates to experience roll-outs using a modified Q-based update equation that introduces a parametric linear rectifier and motivational discount. The significance of this approach is it allows model-free training of agents that take into account risk factors and motivated exploration to gain better path decisions. Experimentations suggest that the proposed method achieved better exploration and path selection resulting to higher episode scores than classic off-policy and on-policy Q-based updates.

신경망을 이용한 PID 제어기의 제어 사양 최적의 이득값 추정 (Optimal Condition Gain Estimation of PID Controller using Neural Networks)

  • 손준혁;서보혁
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2003년도 학술회의 논문집 정보 및 제어부문 B
    • /
    • pp.717-719
    • /
    • 2003
  • Recently Neural Network techniques have widely used in adaptive and learning control schemes for production systems. However, generally it costs a lot of time for learning in the case applied in control system. Furthermore, the physical meaning of neural networks constructed as a result is not obvious. And in practice since it is difficult to the PID gains suitably lots of researches have been reported with respect to turning schemes of PID gains. A Neural Network-based PID control scheme is proposed, which extracts skills of human experts as PID gains. This controller is designed by using three-layered neural networks. The effectiveness of the proposed Neural Network-based PID control scheme is investigated through an application for a production control system. This control method can enable a plant to operate smoothy and obviously as the plant condition varies with any unexpected accident.

  • PDF

초기 오차와 시간 지연을 고려한 선형 플랜트에 대한 강인한 반복 학습 제어기의 설계 (Design of robust iterative learning controller for linear plant with initial error and time-delay)

  • 박광현;변증남;황동환
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1996년도 한국자동제어학술회의논문집(국내학술편); 포항공과대학교, 포항; 24-26 Oct. 1996
    • /
    • pp.335-338
    • /
    • 1996
  • In this paper, we are going to design an iterative learning controller with the robust properties for initial error. For this purpose, the PID-type learning law will be considered and the design guide-line will be presented for the selection of the learning gain. Also, we are going to suggest a condition for the convergence of control input for a plant with input delay. Several simulation results are presented, which shows the effectiveness of the proposed algorithms.

  • PDF

동적 마찰이 있는 다변수 시스템에서의 PID 학습 제어 (PID Learning Controller for Multivariable System with Dynamic Friction)

  • 정병묵
    • 한국정밀공학회지
    • /
    • 제24권12호
    • /
    • pp.57-64
    • /
    • 2007
  • There have been many researches for optimal controllers in multivariable systems, and they generally use accurate linear models of the plant dynamics. Real systems, however, contain nonlinearities and high-order dynamics that may be difficult to model using conventional techniques. Therefore, it is necessary a PID gain tuning method without explicit modeling for the multivariable plant dynamics. The PID tuning method utilizes the sign of Jacobian and gradient descent techniques to iteratively reduce the error-related objective function. This paper, especially, focuses on the role of I-controller when there is a steady state error. However, it is not easy to tune I-gain unlike P- and D-gain because I-controller is mainly operated in the steady state. Simulations for an overhead crane system with dynamic friction show that the proposed PID-LC algorithm improves controller performance, even in the steady state error.

불확실한 로봇 시스템을 위한 P형 반복 학습 제어기 (A P-type Iterative Learning Controller for Uncertain Robotic Systems)

  • 최준영;서원기
    • 전자공학회논문지SC
    • /
    • 제41권3호
    • /
    • pp.17-24
    • /
    • 2004
  • 동일한 작업을 반복하여 수행하는 불확실한 로봇 시스템을 위한 P형 반복 학습 제어기를 제안한다. 제안된 반복 학습 제어기는 조인트 위치 오차로 구성되는 선형 피드백 제어기와 현재의 조인트 속도 오차로 갱신되는 피드포워드 및 피드백 학습 제어기로 구성된다. 반복 작업 동작이 계속 진행됨에 따라 조인트 위치와 속도 오차는 균일하게 0으로 수렴한다. 반복 횟수에 따라 변화하는 학습 이득을 채택함으로서 반복 횟수 영역에서 임의적으로 수렴 비율을 조절할 수 있는 조인트 위치, 속도 오차한계를 제시하고, 조인트 위치와 속도 오차는 그 한계 내에서 반복 횟수 영역에서 0으로 수렴한다. 기존의 P형 반복 학습 제어기와는 달리 제안된 반복 학습 제어 알고리즘은 학습 이득을 적절하게 설계함으로써 반복 횟수 영역에서 오차 수렴 비율의 분석과 조정을 가능하게 하는 장점이 있다.