• 제목/요약/키워드: optimal learning

검색결과 1,242건 처리시간 0.036초

NETLA를 이용한 이진 신경회로망의 최적합성 (Optimal Synthesis of Binary Neural Network using NETLA)

  • 정종원;성상규;지석준;최우진;이준탁
    • 한국마린엔지니어링학회:학술대회논문집
    • /
    • 한국마린엔지니어링학회 2002년도 춘계학술대회논문집
    • /
    • pp.273-277
    • /
    • 2002
  • This paper describes an optimal synthesis method of binary neural network(BNN) for an approximation problem of a circular region and synthetic image having four class using a newly proposed learning algorithm. Our object is to minimize the number of connections and neurons in hidden layer by using a Newly Expanded and Truncated Learning Algorithm(NETLA) based on the multilayer BNN. The synthesis method in the NETLA is based on the extension principle of Expanded and Truncated Learning (ETL) learning algorithm using the multilayer perceptron and is based on Expanded Sum of Product (ESP) as one of the boolean expression techniques. The number of the required neurons in hidden layer can be reduced and fasted for learning pattern recognition.. The superiority of this NETLA to other algorithms was proved by simulation.

  • PDF

Q-learning 모델을 이용한 IoT 기반 주차유도 시스템의 설계 및 구현 (Design and Implementation of Parking Guidance System Based on Internet of Things(IoT) Using Q-learning Model)

  • 지용주;최학희;김동성
    • 대한임베디드공학회논문지
    • /
    • 제11권3호
    • /
    • pp.153-162
    • /
    • 2016
  • This paper proposes an optimal dynamic resource allocation method in IoT (Internet of Things) parking guidance system using Q-learning resource allocation model. In the proposed method, a resource allocation using a forecasting model based on Q-learning is employed for optimal utilization of parking guidance system. To demonstrate efficiency and availability of the proposed method, it is verified by computer simulation and practical testbed. Through simulation results, this paper proves that the proposed method can enhance total throughput, decrease penalty fee issued by SLA (Service Level Agreement) and reduce response time with the dynamic number of users.

Application of reinforcement learning to fire suppression system of an autonomous ship in irregular waves

  • Lee, Eun-Joo;Ruy, Won-Sun;Seo, Jeonghwa
    • International Journal of Naval Architecture and Ocean Engineering
    • /
    • 제12권1호
    • /
    • pp.910-917
    • /
    • 2020
  • In fire suppression, continuous delivery of water or foam to the fire source is essential. The present study concerns fire suppression in a ship under sea condition, by introducing reinforcement learning technique to aiming of fire extinguishing nozzle, which works in a ship compartment with six degrees of freedom movement by irregular waves. The physical modeling of the water jet and compartment motion was provided using Unity 3D engine. In the reinforcement learning, the change of the nozzle angle during the scenario was set as the action, while the reward is proportional to the ratio of the water particle delivered to the fire source area. The optimal control of nozzle aiming for continuous delivery of water jet could be derived. Various algorithms of reinforcement learning were tested to select the optimal one, the proximal policy optimization.

Principles of Learning and the Mathematics Curriculum

  • Ediger, Marlow
    • 한국수학교육학회지시리즈A:수학교육
    • /
    • 제23권2호
    • /
    • pp.13-15
    • /
    • 1985
  • There are selected principles of learning which need adequate emphasis in the mathematics curriculum. These include: 1. Pupils perceiving purpose in learning. 2. Learners being involved in the solving of problems. 3. Meaningful learning experiences being inherent in the mathematics curriculum. 4. Provision being made to guide each learner in achieving optimal gains in ongoing study.

  • PDF

Steering the Dynamics within Reduced Space through Quantum Learning Control

  • Kim, Young-Sik
    • Bulletin of the Korean Chemical Society
    • /
    • 제24권6호
    • /
    • pp.744-750
    • /
    • 2003
  • In quantum dynamics of many-body systems, to identify the Hamiltonian becomes more difficult very rapidly as the number of degrees of freedom increases. In order to simplify the dynamics and to deduce dynamically relevant Hamiltonian information, it is desirable to control the dynamics to lie within a reduced space. With a judicious choice for the cost functional, the closed loop optimal control experiments can be manipulated efficiently to steer the dynamics to lie within a subspace of the system eigenstates without requiring any prior detailed knowledge about the system Hamiltonian. The procedure is simulated for optimally controlled population transfer experiments in the system of two degrees of freedom. To show the feasibility of steering the dynamics to lie in a specified subspace, the learning algorithms guiding the dynamics are presented along with frequency filtering. The results demonstrate that the optimal control fields derive the system to the desired target state through the desired subspace.

A New Solution for Stochastic Optimal Power Flow: Combining Limit Relaxation with Iterative Learning Control

  • Gong, Jinxia;Xie, Da;Jiang, Chuanwen;Zhang, Yanchi
    • Journal of Electrical Engineering and Technology
    • /
    • 제9권1호
    • /
    • pp.80-89
    • /
    • 2014
  • A stochastic optimal power flow (S-OPF) model considering uncertainties of load and wind power is developed based on chance constrained programming (CCP). The difficulties in solving the model are the nonlinearity and probabilistic constraints. In this paper, a limit relaxation approach and an iterative learning control (ILC) method are implemented to solve the S-OPF model indirectly. The limit relaxation approach narrows the solution space by introducing regulatory factors, according to the relationship between the constraint equations and the optimization variables. The regulatory factors are designed by ILC method to ensure the optimality of final solution under a predefined confidence level. The optimization algorithm for S-OPF is completed based on the combination of limit relaxation and ILC and tested on the IEEE 14-bus system.

Optimal deep machine learning framework for vibration mitigation of seismically-excited uncertain building structures

  • Afshin Bahrami Rad;Javad Katebi;Saman Yaghmaei-Sabegh
    • Structural Engineering and Mechanics
    • /
    • 제88권6호
    • /
    • pp.535-549
    • /
    • 2023
  • Deep extreme learning machine (DELM) and multi-verse optimization algorithms (MVO) are hybridized for designing an optimal and adaptive control framework for uncertain buildings. In this approach, first, a robust model predictive control (RMPC) scheme is developed to handle the problem uncertainty. The optimality and adaptivity of the proposed controller are provided by the optimal determination of the tunning weights of the linear programming (LP) cost function for clustered external loads using the MVO. The final control policy is achieved by collecting the clustered data and training them by DELM. The efficiency of the introduced control scheme is demonstrated by the numerical simulation of a ten-story benchmark building subjected to earthquake excitations. The results represent the capability of the proposed framework compared to robust MPC (RMPC), conventional MPC (CMPC), and conventional DELM algorithms in structural motion control.

감독 지식을 융합하는 강화 학습 기법을 사용하는 셀룰러 네트워크에서 동적 채널 할당 기법 (A Dynamic Channel Assignment Method in Cellular Networks Using Reinforcement learning Method that Combines Supervised Knowledge)

  • 김성완;장형수
    • 한국정보과학회논문지:컴퓨팅의 실제 및 레터
    • /
    • 제14권5호
    • /
    • pp.502-506
    • /
    • 2008
  • 최근에 제안된 강화 학습 기법인 "potential-based" reinforcement learning(RL) 기법은 다수 학습들과 expert advice들을 감독 지식으로 강화 학습 알고리즘에 융합하는 것을 가능하게 했고 그 효용성은 최적 정책으로의 이론적 수렴성 보장으로 증명되었다. 본 논문에서는 potential-based RL 기법을 셀룰러 네트워크에서의 채널 할당 문제에 적용한다. Potential-based RL 기반의 동적 채널 할당 기법이 기존의 fixed channel assignment, Maxavail, Q-learning-based dynamic channel assignment 채널 할당 기법들보다 효율적으로 채널을 할당한다. 또한, potential-based RL 기법이 기존의 강화 학습 알고리즘인 Q-learning, SARSA(0)에 비하여 최적 정책에 더 빠르게 수렴함을 실험적으로 보인다.

픽셀 데이터를 이용한 강화 학습 알고리즘 적용에 관한 연구 (A Study on Application of Reinforcement Learning Algorithm Using Pixel Data)

  • 문새마로;최용락
    • 한국IT서비스학회지
    • /
    • 제15권4호
    • /
    • pp.85-95
    • /
    • 2016
  • Recently, deep learning and machine learning have attracted considerable attention and many supporting frameworks appeared. In artificial intelligence field, a large body of research is underway to apply the relevant knowledge for complex problem-solving, necessitating the application of various learning algorithms and training methods to artificial intelligence systems. In addition, there is a dearth of performance evaluation of decision making agents. The decision making agent that can find optimal solutions by using reinforcement learning methods designed through this research can collect raw pixel data observed from dynamic environments and make decisions by itself based on the data. The decision making agent uses convolutional neural networks to classify situations it confronts, and the data observed from the environment undergoes preprocessing before being used. This research represents how the convolutional neural networks and the decision making agent are configured, analyzes learning performance through a value-based algorithm and a policy-based algorithm : a Deep Q-Networks and a Policy Gradient, sets forth their differences and demonstrates how the convolutional neural networks affect entire learning performance when using pixel data. This research is expected to contribute to the improvement of artificial intelligence systems which can efficiently find optimal solutions by using features extracted from raw pixel data.

의료 영상에 최적화된 딥러닝 모델의 개발 (Development of an Optimized Deep Learning Model for Medical Imaging)

  • 김영재;김광기
    • 대한영상의학회지
    • /
    • 제81권6호
    • /
    • pp.1274-1289
    • /
    • 2020
  • 최근, 의료 영상 분야에서 딥러닝은 가장 활발하게 연구되고 있는 기술 중 하나이다. 충분한 데이터와 최신의 딥러닝 알고리즘은 딥러닝 모델의 개발에 중요한 요소이다. 하지만 일반화된 최적의 딥러닝 모델을 개발하기 위해서는 데이터의 양과 최신의 딥러닝 알고리즘 외에도 많은 것을 고려해야 한다. 데이터 수집부터 가공, 전처리, 모델의 학습 및 검증, 경량화까지 모든 과정이 딥러닝 모델의 성능에 영향을 미칠 수 있기 때문이다. 본 종설에서는 의료 영상에 최적화된 딥러닝 모델을 위해 개발 과정 각각에서 고려해야 할 중요한 요소들을 살펴보고자 한다.