• Title/Summary/Keyword: 학습강화

Search Result 1,589, Processing Time 0.03 seconds

A Study on the Teaching Plan Using Multimedia for the Local Community Unit (멀티미디어를 이용한 지역화 단원의 교수-학습 과정안에 관한 연구)

  • 노정권;양단희;정혜정
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.10d
    • /
    • pp.388-390
    • /
    • 2002
  • 21세기에 접어들어 우리 교육은 세계화, 정보화, 다양화라는 큰 흐름 속에서 민족적 주체성을 함양하는 방향으로 변하고 있다. 특히 사회과목에 있어서는 학습자가 살고 있는 지역의 전통과 특수성에 대한 교육을 강화하고 있다. 이런 흐름 속에서 본 연구는 초등학교 3학년 사회과목의 지역화 단원을 지도하는데 필요한 ICT 교수-학습 과정안의 개발에 대해 다루었다. 특히 학습자가 다양한 멀티미디어 자료를 통해 보다 쉽게 학습 내용에 접근할 수 있고, 교사는 교수-학습에 다양한 자료를 손쉽게 투입하고 다양한 방법으로 재구성할 수 있도록 하는 데 주안점을 두었다.

  • PDF

Simple Q-learning using heuristic strategies (휴리스틱 전략을 이용한 Q러닝의 학습 간단화)

  • Park, Jong-cheol;Kim, Hyeon-cheol
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.10a
    • /
    • pp.708-710
    • /
    • 2018
  • 강화학습은 게임의 인공지능을 대체할 수 있는 수단이지만 불완전한 게임에서 학습하기 힘들다. 학습하기 복잡한 불완전안 카드게임에서 휴리스틱한 전략을 만들고 비슷한 상태끼리 묶으면서 학습의 복잡성을 낮추었다. 인공신경망 없이 Q-러닝만으로 게임을 5만판을 통해서 상태에 따른 전략을 학습하였다. 그 결과 동일한 전략만을 사용하는 대결보다 승률이 높게 나왔고, 다양한 상태에서 다른 전략을 선택하는 것을 관찰하였다.

Implementation of the Agent using Universal On-line Q-learning by Balancing Exploration and Exploitation in Reinforcement Learning (강화 학습에서의 탐색과 이용의 균형을 통한 범용적 온라인 Q-학습이 적용된 에이전트의 구현)

  • 박찬건;양성봉
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.7_8
    • /
    • pp.672-680
    • /
    • 2003
  • A shopbot is a software agent whose goal is to maximize buyer´s satisfaction through automatically gathering the price and quality information of goods as well as the services from on-line sellers. In the response to shopbots´ activities, sellers on the Internet need the agents called pricebots that can help them maximize their own profits. In this paper we adopts Q-learning, one of the model-free reinforcement learning methods as a price-setting algorithm of pricebots. A Q-learned agent increases profitability and eliminates the cyclic price wars when compared with the agents using the myoptimal (myopically optimal) pricing strategy Q-teaming needs to select a sequence of state-action fairs for the convergence of Q-teaming. When the uniform random method in selecting state-action pairs is used, the number of accesses to the Q-tables to obtain the optimal Q-values is quite large. Therefore, it is not appropriate for universal on-line learning in a real world environment. This phenomenon occurs because the uniform random selection reflects the uncertainty of exploitation for the optimal policy. In this paper, we propose a Mixed Nonstationary Policy (MNP), which consists of both the auxiliary Markov process and the original Markov process. MNP tries to keep balance of exploration and exploitation in reinforcement learning. Our experiment results show that the Q-learning agent using MNP converges to the optimal Q-values about 2.6 time faster than the uniform random selection on the average.

DeNERT: Named Entity Recognition Model using DQN and BERT

  • Yang, Sung-Min;Jeong, Ok-Ran
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.4
    • /
    • pp.29-35
    • /
    • 2020
  • In this paper, we propose a new structured entity recognition DeNERT model. Recently, the field of natural language processing has been actively researched using pre-trained language representation models with a large amount of corpus. In particular, the named entity recognition, which is one of the fields of natural language processing, uses a supervised learning method, which requires a large amount of training dataset and computation. Reinforcement learning is a method that learns through trial and error experience without initial data and is closer to the process of human learning than other machine learning methodologies and is not much applied to the field of natural language processing yet. It is often used in simulation environments such as Atari games and AlphaGo. BERT is a general-purpose language model developed by Google that is pre-trained on large corpus and computational quantities. Recently, it is a language model that shows high performance in the field of natural language processing research and shows high accuracy in many downstream tasks of natural language processing. In this paper, we propose a new named entity recognition DeNERT model using two deep learning models, DQN and BERT. The proposed model is trained by creating a learning environment of reinforcement learning model based on language expression which is the advantage of the general language model. The DeNERT model trained in this way is a faster inference time and higher performance model with a small amount of training dataset. Also, we validate the performance of our model's named entity recognition performance through experiments.

Influence of Learning Instrument and Self-regulated Learning Strategy on Learning Achievement in Online Learning (온라인 수업에서 학습도구와 자기조절학습, 학업성취 간의 관계 연구)

  • Shim, Sun-Kyung
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.3
    • /
    • pp.456-467
    • /
    • 2013
  • The purpose of this study is to increase the level of learning achievement by on-line learning instruments and self-regulated learning. The study were to investigate the 3 learning instrument-lecture, homework, participation-self related learning strategy, learning achievement. This Survey was conducted 226 responses of e-learning social welfare education learners. For this research, Amos 18 is used. The research findings can be summarized as the following. Lecture, homework and participation is indicated significantly affecting self-regulated learning and self-regulated learning is indicated significantly affecting learning achievement. And learning instruments has not statically significant direct effect on the learning achievement but learning instruments has statically significant indirect effect on the learning achievemen and the self-regulated.

Modeling and Simulation on One-vs-One Air Combat with Deep Reinforcement Learning (깊은강화학습 기반 1-vs-1 공중전 모델링 및 시뮬레이션)

  • Moon, Il-Chul;Jung, Minjae;Kim, Dongjun
    • Journal of the Korea Society for Simulation
    • /
    • v.29 no.1
    • /
    • pp.39-46
    • /
    • 2020
  • The utilization of artificial intelligence (AI) in the engagement has been a key research topic in the defense field during the last decade. To pursue this utilization, it is imperative to acquire a realistic simulation to train an AI engagement agent with a synthetic, but realistic field. This paper is a case study of training an AI agent to operate with a hardware realism in the air-warfare dog-fighting. Particularly, this paper models the pursuit of an opponent in the dog-fighting setting with a gun-only engagement. In this context, the AI agent requires to make a decision on the pursuit style and intensity. We developed a realistic hardware simulator and trained the agent with a reinforcement learning. Our training shows a success resulting in a lead pursuit with a decreased engagement time and a high reward.

Local Path Generation Method for Unmanned Autonomous Vehicles Using Reinforcement Learning (강화학습을 이용한 무인 자율주행 차량의 지역경로 생성 기법)

  • Kim, Moon Jong;Choi, Ki Chang;Oh, Byong Hwa;Yang, Ji Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.9
    • /
    • pp.369-374
    • /
    • 2014
  • Path generation methods are required for safe and efficient driving in unmanned autonomous vehicles. There are two kinds of paths: global and local. A global path consists of all the way points including the source and the destination. A local path is the trajectory that a vehicle needs to follow from a way point to the next in the global path. In this paper, we propose a novel method for local path generation through machine learning, with an effective curve function used for initializing the trajectory. First, reinforcement learning is applied to a set of candidate paths to produce the best trajectory with maximal reward. Then the optimal steering angle with respect to the trajectory is determined by training an artificial neural network. Our method outperformed existing approaches and successfully found quality paths in various experimental settings, including the cases with obstacles.

Prediction Technique of Energy Consumption based on Reinforcement Learning in Microgrids (마이크로그리드에서 강화학습 기반 에너지 사용량 예측 기법)

  • Sun, Young-Ghyu;Lee, Jiyoung;Kim, Soo-Hyun;Kim, Soohwan;Lee, Heung-Jae;Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.3
    • /
    • pp.175-181
    • /
    • 2021
  • This paper analyzes the artificial intelligence-based approach for short-term energy consumption prediction. In this paper, we employ the reinforcement learning algorithms to improve the limitation of the supervised learning algorithms which usually utilize to the short-term energy consumption prediction technologies. The supervised learning algorithm-based approaches have high complexity because the approaches require contextual information as well as energy consumption data for sufficient performance. We propose a deep reinforcement learning algorithm based on multi-agent to predict energy consumption only with energy consumption data for improving the complexity of data and learning models. The proposed scheme is simulated using public energy consumption data and confirmed the performance. The proposed scheme can predict a similar value to the actual value except for the outlier data.

Path selection algorithm for multi-path system based on deep Q learning (Deep Q 학습 기반의 다중경로 시스템 경로 선택 알고리즘)

  • Chung, Byung Chang;Park, Heasook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.1
    • /
    • pp.50-55
    • /
    • 2021
  • Multi-path system is a system in which utilizes various networks simultaneously. It is expected that multi-path system can enhance communication speed, reliability, security of network. In this paper, we focus on path selection in multi-path system. To select optimal path, we propose deep reinforcement learning algorithm which is rewarded by the round-trip-time (RTT) of each networks. Unlike multi-armed bandit model, deep Q learning is applied to consider rapidly changing situations. Due to the delay of RTT data, we also suggest compensation algorithm of the delayed reward. Moreover, we implement testbed learning server to evaluate the performance of proposed algorithm. The learning server contains distributed database and tensorflow module to efficiently operate deep learning algorithm. By means of simulation, we showed that the proposed algorithm has better performance than lowest RTT about 20%.

A study on the effect of startup entrepreneurs' experience of industry-university cooperation through incubator organizations on organizational learning capability and innovation performance (벤처기업 창업가의 배태조직과 산학협력 경험이 조직학습역량과 혁신성과에 미치는 영향)

  • Kim, Deokyong;Bae, Sung Joo
    • Journal of Technology Innovation
    • /
    • v.30 no.2
    • /
    • pp.29-58
    • /
    • 2022
  • Startups lack resources and manpower to build internal capabilities to strengthen market competitiveness; external cooperation such as joint research and networking plays is important. In this study, we analyzed the effect of startups' industry-university cooperation on organizational learning capability and innovation performance. Empirical results demonstrate the mechanism by which government R&D investment strengthens organizational learning capability and creates innovative results by promoting cooperation between startups and universities. First, industry-university cooperation strengthened organizational learning capability. An empirical analysis shows that startups increase internal capabilities through external cooperation. Second, startups' organizational learning capability had a significant effect on innovation performance. We analyze how organizations with high learning capabilities positively develop corporate innovation performance by having a culture of discovery and sharing new ideas. Finally, industry-university cooperation had different effects on organizational learning capability and innovation performance according to the previous experiences of startup founders. In particular, small- and medium-sized (startup) businesses and individual-based experience groups positively affected the creation of organizational learning capabilities and innovation performance through industry-university cooperation. Small- and medium-sized businesses and individual founders have a relatively small cooperative network with the outside world compared to founders of large companies, universities, and research institutes; therefore, they strengthen organizational learning capabilities through cooperation with universities. This study demonstrates that government should create policy inducements for cooperation with universities to maximize the R&D performance of startups. Criticism exists that lending support to startups and universities will hinder innovation performance; nevertheless, government investment plays a role in expanding intangible resources such as accumulating technologies, fostering high-quality human resources, and strengthening innovation networks. Therefore, the government should appropriately utilize the its authority to strengthen investment strategies for startup growth.