Motivation based Behavior Sequence Learning for an Autonomous Agent in Virtual Reality

  • Song, Wei (Dept. of Multimedia, Graduate School of Digital Image & Contents, Dongguk University) ;
  • Cho, Kyung-Eun (Dept. of Game & Multimedia, Dongguk University) ;
  • Um, Ky-Hyun (Dept. of Game & Multimedia, Dongguk University)
  • Published : 2009.12.30

Abstract

To enhance the automatic performance of existing predicting and planning algorithms that require a predefined probability of the states' transition, this paper proposes a multiple sequence generation system. When interacting with unknown environments, a virtual agent needs to decide which action or action order can result in a good state and determine the transition probability based on the current state and the action taken. We describe a sequential behavior generation method motivated from the change in the agent's state in order to help the virtual agent learn how to adapt to unknown environments. In a sequence learning process, the sensed states are grouped by a set of proposed motivation filters in order to reduce the learning computation of the large state space. In order to accomplish a goal with a high payoff, the learning agent makes a decision based on the observation of states' transitions. The proposed multiple sequence behaviors generation system increases the complexity and heightens the automatic planning of the virtual agent for interacting with the dynamic unknown environment. This model was tested in a virtual library to elucidate the process of the system.

Keywords

References

  1. R. Sun and C. Sessions, "Learning plans without a priori knowledge," Journal of Adaptive Behavior, Vol.8, No.3, pp. 225-253, 2000. https://doi.org/10.1177/105971230000800302
  2. M. Kwiatkowska, G. Norman and D. Parker, "Game-based Abstraction for Markov Decision Processes," Proceedings of the 3rd international conference on the Quantitative Evaluation of Systems, pp. 157-166, 2006.
  3. S. Siddiqi and A. Moore, "Fast Inference and Learning in Large-State-Space HMMs," Proceedings of the 22nd International Conference on Machine Learning, Vol.119, pp. 800-807, 2005.
  4. T.M. Gureckis and B.C. Love, "Short-term gains, long-term pains: How cues about state aid learning in dynamic environments," Cognition, Vol.113, No.3, pp. 293-313, 2009. https://doi.org/10.1016/j.cognition.2009.03.013
  5. R.S. Sutton and A.G. Barto, "Reinforcement Learning: An Introduction," MIT Press, A Bradford Book, Cambridge, MA, 1998.