By Topic

A Policy-Improving System with a Mixture of Bayesian Networks Adapting Agents to Continuously Changing Environments

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Kitakoshi, D. ; Graduate Sch. of Eng., Nagoya Inst. of Technol. ; Shioya, H. ; Nakano, R.

A variety of adaptive learning systems which adapt themselves to complicated environments has been studied and developed in the broad field of AI researches. For example, many reinforcement learning (RL) methods have been proposed to adapt agents to the environments. At the same time, Bayesian network (BN), one of the stochastic models, has attracted increasing attention due to its noise robustness, reasoning power, etc. We have proposed a system improving RL agents' policies with a mixture model of RNs, and have evaluated the adapting performance of our system. Each structure of BN can be regarded as a stochastic knowledge representation in the policy acquired through RL. It has been confirmed that the agent with our system could improve their policies by the information derived from the mixture, and then could adequately adapt to dynamically-switched environments. In this research, we propose a method to appropriately normalize mixing parameters of the mixture for the use in common adaptive learning systems, and evaluate the fundamental performance of our system in continuously-changing environment

Published in:

SICE-ICASE, 2006. International Joint Conference

Date of Conference:

18-21 Oct. 2006