By Topic

Aggregation of Reinforcement Learning Algorithms

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Ju Jiang ; Department of Electrical and Computer Engineering, University of Waterloo Waterloo, ON, Canada, Email: ; M. S. Kamel

Reinforcement learning (RL) is a machine learning method that can learn an optimal strategy for a system without knowing the mathematical model of the system. Many RL algorithms are successfully applied in various fields. However, each algorithm has its advantages and disadvantages. With the increasing complexity of environments and tasks, it is difficult for a single learning algorithm to cope with complicated learning problems with high performance. This motivated us to combine some learning algorithms to improve the learning quality. This paper proposes a new multiple learning architecture, "aggregated multiple reinforcement learning system (AMRLS)". AMRLS adopts three different learning algorithms to learn individually and then combines their results with aggregation methods. To evaluate its performance, AMRLS is tested on two different environments: a cart-pole system and a maze environment. The presented simulation results reveal that aggregation not only provides robustness and fault tolerance ability, but also produces more smooth learning curves and needs fewer learning steps than individual learning algorithms.

Published in:

The 2006 IEEE International Joint Conference on Neural Network Proceedings

Date of Conference:

0-0 0