By Topic

Extend Single-agent Reinforcement Learning Approach to a Multi-robot Cooperative Task in an Unknown Dynamic Environment

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Ying Wang ; Mechanical Engineer Department, The University of British Columbia, Vancouver, BC, V6T 1N3, Canada. phone: 604-822-4850; e-mail: ; C. W. de Silva

Machine learning technology helps multi-robot systems to carry out desired tasks in an unknown dynamic environment. In this paper, we extend the single-agent Q-learning algorithm to a multi-robot box-pushing system in an unknown dynamic environment with random obstacle distribution. There are two kinds of extensions available: directly extending MDP (Markov decision process) based Q-learning to the multi-robot domain, and SG-based (stochastic game based) Q-learning. Here, we select the first kind of extension because of its simplicity. The learning space, the box dynamics, and the reward function etc. are presented in this paper. Furthermore, a simulation system is developed and its results show effectiveness, robustness and adaptivity of this learning-based multi-robot system. Our statistical analysis of the results also shows that the robots learned correct cooperative strategy even in a dynamic environment.

Published in:

The 2006 IEEE International Joint Conference on Neural Network Proceedings

Date of Conference:

0-0 0