By Topic

Battlefield Agent Alliance decision-making Two Layer Reinforcement learning algorithm

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Xie Zhi-jun ; Sch. of Autom. Sci. & Electr. Eng., Beihang Univ., Beijing, China ; Dong Chao-yang ; Yang Fei ; Chen Wei

In the background of Agent Alliance combat deduction, here we present a Two Layer Reinforcement learning algorithm, referred to a TLRL algorithm, for the special requirements of battlefield simulation environment Agents offensive and defensive decision-making study. The algorithm model is classified into two layers: one is the global decision-making Agent, called Commandant Agent, learning from the environment as well as both enemies' and friends' actions, the other is the Servant Agents optimizing the action by receiving local environment feedback. Finally the war situation deduction which is carried out on the simulation platform TBS we set up, has showed the fast convergence and effectiveness of this algorithm.

Published in:

Computer Application and System Modeling (ICCASM), 2010 International Conference on  (Volume:1 )

Date of Conference:

22-24 Oct. 2010