By Topic

Model-based reinforcement learning for a multi-player card game with partial observability

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
H. Fujita ; Nara Inst. of Sci. & Technol., Japan ; S. Ishii

This article presents a model-based reinforcement learning (RL) scheme for a card game, "Hearts". Since this is a large-scale multi-player game with partial observability, effective state estimation and optimal control based on an environmental model are required. In our method, the learning agent is controlled by a one-step-ahead utility prediction using opponent agents' models. The computational intractability is overcome by the sampling method over a specific subspace. Simulation results show that our model-based RL method can produce an agent comparable to a human expert for this realistic problem.

Published in:

IEEE/WIC/ACM International Conference on Intelligent Agent Technology

Date of Conference:

19-22 Sept. 2005