By Topic

A Bayesian approach to conceptualization using reinforcement learning

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Saeed Amizadeh ; Control and Intelligent Processing Center of Excellence (CIPCE), Electrical and Computer Engineering Department, University of Tehran, Iran ; Majid Nili Ahmadabadi ; Babak N. Araabi ; Roland Siegwart

Abstraction provides cognition economy and generalization skill in addition to facilitating knowledge communication for learning agents situated in real world. Concept learning introduces a way of abstraction which maps the continuous state and action spaces into entities called concepts. Of computational concept learning approaches, action-based conceptualization is favored because of its simplicity and mirror neuron foundations in neuroscience. In this paper, a new biologically inspired concept learning approach based on the Bayesian framework is proposed. This approach exploits and extends the mirror neuron's role in conceptualization for a reinforcement learning agent in nondeterministic environments. In the proposed method, an agent sequentially learns the concepts from both of its successes and its failures through interaction with the environment. These characteristics as a whole distinguish the proposed learning algorithm from positive sample learning. Simulation results show the correct formation of concepts' distributions in perceptual space in addition to benefits of utilizing both successes and failures in terms of convergence speed as well as asymptotic behavior. Experimental results, on the other hand, show the applicability and effectiveness of our method for a real robotic task such as wall-following.

Published in:

2007 IEEE/ASME international conference on advanced intelligent mechatronics

Date of Conference:

4-7 Sept. 2007