By Topic

Dynamic tuning of PI-controllers based on model-free Reinforcement Learning methods

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Lena Abbasi Brujeni ; Department of Chemical and Materials Engineering, University of Alberta, Edmonton, Canada ; Jong Min Lee ; Sirish L. Shah

A Reinforcement Learning (RL) method called SARSA is used to dynamically tune a PI-controller for a Continuous Stirred Tank Heater (CSTH) experimental setup. In order to start from an acceptable policy, the proposed approach uses an approximate First Order Plus Time Delay (FOPTD) model to train the RL agent in the simulation environment before implementation on the real plant. As a result of the existing plant-model mismatch, the performance of the RL-based PI-controller based on the policy derived from simulations is not as good as the simulation results; however, training on the real plant results in a significant performance improvement. On the other hand, the IMC-tuned PI-controllers, which are the most commonly used feedback controllers, degrade because of the inevitable plant-model mismatch. The experimental tests are carried out for the cases of set-point tracking and disturbance rejection. In both cases, the successful adaptability of the RL-based PI-controller is clearly evident.

Published in:

Control Automation and Systems (ICCAS), 2010 International Conference on

Date of Conference:

27-30 Oct. 2010