Dynamic Actor-Advisor Programming for Scalable Safe Reinforcement Learning | IEEE Conference Publication | IEEE Xplore

Dynamic Actor-Advisor Programming for Scalable Safe Reinforcement Learning


Abstract:

Real-world robots have complex strict constraints. Therefore, safe reinforcement learning algorithms that can simultaneously minimize the total cost and the risk of const...Show More

Abstract:

Real-world robots have complex strict constraints. Therefore, safe reinforcement learning algorithms that can simultaneously minimize the total cost and the risk of constraint violation are crucial. However, almost no algorithms exist that can scale to high-dimensional systems to the best of our knowledge. In this paper, we propose Dynamic Actor-Advisor Programming (DAAP), as an algorithm for sample-efficient and scalable safe reinforcement learning. DAAP employs two control policies, actor and advisor. They are updated to minimize total cost and risk of constraint violation intertwiningly and smoothly towards each other's direction by using the other as the baseline policy in the Kullback-Leibler divergence of Dynamic Policy Programming framework. We demonstrate the scalability and sample efficiency of DAAP through its application on simulated robot arm control tasks with performance comparisons to baselines.
Date of Conference: 31 May 2020 - 31 August 2020
Date Added to IEEE Xplore: 15 September 2020
ISBN Information:

ISSN Information:

Conference Location: Paris, France

Contact IEEE to Subscribe

References

References is not available for this document.