Loading [a11y]/accessibility-menu.js
A Framework of Improving Human Demonstration Efficiency for Goal-Directed Robot Skill Learning | IEEE Journals & Magazine | IEEE Xplore

A Framework of Improving Human Demonstration Efficiency for Goal-Directed Robot Skill Learning


Abstract:

Robot learning from humans allows robots to automatically adjust to stochastic and dynamic environments by learning from nontechnical end user’s demonstrations, which is ...Show More

Abstract:

Robot learning from humans allows robots to automatically adjust to stochastic and dynamic environments by learning from nontechnical end user’s demonstrations, which is best known as robot programming by demonstration, robot learning from demonstration, apprenticeship learning, and imitation learning. Although most of those methods are probabilistic, and their performances intensively depend on the demonstrated data, measuring and evaluating human demonstrations are rarely investigated. A poorly demonstrated data set with useless prior knowledge or redundant demonstrations increases the complexity and time cost of robot learning. To solve these problems, a goal-directed robot skill learning framework named GPm-MOGP is presented. It 1) decides when and where to add a new demonstration by calculating the trajectory uncertainty; 2) determines which demonstration is useless or redundant by Kullback–Leibler (KL) divergence; 3) implements robot skill learning with a minimum number of demonstrations using a multioutput Gaussian process; and 4) learns orientation uncertainty and representation by combining logarithmic and exponential maps. The proposed framework significantly reduces the demonstrated effort of nontechnical end users who lack an understanding of how and what the robot learns during the demonstrating process. To evaluate the proposed framework, a pick-and-place experiment was designed with five unseen goals to verify the effectiveness of our methods. This experiment is well illustrated with two phases: 1) demonstration efficiency and 2) skill representation and reproduction. The results indicate an improvement of 60% in human demonstration efficiency, compared to common learning from demonstrations (LfD) applications that require at least ten demonstrations, and the robot average success rate of pick-and-place task reaches 85%.
Published in: IEEE Transactions on Cognitive and Developmental Systems ( Volume: 14, Issue: 4, December 2022)
Page(s): 1743 - 1754
Date of Publication: 21 December 2021

ISSN Information:

Funding Agency:


I. Introduction

Robots eventually will become a part of our daily lives in human society [1]. They are no longer used to only perform the same task thousands of times; rather, they will be faced with thousands of different tasks that rarely repeat in an ever-changing environment [2]. Robot learning will be necessary to those end users without a programming ability. It will allow robots to automatically adjust to stochastic and dynamic environments by learning from demonstrations (LfD) or interacting with the environment [3]. The LfD methods are best known as robot programming by demonstration, robot learning from/by demonstration, apprenticeship learning, and imitation learning [4]. They highlight several strengths, including nonexpert robot programming, data efficiency, safe learning, performance guarantees, and platform independence, as described in [5]. As pointed out by many surveys in the context of LfD [6]–[8], determining when and where to add a new demonstration, as well as which demonstration is redundant, remains challenging.

Contact IEEE to Subscribe

References

References is not available for this document.