Conference paper Open Access

Active Improvement of Control Policies with Bayesian Gaussian Mixture Model

Girgin, Hakan; Pignat, Emmanuel; Jaquier, Noémie; Calinon, Sylvain

Learning from demonstration (LfD) is an intuitive framework allowing non-expert users to easily (re-)program robots. However, the quality and quantity of demonstrations have a great influence on the generalization performances of LfD approaches. In this paper, we introduce a novel active learning framework in order to improve the generalization capabilities of control policies. The proposed approach is based on the epistemic uncertainties of Bayesian Gaussian mixture models (BGMMs). We determine the new query point location by optimizing a closed-form information-density cost based on the quadratic Rényi entropy. Furthermore, to better represent uncertain regions and to avoid local optima problem, we propose to approximate the active learning cost with a Gaussian mixture model (GMM). We demonstrate our active learning framework in the context of a reaching task in a cluttered environment with an illustrative toy example and a real experiment with a Panda robot.

Files (3.1 MB)
Name Size
active.pdf
md5:f2ad9da613543b3670264fb9a6f0e113
3.1 MB Download
5
2
views
downloads
All versions This version
Views 55
Downloads 22
Data volume 6.1 MB6.1 MB
Unique views 44
Unique downloads 22

Share

Cite as