Authors
Achkan Salehi, Alexandre Coninx, Stephane Doncieux
Publication date
2022/2/7
Journal
IEEE Robotics and Automation Letters
Volume
7
Issue
2
Pages
4424-4431
Publisher
IEEE
Description
In the past few years,a considerable amount of research has been dedicated to the exploitation of previous learning experiences and the design of Few-shot and Meta Learning approaches, in problem domains ranging from Computer Vision to Reinforcement Learning based control. A notable exception, where to the best of our knowledge, little to no effort has been made in this direction is Quality-Diversity (QD) optimization. QD methods have been shown to be effective tools in dealing with deceptive minima and sparse rewards in Reinforcement Learning. However, they remain costly due to their reliance on inherently sample inefficient evolutionary processes. We show that, given examples from a task distribution, information about the paths taken by optimization in parameter space can be leveraged to build a prior population, which when used to initialize QD methods in unseen environments, allows for few-shot …
Total citations
202220232024463
Scholar articles
A Salehi, A Coninx, S Doncieux - IEEE Robotics and Automation Letters, 2022