V. Mnih, Human-level control through deep reinforcement learning, Nature, vol.518, issue.7540, pp.529-533, 2015.

D. Silver, Mastering the game of go with deep neural networks and tree search, Nature, vol.529, issue.7587, pp.484-489, 2016.

N. Heess, Emergence of locomotion behaviours in rich environments, 2017.

M. P. Deisenroth and C. E. Rasmussen, PILCO: A model-based and data-efficient approach to policy search, Proc. of ICML, 2011.

K. Chatzilygeroudis, R. Rama, R. Kaushik, D. Goepp, V. Vassiliades et al., Black-Box Data-efficient Policy Search for Robotics, Proc. of IROS, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01576683

R. Kaushik, K. Chatzilygeroudis, and J. Mouret, Multi-objective model-based policy search for data-efficient learning with sparse rewards, Conference on Robot Learning, pp.839-855, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01884294

G. Williams, N. Wagener, B. Goldfain, P. Drews, J. M. Rehg et al., Information theoretic mpc for model-based reinforcement learning, Proc. of ICRA, 2017.

A. Nagabandi, I. Clavera, S. Liu, R. S. Fearing, P. Abbeel et al., Learning to adapt in dynamic, real-world environment through meta-reinforcement learning, Proc. of ICLR, 2019.

K. Chua, R. Calandra, R. Mcallister, and S. Levine, Deep reinforcement learning in a handful of trials using probabilistic dynamics models, Proc. of NIPS, pp.4754-4765, 2018.

E. Keogh and A. Mueen, Curse of dimensionality, Encyclopedia of Machine Learning and Data Mining, pp.314-315, 2017.

,

A. Cully and J. Mouret, Evolving a behavioral repertoire for a walking robot, Evolutionary Computation, vol.24, issue.1, pp.59-88, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01095543

A. Cully, J. Clune, D. Tarapore, and J. Mouret, Robots that can adapt like animals, Nature, vol.521, issue.7553, pp.503-507, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01158243

K. Chatzilygeroudis, V. Vassiliades, and J. Mouret, Reset-free trial-and-error learning for robot damage recovery, Robotics and Autonomous Systems, vol.100, pp.236-250, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01654641

M. Duarte, J. Gomes, S. M. Oliveira, and A. L. Christensen, Evolution of repertoire-based control for robots with complex locomotor systems, IEEE Transactions on Evolutionary Computation, vol.22, issue.2, pp.314-328, 2017.

A. Sharma, S. Gu, S. Levine, V. Kumar, and K. Hausman, Dynamics-aware unsupervised discovery of skills, 2019.

J. Mouret and J. Clune, Illuminating search spaces by mapping elites, 2015.

C. E. Rasmussen and C. K. Williams, Gaussian processes for machine learning, 2006.

K. Chatzilygeroudis, V. Vassiliades, F. Stulp, S. Calinon, and J. Mouret, A survey on policy search algorithms for learning robot controllers in a handful of trials, 2018.
URL : https://hal.archives-ouvertes.fr/hal-02393432

M. P. Deisenroth, D. Fox, and C. E. Rasmussen, Gaussian processes for data-efficient learning in robotics and control, IEEE Trans. Pattern Anal. Mach. Intell, vol.37, issue.2, pp.408-423, 2015.

J. Hollerbach, W. Khalil, and M. Gautier, Model Identification, pp.113-138, 2016.

K. Chatzilygeroudis and J. Mouret, Using Parameterized Black-Box Priors to Scale Up Model-Based Policy Search for Robotics, Proc. of ICRA, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01768285

M. Cutler and J. P. How, Efficient reinforcement learning for robots using informative simulated priors, Proc. of ICRA, 2015.

V. Papaspyros, K. Chatzilygeroudis, V. Vassiliades, and J. Mouret, Safety-aware robot damage recovery using constrained bayesian optimization and simulated priors, BayesOpt '16 Workshop at NIPS, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01407757

R. Pautrat, K. Chatzilygeroudis, and J. Mouret, Bayesian optimization with automatic prior selection for data-efficient direct policy search, Proc. of ICRA, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01768279

M. Saveriano, Y. Yin, P. Falco, and D. Lee, Data-efficient control policy search using residual dynamics learning, Proc. of IROS, 2017.

M. P. Deisenroth, G. Neumann, and J. Peters, A survey on policy search for robotics, Foundations and Trends in Robotics, vol.2, issue.1, pp.1-142, 2013.

J. K. Pugh, L. B. Soros, and K. O. Stanley, Quality diversity: A new frontier for evolutionary computation, Frontiers in Robotics and AI, vol.3, p.40, 2016.

A. Cully and Y. Demiris, Quality and diversity optimization: A unifying modular framework, IEEE Trans. on Evolutionary Computation, vol.22, issue.2, pp.245-259, 2018.

V. Vassiliades, K. Chatzilygeroudis, and J. Mouret, Using centroidal voronoi tessellations to scale up the multidimensional archive of phenotypic elites algorithm, IEEE Transactions on Evolutionary Computation, vol.22, issue.4, pp.623-630, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01630627

M. Duarte, J. C. Gomes, S. Oliveira, and A. L. Christensen, Evolution of repertoire-based control for robots with complex locomotor systems, IEEE Transactions on Evolutionary Computation, vol.22, pp.314-328, 2018.

R. S. Sutton and A. G. Barto, Reinforcement learning: An introduction, 1998.

P. Auer, N. Cesa-bianchi, and P. Fischer, Finite-time analysis of the multiarmed bandit problem, Machine learning, vol.47, issue.2-3, pp.235-256, 2002.

E. Coumans, Bullet physics library, Open source: bulletphysics. org, vol.15, p.5, 2013.

. Gpy, GPy: A gaussian process framework in python

J. , Q. Candela, and C. E. Rasmussen, A unifying view of sparse approximate gaussian process regression, JMLR, vol.6, pp.1939-1959, 2005.

C. Park and D. Apley, Patchwork kriging for large-scale gaussian process regression, 2017.

Y. Gal and Z. Ghahramani, Dropout as a bayesian approximation: Representing model uncertainty in deep learning, Proc. of ICML, 2015.