A. Baranes and P. Y. Oudeyer, Active learning of inverse models with intrinsically motivated goal exploration in robots, Robotics and Autonomous Systems, vol.61, issue.1, 2012.
DOI : 10.1016/j.robot.2012.05.008

URL : https://hal.archives-ouvertes.fr/hal-00788440

Y. Bengio and Y. Grandvalet, No unbiased estimator of the variance of k-fold crossvalidation, Journal of Machine Learning Research (JMLR), vol.5, pp.1089-1105, 2004.

Y. Bengio, J. Louradour, R. Collobert, and J. Weston, Curriculum learning, Proceedings of the 26th Annual International Conference on Machine Learning, ICML '09, pp.41-48, 2009.
DOI : 10.1145/1553374.1553380

R. I. Brafman and M. Tennenholtz, R-max -a general polynomial time algorithm for near-optimal reinforcement learning, Journal of Machine Learning Research (JMLR), vol.3, pp.213-231, 2002.

M. Milani, F. , and J. Pineau, Pac-bayesian model selection for reinforcement learning, Conf. on Neural Information Processing Systems (NIPS), 2010.

T. Hester and P. Stone, Intrinsically motivated model learning for a developing curious agent, 2012 IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL), 2012.
DOI : 10.1109/DevLrn.2012.6400802

S. Kakade, On the Sample Complexity of Reinforcement Learning, 2003.

M. Kearns and S. Singh, Near-optimal reinforcement learning in polynomial time, Machine Learning, vol.49, issue.2/3, pp.209-232, 2002.
DOI : 10.1023/A:1017984413808

J. , Z. Kolter, and A. Ng, Near-Bayesian exploration in polynomial time, Int. Conf. on Machine Learning (ICML), pp.513-520, 2009.
DOI : 10.1145/1553374.1553441

P. Y. Oudeyer, F. Kaplan, and V. V. Hafner, Intrinsic Motivation Systems for Autonomous Mental Development, IEEE Transactions on Evolutionary Computation, vol.11, issue.2, pp.265-286, 2007.
DOI : 10.1109/TEVC.2006.890271

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.107.666

P. Poupart, N. Vlassis, J. Hoey, and K. Regan, An analytic solution to discrete Bayesian reinforcement learning, Proceedings of the 23rd international conference on Machine learning , ICML '06, 2006.
DOI : 10.1145/1143844.1143932

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.126.1774

J. Schmidhuber, Curious model-building control systems, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks, pp.1458-1463, 1991.
DOI : 10.1109/IJCNN.1991.170605

S. Singh, A. G. Barto, and N. Chentanez, Intrinsically motivated reinforcement learning, Conf. on Neural Information Processing Systems (NIPS), pp.1281-1288, 2005.

A. L. Strehl, L. Li, and M. Littman, Reinforcement learning in finite MDPs: PAC analysis, Journal of Machine Learning Research, 2009.

L. Alexander, M. L. Strehl, and . Littman, An analysis of model-based interval estimation for markov decision processes, J. Comput. Syst. Sci, vol.74, issue.8, pp.1309-1331, 2008.

M. Christopher, A. G. Vigorito, and . Barto, Intrinsically motivated hierarchical skill learning in structured environments, IEEE Transactions on Autonomous Mental Development (TAMD), vol.2, issue.2, p.2010

M. Wiering and J. Schmidhuber, Efficient model-based exploration, International Conference on Simulation of Adaptive Behavior: From Animals to Animats 6, 1998.