M. Araya-lópez, Near-Optimal Algorithms for Sequential Information-Gathering Decision Problems, 2013.

M. Araya-lópez, O. Buffet, V. Thomas, and F. Charpillet, A POMDP extension with belief-dependent rewards, NIPS-10, 2010.

E. Bargiacchi, Dynamic resource allocation for multi-camera systems, 2016.

R. Bellman, A Markovian decision process, Journal of Mathematics and Mechanics, vol.6, issue.5, p.1957

M. Bonneau, N. Peyrard, and R. Sabbadin, A reinforcementlearning algorithm for sampling design in Markov random fields, ECAI-12, 2012.
URL : https://hal.archives-ouvertes.fr/hal-02748492

C. Browne, E. Powley, D. Whitehouse, S. Lucas, P. Cowling et al., A survey of Monte Carlo tree search methods, IEEE Transactions on Computational Intelligence and AI in Games, vol.4, issue.1, 2012.

R. Coulom, Efficient selectivity and backup operators in Monte-Carlo tree search, Proceedings of the Fifth International Conference on Computer and Games (CG-06), 2006.
URL : https://hal.archives-ouvertes.fr/inria-00116992

A. Doucet, S. Godsill, and A. Christophe, On sequential monte carlo sampling methods for bayesian filtering, Statistics and Computing, vol.10, issue.3, pp.197-208, 2000.

M. Egorov, M. J. Kochenderfer, and J. J. Uudmae, Target surveillance in adversarial environments using POMDPs, AAAI-16, 2016.

M. Fehr, O. Buffet, V. Thomas, and J. Dibangoye, ?-POMDPs have Lipschitz-continuous -optimal value functions, NIPS-18, 2018.

D. Fox, W. Burgard, and S. Thrun, Active Markov localization for mobile robots, Robotics and Autonomous Systems, vol.25, issue.3-4, 1998.

J. Grill, M. Valko, and R. Munos, Blazing the trails before beating the path: Sample-efficient Monte-Carlo planning, NIPS-16, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01389107

M. Henrion, Propagating uncertainty in Bayesian networks by probabilistic logic sampling, Uncertainty in Artificial Intelligence, volume 5 of Machine Intelligence and Pattern Recognition, pp.149-163, 1988.

L. Kocsis and C. Szepesvari, Bandit based Monte-Carlo planning, ECML-06, 2006.

M. Lauri and R. Ritala, Planning for robotic exploration based on forward simulation, Robotics and Autonomous Systems, vol.83, pp.15-31, 2016.

M. Littman, A. Cassandra, and L. Kaelbling, Learning policies for partially observable environments: Scaling up, ICML-95, 1995.

L. Mihaylova, T. Lefebvre, H. Bruyninckx, and J. D. Schutter, Active robotic sensing as decision making with statistical methods, in Data Fusion for Situation Monitoring, Incident Detection, Alert and Response Management, 2006.

K. Åström, Optimal control of Markov processes with incomplete state information, Journal of Mathematical Analysis and Applications, vol.10, issue.1, 1965.

Y. Satsangi, S. Whiteson, and M. T. Spaan, An analysis of piecewise-linear and convex value functions for active perception POMDPs, IAS, 2015.

Y. Satsangi, S. Whiteson, F. A. Oliehoek, and M. T. Spaan, Exploiting submodular value functions for scaling up active perception, Autonomous Robots, vol.42, pp.209-233, 2018.

D. Silver and J. Veness, Monte-Carlo planning in large POMDPs, NIPS-10, 2010.

T. Smith, Probabilistic Planning for Robotic Exploration, 2007.

T. Smith and R. Simmons, Heuristic search value iteration for POMDPs, UAI-04, 2004.

M. T. Spaan, T. S. Veiga, and P. U. Lima, Decision-theoretic planning under uncertainty with information rewards for active cooperative perception, JAAMAS, vol.29, issue.6, 2015.

Z. N. Sunberg and M. J. Kochenderfer, POMCPOW: an online algorithm for POMDPs with continuous state, action, and observation spaces, CoRR, 2017.

Z. N. Sunberg and M. J. Kochenderfer, Online algorithms for POMDPs with continuous state, action, and observation spaces, ICAPS-18, 2018.