A POMDP extension with belief-dependent rewards, Advances in Neural Information Processing Systems 23 (NIPS-10), 2010. ,
Optimal control of Markov processes with incomplete state information, Journal of Mathematical Analysis and Applications, vol.10, issue.1 ,
A Markovian decision process, Journal of Mathematics and Mechanics, vol.6, issue.5, p.1957 ,
DOI : 10.1512/iumj.1957.6.56038
URL : http://www.iumj.indiana.edu/IUMJ/FTDLOAD/1957/6/56038/pdf
Optimally solving Dec-POMDPs as continuousstate MDPs, Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence (IJCAI-13), 2013. ,
DOI : 10.1613/jair.4623
URL : https://hal.archives-ouvertes.fr/hal-00907338
Optimally solving Dec-POMDPs as continuousstate MDPs, Journal of Artificial Intelligence Research, vol.55, 2016. ,
DOI : 10.1613/jair.4623
URL : https://hal.archives-ouvertes.fr/hal-00907338
Approximation of Markov decision processes with general state space, Journal of Mathematical Analysis and Applications, vol.388, issue.2, 2012. ,
DOI : 10.1016/j.jmaa.2011.11.015
URL : https://hal.archives-ouvertes.fr/hal-00648223
Target surveillance in adversarial environments using POMDPs, Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16), 2016. ,
Inferring bounds on the performance of a control policy from a sample of trajectories, Proceedings of the IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning, 2009. ,
Active Markov localization for mobile robots, Robotics and Autonomous Systems, vol.25, pp.49-58, 1998. ,
DOI : 10.1016/s0921-8890(98)00049-9
URL : http://robots.stanford.edu/papers/fox.ras_act_local.ps.gz
Dynamic programming for partially observable stochastic games, Proceedings of the Nineteenth National Conference on Artificial Intelligence (AAAI-04), 2004. ,
Lipschitz continuity of value functions in Markovian decision processes, Mathematical Methods of Operations Research, vol.62, issue.1, 2005. ,
DOI : 10.1007/s00186-005-0438-1
Near-optimal search in continuous domains, Proceedings of the National Conference on Artificial Intelligence (AAAI-07), 2007. ,
SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces, Robotics: Science and Systems IV, 2008. ,
DOI : 10.15607/rss.2008.iv.009
URL : https://doi.org/10.15607/rss.2008.iv.009
The preservation of continuity and Lipschitz continuity by optimal reward operators, Mathematics of Operations Research, vol.29, issue.3, 2004. ,
DOI : 10.1287/moor.1030.0085
, NATO Science Series on Data Fusion for Situation Monitoring, Incident Detection, Alert and Response Management, vol.198, 2006.
Playing atari with deep reinforcement learning, NIPS Deep Learning Workshop, 2013. ,
Point-based value iteration: An anytime algorithm for POMDPs, Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence (IJCAI03), 2003. ,
Anytime point-based approximations for large POMDPs, Journal of Artificial Intelligence Research, p.27, 2006. ,
Finite Memory Estimation and Control of Finite Probabilistic Systems, 1977. ,
Closing the gap: Improved bounds on optimal POMDP solutions, Proceedings of the Twenty-First International Conference on Automated Planning and Scheduling (ICAPS-11), 2011. ,
On the locality of action domination in sequential decision making, Proc. of the International Symposium on Artificial Intelligence and Mathematics (ISAIM-10), 2010. ,
An analysis of piecewise-linear and convex value functions for active perception POMDPs, IAS, 2015. ,
The optimal control of partially observable Markov decision processes over a finite horizon, Operation Research, vol.21, 1973. ,
Probabilistic Planning for Robotic Exploration, 2007. ,
Heuristic search value iteration for POMDPs, Proceedings of the Annual Conference on Uncertainty in Artificial Intelligence (UAI-04), 2004. ,
Point-based POMDP algorithms: Improved analysis and implementation, Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence (UAI-05), 2005. ,
The Optimal Control of Partially Observable Markov Decision Processes, 1971. ,
Decision-theoretic planning under uncertainty with information rewards for active cooperative perception, Autonomous Agents and Multi-Agent Systems, vol.29, issue.6, 2015. ,
Speeding up the convergence of value iteration in partially observable Markov decision processes, Journal of Artificial Intelligence Research, vol.14, 2001. ,
Covering number for efficient heuristic-based pomdp planning, Proceedings of the 31st International Conference on Machine Learning (ICML-14), 2014. ,