M. Araya-lópez, O. Buffet, V. Thomas, and F. C. Pomdp, Extension with Beliefdependent Rewards, Advances in Neural Information Processing Systems, vol.23, 2010.

M. Araya-lópez, « Near-Optimal Algorithms for Sequential Information-Gathering Decision Problems, 2013.

K. Å. , Optimal control of Markov processes with incomplete state information, Journal of Mathematical Analysis and Applications, vol.10, issue.1, pp.22-247, 1965.

R. B. , A Markovian Decision Process, Journal of Mathematics and Mechanics, vol.6, 1957.

M. Bonneau, N. Peyrard, and R. Sabbadin, « A Reinforcement-Learning Algorithm for Sampling Design in Markov Random Fields, p.2012

C. Browne, « A Survey of Monte Carlo Tree Search Methods, IEEE Transactions on Computational Intelligence and AI in Games, vol.4, p.1, 2012.

R. C. , Efficient Selectivity and Backup Operators in Monte-Carlo Tree Search, Proceedings of the Fifth International Conference on Computer and Games (CG-2006), 2006.

A. Doucet, S. Godsill, and A. Christophe, « On sequential Monte Carlo sampling methods for Bayesian filtering, Statistics and Computing, vol.10, pp.197-208, 2000.

M. Egorov, M. J. Kochenderfer, and J. J. Uud-mae, Target Surveillance in Adversarial Environments using POMDPs ». In : AAAI-16, 2016.

M. Fehr, O. Buffet, V. Thomas, and J. Diban-goye, « ?-POMDPs have Lipschitz-Continuous -Optimal Value Functions, Advances in Neural Information Processing Systems 32 (NIPS-18), 2018.

D. Fox, W. Burgard, and S. Thrun, « Active Markov Localization for Mobile Robots, Robotics and Autonomous Systems, vol.25, pp.3-4, 1998.

M. Kearns, Y. Mansour, and A. Ng, « A Sparse Sampling Algorithm for Near-Optimal Planning in Large Markov Decision Processes, Machine Learning, p.49, 2002.

L. Kocsis and C. Szepesvari, « Bandit based Monte-Carlo Planning, ECML-06, 2006.

L. Mihaylova, T. Lefebvre, H. Bruyninckx, and J. D. Schutter, NATO Science Series on Data Fusion for Situation Monitoring, Incident Detection, Alert and Response Management. T. 198. 2006. Chap. Active Robotic Sensing as Decision Making with Statistical Methods

M. L. Puterman, Markov Decision Processes -Discrete Stochastic Dynamic Programming, 1994.

Y. Satsangi, S. Whiteson, and M. T. Spaan, An Analysis of Piecewise-Linear and Convex Value Functions for Active Perception POMDPs, 2015.

D. Silver and J. Veness, « Monte-Carlo Planning in Large POMDPs, NIPS-10, 2010.

T. Smith and R. Simmons, Heuristic Search Value Iteration for POMDPs, UAI-04, 2004.

M. T. Spaan, T. S. Veiga, and P. U. Lima, « Decision-theoretic planning under uncertainty with information rewards for active cooperative perception, JAAMAS, vol.29, issue.6, 2015.

Z. N. Sunberg and M. J. Kochenderfer, « Online Algorithms for POMDPs with Continuous State, Action, and Observation Spaces, p.18, 2018.

Z. Sunberg, M. J. Kochenderfer, and . Pomc-pow, An online algorithm for POMDPs with continuous state, action, and observation spaces, 2017.

, RockSampling, vol.44, p.104