. Boger, A Planning System Based on Markov Decision Processes to Guide People With Dementia Through Activities of Daily Living, IEEE Transactions on Information Technology in Biomedicine, vol.10, issue.2, pp.323-333, 2006.
DOI : 10.1109/TITB.2006.864480

. Boutilier, Decision-theoretic planning: Structural assumptions and computational leverage, Journal of Artificial Intelligence Research, vol.11, pp.1-94, 1999.

. Boutilier, Cooperative negotiation in autonomic systems using incremental utility elicitation, UAI, 2003.

. Boyd, ]. S. Vandenberghe, L. Boyd, and . Vandenberghe, Convex Optimization, 2004.

. Fürnkranz, Preference-based reinforcement learning: a formal framework and a policy iteration algorithm, Machine Learning, pp.123-156, 2012.
DOI : 10.1007/s10994-012-5313-8

. Krantz, Foundations of measurement, volume Additive and Polynomial Representations, 1971.

. Littman, On the complexity of solving markov decision problems, UAI, pp.394-402, 1995.

R. Y. Ng, S. Ng, and . Russell, Algorithms for inverse reinforcement learning, ICML, 2000.

]. M. Puterman, Markov decision processes: discrete stochastic dynamic programming, 1994.
DOI : 10.1002/9780470316887

. Regan, . K. Boutilier, C. Regan, and . Boutilier, Online feature elicitation in interactive optimization, ICML, 2009.

. Regan, . K. Boutilier, C. Regan, and . Boutilier, Regret based reward elicitation for Markov decision processes, UAI, pp.444-451, 2009.

R. , B. Regan, and C. Boutilier, Robust policy computation in reward-uncertain MDPs using nondominated policies, AAAI, 2010.

. Regan, . K. Boutilier, C. Regan, and . Boutilier, Eliciting additive reward functions for Markov decision processes, IJCAI, 2011.

. Regan, . K. Boutilier, C. Regan, and . Boutilier, Robust online optimization of reward-uncertain MDPs, IJ- CAI, 2011.

S. Shaked, J. G. Shaked, . Shanthikumarviappiani, ]. P. Boutilier, C. Viappiani et al., Stochastic Orders and Their Applications Academic press Recommendation sets and choice queries: There is no exploration/exploitation tradeoff! In AAAI Markov decision processes with ordinal rewards: Reference point-based preferences Multi-objective infinite-horizon discounted Markov decision processes, ICAPS, pp.282-289639, 1982.

M. Xu, S. Xu, and . Mannor, Parametric regret in uncertain Markov decision processes, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference, 2009.
DOI : 10.1109/CDC.2009.5400796