Skip to Main content Skip to Navigation
Conference papers

On Minimizing Ordered Weighted Regrets in Multiobjective Markov Decision Processes

Wlodzimierz Ogryczak Patrice Perny 1 Paul Weng 1
LIP6 - Laboratoire d'Informatique de Paris 6
Abstract : In this paper, we propose an exact solution method to generate fair policies in Multiobjective Markov Decision Processes (MMDPs). MMDPs consider n immediate reward functions, representing either individual payoffs in a multiagent problem or rewards with respect to different objectives. In this context, we focus on the determination of a policy that fairly shares regrets among agents or objectives, the regret being defined on each dimension as the opportunity loss with respect to optimal expected rewards. To this end, we propose to minimize the ordered weighted average of regrets (OWR). The OWR criterion indeed extends the minimax regret, relaxing egalitarianism for a milder notion of fairness. After showing that OWR-optimality is state-dependent and that the Bellman principle does not hold for OWR-optimal policies, we propose a linear programming reformulation of the problem. We also provide experimental results showing the efficiency of our approach.
Document type :
Conference papers
Complete list of metadata
Contributor : Lip6 Publications <>
Submitted on : Wednesday, March 9, 2016 - 5:05:52 PM
Last modification on : Friday, January 8, 2021 - 5:32:06 PM

Links full text



Wlodzimierz Ogryczak, Patrice Perny, Paul Weng. On Minimizing Ordered Weighted Regrets in Multiobjective Markov Decision Processes. 2nd International Conference on Algorithmic Decision Theory (ADT'11), Oct 2011, Piscataway, NJ, United States. pp.190-204, ⟨10.1007/978-3-642-24873-3_15⟩. ⟨hal-01285802⟩



Record views