Policy Iteration Algorithms for DEC-POMDPs with discounted rewards - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2009

Policy Iteration Algorithms for DEC-POMDPs with discounted rewards

Jilles Dibangoye
  • Fonction : Auteur
  • PersonId : 954596
Braim Chaid-Draa
  • Fonction : Auteur
  • PersonId : 954597

Résumé

Over the past seven years, researchers have been trying to find algorithms for the decentralized control of multiple agent under uncertainty. Unfortunately, most of the standard methods are unable to scale to real-world-size domains. In this paper, we come up with promising new theoretical insights to build scalable algorithms with provable error bounds. In the light of the new theoretical insights, this research revisits the policy iteration algorithm for the decentralized partially observable Markov decision process (DEC-POMDP). We derive and analyze the first point-based policy iteration algorithmswith provable error bounds. Our experimental results show that we are able to successfully solve all tested DEC-POMDP benchmarks: outperforming standard algorithms, both in solution time and policy quality.
Fichier principal
Vignette du fichier
acti-dibangoye-2009-3.pdf (266.16 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-00968744 , version 1 (01-04-2014)

Identifiants

  • HAL Id : hal-00968744 , version 1

Citer

Jilles Dibangoye, Braim Chaid-Draa, Abdel-Illah Mouaddib. Policy Iteration Algorithms for DEC-POMDPs with discounted rewards. Proc. AAMAS 2009 Workshop on Multi-agent Sequential Decision-Making in Uncertain Domains (MSDM~2009), 2009, Budapest, Hungary. ⟨hal-00968744⟩
88 Consultations
65 Téléchargements

Partager

Gmail Facebook X LinkedIn More