Exploration vs Exploitation vs Safety: Risk-averse Multi-Armed Bandits - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Journal of Machine Learning Research Année : 2013

Exploration vs Exploitation vs Safety: Risk-averse Multi-Armed Bandits

Résumé

Motivated by applications in energy management, this paper presents the Multi-Armed Risk-Aware Bandit (MARAB) algorithm. With the goal of limiting the exploration of risky arms, MARAB takes as arm quality its conditional value at risk. When the user-supplied risk level goes to 0, the arm quality tends toward the essential infimum of the arm distribution density, and MARAB tends toward the MIN multi-armed bandit algorithm, aimed at the arm with maximal minimal value. As a first contribution, this paper presents a theoretical analysis of the MIN algorithm under mild assumptions, establishing its robustness comparatively to UCB. The analysis is supported by extensive experimental validation of MIN and MARAB compared to UCB and state-of-art risk-aware MAB algorithms on artificial and real-world problems.
Fichier principal
Vignette du fichier
acml2013_marab.pdf (916.47 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-00924062 , version 1 (06-01-2014)
hal-00924062 , version 2 (06-01-2014)

Identifiants

Citer

Nicolas Galichet, Michèle Sebag, Olivier Teytaud. Exploration vs Exploitation vs Safety: Risk-averse Multi-Armed Bandits. Asian Conference on Machine Learning 2013, Nov 2013, Canberra, Australia. pp.245-260. ⟨hal-00924062v2⟩
625 Consultations
316 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More