On the analysis of adaptability in multi-source domain adaptation - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue Machine Learning Année : 2019

On the analysis of adaptability in multi-source domain adaptation

Ievgen Redko
Amaury Habrard
Marc Sebban

Résumé

In many real-world applications, it may be desirable to benefit from a classi-fier trained on a given source task from some largely annotated dataset in order to address a different but related target task for which only weakly labeled data are available. Domain adaptation (DA) is the framework which aims at leveraging the statistical similarities between the source and target distributions to learn well. Current theoretical results show that the efficiency of DA algorithms depends on (i) their capacity of minimizing the divergence between the source and target domains and (ii) the existence of a good hypothesis that commits few errors in both domains. While most of the work in DA has focused on new divergence measures, the second aspect, often modeled as the capability term, remains surprisingly under investigated. In this paper, we show that the problem of the best joint hypothesis estimation can be reformulated using a Wasserstein distance-based error function in the context of multi-source DA. Based on this idea, we provide a theoretical analysis of the capability term and derive inequalities allowing us to estimate it from finite samples. We empirically illustrate the proposed idea on different data sets.
Fichier principal
Vignette du fichier
MLJ_lambda.pdf (1.31 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-02168941 , version 1 (30-06-2019)

Identifiants

Citer

Ievgen Redko, Amaury Habrard, Marc Sebban. On the analysis of adaptability in multi-source domain adaptation. Machine Learning, 2019, ⟨10.1007/s10994-019-05823-0⟩. ⟨hal-02168941⟩
109 Consultations
181 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More