A New Boosting Algorithm for Provably Accurate Unsupervised Domain Adaptation

Abstract : Domain Adaptation (DA) is a new learning framework dealing with learning problems where the target test data are drawn from a distribution different from the one that has generated the learning source data. In this article, we introduce SLDAB (Self-Labeling Domain Adaptation Boosting), a new DA algorithm that falls both within the theory of DA and the theory of Boosting, allowing us to derive strong theoretical properties. SLDAB stands in the unsupervised DA setting where labeled data are only available in the source domain. To deal with this more complex situation, the strategy of SLDAB consists in jointly minimizing the empirical error on the source domain while limiting the violations of a natural notion of pseudo-margin over the target domain instances. Another contribution of this paper is the definition of a new divergence measure aiming at penalizing models that induce a large discrepancy between the two domains, reducing the production of degenerate models. We provide several theoretical results that justify this strategy. The practical efficiency of our model is assessed on two widely used datasets.
Document type :
Journal articles
Complete list of metadatas

Contributor : Marc Sebban <>
Submitted on : Thursday, April 23, 2015 - 8:57:18 AM
Last modification on : Thursday, July 26, 2018 - 1:10:46 AM



Amaury Habrard, Jean-Philippe Peyrache, Marc Sebban. A New Boosting Algorithm for Provably Accurate Unsupervised Domain Adaptation. Knowledge and Information Systems (KAIS), Springer, 2015, pp.1. ⟨10.1007/s10115-015-0839-2⟩. ⟨hal-01144896⟩



Record views