Skip to Main content Skip to Navigation
Preprints, Working Papers, ...

Dissimilarity Measure Machines

Abstract : This paper presents a dissimilarity-based discriminative framework for learning from data coming in the form of probability distributions. Departing from the use of positive kernel-based methods, we build upon embeddings based on dissimilarities tailored for distribution. We enable this by extending \citet{balcan2008theory}'s theory of learning with similarity functions to the case of distribution-shaped data. Then, we show that several learning guarantees of the dissimilarity still hold when estimated from empirical distributions. Algorithmically, the proposed approach consists in building features from pairwise dissimilarities and in learning a linear decision function in this new feature space. Our experimental results show that this dissimilarity-based approach works better than the so-called support measure machines or the sliced Wasserstein kernel, and that among several dissimilarities including Kullback-Leibler divergence and Maximum Mean Discrepancy, the entropy-regularized Wasserstein distance provides the best compromise between computational efficiency and accuracy.
Complete list of metadata
Contributor : Alain Rakotomamonjy Connect in order to contact the contributor
Submitted on : Wednesday, November 7, 2018 - 11:08:29 PM
Last modification on : Wednesday, November 3, 2021 - 5:17:29 AM
Long-term archiving on: : Friday, February 8, 2019 - 4:12:15 PM


Files produced by the author(s)


  • HAL Id : hal-01717940, version 2
  • ARXIV : 1803.00250


Alain Rakotomamonjy, Abraham Traoré, Maxime Berar, Rémi Flamary, Nicolas Courty, et al.. Dissimilarity Measure Machines. 2018. ⟨hal-01717940v2⟩



Record views


Files downloads