Learning maximum excluding ellipsoids from imbalanced data with theoretical guarantees

Abstract : In this paper, we address the problem of learning from imbalanced data. We consider the scenario where the number of negative examples is much larger than the number of positive ones. We propose a theoretically-founded method which learns a set of local ellipsoids centered at the minority class examples while excluding the negative examples of the majority class. We address this task from a Mahalanobis-like metric learning point of view and we derive generalization guarantees on the learned metric using the uniform stability framework. Our experimental evaluation on classic benchmarks and on a proprietary dataset in bank fraud detection shows the effectiveness of our approach, particularly when the imbalancy is huge.
Complete list of metadatas

Contributor : Guillaume Metzler <>
Submitted on : Friday, November 9, 2018 - 6:52:37 PM
Last modification on : Thursday, February 14, 2019 - 3:52:02 PM
Long-term archiving on : Sunday, February 10, 2019 - 3:14:36 PM


Files produced by the author(s)



Guillaume Metzler, Xavier Badiche, Brahim Belkasmi, Elisa Fromont, Amaury Habrard, et al.. Learning maximum excluding ellipsoids from imbalanced data with theoretical guarantees. Pattern Recognition Letters, Elsevier, 2018, 112, pp.310-316. ⟨10.1016/j.patrec.2018.08.016⟩. ⟨hal-01878830⟩



Record views


Files downloads