Learning maximum excluding ellipsoids from imbalanced data with theoretical guarantees

Abstract : In this paper, we address the problem of learning from imbalanced data. We consider the scenario where the number of negative examples is much larger than the number of positive ones. We propose a theoretically-founded method which learns a set of local ellipsoids centered at the minority class examples while excluding the negative examples of the majority class. We address this task from a Mahalanobis-like metric learning point of view and we derive generalization guarantees on the learned metric using the uniform stability framework. Our experimental evaluation on classic benchmarks and on a proprietary dataset in bank fraud detection shows the effectiveness of our approach, particularly when the imbalancy is huge.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-01878830
Contributor : Guillaume Metzler <>
Submitted on : Friday, November 9, 2018 - 6:52:37 PM
Last modification on : Thursday, February 14, 2019 - 3:52:02 PM
Long-term archiving on : Sunday, February 10, 2019 - 3:14:36 PM

File

PRL_format.pdf
Files produced by the author(s)

Identifiers

Citation

Guillaume Metzler, Xavier Badiche, Brahim Belkasmi, Elisa Fromont, Amaury Habrard, et al.. Learning maximum excluding ellipsoids from imbalanced data with theoretical guarantees. Pattern Recognition Letters, Elsevier, 2018, 112, pp.310-316. ⟨10.1016/j.patrec.2018.08.016⟩. ⟨hal-01878830⟩

Share

Metrics

Record views

122

Files downloads

118