k-NN Boosting Prototype Learning for Object Classification

Abstract : Object classification is a challenging task in computer vision. Many approaches have been proposed to extract meaningful descriptors from images and classifying them in a supervised learning framework. In this paper, we revisit the classic k-nearest neighbors (k-NN) classification rule, which has shown to be very effective when dealing with local image descriptors. However, k-NN still features some major drawbacks, mainly due to the uniform voting among the nearest prototypes in the feature space. In this paper, we propose a generalization of the classic k-NN rule in a supervised learning (boosting) framework. Namely, we redefine the voting rule as a strong classifier that linearly combines predictions from the k closest prototypes. To induce this classifier, we propose a novel learning algorithm, MLNN (Multiclass Leveraged Nearest Neighbors), which gives a simple procedure for performing prototype selection very efficiently. We tested our method on 12 categories of objects, and observed significant improvement over classic k-NN in terms of classification performances.
Document type :
Conference papers
Liste complète des métadonnées

Cited literature [10 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-00481725
Contributor : Paolo Piro <>
Submitted on : Friday, May 7, 2010 - 11:14:32 AM
Last modification on : Wednesday, March 27, 2019 - 4:41:27 PM
Document(s) archivé(s) le : Thursday, September 16, 2010 - 1:49:39 PM

File

pbnn_wiamis_10.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00481725, version 1

Citation

Paolo Piro, Michel Barlaud, Richard Nock, Frank Nielsen. k-NN Boosting Prototype Learning for Object Classification. WIAMIS 2010 - 11th Workshop on Image Analysis for Multimedia Interactive Services, Apr 2010, Desenzano del Garda, Italy. pp.1-4. ⟨hal-00481725⟩

Share

Metrics

Record views

329

Files downloads

103