Skip to Main content Skip to Navigation
Conference papers

k-NN Boosting Prototype Learning for Object Classification

Abstract : Object classification is a challenging task in computer vision. Many approaches have been proposed to extract meaningful descriptors from images and classifying them in a supervised learning framework. In this paper, we revisit the classic k-nearest neighbors (k-NN) classification rule, which has shown to be very effective when dealing with local image descriptors. However, k-NN still features some major drawbacks, mainly due to the uniform voting among the nearest prototypes in the feature space. In this paper, we propose a generalization of the classic k-NN rule in a supervised learning (boosting) framework. Namely, we redefine the voting rule as a strong classifier that linearly combines predictions from the k closest prototypes. To induce this classifier, we propose a novel learning algorithm, MLNN (Multiclass Leveraged Nearest Neighbors), which gives a simple procedure for performing prototype selection very efficiently. We tested our method on 12 categories of objects, and observed significant improvement over classic k-NN in terms of classification performances.
Complete list of metadatas

Cited literature [10 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-00481725
Contributor : Paolo Piro <>
Submitted on : Friday, May 7, 2010 - 11:14:32 AM
Last modification on : Tuesday, May 26, 2020 - 6:50:35 PM
Document(s) archivé(s) le : Thursday, September 16, 2010 - 1:49:39 PM

File

pbnn_wiamis_10.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00481725, version 1

Collections

Citation

Paolo Piro, Michel Barlaud, Richard Nock, Frank Nielsen. k-NN Boosting Prototype Learning for Object Classification. WIAMIS 2010 - 11th Workshop on Image Analysis for Multimedia Interactive Services, Apr 2010, Desenzano del Garda, Italy. pp.1-4. ⟨hal-00481725⟩

Share

Metrics

Record views

394

Files downloads

225