Skip to Main content Skip to Navigation
Conference papers

k-NN Boosting Prototype Learning for Object Classification

Abstract : Object classification is a challenging task in computer vision. Many approaches have been proposed to extract meaningful descriptors from images and classifying them in a supervised learning framework. In this paper, we revisit the classic k-nearest neighbors (k-NN) classification rule, which has shown to be very effective when dealing with local image descriptors. However, k-NN still features some major drawbacks, mainly due to the uniform voting among the nearest prototypes in the feature space. In this paper, we propose a generalization of the classic k-NN rule in a supervised learning (boosting) framework. Namely, we redefine the voting rule as a strong classifier that linearly combines predictions from the k closest prototypes. To induce this classifier, we propose a novel learning algorithm, MLNN (Multiclass Leveraged Nearest Neighbors), which gives a simple procedure for performing prototype selection very efficiently. We tested our method on 12 categories of objects, and observed significant improvement over classic k-NN in terms of classification performances.
Document type :
Conference papers
Complete list of metadata

Cited literature [10 references]  Display  Hide  Download
Contributor : Paolo Piro Connect in order to contact the contributor
Submitted on : Friday, May 7, 2010 - 11:14:32 AM
Last modification on : Thursday, August 4, 2022 - 4:53:49 PM
Long-term archiving on: : Thursday, September 16, 2010 - 1:49:39 PM


Files produced by the author(s)


  • HAL Id : hal-00481725, version 1



Paolo Piro, Michel Barlaud, Richard Nock, Frank Nielsen. k-NN Boosting Prototype Learning for Object Classification. WIAMIS 2010 - 11th Workshop on Image Analysis for Multimedia Interactive Services, Apr 2010, Desenzano del Garda, Italy. pp.1-4. ⟨hal-00481725⟩



Record views


Files downloads