Data fusion through cross-modality metric learning using similarity-sensitive hashing

Abstract : Visual understanding is often based on measuring similarity between observations. Learning similarities specific to a certain perception task from a set of examples has been shown advantageous in various computer vision and pattern recognition problems. In many important applications, the data that one needs to compare come from different representations or modalities, and the similarity between such data operates on objects that may have different and often incommensurable structure and dimensionality. In this paper, we propose a framework for supervised similarity learning based on embedding the input data from two arbitrary spaces into the Hamming space. The mapping is expressed as a binary classification problem with positive and negative examples, and can be efficiently learned using boosting algorithms. The utility and efficiency of such a generic approach is demonstrated on several challenging applications including cross-representation shape retrieval and alignment of multi-modal medical images.
Document type :
Conference papers
Liste complète des métadonnées
Contributor : Vivien Fécamp <>
Submitted on : Friday, August 30, 2013 - 1:56:37 PM
Last modification on : Tuesday, February 5, 2019 - 1:52:14 PM




Michael Bronstein, Alexander Bronstein, Fabrice Michel, Nikos Paragios. Data fusion through cross-modality metric learning using similarity-sensitive hashing. 23rd IEEE Conference on Computer Vision and Pattern Recognition - CVPR 2010, Jun 2010, San Francisco, United States. pp.3594-3601, ⟨10.1109/CVPR.2010.5539928⟩. ⟨hal-00856061⟩



Record views