Multiview self-learning

Abstract : In many applications, observations are available with different views. This is, for example, the case with image-text classification, multilingual document classification or document classification on the web. In addition, unlabeled multiview examples can be easily acquired, but assigning labels to these examples is usually a time consuming task. We describe a multiview self-learning strategy which trains different voting classifiers on different views. The margin distributions over the unlabeled training data, obtained with each view-specific classifier are then used to estimate an upper-bound on their transductive Bayes error. Minimizing this upper-bound provides an automatic margin-threshold which is used to assign pseudo-labels to unlabeled examples. Final class labels are then assigned to these examples, by taking a vote on the pool of the previous pseudo-labels. New view-specific classifiers are then trained using the labeled and pseudo-labeled training data. We consider applications to image-text classification and to multilingual document classification. We present experimental results on the NUS-WIDE collection and on Reuters RCV1-RCV2 which show that despite its simplicity, our approach is competitive with other state-of-the-art techniques.
Type de document :
Article dans une revue
Liste complète des métadonnées

https://hal.archives-ouvertes.fr/hal-01211216
Contributeur : Massih-Reza Amini <>
Soumis le : dimanche 4 octobre 2015 - 07:54:17
Dernière modification le : mercredi 29 novembre 2017 - 15:25:01

Identifiants

Collections

Citation

Ali Fakeri-Tabrizi, Massih-Reza Amini, Goutte Cyril, Nicolas Usunier. Multiview self-learning. Neurocomputing, Elsevier, 2015, 155, pp.117-127. 〈http://www.sciencedirect.com/science/article/pii/S0925231214017056〉. 〈10.1016/j.neucom.2014.12.041〉. 〈hal-01211216〉

Partager

Métriques

Consultations de la notice

200