Skip to Main content Skip to Navigation
Conference papers

Using Sentences as Semantic Representations in Large Scale Zero-Shot Learning

Abstract : Zero-shot learning aims to recognize instances of unseen classes, for which no visual instance is available during training, by learning mul-timodal relations between samples from seen classes and corresponding class semantic representations. These class representations usually consist of either attributes, which do not scale well to large datasets, or word embeddings, which lead to poorer performance. A good trade-off could be to employ short sentences in natural language as class descriptions. We explore different solutions to use such short descriptions in a ZSL setting and show that while simple methods cannot achieve very good results with sentences alone, a combination of usual word embeddings and sentences can significantly outperform current state-of-the-art 3 .
Document type :
Conference papers
Complete list of metadata
Contributor : Michel Crucianu Connect in order to contact the contributor
Submitted on : Friday, November 13, 2020 - 12:43:50 PM
Last modification on : Saturday, June 26, 2021 - 3:41:08 AM


Files produced by the author(s)


  • HAL Id : hal-03003689, version 1


Yannick Le Cacheux, Hervé Le Borgne, Michel Crucianu. Using Sentences as Semantic Representations in Large Scale Zero-Shot Learning. ECCV 2020 workshop Transferring and adapting source knowledge in computer vision (TASK-CV), Aug 2020, Glasgow, United Kingdom. ⟨hal-03003689⟩



Les métriques sont temporairement indisponibles