Skip to Main content Skip to Navigation
Conference papers

Multi-View Data Generation Without View Supervision

Abstract : The development of high-dimensional generative models has recently gained a great surge of interest with the introduction of variational auto-encoders and generative adversarial neural networks. Different variants have been proposed where the underlying latent space is structured, for example, based on attributes describing the data to generate. We focus on a particular problem where one aims at generating samples corresponding to a number of objects under various views. We assume that the distribution of the data is driven by two independent latent factors: the content, which represents the intrinsic features of an object, and the view, which stands for the settings of a particular observation of that object. Therefore, we propose a generative model and a conditional variant built on such a disentangled latent space. This approach allows us to generate realistic samples corresponding to various objects in a high variety of views. Unlike many multi-view approaches, our model doesn't need any supervision on the views but only on the content. Compared to other conditional generation approaches that are mostly based on binary or categorical attributes, we make no such assumption about the factors of variations. Our model can be used on problems with a huge, potentially infinite, number of categories. We experiment it on four image datasets on which we demonstrate the effectiveness of the model and its ability to generalize.
Complete list of metadatas

Cited literature [30 references]  Display  Hide  Download
Contributor : Mickael Chen <>
Submitted on : Tuesday, April 16, 2019 - 4:37:40 PM
Last modification on : Thursday, January 23, 2020 - 6:22:13 PM


multi-view-data (1).pdf
Files produced by the author(s)


  • HAL Id : hal-02101404, version 1
  • ARXIV : 1711.00305


Mickaël Chen, Ludovic Denoyer, Thierry Artières. Multi-View Data Generation Without View Supervision. 6th International Conference on Learning Representations (ICLR 2018), Apr 2018, Vancouver, Canada. ⟨hal-02101404⟩



Record views


Files downloads