Fader Networks: Generating Image Variations by Sliding Attribute Values - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2017

Fader Networks: Generating Image Variations by Sliding Attribute Values

Guillaume Lample
  • Fonction : Auteur
  • PersonId : 1040249
Neil Zeghidour
  • Fonction : Auteur
Nicolas Usunier
  • Fonction : Auteur
Antoine Bordes
  • Fonction : Auteur
Marc'Aurelio Ranzato
  • Fonction : Auteur

Résumé

This paper introduces a new encoder-decoder architecture that is trained to reconstruct images by disentangling the salient information of the image and the values of attributes directly in the latent space. As a result, after training, our model can generate different realistic versions of an input image by varying the attribute values. By using continuous attribute values, we can choose how much a specific attribute is perceivable in the generated image. This property could allow for applications where users can modify an image using sliding knobs, like faders on a mixing console, to change the facial expression of a portrait, or to update the color of some objects. Compared to the state-of-the-art which mostly relies on training adversarial networks in pixel space by altering attribute values at train time, our approach results in much simpler training schemes and nicely scales to multiple attributes. We present evidence that our model can significantly change the perceived value of the attributes while preserving the naturalness of images.
Fichier non déposé

Dates et versions

hal-01949501 , version 1 (10-12-2018)

Identifiants

  • HAL Id : hal-01949501 , version 1

Citer

Guillaume Lample, Neil Zeghidour, Nicolas Usunier, Antoine Bordes, Ludovic Denoyer, et al.. Fader Networks: Generating Image Variations by Sliding Attribute Values. 31st Conference on Neural Information Processing Systems (NIPS 2017), Dec 2017, Long Beach, CA, United States. pp.5969-5978. ⟨hal-01949501⟩
111 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More