Bilevel Learning of Deep Representations - Archive ouverte HAL Access content directly
Preprints, Working Papers, ... Year : 2021

Bilevel Learning of Deep Representations

Abstract

We present a framework based on bilevel optimization for learning multilayer, deep data representations. While the lower-level problem implicitly defines the representation through the critical point of a learnable objective, the upper-level problem optimizes the representation mapping. We reformulate the problem via a majorizationminimization algorithm. On one hand, for some quadratic majorants, we show that the bilevel problem reduces down to the training of a feedforward neural network. On the other hand, for majorants based on Bregman distances, we introduce a new neural network architecture involving the inverse of the activation function. We argue theoretically that the novel architecture may have better mathematical properties than standard networks. Numerical experiments show that the proposed variant benefits from better training behaviors, resulting in more accurate prediction.
Fichier principal
Vignette du fichier
2021_Bregmanet_HAL.pdf (996.84 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-03132512 , version 1 (05-02-2021)
hal-03132512 , version 2 (10-02-2022)
hal-03132512 , version 3 (16-02-2022)

Identifiers

  • HAL Id : hal-03132512 , version 1

Cite

Jordan Frecon, Saverio Salzo, Massimiliano Pontil. Bilevel Learning of Deep Representations. 2021. ⟨hal-03132512v1⟩
178 View
298 Download

Share

Gmail Facebook X LinkedIn More