Block-wise Training of Residual Networks via the Minimizing Movement Scheme - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

Block-wise Training of Residual Networks via the Minimizing Movement Scheme

Résumé

End-to-end backpropagation has a few shortcomings: it requires loading the entire model during training, which can be impossible in constrained settings, and suffers from three locking problems (forward locking, update locking and backward locking), which prohibit training the layers in parallel. Solving layer-wise optimization problems can address these problems and has been used in on-device training of neural networks. We develop a layer-wise training method, particularly welladapted to ResNets, inspired by the minimizing movement scheme for gradient flows in distribution space. The method amounts to a kinetic energy regularization of each block that makes the blocks optimal transport maps and endows them with regularity. It works by alleviating the stagnation problem observed in layer-wise training, whereby greedily-trained early layers overfit and deeper layers stop increasing test accuracy after a certain depth. We show on classification tasks that the test accuracy of block-wise trained ResNets is improved when using our method, whether the blocks are trained sequentially or in parallel.
Fichier principal
Vignette du fichier
blockwise.pdf (901.75 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-04108676 , version 1 (28-05-2023)

Identifiants

  • HAL Id : hal-04108676 , version 1

Citer

Skander Karkar, Ibrahim Ayed, Emmanuel de Bezenac, Patrick Gallinari. Block-wise Training of Residual Networks via the Minimizing Movement Scheme. 1st International Workshop on Practical Deep Learning in the Wild at 26th AAAI Conference on Artificial Intelligence 2022, AAAI, Feb 2022, Vancouver, Canada. ⟨hal-04108676⟩
32 Consultations
28 Téléchargements

Partager

Gmail Facebook X LinkedIn More