# An Inertial Newton Algorithm for Deep Learning

1 IRIT-SC - Signal et Communications
IRIT - Institut de recherche en informatique de Toulouse
4 IRIT-ADRIA - Argumentation, Décision, Raisonnement, Incertitude et Apprentissage
IRIT - Institut de recherche en informatique de Toulouse
Abstract : We introduce a new second-order inertial optimization method for machine learning called INNA. It exploits the geometry of the loss function while only requiring stochastic approximations of the function values and the generalized gradients. This makes INNA fully implementable and adapted to large-scale optimization problems such as the training of deep neural networks. The algorithm combines both gradient-descent and Newton-like behaviors as well as inertia. We prove the convergence of INNA for most deep learning problems. To do so, we provide a well-suited framework to analyze deep learning loss functions involving tame optimization in which we study a continuous dynamical system together with its discrete stochastic approximations. We prove sublinear convergence for the continuous-time differential inclusion which underlies our algorithm. Additionally, we also show how standard optimization mini-batch methods applied to non-smooth non-convex problems can yield a certain type of spurious stationary points never discussed before. We address this issue by providing a theoretical framework around the new idea of $D$-criticality; we then give a simple asymptotic analysis of INNA. Our algorithm allows for using an aggressive learning rate of $o(1/\log k)$. From an empirical viewpoint, we show that INNA returns competitive results with respect to state of the art (stochastic gradient descent, ADAGRAD, ADAM) on popular deep learning benchmark problems.
Keywords :
Document type :
Journal articles
Domain :

https://hal.archives-ouvertes.fr/hal-02140748
Contributor : Camille Castera Connect in order to contact the contributor
Submitted on : Friday, August 20, 2021 - 2:28:19 PM
Last modification on : Monday, July 4, 2022 - 9:13:27 AM

### File

arxiv.pdf
Files produced by the author(s)

### Identifiers

• HAL Id : hal-02140748, version 6
• ARXIV : 1905.12278

### Citation

Camille Castera, Jérôme Bolte, Cédric Févotte, Edouard Pauwels. An Inertial Newton Algorithm for Deep Learning. Journal of Machine Learning Research, Microtome Publishing, 2021, 22 (134). ⟨hal-02140748v6⟩

Record views