Convergence and Dynamical Behavior of the Adam Algorithm for Non Convex Stochastic Optimization - Département Image, Données, Signal Accéder directement au contenu
Article Dans Une Revue SIAM Journal on Optimization Année : 2021

Convergence and Dynamical Behavior of the Adam Algorithm for Non Convex Stochastic Optimization

Résumé

Adam is a popular variant of the stochastic gradient descent for finding a local minimizer of a function. The objective function is unknown but a random estimate of the current gradient vector is observed at each round of the algorithm. Assuming that the objective function is differentiable and non-convex, we establish the convergence in the long run of the iterates to a stationary point. The key ingredient is the introduction of a continuous-time version of Adam, under the form of a non-autonomous ordinary differential equation. The existence and the uniqueness of the solution are established, as well as the convergence of the solution towards the stationary points of the objective function. The continuous-time system is a relevant approximation of the Adam iterates, in the sense that the interpolated Adam process converges weakly to the solution to the ODE.
Fichier principal
Vignette du fichier
1810.02263.pdf (526.54 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-02366280 , version 1 (15-11-2019)
hal-02366280 , version 2 (18-11-2022)

Identifiants

  • HAL Id : hal-02366280 , version 2

Citer

Anas Barakat, Pascal Bianchi. Convergence and Dynamical Behavior of the Adam Algorithm for Non Convex Stochastic Optimization. SIAM Journal on Optimization, 2021. ⟨hal-02366280v2⟩
97 Consultations
429 Téléchargements

Partager

Gmail Facebook X LinkedIn More