Convergence beyond the over-parameterized regime using Rayleigh quotients - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

Convergence beyond the over-parameterized regime using Rayleigh quotients

David Robin
  • Fonction : Auteur
  • PersonId : 1216866
  • IdHAL : robindar
Kevin Scaman
  • Fonction : Auteur
  • PersonId : 1062981
Marc Lelarge
  • Fonction : Auteur
  • PersonId : 833445

Résumé

In this paper, we present a new strategy to prove the convergence of deep learning architectures to a zero training (or even testing) loss by gradient flow. Our analysis is centered on the notion of Rayleigh quotients in order to prove Kurdyka-Łojasiewicz inequalities for a broader set of neural network architectures and loss functions. We show that Rayleigh quotients provide a unified view for several convergence analysis techniques in the literature. Our strategy produces a proof of convergence for various examples of parametric learning. In particular, our analysis does not require the number of parameters to tend to infinity, nor the number of samples to be finite, thus extending to test loss minimization and beyond the over-parameterized regime.
Fichier principal
Vignette du fichier
3433_convergence_beyond_the_over_pa.pdf (2.22 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03896153 , version 1 (13-12-2022)

Identifiants

  • HAL Id : hal-03896153 , version 1

Citer

David Robin, Kevin Scaman, Marc Lelarge. Convergence beyond the over-parameterized regime using Rayleigh quotients. NeurIPS 2022 - 36th Conference on Neural Information Processing System, Nov 2022, New Orleans, United States. ⟨hal-03896153⟩
35 Consultations
55 Téléchargements

Partager

Gmail Facebook X LinkedIn More