Optimal Memory-aware Backpropagation of Deep Join Networks - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Article Dans Une Revue Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences Année : 2019

Optimal Memory-aware Backpropagation of Deep Join Networks

Résumé

Deep Learning training memory needs can prevent the user to consider large models and large batch sizes. In this work, we propose to use techniques from memory-aware scheduling and Automatic Differentiation (AD) to execute a backpropagation graph with a bounded memory requirement at the cost of extra recomputations. The case of a single homogeneous chain, i.e. the case of a network whose all stages are identical and form a chain, is well understood and optimal solutions have been proposed in the AD literature. The networks encountered in practice in the context of Deep Learning are much more diverse, both in terms of shape and heterogeneity. In this work, we define the class of backpropagation graphs, and extend those on which one can compute in polynomial time a solution that minimizes the total number of recomputations. In particular we consider join graphs which correspond to models such as Siamese or Cross Modal Networks.
Fichier principal
Vignette du fichier
rsta_rev2.pdf (336.26 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-02401105 , version 1 (09-12-2019)

Identifiants

  • HAL Id : hal-02401105 , version 1

Citer

Olivier Beaumont, Julien Herrmann, Guillaume Pallez, Alena Shilova. Optimal Memory-aware Backpropagation of Deep Join Networks. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, In press. ⟨hal-02401105⟩
71 Consultations
167 Téléchargements

Partager

Gmail Facebook X LinkedIn More