Skip to Main content Skip to Navigation
Conference papers

Greedy Layerwise Learning Can Scale to ImageNet

Abstract : Shallow supervised 1-hidden layer neural networks have a number of favorable properties that make them easier to interpret, analyze, and optimize than their deep counterparts, but lack their representational power. Here we use 1-hidden layer learning problems to sequentially build deep networks layer by layer, which can inherit properties from shallow networks. Contrary to previous approaches using shallow networks, we focus on problems where deep learning is reported as critical for success. We thus study CNNs on image classification tasks using the large-scale ImageNet dataset and the CIFAR-10 dataset. Using a simple set of ideas for architecture and training we find that solving sequential 1-hidden-layer auxiliary problems lead to a CNN that exceeds AlexNet performance on ImageNet. Extending this training methodology to construct individual layers by solving 2-and-3-hidden layer auxiliary problems , we obtain an 11-layer network that exceeds several members of the VGG model family on ImageNet, and can train a VGG-11 model to the same accuracy as end-to-end learning. To our knowledge, this is the first competitive alternative to end-to-end training of CNNs that can scale to ImageNet. We illustrate several interesting properties of these models theoretically and conduct a range of experiments to study the properties this training induces on the intermediate representations .
Complete list of metadatas
Contributor : Eugene Belilovsky <>
Submitted on : Friday, May 3, 2019 - 5:15:55 PM
Last modification on : Monday, July 20, 2020 - 12:34:52 PM


Files produced by the author(s)


  • HAL Id : hal-02119398, version 1
  • ARXIV : 1812.11446


Eugene Belilovsky, Michael Eickenberg, Edouard Oyallon. Greedy Layerwise Learning Can Scale to ImageNet. ICML 2019 - 36th International Conference on Machine Learning, Jun 2019, Long Beach, CA, United States. ⟨hal-02119398⟩



Record views


Files downloads