Skip to Main content Skip to Navigation
Preprints, Working Papers, ...

Distilled Hierarchical Neural Ensembles with Adaptive Inference Cost

Abstract : Deep neural networks form the basis of state-of-the-art models across a variety of application domains. Moreover, networks that are able to dynamically adapt the computational cost of inference are important in scenarios where the amount of compute or input data varies over time. In this paper, we propose Hierarchical Neural Ensembles (HNE), a novel framework to embed an ensemble of multiple networks by sharing intermediate layers using a hierarchical structure. In HNE we control the inference cost by evaluating only a subset of models, which are organized in a nested manner. Our second contribution is a novel co-distillation method to boost the performance of ensemble predictions with low inference cost. This approach leverages the nested structure of our ensembles, to optimally allocate accuracy and diversity across the ensemble members. Comprehensive experiments over the CIFAR and Ima-geNet datasets confirm the effectiveness of HNE in building deep networks with adaptive inference cost for image classification.
Complete list of metadatas

Cited literature [54 references]  Display  Hide  Download

https://hal.inria.fr/hal-02500660
Contributor : Jakob Verbeek <>
Submitted on : Friday, March 6, 2020 - 11:31:37 AM
Last modification on : Friday, July 3, 2020 - 4:52:52 PM
Document(s) archivé(s) le : Sunday, June 7, 2020 - 1:26:29 PM

File

Distilled_Hierarchical_Neural_...
Files produced by the author(s)

Identifiers

  • HAL Id : hal-02500660, version 1

Collections

Citation

Adrià Ruiz, Jakob Verbeek. Distilled Hierarchical Neural Ensembles with Adaptive Inference Cost. 2020. ⟨hal-02500660⟩

Share

Metrics

Record views

119

Files downloads

300