Skip to Main content Skip to Navigation
Conference papers

Combinatorial Tiling for Sparse Neural Networks

Abstract : Sparse deep neural networks (DNNs) emerged as the result of search for networks with less storage and lower computational complexity. The sparse DNN inference is the task of using such trained DNN networks to classify a batch of input data. We propose an efficient, hybrid model- and data-parallel DNN inference using hypergraph models and partitioners. We exploit tiling and weak synchronization to increase cache reuse, hide load imbalance, and hide synchronization costs. Finally, a blocking approach allows application of this new hybrid inference procedure for deep neural networks. We initially experiment using the hybrid tiled inference approach only, using the first five layers of networks from the IEEE HPEC 2019 Graph Challenge, and attain up to 2x speedup versus a data-parallel baseline
Complete list of metadata

Cited literature [20 references]  Display  Hide  Download
Contributor : Bora Uçar Connect in order to contact the contributor
Submitted on : Thursday, September 3, 2020 - 9:23:32 PM
Last modification on : Friday, September 30, 2022 - 4:12:24 AM


Files produced by the author(s)


  • HAL Id : hal-02910997, version 3



Filip Pawłowski, Rob H Bisseling, Bora Uçar, Albert-Jan N Yzelman. Combinatorial Tiling for Sparse Neural Networks. 2020 IEEE High Performance Extreme Computing (virtual conference), Sep 2020, Waltham, MA, United States. ⟨hal-02910997v3⟩



Record views


Files downloads