Evaluating the Impact of Mixed-Precision on Fault Propagation for Deep Neural Networks on GPUs - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

Evaluating the Impact of Mixed-Precision on Fault Propagation for Deep Neural Networks on GPUs

Résumé

Graphics Processing Units (GPUs) offer the possibility to execute floating-point operations (FLOP) with mixed-precisions such as INT8, FP16, Bfloat, FP32, and FP64. For Deep Neural Networks (DNNs), a reduced precision is likely to lower the execution time and power consumption as it requires a smaller hardware area and fewer clock cycles to perform instructions than the standard FP32 and FP64 precisions. As less area is needed for reduced precision, the circuit error rate is also expected to be lower [1]. NVIDIA GPUs also have tensor cores that perform matrix multiplication on hardware. The tensor cores are capable to perform a 4 ×4 FP16 matrix multiplication in one clock cycle [2]. The tensor cores can deliver up to 9 × higher performance than the software implementation of matrix multiplication (sequence of sums and multiplications) on GPUs and up to 47 ×than a CPU-based system [2].
Fichier principal
Vignette du fichier
isvlsi_2022_sps.pdf (122.08 Ko) Télécharger le fichier
isvlsi_2022_sps.zip (136.64 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03903347 , version 1 (16-12-2022)

Licence

Paternité

Identifiants

Citer

Fernando Fernandes dos Santos, Paolo Rech, Angeliki Kritikakou, Olivier Sentieys. Evaluating the Impact of Mixed-Precision on Fault Propagation for Deep Neural Networks on GPUs. ISVLSI 2022 - IEEE Computer Society Annual Symposium on VLSI, Jul 2022, Nicosia, Italy. pp.327-327, ⟨10.1109/ISVLSI54635.2022.00071⟩. ⟨hal-03903347⟩
18 Consultations
51 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More