Efficient Compression Technique for NoC-based Deep Neural Network Accelerators - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Efficient Compression Technique for NoC-based Deep Neural Network Accelerators

Résumé

Deep Neural Networks (DNNs) are very powerful neural networks, widely used in many applications. On the other hand, such networks are computation and memory intensive, which makes their implementation difficult onto hardware-constrained systems, that could use network-on-chip as interconnect infrastructure. A way to reduce the traffic generated among memory and the processing elements is to compress the information before their exchange inside the network. In particular, our work focuses on reducing the huge number of DNN parameters, i.e., weights. In this paper, we propose a flexible and low-complexity compression technique which preserves the DNN performance, allowing to reduce the memory footprint and the volume of data to be exchanged while necessitating few hardware resources. The technique is evaluated on several DNN models, achieving a compression rate close to 80% without significant loss in accuracy on AlexNet, ResNet, or LeNet-5.
Fichier non déposé

Dates et versions

hal-02903208 , version 1 (20-07-2020)

Identifiants

  • HAL Id : hal-02903208 , version 1

Citer

Jordane Lorandel, Habiba Lahdhiri, Emmanuelle Bourdel, Salvatore Monteleone, Maurizio Palesi. Efficient Compression Technique for NoC-based Deep Neural Network Accelerators. Euromicro Conference on Digital System Design (DSD 2020), Aug 2020, Portoroz, Slovenia. ⟨hal-02903208⟩
86 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More