Ternary Neural Networks for Resource-Efficient AI Applications - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2017

Ternary Neural Networks for Resource-Efficient AI Applications

Résumé

—The computation and storage requirements for Deep Neural Networks (DNNs) are usually high. This issue limits their deployability on ubiquitous computing devices such as smart phones, wearables and autonomous drones. In this paper, we propose ternary neural networks (TNNs) in order to make deep learning more resource-efficient. We train these TNNs using a teacher-student approach based on a novel, layer-wise greedy methodology. Thanks to our two-stage training procedure, the teacher network is still able to use state-of-the-art methods such as dropout and batch normalization to increase accuracy and reduce training time. Using only ternary weights and activations, the student ternary network learns to mimic the behavior of its teacher network without using any multiplication. Unlike its {-1,1} binary counterparts, a ternary neural network inherently prunes the smaller weights by setting them to zero during training. This makes them sparser and thus more energy-efficient. We design a purpose-built hardware architecture for TNNs and implement it on FPGA and ASIC. We evaluate TNNs on several benchmark datasets and demonstrate up to 3.1× better energy efficiency with respect to the state of the art while also improving accuracy.
Fichier principal
Vignette du fichier
paper_TNN_arxiv.pdf (364.62 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01481478 , version 1 (02-03-2017)

Identifiants

Citer

Hande Alemdar, Vincent Leroy, Adrien Prost-Boucle, Frédéric Pétrot. Ternary Neural Networks for Resource-Efficient AI Applications. International Joint Conference on Neural Networks, May 2017, Anchorage, United States. ⟨hal-01481478⟩
262 Consultations
1352 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More