High-Efficiency Convolutional Ternary Neural Networks with Custom Adder Trees and Weight Compression
Résumé
Although performing inference with artiicial neural networks (ANN) was until quite recently considered as essentially compute intensive, the emergence of deep neural networks coupled with the evolution of the integration technology transformed inference into a memory bound problem. This ascertainment being established, many works have lately focused on minimizing memory accesses, either by enforcing and exploiting sparsity on weights or by using few bits for representing activations and weights, so as to be able to use ANNs inference in embedded devices. In this work, we detail an architecture dedicated to inference using ternary {−1, 0, 1} weights and activations. This architecture is conngurable at design time to provide throughput vs power trade-oos to choose from. It is also generic in the sense that it uses information drawn for the target technologies (memory geometries and cost, number of available cuts, etc) to adapt at best to the FPGA resources. This allows to achieve up to 5.2k fps per Watt for classiication on a VC709 board using approximately half of the resources of the FPGA. Additional Key Words and Phrases: Ternary CNN, low power inference, hardware acceleration, FPGA ACM Reference format: Adrien Prost-Boucle, Alban Bourge, and Frédéric Pétrot. 2018. High-EEciency Convolutional Ternary Neural Networks with Custom Adder Trees and Weight Compression.
Origine : Fichiers produits par l'(les) auteur(s)
Loading...