Abstract
In this work, we propose an energy-efficient hardware accelerator for Deep Neural Network (DNN) using Low Rank Approximation (LRADNN). Using this scheme, inactive neurons in each layer of the DNN are dynamically identified and the corresponding computations are then bypassed. Accordingly, both the memory accesses and the arithmetic operations associated with these inactive neurons can be saved. Therefore, compared to the architectures using the direct feed-forward algorithm, LRADNN can achieve a higher throughput as well as a lower energy consumption with negligible prediction accuracy loss (within 0.1%). We implement and synthesize the proposed accelerator using TSMC 65nm technology. From the experimental results, a 31% to 53% energy reduction together with a 22% to 43% throughput increase can be achieved.
| Original language | English |
|---|---|
| Title of host publication | 2016 21st Asia and South Pacific Design Automation Conference, ASP-DAC 2016 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 581-586 |
| Number of pages | 6 |
| ISBN (Electronic) | 9781467395694 |
| DOIs | |
| Publication status | Published - 7 Mar 2016 |
| Event | 21st Asia and South Pacific Design Automation Conference, ASP-DAC 2016 - Macao, Macao Duration: 25 Jan 2016 → 28 Jan 2016 |
Publication series
| Name | Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC |
|---|---|
| Volume | 25-28-January-2016 |
Conference
| Conference | 21st Asia and South Pacific Design Automation Conference, ASP-DAC 2016 |
|---|---|
| Country/Territory | Macao |
| City | Macao |
| Period | 25/01/16 → 28/01/16 |
Bibliographical note
Publisher Copyright:© 2016 IEEE.
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
Fingerprint
Dive into the research topics of 'LRADNN: High-throughput and energy-efficient Deep Neural Network accelerator using Low Rank Approximation'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver