Abstract
Supporting variable precision for computing quantized neural network in a hardware accelerator is an efficient way to reduce overall computation time and energy. However, in the previous precision-scalable hardware, bit-reconfiguration logic increases the chip area significantly. In this paper, we demonstrate a compact precision-scalable accelerator chip using bitwise summation and channel-wise aligning schemes. The measurement results show that the peak performance per compute area is improved by 5.1-7.7x and system-level energy-efficiency is improved by up to 64% compared to previous precision-scalable accelerators.
| Original language | English |
|---|---|
| Title of host publication | 2020 IEEE Custom Integrated Circuits Conference, CICC 2020 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| ISBN (Electronic) | 9781728160313 |
| DOIs | |
| State | Published - Mar 2020 |
| Externally published | Yes |
| Event | 2020 IEEE Custom Integrated Circuits Conference, CICC 2020 - Boston, United States Duration: 22 Mar 2020 → 25 Mar 2020 |
Publication series
| Name | Proceedings of the Custom Integrated Circuits Conference |
|---|---|
| Volume | 2020-March |
| ISSN (Print) | 0886-5930 |
Conference
| Conference | 2020 IEEE Custom Integrated Circuits Conference, CICC 2020 |
|---|---|
| Country/Territory | United States |
| City | Boston |
| Period | 22/03/20 → 25/03/20 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
Keywords
- Deep neural network
- hardware accelerator. bit-precision scaling
- multiply-accumulate unit
Fingerprint
Dive into the research topics of 'A 44.1TOPS/W Precision-Scalable Accelerator for Quantized Neural Networks in 28nm CMOS'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver