Abstract
Deep neural networks (DNN) have demonstrated effectiveness for various applications such as image processing, video segmentation, and speech recognition. Running state-of-theart DNNs on current systems mostly relies on either generalpurpose processors, ASIC designs, or FPGA accelerators, all of which suffer from data movements due to the limited on-chip memory and data transfer bandwidth. In this work, we propose a novel framework, called RAPIDNN, which performs neuron-to-memory transformation in order to accelerate DNNs in a highly parallel architecture. RAPIDNN reinterprets a DNN model and maps it into a specialized accelerator, which is designed using non-volatile memory blocks that model four fundamental DNN operations, i.e., multiplication, addition, activation functions, and pooling. The framework extracts representative operands of a DNN model, e.g., weights and input values, using clustering methods to optimize the model for in-memory processing. Then, it maps the extracted operands and their pre-computed results into the accelerator memory blocks. At runtime, the accelerator identifies computation results based on efficient in-memory search capability which also provides tunability of approximation to improve computation efficiency further. Our evaluation shows that RAPIDNN achieves 68.4×, 49.5× energy efficiency improvement and 48.1×, 10.9× speedup as compared to ISAAC and PipeLayer, the state-of-the-art DNN accelerators, while ensuring less than 0.5% quality loss.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2020 IEEE International Symposium on High Performance Computer Architecture, HPCA 2020 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 1-14 |
| Number of pages | 14 |
| ISBN (Electronic) | 9781728161495 |
| DOIs | |
| State | Published - Feb 2020 |
| Event | 26th IEEE International Symposium on High Performance Computer Architecture, HPCA 2020 - San Diego, United States Duration: 22 Feb 2020 → 26 Feb 2020 |
Publication series
| Name | Proceedings - 2020 IEEE International Symposium on High Performance Computer Architecture, HPCA 2020 |
|---|
Conference
| Conference | 26th IEEE International Symposium on High Performance Computer Architecture, HPCA 2020 |
|---|---|
| Country/Territory | United States |
| City | San Diego |
| Period | 22/02/20 → 26/02/20 |
Bibliographical note
Publisher Copyright:© 2020 IEEE.
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
Fingerprint
Dive into the research topics of 'Deep learning acceleration with neuron-to-memory transformation'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver