FloatPIM: In-memory acceleration of deep neural network training with high precision

Mohsen Imani, Saransh Gupta, Yeseong Kim, Tajana Rosing

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

187 Scopus citations

Abstract

Processing In-Memory (PIM) has shown a great potential to accelerate inference tasks of Convolutional Neural Network (CNN). However, existing PIM architectures do not support high precision computation, e.g., in floating point precision, which is essential for training accurate CNN models. In addition, most of the existing PIM approaches require analog/mixed-signal circuits, which do not scale, exploiting insufficiently reliable multi-bit Non-Volatile Memory (NVM). In this paper, we propose FloatPIM, a fully-digital scalable PIM architecture that accelerates CNN in both training and testing phases. FloatPIM natively supports floating-point representation, thus enabling accurate CNN training. FloatPIM also enables fast communication between neighboring memory blocks to reduce internal data movement of the PIM architecture. We evaluate the efficiency of FloatPIM on ImageNet dataset using popular large-scale neural networks. Our evaluation shows that FloatPIM supporting floating point precision can achieve up to 5.1% higher classification accuracy as compared to existing PIM architectures with limited fixed-point precision. FloatPIM training is on average 303.2× and 48.6× (4.3× and 15.8×) faster and more energy efficient as compared to GTX 1080 GPU (PipeLayer [1] PIM accelerator). For testing, FloatPIM also provides 324.8× and 297.9× (6.3× and 21.6×) speedup and energy efficiency as compared to GPU (ISAAC [2] PIM accelerator) respectively.

Original languageEnglish
Title of host publicationISCA 2019 - Proceedings of the 2019 46th International Symposium on Computer Architecture
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages802-815
Number of pages14
ISBN (Electronic)9781450366694
DOIs
StatePublished - 22 Jun 2019
Event46th International Symposium on Computer Architecture, ISCA 2019 - Phoenix, United States
Duration: 22 Jun 201926 Jun 2019

Publication series

NameProceedings - International Symposium on Computer Architecture
ISSN (Print)1063-6897

Conference

Conference46th International Symposium on Computer Architecture, ISCA 2019
Country/TerritoryUnited States
CityPhoenix
Period22/06/1926/06/19

Bibliographical note

Publisher Copyright:
© 2019 ACM.

Keywords

  • Deep neural network
  • Machine learning acceleration
  • Non-volatile memory
  • Processing in-memory

Fingerprint

Dive into the research topics of 'FloatPIM: In-memory acceleration of deep neural network training with high precision'. Together they form a unique fingerprint.

Cite this