GradPIM: A Practical Processing-in-DRAM Architecture for Gradient Descent
In this paper, we present GradPIM, a processing-in-memory architecture which accelerates parameter updates of deep neural networks training. As one of processing-in-memory techniques that could be realized in the near future, we propose an incremental, simple architectural design that does not invad...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In this paper, we present GradPIM, a processing-in-memory architecture which
accelerates parameter updates of deep neural networks training. As one of
processing-in-memory techniques that could be realized in the near future, we
propose an incremental, simple architectural design that does not invade the
existing memory protocol. Extending DDR4 SDRAM to utilize bank-group
parallelism makes our operation designs in processing-in-memory (PIM) module
efficient in terms of hardware cost and performance. Our experimental results
show that the proposed architecture can improve the performance of DNN training
and greatly reduce memory bandwidth requirement while posing only a minimal
amount of overhead to the protocol and DRAM area. |
---|---|
DOI: | 10.48550/arxiv.2102.07511 |