TY - GEN
T1 - CoDG-ReRAM
T2 - 40th IEEE International Conference on Computer Design, ICCD 2022
AU - Luo, Yixuan
AU - Behnam, Payman
AU - Thorat, Kiran
AU - Liu, Zhuo
AU - Peng, Hongwu
AU - Huang, Shaoyi
AU - Zhou, Shu
AU - Khan, Omer
AU - Tumanov, Alexey
AU - Ding, Caiwen
AU - Geng, Tong
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Graph Neural Networks (GCNs) have attracted wide attention and are applied to the real world. However, due to the ever-growing graph data with significant irregularities, off-chip communication with poor data locality has become the major bottleneck hurdling the development of GCNs. Fortunately, recent works demonstrate Resistive Random Access Memory (ReRAM) has the potential to perform inherently parallel in-situ computation of Matrix-Vector Multiplication (MVM) in the analog regime fundamentally breaking the communication bottleneck.Inspired by this observation, we propose a novel ReRAM-based GCN acceleration co-design (i.e. algorithm-hardware) framework, CoDG-ReRAM, that can deliver real-time GCN inference with high accuracy. On the algorithm side, we propose a novel model optimization pipeline that simultaneously and efficiently sparsifies and regularizes both graph and parameter matrices in GCNs and creates ReRAM-friendly models. On the hardware side, we take advantage of the software optimization results to provide a more systematic mapping scheme and in-crease computation efficiency to have an energy-efficient ReRAM-based GCN acceleration with low latency. Experimental results show that the proposed work improves performance and energy efficiency by 4× and 5.1 × respectively over SOTA ReRAM-based accelerators of GCNs with a negligible accuracy loss.
AB - Graph Neural Networks (GCNs) have attracted wide attention and are applied to the real world. However, due to the ever-growing graph data with significant irregularities, off-chip communication with poor data locality has become the major bottleneck hurdling the development of GCNs. Fortunately, recent works demonstrate Resistive Random Access Memory (ReRAM) has the potential to perform inherently parallel in-situ computation of Matrix-Vector Multiplication (MVM) in the analog regime fundamentally breaking the communication bottleneck.Inspired by this observation, we propose a novel ReRAM-based GCN acceleration co-design (i.e. algorithm-hardware) framework, CoDG-ReRAM, that can deliver real-time GCN inference with high accuracy. On the algorithm side, we propose a novel model optimization pipeline that simultaneously and efficiently sparsifies and regularizes both graph and parameter matrices in GCNs and creates ReRAM-friendly models. On the hardware side, we take advantage of the software optimization results to provide a more systematic mapping scheme and in-crease computation efficiency to have an energy-efficient ReRAM-based GCN acceleration with low latency. Experimental results show that the proposed work improves performance and energy efficiency by 4× and 5.1 × respectively over SOTA ReRAM-based accelerators of GCNs with a negligible accuracy loss.
KW - Computer Architecture
KW - Graph Neural Network
KW - Processing In Memory
KW - Resistive Random Access Memory
UR - http://www.scopus.com/inward/record.url?scp=85145883603&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85145883603&partnerID=8YFLogxK
U2 - 10.1109/ICCD56317.2022.00049
DO - 10.1109/ICCD56317.2022.00049
M3 - Conference contribution
AN - SCOPUS:85145883603
T3 - Proceedings - IEEE International Conference on Computer Design: VLSI in Computers and Processors
SP - 280
EP - 289
BT - Proceedings - 2022 IEEE 40th International Conference on Computer Design, ICCD 2022
Y2 - 23 October 2022 through 26 October 2022
ER -