TY - GEN
T1 - Towards Sparsification of Graph Neural Networks
AU - Peng, Hongwu
AU - Gurevin, Deniz
AU - Huang, Shaoyi
AU - Geng, Tong
AU - Jiang, Weiwen
AU - Khan, Orner
AU - Ding, Caiwen
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - As real-world graphs expand in size, larger GNN models with billions of parameters are deployed. High parameter count in such models makes training and inference on graphs expensive and challenging. To reduce the computational and memory costs of GNNs, optimization methods such as pruning the redundant nodes and edges in input graphs have been commonly adopted. However, model compression, which directly targets the sparsification of model layers, has been mostly limited to traditional Deep Neural Networks (DNNs) used for tasks such as image classification and object detection. In this paper, we utilize two state-of-the-art model compression methods (1) train and prune and (2) sparse training for the sparsification of weight layers in GNNs. We evaluate and compare the efficiency of both methods in terms of accuracy, training sparsity, and training FLOPs on real-world graphs. Our experimental results show that on the ia-email, wiki-talk, and stackoverflow datasets for link prediction, sparse training with much lower training FLOPs achieves a comparable accuracy with the train and prune method. On the brain dataset for node classification, sparse training uses a lower number FLOPs Oess than 1/7 FLOPs of train and prune method) and preserves a much better accuracy performance under extreme model sparsity. Our model sparsification code is publicly available on GitHubl1.
AB - As real-world graphs expand in size, larger GNN models with billions of parameters are deployed. High parameter count in such models makes training and inference on graphs expensive and challenging. To reduce the computational and memory costs of GNNs, optimization methods such as pruning the redundant nodes and edges in input graphs have been commonly adopted. However, model compression, which directly targets the sparsification of model layers, has been mostly limited to traditional Deep Neural Networks (DNNs) used for tasks such as image classification and object detection. In this paper, we utilize two state-of-the-art model compression methods (1) train and prune and (2) sparse training for the sparsification of weight layers in GNNs. We evaluate and compare the efficiency of both methods in terms of accuracy, training sparsity, and training FLOPs on real-world graphs. Our experimental results show that on the ia-email, wiki-talk, and stackoverflow datasets for link prediction, sparse training with much lower training FLOPs achieves a comparable accuracy with the train and prune method. On the brain dataset for node classification, sparse training uses a lower number FLOPs Oess than 1/7 FLOPs of train and prune method) and preserves a much better accuracy performance under extreme model sparsity. Our model sparsification code is publicly available on GitHubl1.
KW - GNN
KW - graph
KW - model compression
KW - sparse training
KW - sparsification
KW - Surrogate Lagrangian Relaxation (SLR)
UR - http://www.scopus.com/inward/record.url?scp=85145879210&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85145879210&partnerID=8YFLogxK
U2 - 10.1109/ICCD56317.2022.00048
DO - 10.1109/ICCD56317.2022.00048
M3 - Conference contribution
AN - SCOPUS:85145879210
T3 - Proceedings - IEEE International Conference on Computer Design: VLSI in Computers and Processors
SP - 272
EP - 279
BT - Proceedings - 2022 IEEE 40th International Conference on Computer Design, ICCD 2022
T2 - 40th IEEE International Conference on Computer Design, ICCD 2022
Y2 - 23 October 2022 through 26 October 2022
ER -