TY - GEN
T1 - DR-CircuitGNN
T2 - 39th ACM International Conference on Supercomputing, ICS 2025
AU - Luo, Yuebo
AU - Li, Shiyang
AU - Tao, Junran
AU - Thorat, Kiran Gautam
AU - Xie, Xi
AU - Peng, Hongwu
AU - Xu, Nuo
AU - Ding, Caiwen
AU - Huang, Shaoyi
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2025/8/22
Y1 - 2025/8/22
N2 - The increasing scale and complexity of integrated circuit design have led to increased challenges in Electronic Design Automation (EDA). Graph Neural Networks (GNNs), have emerged as a promising approach to assist EDA design as circuits can be naturally represented as graph. While GNNs offer a foundation for circuit analysis, they often fail to capture the full complexity of EDA designs. Heterogeneous Graph Neural Networks (HGNNs) can better interpret EDA circuit graphs as they capture both topological relationships and geometric features. However, the improved representation capability comes at the cost of even higher computational complexity and processing cost due to their serial module-wise message-passing scheme, creating a significant performance bottleneck. In this paper, we propose DR-CircuitGNN, a fast GPU kernel design by leveraging row-wise sparsity-aware Dynamic-ReLU and optimizing SpMM kernels during heterogeneous message-passing to accelerate HGNNs training on EDA-related circuit graph datasets. To further enhance performance, we propose a parallel optimization strategy that maximizes CPU-GPU concurrency by concurrently processing independent subgraphs using multi-threaded CPU initialization and GPU kernel execution via multiple cudaStreams. Our experiments show that on three representative CircuitNet designs (small, medium, large), the proposed method can achieve up to 3.51 × and 4.09 × speedup compared to the SOTA for forward and backward propagation, respectively. On full-size CircuitNet and sampled Mini-CircuitNet, our parallel design enables up to 2.71 × speed up over the official DGL implementation cuSPARSE with negligible impact on correlation scores and error rates.
AB - The increasing scale and complexity of integrated circuit design have led to increased challenges in Electronic Design Automation (EDA). Graph Neural Networks (GNNs), have emerged as a promising approach to assist EDA design as circuits can be naturally represented as graph. While GNNs offer a foundation for circuit analysis, they often fail to capture the full complexity of EDA designs. Heterogeneous Graph Neural Networks (HGNNs) can better interpret EDA circuit graphs as they capture both topological relationships and geometric features. However, the improved representation capability comes at the cost of even higher computational complexity and processing cost due to their serial module-wise message-passing scheme, creating a significant performance bottleneck. In this paper, we propose DR-CircuitGNN, a fast GPU kernel design by leveraging row-wise sparsity-aware Dynamic-ReLU and optimizing SpMM kernels during heterogeneous message-passing to accelerate HGNNs training on EDA-related circuit graph datasets. To further enhance performance, we propose a parallel optimization strategy that maximizes CPU-GPU concurrency by concurrently processing independent subgraphs using multi-threaded CPU initialization and GPU kernel execution via multiple cudaStreams. Our experiments show that on three representative CircuitNet designs (small, medium, large), the proposed method can achieve up to 3.51 × and 4.09 × speedup compared to the SOTA for forward and backward propagation, respectively. On full-size CircuitNet and sampled Mini-CircuitNet, our parallel design enables up to 2.71 × speed up over the official DGL implementation cuSPARSE with negligible impact on correlation scores and error rates.
KW - congestion prediction
KW - Electronic Design Automation
KW - Heterogeneous Graph Neural Network
KW - Sparse Matrix Multiplication kernels
UR - https://www.scopus.com/pages/publications/105021470873
UR - https://www.scopus.com/pages/publications/105021470873#tab=citedBy
U2 - 10.1145/3721145.3734528
DO - 10.1145/3721145.3734528
M3 - Conference contribution
AN - SCOPUS:105021470873
T3 - Proceedings of the International Conference on Supercomputing
SP - 221
EP - 235
BT - ACM ICS 2025 - Proceedings of the 39th ACM International Conference on Supercomputing
Y2 - 8 June 2025 through 11 June 2025
ER -