Scalable and Versatile Hardware Acceleration of Graph Neural Networks
2024-06
Loading...
View/Download File
Persistent link to this item
Statistics
View StatisticsJournal Title
Journal ISSN
Volume Title
Title
Scalable and Versatile Hardware Acceleration of Graph Neural Networks
Alternative title
Authors
Published Date
2024-06
Publisher
Type
Thesis or Dissertation
Abstract
Graph neural networks (GNN) are vital for analyzing real-world problems (e.g., networkanalysis, drug interaction, electronic design automation, e-commerce) that use graph models. However, efficient GNN acceleration faces with multiple challenges related to high and variable sparsity of input feature vectors, power-law degree distribution in the adjacency matrix, and maintaining load-balanced computation with minimal random memory accesses. This thesis addresses the problems of building fast, energy-efficient
inference and training accelerators for GNNs, addressing both static and dynamic graphs. For inference, this thesis proposes GNNIE, a versatile GNN inference acceleratorcapable of handling a diverse set of GNNs, including graph attention networks (GATs), graph convolutional networks (GCNs), GraphSAGE, GINConv, and DiffPool. It mitigates workload imbalance by (i) splitting vertex feature operands into blocks, (ii) reordering and redistributing computations, (iii) using a novel “flexible MAC” architecture. To maximize on-chip data reuse and reduce random DRAM fetches, GNNIE adopts a novel graph-specific, degree-aware caching policy. GNNIE attains substantial speedup over CPU (7197×), GPU (17.81×), and prior works, e.g., HyGCN (5×), AWB-GCN (1.3×) over multiple datasets on GCN, GAT, GraphSAGE, and GINConv. For training GNNs for large graphs, this research develops a GNNIE-based multicoreaccelerator. A novel feature vector segmentation approach is proposed to scale onlarge graphs using small on-chip buffers. A multicore-specific graph specific caching is also implemented to reduce off-chip and on-chip communication and to alleviate random DRAM accesses. Experiments over multiple large datasets and multiple GNNs demonstrate an average training speedup and energy efficiency improvement of 17× and 322×, respectively, over DGL on a GPU, and a speedup of 14× with 268× lower energy over the GPU-based GNNAdvisor approach. Overall, this research tackles scalability and versatility issues of building GNN accelerators while delivering significant speedup and energy efficiency. Finally, this thesis addresses the acceleration of dynamic graph neural networks(DGNNs), which play a crucial role in applications such as social network analytics and urban traffic prediction that require inferencing on graph-structured data, where the connectivity and features of the underlying graph evolve over time. The proposed platform integrates GNN and Recurrent Neural Network (RNN) components of DGNNs, providing a unified platform for spatial and temporal information capture, respectively. The contributions encompass optimized cache reuse strategies, a novel caching policy, and an efficient pipelining mechanism. Evaluation across multiple graph datasets and multiple DGNNs demonstrates average energy efficiency gains of 8393×, 183×, and 87× – 10×, and inference speedups of 1796×, 77×, and 21× – 2.4× , over Intel Xeon Gold CPU, NVIDIA V100 GPU, and prior state-of-the-art DGNN accelerators, respectively, are demonstrated across multiple graph datasets and multiple DGNNs.
Description
University of Minnesota Ph.D. dissertation. June 2024. Major: Electrical Engineering. Advisor: Sachin Suresh Sapatnekar. 1 computer file (PDF); xi, 98 pages.
Related to
Replaces
License
Collections
Series/Report Number
Funding information
Isbn identifier
Doi identifier
Previously Published Citation
Other identifiers
Suggested citation
Mondal, Sudipta. (2024). Scalable and Versatile Hardware Acceleration of Graph Neural Networks. Retrieved from the University Digital Conservancy, https://hdl.handle.net/11299/269222.
Content distributed via the University Digital Conservancy may be subject to additional license and use restrictions applied by the depositor. By using these files, users agree to the Terms of Use. Materials in the UDC may contain content that is disturbing and/or harmful. For more information, please see our statement on harmful content in digital repositories.