SUGAR: Efficient Subgraph-Level Training via Resource-Aware Graph Partitioning | IEEE Journals & Magazine | IEEE Xplore

SUGAR: Efficient Subgraph-Level Training via Resource-Aware Graph Partitioning


Abstract:

Graph Neural Networks (GNNs) have demonstrated a great potential in a variety of graph-based applications, such as recommender systems, drug discovery, and object recogni...Show More

Abstract:

Graph Neural Networks (GNNs) have demonstrated a great potential in a variety of graph-based applications, such as recommender systems, drug discovery, and object recognition. Nevertheless, resource-efficient GNN learning is a rarely explored topic despite its many benefits for edge computing and Internet of Things (IoT) applications. To improve this state of affairs, this work proposes efficient subgraph-level training via resource-aware graph partitioning (SUGAR). SUGAR first partitions the initial graph into a set of disjoint subgraphs and then performs local training at the subgraph-level We provide a theoretical analysis and conduct extensive experiments on five graph benchmarks to verify its efficacy in practice. Our results across five different hardware platforms demonstrate great runtime speedup and memory reduction of SUGAR on large-scale graphs. We believe SUGAR opens a new research direction towards developing GNN methods that are resource-efficient, hence suitable for IoT deployment.
Published in: IEEE Transactions on Computers ( Volume: 72, Issue: 11, November 2023)
Page(s): 3167 - 3177
Date of Publication: 29 June 2023

ISSN Information:

Funding Agency:

References is not available for this document.

References is not available for this document.

Contact IEEE to Subscribe

References

References is not available for this document.