Abstract:
Recently, graph neural networks (GNNs) have achieved great success for graph representation learning tasks. Enlightened by the fact that numerous message passing redundan...Show MoreMetadata
Abstract:
Recently, graph neural networks (GNNs) have achieved great success for graph representation learning tasks. Enlightened by the fact that numerous message passing redundancies exist in GNNs, we propose DyGNN, which speeds up GNNs by reducing redundancies. DyGNN is supported by an algorithm and architecture co-design. The proposed algorithm can dynamically prune vertices and edges during execution without accuracy loss. An architecture is designed to support dynamic pruning and transform it into performance improvement. DyGNN opens new directions for accelerating GNNs by pruning vertices and edges. DyGNN gains average 2\times speedup with accuracy improvement of 4% compared with state-of-the-art GNN accelerators.
Published in: 2021 58th ACM/IEEE Design Automation Conference (DAC)
Date of Conference: 05-09 December 2021
Date Added to IEEE Xplore: 08 November 2021
ISBN Information:
Print on Demand(PoD) ISSN: 0738-100X