• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Chen, Ruiqi (Chen, Ruiqi.) [1] | Zhang, Haoyang (Zhang, Haoyang.) [2] | Li, Shun (Li, Shun.) [3] | Tang, Enhao (Tang, Enhao.) [4] | Yu, Jun (Yu, Jun.) [5] | Wang, Kun (Wang, Kun.) [6]

Indexed by:

EI

Abstract:

Field-programmable gate array (FPGA) is an ideal candidate for accelerating graph neural networks (GNNs). However, FPGA reconfiguration is a time-consuming process when updating or switching between diverse GNN models across different applications. This paper proposes a highly integrated FPGA-based overlay processor for GNN accelerations. Graph-OPU provides excellent flexibility and software-like programmability for GNN end-users, as the executable code of GNN models are automatically compiled and reloaded without requiring FPGA reconfiguration. First, we customize the instruction sets for the inference qprocess of different GNN models. Second, we propose a microarchitecture ensuring a fully-pipelined process for GNN inference. Third, we design a unified matrix multiplication to process sparse-dense matrix multiplication and general matrix multiplication to increase the Graph-OPU performance. Finally, we implement a hardware prototype on the Xilinx Alveo U50 and test the mainstream GNN models using various datasets. Graph-OPU takes an average of only 2 minutes to switch between different GNN models, exhibiting average 128× speedup compared to related works. In addition, Graph-OPU outperforms state-of-the-art end-to-end overlay accelerators for GNN, reducing latency by an average of 1.36× and improving energy efficiency by an average of 1.41×. Moreover, Graph-OPU achieves up to 1654× and 63× speedup, as well as up to 5305× and 422× energy efficiency boosts, compared to implementations on CPU and GPU, respectively. To the best of our knowledge, Graph-OPU represents the first in-depth study of an FPGA-based overlay processor for GNNs, offering high flexibility, speedup, and energy efficiency. © 2023 IEEE.

Keyword:

Energy efficiency Field programmable gate arrays (FPGA) Graph neural networks Matrix algebra Neural network models Reconfigurable hardware

Community:

  • [ 1 ] [Chen, Ruiqi]Fudan University, State Key Lab of Asic & System, Shanghai, China
  • [ 2 ] [Zhang, Haoyang]Fudan University, State Key Lab of Asic & System, Shanghai, China
  • [ 3 ] [Li, Shun]College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
  • [ 4 ] [Tang, Enhao]College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
  • [ 5 ] [Yu, Jun]Fudan University, State Key Lab of Asic & System, Shanghai, China
  • [ 6 ] [Wang, Kun]Fudan University, State Key Lab of Asic & System, Shanghai, China

Reprint 's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

Year: 2023

Page: 228-234

Language: English

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count: 8

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 4

Affiliated Colleges:

Online/Total:118/10067445
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1