• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索

author:

Chen, Ruiqi (Chen, Ruiqi.) [1] | Zhang, Haoyang (Zhang, Haoyang.) [2] | Li, Shun (Li, Shun.) [3] | Tang, Enhao (Tang, Enhao.) [4] | Yu, Jun (Yu, Jun.) [5] | Wang, Kun (Wang, Kun.) [6]

Indexed by:

CPCI-S EI

Abstract:

Field-programmable gate array (FPGA) is an ideal candidate for accelerating graph neural networks (GNNs). However, FPGA reconfiguration is a time-consuming process when updating or switching between diverse GNN models across different applications. This paper proposes a highly integrated FPGA-based overlay processor for GNN accelerations. Graph-OPU provides excellent flexibility and software-like programmability for GNN end-users, as the executable code of GNN models are automatically compiled and reloaded without requiring FPGA reconfiguration. First, we customize the instruction sets for the inference qprocess of different GNN models. Second, we propose a microarchitecture ensuring a fully-pipelined process for GNN inference. Third, we design a unified matrix multiplication to process sparse-dense matrix multiplication and general matrix multiplication to increase the Graph-OPU performance. Finally, we implement a hardware prototype on the Xilinx Alveo U50 and test the mainstream GNN models using various datasets. Graph-OPU takes an average of only 2 minutes to switch between different GNN models, exhibiting average 128x speedup compared to related works. In addition, Graph-OPU outperforms state-of-the-art end-to-end overlay accelerators for GNN, reducing latency by an average of 1.36x and improving energy efficiency by an average of 1.41x. Moreover, Graph-OPU achieves up to 1654x and 63x speedup, as well as up to 5305x and 422x energy efficiency boosts, compared to implementations on CPU and GPU, respectively. To the best of our knowledge, Graph-OPU represents the first in-depth study of an FPGA-based overlay processor for GNNs, offering high flexibility, speedup, and energy efficiency.

Keyword:

Community:

  • [ 1 ] [Chen, Ruiqi]Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
  • [ 2 ] [Zhang, Haoyang]Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
  • [ 3 ] [Yu, Jun]Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
  • [ 4 ] [Wang, Kun]Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
  • [ 5 ] [Li, Shun]Fuzhou Univ, Coll Phys & Informat Engn, Fuzhou, Peoples R China
  • [ 6 ] [Tang, Enhao]Fuzhou Univ, Coll Phys & Informat Engn, Fuzhou, Peoples R China

Reprint 's Address:

Show more details

Related Keywords:

Related Article:

Source :

2023 33RD INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS, FPL

ISSN: 1946-1488

Year: 2023

Page: 228-234

Cited Count:

WoS CC Cited Count: 2

SCOPUS Cited Count: 3

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 0

Online/Total:75/10060617
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1