CSpace  > 中国科学院计算技术研究所期刊论文  > 英文
HiHGNN: Accelerating HGNNs Through Parallelism and Data Reusability Exploitation
Xue, Runzhen1; Han, Dengke1; Yan, Mingyu1,2,3; Zou, Mo4; Yang, Xiaocheng; Wang, Duo1; Li, Wenming1; Tang, Zhimin1; Kim, John4; Ye, Xiaochun1; Fan, Dongrui1
2024-07-01
发表期刊IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS
ISSN1045-9219
卷号35期号:7页码:1122-1138
摘要Heterogeneous graph neural networks (HGNNs) have emerged as powerful algorithms for processing heterogeneous graphs (HetGs), widely used in many critical fields. To capture both structural and semantic information in HetGs, HGNNs first aggregate the neighboring feature vectors for each vertex in each semantic graph and then fuse the aggregated results across all semantic graphs for each vertex. Unfortunately, existing graph neural network accelerators are ill-suited to accelerate HGNNs. This is because they fail to efficiently tackle the specific execution patterns and exploit the high-degree parallelism as well as data reusability inside and across the processing of semantic graphs in HGNNs. In this work, we first quantitatively characterize a set of representative HGNN models on GPU to disclose the execution bound of each stage, inter-semantic-graph parallelism, and inter-semantic-graph data reusability in HGNNs. Guided by our findings, we propose a high-performance HGNN accelerator, HiHGNN, to alleviate the execution bound and exploit the newfound parallelism and data reusability in HGNNs. Specifically, we first propose a bound-aware stage-fusion methodology that tailors to HGNN acceleration, to fuse and pipeline the execution stages being aware of their execution bounds. Second, we design an independency-aware parallel execution design to exploit the inter-semantic-graph parallelism. Finally, we present a similarity-aware execution scheduling to exploit the inter-semantic-graph data reusability. Compared to the state-of-the-art software framework running on NVIDIA GPU T4 and GPU A100, HiHGNN respectively achieves an average 40.0x and 8.3x speedup as well as 99.59% and 99.74% energy reduction with quintile the memory bandwidth of GPU A100.
关键词Semantics Parallel processing Graph neural networks Vectors Graphics processing units Fuses Hardware GNN GNN accelerator graph neural network HGNN HGNN accelerator heterogeneous graph neural network
DOI10.1109/TPDS.2024.3394841
收录类别SCI
语种英语
资助项目National Key Research and Development Program of China
WOS研究方向Computer Science ; Engineering
WOS类目Computer Science, Theory & Methods ; Engineering, Electrical & Electronic
WOS记录号WOS:001224174400002
出版者IEEE COMPUTER SOC
引用统计
被引频次:1[WOS]   [WOS记录]     [WOS相关记录]
文献类型期刊论文
条目标识符http://119.78.100.204/handle/2XEOYT63/38955
专题中国科学院计算技术研究所期刊论文_英文
通讯作者Yan, Mingyu
作者单位1.Chinese Acad Sci, Inst Comp Technol, State Key Lab Processors, Beijing 100045, Peoples R China
2.Univ Chinese Acad Sci, Beijing 101408, Peoples R China
3.Shanghai Innovat Ctr Processor Technol, Shanghai 201203, Peoples R China
4.Korea Adv Inst Sci & Technol, Sch Elect Engn, Daejeon 305701, South Korea
推荐引用方式
GB/T 7714
Xue, Runzhen,Han, Dengke,Yan, Mingyu,et al. HiHGNN: Accelerating HGNNs Through Parallelism and Data Reusability Exploitation[J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS,2024,35(7):1122-1138.
APA Xue, Runzhen.,Han, Dengke.,Yan, Mingyu.,Zou, Mo.,Yang, Xiaocheng.,...&Fan, Dongrui.(2024).HiHGNN: Accelerating HGNNs Through Parallelism and Data Reusability Exploitation.IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS,35(7),1122-1138.
MLA Xue, Runzhen,et al."HiHGNN: Accelerating HGNNs Through Parallelism and Data Reusability Exploitation".IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS 35.7(2024):1122-1138.
条目包含的文件
条目无相关文件。
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Xue, Runzhen]的文章
[Han, Dengke]的文章
[Yan, Mingyu]的文章
百度学术
百度学术中相似的文章
[Xue, Runzhen]的文章
[Han, Dengke]的文章
[Yan, Mingyu]的文章
必应学术
必应学术中相似的文章
[Xue, Runzhen]的文章
[Han, Dengke]的文章
[Yan, Mingyu]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。