Learning to pre-train graph neural networks

Graph neural networks (GNNs) have become the de facto standard for representation learning on graphs, which derive effective node representations by recursively aggregating information from graph neighborhoods. While GNNs can be trained from scratch, pre-training GNNs to learn transferable knowledge...

Full description

Saved in:
Bibliographic Details
Main Authors: LU, Yuanfu, JIANG, Xunqiang, FANG, Yuan, SHI, Chuan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/6125
https://ink.library.smu.edu.sg/context/sis_research/article/7128/viewcontent/16552_Article_Text_20046_1_2_20210518.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-7128
record_format dspace
spelling sg-smu-ink.sis_research-71282021-09-29T12:18:44Z Learning to pre-train graph neural networks LU, Yuanfu JIANG, Xunqiang FANG, Yuan SHI, Chuan Graph neural networks (GNNs) have become the de facto standard for representation learning on graphs, which derive effective node representations by recursively aggregating information from graph neighborhoods. While GNNs can be trained from scratch, pre-training GNNs to learn transferable knowledge for downstream tasks has recently been demonstrated to improve the state of the art. However, conventional GNN pre-training methods follow a two-step paradigm: 1) pre-training on abundant unlabeled data and 2) fine-tuning on downstream labeled data, between which there exists a significant gap due to the divergence of optimization objectives in the two steps. In this paper, we conduct an analysis to show the divergence between pre-training and fine-tuning, and to alleviate such divergence, we propose L2P-GNN, a self-supervised pre-training strategy for GNNs. The key insight is that L2PGNN attempts to learn how to fine-tune during the pre-training process in the form of transferable prior knowledge. To encode both local and global information into the prior, L2P-GNN is further designed with a dual adaptation mechanism at both node and graph levels. Finally, we conduct a systematic empirical study on the pre-training of various GNN models, using both a public collection of protein graphs and a new compilation of bibliographic graphs for pre-training. Experimental results show that L2P-GNN is capable of learning effective and transferable prior knowledge that yields powerful representations for downstream tasks. (Code and datasets are available at https://github.com/rootlu/L2P-GNN.) 2021-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6125 https://ink.library.smu.edu.sg/context/sis_research/article/7128/viewcontent/16552_Article_Text_20046_1_2_20210518.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University pre-training graph neural networks meta-learning MAML Databases and Information Systems OS and Networks
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic pre-training
graph neural networks
meta-learning
MAML
Databases and Information Systems
OS and Networks
spellingShingle pre-training
graph neural networks
meta-learning
MAML
Databases and Information Systems
OS and Networks
LU, Yuanfu
JIANG, Xunqiang
FANG, Yuan
SHI, Chuan
Learning to pre-train graph neural networks
description Graph neural networks (GNNs) have become the de facto standard for representation learning on graphs, which derive effective node representations by recursively aggregating information from graph neighborhoods. While GNNs can be trained from scratch, pre-training GNNs to learn transferable knowledge for downstream tasks has recently been demonstrated to improve the state of the art. However, conventional GNN pre-training methods follow a two-step paradigm: 1) pre-training on abundant unlabeled data and 2) fine-tuning on downstream labeled data, between which there exists a significant gap due to the divergence of optimization objectives in the two steps. In this paper, we conduct an analysis to show the divergence between pre-training and fine-tuning, and to alleviate such divergence, we propose L2P-GNN, a self-supervised pre-training strategy for GNNs. The key insight is that L2PGNN attempts to learn how to fine-tune during the pre-training process in the form of transferable prior knowledge. To encode both local and global information into the prior, L2P-GNN is further designed with a dual adaptation mechanism at both node and graph levels. Finally, we conduct a systematic empirical study on the pre-training of various GNN models, using both a public collection of protein graphs and a new compilation of bibliographic graphs for pre-training. Experimental results show that L2P-GNN is capable of learning effective and transferable prior knowledge that yields powerful representations for downstream tasks. (Code and datasets are available at https://github.com/rootlu/L2P-GNN.)
format text
author LU, Yuanfu
JIANG, Xunqiang
FANG, Yuan
SHI, Chuan
author_facet LU, Yuanfu
JIANG, Xunqiang
FANG, Yuan
SHI, Chuan
author_sort LU, Yuanfu
title Learning to pre-train graph neural networks
title_short Learning to pre-train graph neural networks
title_full Learning to pre-train graph neural networks
title_fullStr Learning to pre-train graph neural networks
title_full_unstemmed Learning to pre-train graph neural networks
title_sort learning to pre-train graph neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/6125
https://ink.library.smu.edu.sg/context/sis_research/article/7128/viewcontent/16552_Article_Text_20046_1_2_20210518.pdf
_version_ 1770575825282793472