Heterogeneous graph transformer with poly-tokenization
Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph tr...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9678 https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-10678 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-106782024-11-28T09:14:19Z Heterogeneous graph transformer with poly-tokenization LU, Zhiyuan FANG, Yuan YANG, Cheng SHI, Chuan Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph transformers primarily cater to homogeneous graphs and are unable to model the intricate semantics of heterogeneous graphs. Moreover, unlike small molecular graphs where the entire graph can be considered as the receptive field in graph transformers, real-world heterogeneous graphs comprise a significantly larger number of nodes and cannot be entirely treated as such. Consequently, existing graph transformers struggle to capture the long-range dependencies in these complex heterogeneous graphs. To address these two limitations, we present Poly-tokenized Heterogeneous Graph Transformer (PHGT), a novel transformer-based heterogeneous graph model. In addition to traditional node tokens, PHGT introduces a novel poly-token design with two more token types: semantic tokens and global tokens. Semantic tokens encapsulate high-order heterogeneous semantic relationships, while global tokens capture semantic-aware long-range interactions. We validate the effectiveness of PHGT through extensive experiments on standardized heterogeneous graph benchmarks, demonstrating significant improvements over state-of-the-art heterogeneous graph representation learning models. 2024-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9678 info:doi/10.24963/ijcai.2024/247 https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Data Mining Mining heterogenous data Mining graphs Graph transformer Graphic model Artificial Intelligence and Robotics Graphics and Human Computer Interfaces |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Data Mining Mining heterogenous data Mining graphs Graph transformer Graphic model Artificial Intelligence and Robotics Graphics and Human Computer Interfaces |
spellingShingle |
Data Mining Mining heterogenous data Mining graphs Graph transformer Graphic model Artificial Intelligence and Robotics Graphics and Human Computer Interfaces LU, Zhiyuan FANG, Yuan YANG, Cheng SHI, Chuan Heterogeneous graph transformer with poly-tokenization |
description |
Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph transformers primarily cater to homogeneous graphs and are unable to model the intricate semantics of heterogeneous graphs. Moreover, unlike small molecular graphs where the entire graph can be considered as the receptive field in graph transformers, real-world heterogeneous graphs comprise a significantly larger number of nodes and cannot be entirely treated as such. Consequently, existing graph transformers struggle to capture the long-range dependencies in these complex heterogeneous graphs. To address these two limitations, we present Poly-tokenized Heterogeneous Graph Transformer (PHGT), a novel transformer-based heterogeneous graph model. In addition to traditional node tokens, PHGT introduces a novel poly-token design with two more token types: semantic tokens and global tokens. Semantic tokens encapsulate high-order heterogeneous semantic relationships, while global tokens capture semantic-aware long-range interactions. We validate the effectiveness of PHGT through extensive experiments on standardized heterogeneous graph benchmarks, demonstrating significant improvements over state-of-the-art heterogeneous graph representation learning models. |
format |
text |
author |
LU, Zhiyuan FANG, Yuan YANG, Cheng SHI, Chuan |
author_facet |
LU, Zhiyuan FANG, Yuan YANG, Cheng SHI, Chuan |
author_sort |
LU, Zhiyuan |
title |
Heterogeneous graph transformer with poly-tokenization |
title_short |
Heterogeneous graph transformer with poly-tokenization |
title_full |
Heterogeneous graph transformer with poly-tokenization |
title_fullStr |
Heterogeneous graph transformer with poly-tokenization |
title_full_unstemmed |
Heterogeneous graph transformer with poly-tokenization |
title_sort |
heterogeneous graph transformer with poly-tokenization |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2024 |
url |
https://ink.library.smu.edu.sg/sis_research/9678 https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf |
_version_ |
1819113100403539968 |