Heterogeneous graph transformer with poly-tokenization

Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph tr...

Full description

Saved in:
Bibliographic Details
Main Authors: LU, Zhiyuan, FANG, Yuan, YANG, Cheng, SHI, Chuan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9678
https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10678
record_format dspace
spelling sg-smu-ink.sis_research-106782024-11-28T09:14:19Z Heterogeneous graph transformer with poly-tokenization LU, Zhiyuan FANG, Yuan YANG, Cheng SHI, Chuan Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph transformers primarily cater to homogeneous graphs and are unable to model the intricate semantics of heterogeneous graphs. Moreover, unlike small molecular graphs where the entire graph can be considered as the receptive field in graph transformers, real-world heterogeneous graphs comprise a significantly larger number of nodes and cannot be entirely treated as such. Consequently, existing graph transformers struggle to capture the long-range dependencies in these complex heterogeneous graphs. To address these two limitations, we present Poly-tokenized Heterogeneous Graph Transformer (PHGT), a novel transformer-based heterogeneous graph model. In addition to traditional node tokens, PHGT introduces a novel poly-token design with two more token types: semantic tokens and global tokens. Semantic tokens encapsulate high-order heterogeneous semantic relationships, while global tokens capture semantic-aware long-range interactions. We validate the effectiveness of PHGT through extensive experiments on standardized heterogeneous graph benchmarks, demonstrating significant improvements over state-of-the-art heterogeneous graph representation learning models. 2024-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9678 info:doi/10.24963/ijcai.2024/247 https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Data Mining Mining heterogenous data Mining graphs Graph transformer Graphic model Artificial Intelligence and Robotics Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Data Mining
Mining heterogenous data
Mining graphs
Graph transformer
Graphic model
Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
spellingShingle Data Mining
Mining heterogenous data
Mining graphs
Graph transformer
Graphic model
Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
LU, Zhiyuan
FANG, Yuan
YANG, Cheng
SHI, Chuan
Heterogeneous graph transformer with poly-tokenization
description Graph neural networks have shown widespread success for learning on graphs, but they still face fundamental drawbacks, such as limited expressive power, over-smoothing, and over-squashing. Meanwhile, the transformer architecture offers a potential solution to these issues. However, existing graph transformers primarily cater to homogeneous graphs and are unable to model the intricate semantics of heterogeneous graphs. Moreover, unlike small molecular graphs where the entire graph can be considered as the receptive field in graph transformers, real-world heterogeneous graphs comprise a significantly larger number of nodes and cannot be entirely treated as such. Consequently, existing graph transformers struggle to capture the long-range dependencies in these complex heterogeneous graphs. To address these two limitations, we present Poly-tokenized Heterogeneous Graph Transformer (PHGT), a novel transformer-based heterogeneous graph model. In addition to traditional node tokens, PHGT introduces a novel poly-token design with two more token types: semantic tokens and global tokens. Semantic tokens encapsulate high-order heterogeneous semantic relationships, while global tokens capture semantic-aware long-range interactions. We validate the effectiveness of PHGT through extensive experiments on standardized heterogeneous graph benchmarks, demonstrating significant improvements over state-of-the-art heterogeneous graph representation learning models.
format text
author LU, Zhiyuan
FANG, Yuan
YANG, Cheng
SHI, Chuan
author_facet LU, Zhiyuan
FANG, Yuan
YANG, Cheng
SHI, Chuan
author_sort LU, Zhiyuan
title Heterogeneous graph transformer with poly-tokenization
title_short Heterogeneous graph transformer with poly-tokenization
title_full Heterogeneous graph transformer with poly-tokenization
title_fullStr Heterogeneous graph transformer with poly-tokenization
title_full_unstemmed Heterogeneous graph transformer with poly-tokenization
title_sort heterogeneous graph transformer with poly-tokenization
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9678
https://ink.library.smu.edu.sg/context/sis_research/article/10678/viewcontent/IJCAI24_PHGT.pdf
_version_ 1819113100403539968