Low-rank and global-representation-key-based attention for graph transformer

Transformer architectures have been applied to graph-specific data such as protein structure and shopper lists, and they perform accurately on graph/node classification and prediction tasks. Researchers have proved that the attention matrix in Transformers has low-rank properties, and the self-atten...

Full description

Saved in:
Bibliographic Details
Main Authors: Kong, Lingping, Ojha, Varun, Gao, Ruobin, Suganthan, Ponnuthurai Nagaratnam, Snášel, Václav
Other Authors: School of Civil and Environmental Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/170863
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-170863
record_format dspace
spelling sg-ntu-dr.10356-1708632023-10-06T15:33:33Z Low-rank and global-representation-key-based attention for graph transformer Kong, Lingping Ojha, Varun Gao, Ruobin Suganthan, Ponnuthurai Nagaratnam Snášel, Václav School of Civil and Environmental Engineering Engineering::Civil engineering Engineering::Computer science and engineering Graph Transformer Graph Representation Transformer architectures have been applied to graph-specific data such as protein structure and shopper lists, and they perform accurately on graph/node classification and prediction tasks. Researchers have proved that the attention matrix in Transformers has low-rank properties, and the self-attention plays a scoring role in the aggregation function of the Transformers. However, it can not solve the issues such as heterophily and over-smoothing. The low-rank properties and the limitations of Transformers inspire this work to propose a Global Representation (GR) based attention mechanism to alleviate the two heterophily and over-smoothing issues. First, this GR-based model integrates geometric information of the nodes of interest that conveys the structural properties of the graph. Unlike a typical Transformer where a node feature forms a Key, we propose to use GR to construct the Key, which discovers the relation between the nodes and the structural representation of the graph. Next, we present various compositions of GR emanating from nodes of interest and α-hop neighbors. Then, we explore this attention property with an extensive experimental test to assess the performance and the possible direction of improvements for future works. Additionally, we provide mathematical proof showing the efficient feature update in our proposed method. Finally, we verify and validate the performance of the model on eight benchmark datasets that show the effectiveness of the proposed method. Published version Open Access funding provided by the Qatar National Library; This work was supported by The Ministry of Education, Youth and Sports of the Czech Republic under project META MO-COP; DST/INT/Czech/P-12/2019. 2023-10-04T00:40:41Z 2023-10-04T00:40:41Z 2023 Journal Article Kong, L., Ojha, V., Gao, R., Suganthan, P. N. & Snášel, V. (2023). Low-rank and global-representation-key-based attention for graph transformer. Information Sciences, 642, 119108-. https://dx.doi.org/10.1016/j.ins.2023.119108 0020-0255 https://hdl.handle.net/10356/170863 10.1016/j.ins.2023.119108 2-s2.0-85159566842 642 119108 en Information Sciences © 2023 The Authors. Published by Elsevier Inc. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Civil engineering
Engineering::Computer science and engineering
Graph Transformer
Graph Representation
spellingShingle Engineering::Civil engineering
Engineering::Computer science and engineering
Graph Transformer
Graph Representation
Kong, Lingping
Ojha, Varun
Gao, Ruobin
Suganthan, Ponnuthurai Nagaratnam
Snášel, Václav
Low-rank and global-representation-key-based attention for graph transformer
description Transformer architectures have been applied to graph-specific data such as protein structure and shopper lists, and they perform accurately on graph/node classification and prediction tasks. Researchers have proved that the attention matrix in Transformers has low-rank properties, and the self-attention plays a scoring role in the aggregation function of the Transformers. However, it can not solve the issues such as heterophily and over-smoothing. The low-rank properties and the limitations of Transformers inspire this work to propose a Global Representation (GR) based attention mechanism to alleviate the two heterophily and over-smoothing issues. First, this GR-based model integrates geometric information of the nodes of interest that conveys the structural properties of the graph. Unlike a typical Transformer where a node feature forms a Key, we propose to use GR to construct the Key, which discovers the relation between the nodes and the structural representation of the graph. Next, we present various compositions of GR emanating from nodes of interest and α-hop neighbors. Then, we explore this attention property with an extensive experimental test to assess the performance and the possible direction of improvements for future works. Additionally, we provide mathematical proof showing the efficient feature update in our proposed method. Finally, we verify and validate the performance of the model on eight benchmark datasets that show the effectiveness of the proposed method.
author2 School of Civil and Environmental Engineering
author_facet School of Civil and Environmental Engineering
Kong, Lingping
Ojha, Varun
Gao, Ruobin
Suganthan, Ponnuthurai Nagaratnam
Snášel, Václav
format Article
author Kong, Lingping
Ojha, Varun
Gao, Ruobin
Suganthan, Ponnuthurai Nagaratnam
Snášel, Václav
author_sort Kong, Lingping
title Low-rank and global-representation-key-based attention for graph transformer
title_short Low-rank and global-representation-key-based attention for graph transformer
title_full Low-rank and global-representation-key-based attention for graph transformer
title_fullStr Low-rank and global-representation-key-based attention for graph transformer
title_full_unstemmed Low-rank and global-representation-key-based attention for graph transformer
title_sort low-rank and global-representation-key-based attention for graph transformer
publishDate 2023
url https://hdl.handle.net/10356/170863
_version_ 1779171092078788608