论文标题
PatchGT:用于学习图表表示的不可训练的群集上的变压器
PatchGT: Transformer over Non-trainable Clusters for Learning Graph Representations
论文作者
论文摘要
最近,变压器结构在图形学习任务中表现出良好的表现。但是,这些变压器模型直接在图节点上起作用,并且可能难以学习高级信息。受视觉变压器适用于图像贴片的启发,我们提出了一个新的基于变压器的图形神经网络:Patch Graph Transformer(PatchGT)。与以前的学习图表示模型不同,PatchGT从不可训练的图形补丁中学习,而不是直接从节点中学习。它可以帮助节省计算并改善模型性能。关键思想是基于光谱聚类将图形段分为贴片,而无需任何可训练的参数,该模型可以首先使用GNN层来学习补丁级表示表示,然后使用变压器来获得图形级表示。该体系结构利用图形的光谱信息并结合了GNN和变形金刚的优势。此外,我们从理论和经验上展示了以前可训练的群集的局限性。我们还证明了所提出的不可训练的光谱聚类方法是置换不变的,可以帮助解决图中的信息瓶颈。 PatchGT比1-WL型GNN具有更高的表现力,而经验研究表明,PatchGT在基准数据集上实现了竞争性能,并为其预测提供了解释性。我们的算法的实现在我们的GitHub repo:https://github.com/tufts-ml/patchgt上发布。
Recently the Transformer structure has shown good performances in graph learning tasks. However, these Transformer models directly work on graph nodes and may have difficulties learning high-level information. Inspired by the vision transformer, which applies to image patches, we propose a new Transformer-based graph neural network: Patch Graph Transformer (PatchGT). Unlike previous transformer-based models for learning graph representations, PatchGT learns from non-trainable graph patches, not from nodes directly. It can help save computation and improve the model performance. The key idea is to segment a graph into patches based on spectral clustering without any trainable parameters, with which the model can first use GNN layers to learn patch-level representations and then use Transformer to obtain graph-level representations. The architecture leverages the spectral information of graphs and combines the strengths of GNNs and Transformers. Further, we show the limitations of previous hierarchical trainable clusters theoretically and empirically. We also prove the proposed non-trainable spectral clustering method is permutation invariant and can help address the information bottlenecks in the graph. PatchGT achieves higher expressiveness than 1-WL-type GNNs, and the empirical study shows that PatchGT achieves competitive performances on benchmark datasets and provides interpretability to its predictions. The implementation of our algorithm is released at our Github repo: https://github.com/tufts-ml/PatchGT.