论文标题
异构信息网络的预训练模型
Pre-Trained Models for Heterogeneous Information Networks
论文作者
论文摘要
在网络表示学习中,我们学习如何在低维空间中表示异质信息网络,以促进有效的搜索,分类和预测解决方案。以前的网络表示方法通常需要足够的特定任务标记数据来解决特定于域的问题。训练有素的模型通常不能转移到室外数据集中。我们提出了一个自制的预训练和微调框架PF-hin,以捕获异质信息网络的特征。与传统的网络表示学习模型不同,必须重新训练整个模型的下游任务和数据集,PF-hin只需要微调模型和少量额外的特定任务特定参数,从而提高了模型效率和有效性。在预训练期间,我们首先将给定节点的邻居转换为序列。 PF-HIN是基于两个自制任务,掩盖节点建模和相邻节点预测的预训练。我们采用深双向变压器编码器来训练模型,并利用分解的嵌入参数化和跨层参数共享以减少参数。在微调阶段,我们选择四个基准下游任务,即链接预测,相似性搜索,节点分类和节点群集。在四个数据集上,PF-HIN始终如一地超过这些任务的最先进的替代方案。
In network representation learning we learn how to represent heterogeneous information networks in a low-dimensional space so as to facilitate effective search, classification, and prediction solutions. Previous network representation learning methods typically require sufficient task-specific labeled data to address domain-specific problems. The trained model usually cannot be transferred to out-of-domain datasets. We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network. Unlike traditional network representation learning models that have to train the entire model all over again for every downstream task and dataset, PF-HIN only needs to fine-tune the model and a small number of extra task-specific parameters, thus improving model efficiency and effectiveness. During pre-training, we first transform the neighborhood of a given node into a sequence. PF-HIN is pre-trained based on two self-supervised tasks, masked node modeling and adjacent node prediction. We adopt deep bi-directional transformer encoders to train the model, and leverage factorized embedding parameterization and cross-layer parameter sharing to reduce the parameters. In the fine-tuning stage, we choose four benchmark downstream tasks, i.e., link prediction, similarity search, node classification, and node clustering. PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.