论文标题
文本网络的对抗上下文意识网络嵌入
Adversarial Context Aware Network Embeddings for Textual Networks
论文作者
论文摘要
文本网络的表示形式构成了一个重大挑战,因为它涉及从两种模式中捕获合并的信息:(i)基础网络结构和(ii)节点文本属性。为此,大多数现有方法通过强制连接节点的嵌入方式相似,以学习文本和网络结构的嵌入。然后,为了实现模态融合,他们使用节点的文本嵌入之间的相似性与其连接节点的结构嵌入,反之亦然。这意味着这些方法需要用于学习嵌入的边缘信息,并且无法学习看不见的节点的嵌入。在本文中,我们提出了一种方法,可以实现融合方式和学习看不见节点的嵌入的能力。我们模型的主要特征是它在基于文本嵌入的歧视器和基于结构嵌入的生成器之间使用对抗机制来学习有效的表示。然后,为了学习看不见的节点的嵌入,我们使用基于文本嵌入歧视器提供的监督。此外,我们还提出了一种用于学习文本嵌入的新型架构,可以将相互关注和拓扑注意机制结合起来,从而提供更灵活的文本嵌入。通过对现实世界数据集的广泛实验,我们证明了我们的模型对几个最新基准测试取得了可观的收益。与以前的最先进相比,在预测培训中看到的节点之间的联系和性能提高了高达12%的性能,在预测涉及培训中未见的节点的链接方面,它的性能提高了7%。此外,在节点分类任务中,它的性能提高了2%。
Representation learning of textual networks poses a significant challenge as it involves capturing amalgamated information from two modalities: (i) underlying network structure, and (ii) node textual attributes. For this, most existing approaches learn embeddings of text and network structure by enforcing embeddings of connected nodes to be similar. Then for achieving a modality fusion they use the similarities between text embedding of a node with the structure embedding of its connected node and vice versa. This implies that these approaches require edge information for learning embeddings and they cannot learn embeddings of unseen nodes. In this paper we propose an approach that achieves both modality fusion and the capability to learn embeddings of unseen nodes. The main feature of our model is that it uses an adversarial mechanism between text embedding based discriminator, and structure embedding based generator to learn efficient representations. Then for learning embeddings of unseen nodes, we use the supervision provided by the text embedding based discriminator. In addition this, we propose a novel architecture for learning text embedding that can combine both mutual attention and topological attention mechanism, which give more flexible text embeddings. Through extensive experiments on real-world datasets, we demonstrate that our model makes substantial gains over several state-of-the-art benchmarks. In comparison with previous state-of-the-art, it gives up to 7% improvement in performance in predicting links among nodes seen in the training and up to 12% improvement in performance in predicting links involving nodes not seen in training. Further, in the node classification task, it gives up to 2% improvement in performance.