论文标题

选择偏好获取的多重单词嵌入

Multiplex Word Embeddings for Selectional Preference Acquisition

论文作者

Zhang, Hongming, Bai, Jiaxin, Song, Yan, Xu, Kun, Yu, Changlong, Song, Yangqiu, Ng, Wilfred, Yu, Dong

论文摘要

传统的单词嵌入表示具有固定向量的单词,通常根据单词之间的共发生模式进行训练。但是,这样做的力量是有限的,在不同的句法关系下,相同的单词可能会分别官能化。为了解决这一限制,一种解决方案是将不同单词的关系依赖性纳入其嵌入。因此,在本文中,我们提出了一个多重词嵌入模型,可以根据单词之间的各种关系轻松扩展。结果,每个单词都有一个中心嵌入以表示其整体语义,并且几个关系嵌入代表其关系依赖性。与现有模型相比,我们的模型可以有效地区分不同关系的单词,而无需引入不必要的稀疏性。此外,为了适应各种关系,我们使用一个小维度来进行关系嵌入,我们的模型能够保持其有效性。关于选择性偏好的实验和单词相似性的实验证明了所提出的模型的有效性,并且对可伸缩性的进一步研究也证明,我们的嵌入只需要原始嵌入大小的1/20即可实现更好的性能。

Conventional word embeddings represent words with fixed vectors, which are usually trained based on co-occurrence patterns among words. In doing so, however, the power of such representations is limited, where the same word might be functionalized separately under different syntactic relations. To address this limitation, one solution is to incorporate relational dependencies of different words into their embeddings. Therefore, in this paper, we propose a multiplex word embedding model, which can be easily extended according to various relations among words. As a result, each word has a center embedding to represent its overall semantics, and several relational embeddings to represent its relational dependencies. Compared to existing models, our model can effectively distinguish words with respect to different relations without introducing unnecessary sparseness. Moreover, to accommodate various relations, we use a small dimension for relational embeddings and our model is able to keep their effectiveness. Experiments on selectional preference acquisition and word similarity demonstrate the effectiveness of the proposed model, and a further study of scalability also proves that our embeddings only need 1/20 of the original embedding size to achieve better performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源