[1]张潇鲲,刘琰,陈静.引入外部词向量的文本信息网络表示学习[J].智能系统学报,2019,14(5):1056-1063.[doi:10.11992/tis.201809037]
ZHANG Xiaokun,LIU Yan,CHEN Jing.Representation learning using network embedding based on external word vectors[J].CAAI Transactions on Intelligent Systems,2019,14(5):1056-1063.[doi:10.11992/tis.201809037]
点击复制
《智能系统学报》[ISSN 1673-4785/CN 23-1538/TP] 卷:
14
期数:
2019年第5期
页码:
1056-1063
栏目:
学术论文—自然语言处理与理解
出版日期:
2019-09-05
- Title:
-
Representation learning using network embedding based on external word vectors
- 作者:
-
张潇鲲, 刘琰, 陈静
-
数学工程与先进计算国家重点实验室, 河南 郑州 450000
- Author(s):
-
ZHANG Xiaokun, LIU Yan, CHEN Jing
-
Mathematical Engineering and Advanced Computing State Key Laboratory, Zhengzhou 450000
-
- 关键词:
-
网络表示学习; 文本信息网络; 自编码器; 外部词向量; 节点分类; 词向量; 分布式表示; 表示学习
- Keywords:
-
network embedding; content information network; auto-encoder; external word vectors; vertex classification; word vectors; distributed representation; representation learning
- 分类号:
-
TP181
- DOI:
-
10.11992/tis.201809037
- 摘要:
-
针对信息网络(text-based information network)现有研究多基于网络自身信息建模,受限于任务语料规模,只使用任务相关文本进行建模容易产生语义漂移或语义残缺的问题,本文将外部语料引入建模过程中,利用外部语料得到的词向量对建模过程进行优化,提出基于外部词向量的网络表示模型NE-EWV(network embedding based on external word vectors),从语义特征空间以及结构特征空间两个角度学习特征融合的网络表示。通过实验,在现实网络数据集中对模型有效性进行了验证。实验结果表明,在链接预测任务中的AUC指标,相比只考虑结构特征的模型提升7%~19%,相比考虑结构与文本特征的模型在大部分情况下有1%~12%提升;在节点分类任务中,与基线方法中性能最好的CANE性能相当。证明引入外部词向量作为外部知识能够有效提升网络表示能力。
- Abstract:
-
Network embedding, which preserves a network’s sophisticated features, can effectively learn the low-dimensional embedding of vertices in order to lower the computing and storage costs. Content information networks (such as Twitter), which contain rich text information, are commonly used in daily life. Most studies on content information network are based on the information of the network itself. Distributed word vectors are becoming increasingly popular in natural language processing tasks. As a low-dimensional representation of the semantic feature space, word vectors can preserve syntactic and semantic regularities. By introducing external word vectors into the modeling process, we can use the external syntactic and semantic features. Hence, in this paper, we propose network embedding based on external word vectors (NE-EWV), whereby the feature fusion representation is learned from both semantic feature space as well as structural feature space. Empirical experiments were conducted using real-world content information network datasets to validate the effectiveness of the model. The results show that in link prediction task, the AUC of the model was 7% to 19% higher than that of the model that considers only the structural features, and in most cases was 1% to 12% higher than the model that considers structural and text features. In node classification tasks, the performance is comparable with that of context-aware network embedding (CANE), which was the state-of-the-art baseline model.
备注/Memo
收稿日期:2019-09-19。
基金项目:国家自然科学基金项目(61309007,U1636219);国家重点研发计划课题资助项目(2016YFB0801303,2016QY01W0105).
作者简介:张潇鲲,男,1991年生,硕士研究生,主要研究方向为网络表示学习;刘琰,女,1979年生,副教授,博士,主要研究方向为网络信息安全、网络资源测绘。申请发明专利10项,授权5项,发表学术论文40余篇;陈静,女,1990年生,讲师,主要研究方向为数据挖掘、自然语言处理和社会网络分析。授权发明专利2项,发表学术论文10篇。
通讯作者:刘琰.E-mail:ms_liuyan@aliyun.com
更新日期/Last Update:
1900-01-01