[1]LI Yunjie,WANG Danyang,LIU Haitao,et al.Document-level relation extraction of a graph reasoning embedded dynamic self-attention network[J].CAAI Transactions on Intelligent Systems,2025,20(1):52-63.[doi:10.11992/tis.202311021]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
20
Number of periods:
2025 1
Page number:
52-63
Column:
学术论文—机器学习
Public date:
2025-01-05
- Title:
-
Document-level relation extraction of a graph reasoning embedded dynamic self-attention network
- Author(s):
-
LI Yunjie1; WANG Danyang2; LIU Haitao1; 3; WANG Huadong4; WANG Peizhuang3
-
1. Institute of Mathematics and Systems Science, Liaoning Technical University, Fuxin 123000, China;
2. Institute of Scientific and Technical Information, Chinese Academy of Tropical Agricultural Sciences, Haikou 571000, China;
3. Institute of Intelligence Engineering and Mathematics, Liaoning Technical University, Fuxin 123000, China;
4. Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
-
- Keywords:
-
document-level relation extraction; graph reasoning; dynamic self-attention network; self-attention mechanism; gated token selection mechanism; document graph; graph attention network; key word
- CLC:
-
TP391
- DOI:
-
10.11992/tis.202311021
- Abstract:
-
Document-level relation extraction refers to the extraction of all entity pairs with semantic relationships from documents and judging their relationship categories. It is different from sentence-level relation extraction, where the determination of entity relationships needs to be inferred from multiple sentences in the document. The existing methods mainly use self-attention for document-level relation extraction, but the use of self-attention for document-level relation extraction needs to address two technical challenges: the high computational complexity of long text semantic encoding and the complex reasoning modeling required for relationship prediction. Therefore, a graph reasoning embedded dynamic self-attention network model (GSAN) is proposed. With the aid of gated word selection mechanism, GSAN dynamically selects important words to calculate self attention, achieving high-efficiency modeling for semantic dependency of long text sequences. At the same time, it is considered to construct a document graph with the word selection as the global semantic background, entity candidates and document nodes. Then, the graph reasoning aggregation information of the document graph being embedded into the dynamic self-attention module enables the model to model complex reasoning. The experimental results demonstrate that the proposed model is a significant improvement over other baseline models on the public document-level relation dataset CDR and DocRED.