[1]CHENG Yan,HU Jiansheng,ZHAO Songhua,et al.Aspect-level sentiment classification model combining Transformer and interactive attention network[J].CAAI Transactions on Intelligent Systems,2024,19(3):728-737.[doi:10.11992/tis.202303016]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
19
Number of periods:
2024 3
Page number:
728-737
Column:
学术论文—自然语言处理与理解
Public date:
2024-05-05
- Title:
-
Aspect-level sentiment classification model combining Transformer and interactive attention network
- Author(s):
-
CHENG Yan1; 2; HU Jiansheng1; ZHAO Songhua1; LUO Pin1; ZOU Haifeng1; ZHAN Yongxin1; FU Yan3; LIU Chunlei4
-
1. School of Computer Information Engineering, Jiangxi Normal University, Nanchang 330022, China;
2. Key Laboratory of Intelligent Information Processing and Emotional Computing in Jiangxi Province, Nanchang 330022, China;
3. Jiangxi Ruanyun Technol
-
- Keywords:
-
aspect term; sentiment classification; recurrent neural network; transformer; interactive attention network; BERT; local feature; deep learning
- CLC:
-
TP391
- DOI:
-
10.11992/tis.202303016
- Abstract:
-
At present, most researchers use a combination of recurrent neural networks and attention mechanisms for aspect-level sentiment classification tasks. However, the recurrent neural network cannot be computed in parallel, and the models encounter problems, such as truncated backpropagation, gradient vanishing, and gradient exploration, in the training process. Traditional attention mechanisms may assign reduced attention weights to important sentiment words in sentences. An aspect-level sentiment classification model combining Transformer and interactive attention network is proposed to solve these problems. In this approach, the pretrained model, which considers bidirectional encoder representation from Transformers (BERT), is initially used to construct word embedding vectors. Then, Transformer encoders are used to perform parallel encoding for input sentences. Subsequently, the contextual dynamic mask-off code and the contextual dynamic weighting mechanisms are applied to focus on local context information semantically relevant to specific aspect words. Finally, the model is tested on five English datasets and four Chinese review datasets. Experimental results demonstrate that the proposed model outperforms others in terms of accuracy and F1.