Back to Search Start Over

Application of Attention Mechanism with Prior Information in Natural Language Processing.

Authors :
Zhang, Lingling
Zhou, Zhenxiong
Ji, Pengyu
Mei, Aoxue
Source :
International Journal on Artificial Intelligence Tools; Jun2022, Vol. 31 Issue 4, p1-18, 18p
Publication Year :
2022

Abstract

When using deep learning methods to model natural language, a recurrent neural network that can map input sequences to output sequences is usually used. Considering that natural language contains more complicated syntactic structures, and the performance of cyclic neural networks in long sentence processing will decrease, scholars have introduced an attention mechanism into the model, which has improved the above problems to a certain extent. The existing attention mechanism still has some shortcomings, such as the inability to explicitly obtain the known syntactic structure information in the sentence, and the poor interpretability of the output probability. In response to the above problems, this article will improve the attention mechanism in the recurrent neural network model. Firstly, the prior information in the natural language sequence is constructed as a graph model through syntactic analysis and other means, and then the graph structure regularization term is introduced into the sparse mapping. A new function netmax is constructed to replace the softmax function in the traditional attention mechanism, thereby improving the performance of the model and making the degree of association. The input values corresponding to larger input samples are closer, making the output of the attention mechanism easier to understand. The innovation of this paper mainly lies in that the weight calculation method which can be widely used in the attention mechanism is proposed by combining the deep learning model with statistical knowledge, which opens a channel to introduce the prior information for the deep learning model in natural language processing tasks. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
02182130
Volume :
31
Issue :
4
Database :
Complementary Index
Journal :
International Journal on Artificial Intelligence Tools
Publication Type :
Academic Journal
Accession number :
157517778
Full Text :
https://doi.org/10.1142/S0218213022400085