首页 | 本学科首页   官方微博 | 高级检索  
     检索      

融合多路注意力机制的语句匹配模型
引用本文:王进,刘麒麟,马樱仪,孙开伟,胡珂.融合多路注意力机制的语句匹配模型[J].重庆邮电大学学报(自然科学版),2023,35(3):520-527.
作者姓名:王进  刘麒麟  马樱仪  孙开伟  胡珂
作者单位:重庆邮电大学 数据工程与可视计算重庆市重点实验室, 重庆 400065
基金项目:国家重点研发计划专项(SQ2021YFE010559)
摘    要:为了增强语句内关键信息和语句间交互信息的表征能力,有效整合匹配特征,提出一种融合多路注意力机制的语句匹配模型。采用双向长短时记忆网络获取文本的编码信息,通过自注意力机制学习语句内部的关键信息;将编码信息和自注意力信息拼接之后,通过多路注意力机制学习语句间的交互信息;结合并压缩多路注意力层之前和之后的信息,通过双向长短时记忆网络进行池化获得最终的语句特征,经过全连接层完成语句匹配。实验结果表明,该模型在SNLI和MultiNLI数据集上进行的自然语言推理任务、在Quora Question Pairs数据集上进行的释义识别任务和在SQuAD数据集上进行的问答语句选择任务中均表现出了更好效果,能够有效提升语句匹配的性能。

关 键 词:语句匹配  注意力机制  Bi-LSTM  深度学习
收稿时间:2021/12/15 0:00:00
修稿时间:2023/3/3 0:00:00

Sentence matching model fused with multi-channel attention mechanism
WANG Jin,LIU Qilin,MA Yingyi,SUN Kaiwei,HU Ke.Sentence matching model fused with multi-channel attention mechanism[J].Journal of Chongqing University of Posts and Telecommunications,2023,35(3):520-527.
Authors:WANG Jin  LIU Qilin  MA Yingyi  SUN Kaiwei  HU Ke
Institution:Key Laboratory of Data Engineering and Visual Computing, Chongqing University of Posts and Telecommunications, Chongqing 400065, P.R. China
Abstract:A sentence matching model fused with multi-channel attention mechanism is proposed to enhance the characterization of key information within sentence and interaction information between sentences, and to effectively integrate matching features. The model first uses a bi-directional LSTM to obtain the encoding information of the text, learns the key information within the sentence through the self-attentive mechanism, and then puts the encoding and self-attentive information together to learn the interaction information between sentences through the multi-channel attention mechanism. Combining and compressing the information before and after the multi-channel attention layer, the final sentence features are obtained by bi-directional LSTM and pooling layer. Finally, the sentence matching is completed by full connection layer. The model is compared with other classical deep sentence matching models on SNLI, MultiNLI, Quora Question Pairs and SQuAD datasets. Experimental results show that the model shows better results in natural language inference tasks on SNLI and MultiNLI datasets, paraphrase recognition tasks on Quora Question Pairs dataset and question-answer sentence selection tasks on SQuAD dataset, and can effectively improve the performance of sentence matching.
Keywords:sentence matching  attention  Bi-LSTM  deep learning
点击此处可从《重庆邮电大学学报(自然科学版)》浏览原始摘要信息
点击此处可从《重庆邮电大学学报(自然科学版)》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号