首页 | 本学科首页   官方微博 | 高级检索  
     检索      

基于改进Transformer的生成式文本摘要模型
引用本文:赵伟,王文娟,任彦凝,刘群,胥钟予,彭露.基于改进Transformer的生成式文本摘要模型[J].重庆邮电大学学报(自然科学版),2023,35(1):185-192.
作者姓名:赵伟  王文娟  任彦凝  刘群  胥钟予  彭露
作者单位:重庆邮电大学 国际合作与交流处,重庆 400065;国网重庆市电力公司信息通信分公司 调控中心,重庆 401121;重庆邮电大学 计算机科学与技术学院,重庆 400065
基金项目:国家自然科学基金(61772096);国家重点研发计划(2018YFC0832100,2018YFC0832102)
摘    要:基于循环神经网络(recurrent neural network,RNN)注意力机制的序列到序列模型在摘要信息提取服务中已经取得了较好的应用,但RNN不能较好地捕捉长时序信息,这使现有模型受限。为此,提出了基于改进Transformer的生成式文本摘要模型。该模型利用Transformer提取全局语义,使用局部卷积提取器提取原文细粒度特征,并设计全局门控单元以防止信息冗余和弥补语义不足,筛选出利于摘要生成的文本关键信息。实验结果表明,所提模型在大规模中文短文本摘要数据集(large scale Chinese short text summarization,LCSTS)上的效果有明显的提升,在电力运维摘要数据集上,也有良好效果,具有可扩展性。

关 键 词:生成式摘要  序列到序列  改进Transformer  局部卷积
收稿时间:2021/9/13 0:00:00
修稿时间:2023/1/10 0:00:00

A generative abstractive summarization method based on the improved Transformer
ZHAO Wei,WANG Wenjuan,REN Yanning,LIU Qun,XU Zhongyu,PENG Lu.A generative abstractive summarization method based on the improved Transformer[J].Journal of Chongqing University of Posts and Telecommunications,2023,35(1):185-192.
Authors:ZHAO Wei  WANG Wenjuan  REN Yanning  LIU Qun  XU Zhongyu  PENG Lu
Institution:Office of International Cooperation and Exchanges, Chongqing University of Posts and Telecommunications, Chongqing 400065, P. R. China;Control Center, State Grid Chongqing Information and Telecommunication Company, Chongqing 401121, P. R. China;School of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, P. R. China
Abstract:The sequence-to-sequence model based on recurrent neural network attention mechanism has been well applied in summary information extraction service. But the existing models are limited by the fact that RNN cannot capture long time sequence information. Therefore, this paper proposes a generative text summary method based on the improved Transformer. Firstly, the model uses Transformer to extract global semantics, uses local convolution extractor to extract fine-grained features of the original text, and designs a global gating unit to prevent information redundancy and make up for semantic deficiencies, and screens out key text information that is conducive to summary generation. The experimental results show that the proposed model has significantly improved the effect on large scale Chinese short text summarization (LCSTS), and also has good effect and scalability on power operation and maintenance summary dataset.
Keywords:abstractive summarization  sequence-to-sequence  improved Transformer  local convolution
点击此处可从《重庆邮电大学学报(自然科学版)》浏览原始摘要信息
点击此处可从《重庆邮电大学学报(自然科学版)》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号