电子科技 ›› 2024, Vol. 37 ›› Issue (7): 16-24.doi: 10.16180/j.cnki.issn1007-7820.2024.07.003

• • 上一篇    下一篇

基于强化正则的小样本自动摘要方法

李清, 万卫兵   

  1. 上海工程技术大学 电子电气工程学院,上海 201620
  • 收稿日期:2023-02-04 出版日期:2024-07-15 发布日期:2024-07-17
  • 作者简介:李清(1995-),男,硕士研究生。研究方向:自然语言处理。
    万卫兵(1969-),男,博士,副教授。研究方向:基于知识增强的协同智能决策、自然语言处理。
  • 基金资助:
    科技创新2030“新一代人工智能”重大项目(2020AAA0109300)

Automatic Summarization of Small Samples Based on Enhanced Regularization

LI Qing, WAN Weibing   

  1. School of Electronic and Electrical Engineering,Shanghai University of Engineering Science, Shanghai 201620,China
  • Received:2023-02-04 Online:2024-07-15 Published:2024-07-17
  • Supported by:
    Scientific and Technological Innovation 2030-Major Project of New Generation Artificial Intelligence(2020AAA0109300)

摘要:

文本自动摘要旨在从文本信息中提取主要语句以压缩信息。现有生成式自动摘要方法无法充分利用预训练模型对原文语义进行学习,导致生成内容易丢失重要信息,当面对样本数量较少的数据集时容易发生过拟合。为了解决此类问题并获得更好的微调性能,文中使用预训练模型mT5(multilingual T5)作为基线,通过结合R-drop(Regularized dropout)对模型微调进行强化正则来提高模型学习能力,同时利用Sparse softmax减少预测生成的模糊性来确保输出准确度。模型在中文数据集LCSTS和CSL上通过计算BLEU(Bilingual Evaluation Understudy)进行优化方法超参数测试,并采用Rouge作为评测指标分别对数据集进行了不同数量级的评测。实验结果表明,经过优化的预训练模型能够更好地学习原文语义表征,在小样本情况下模型能够保持较好的拟合效果,并且能够生成实用性较高的结果。

关键词: 文本自动摘要, 文本生成, 预训练模型, 小样本数据, 强化正则, 稀疏化输出, 语义表征学习, mT5

Abstract:

Automatic text summarization aims to extract the main statements from text information for the purpose of compressing information. Existing generative automatic summarization methods do not take full advantage of the pre-trained model to learn the semantics of the original text, resulting in the loss of important information in the generated content, when the data set with a small number of samples is often prone to overfitting. In order to solve such problems and obtain better fine-tuning performance, the pre-trained model mT5(multilingual T5) is used as a baseline to improve the learning ability of the model by combining R-drop(Regularized dropout) with reinforced regularity for model fine-tuning, and Sparse softmax is used to reduce the ambiguity of prediction generation to ensure the accuracy of the output. The model calculates BLEU(Bilingual Evaluation Understudy) for hyperparameter test on Chinese data sets LCSTS and CSL, and uses Rouge as evaluation index to evaluate data sets of different orders of magnitude. The experimental results show that the optimized pre-trained model can better learn the semantic representation of the original text, and the model can maintain a good fit in the small samples and generate more practical results.

Key words: automatic text summarization, text generation, pre-trained model, small sample data, reinforced regularity, sparse output, semantic representation learning, mT5

中图分类号: 

  • TP391.1