Electronic Science and Technology ›› 2024, Vol. 37 ›› Issue (9): 87-94.doi: 10.16180/j.cnki.issn1007-7820.2024.09.013

Previous Articles    

Researchon Short Text News Title Generation Method

ZHAO Ming   

  1. Network Center,Fujian Longyan Xinluo District Education Bureau,Longyan 364099,China
  • Received:2023-01-08 Online:2024-09-15 Published:2024-09-20
  • Supported by:
    National Key R&D Program of China(2022YFF0903404)

Abstract:

Today's news has the characteristics of short text, frequent release, timeliness, etc. A media account releases dozens of news in a day. Developing suitable and attractive headlines for large volumes of news has become a major part of the work of media workers. Media workers need a system that automatically generates short text headlines to relieve their stress. To solve this problem, this study proposes a short text news title generation model. The model adopts sequence-to-sequence structure, using pre-trained language model and layered self-attention decoder in encoder and decoder respectively. In order to make the generated headlines contain the key information of the original news, a staged training method based on LCSTS data set and Weibo4 data set is proposed, and the model learns to extract the key news information and construct a stylized expression from the two data sets respectively, so that the generated headlines can accurately express the core content of the news and attract readers.

Key words: news headline generation, pre-training language model,, layered self-attention decoder, encoder, text extraction, text generation

CLC Number: 

  • TP391