Computer and Modernization

Previous Articles     Next Articles

An Automatic Text Summarization Model Construction Method Based on BERT Embedding

  

  1. (North China Institute of Computing Technology, Beijing 100083, China)
  • Received:2019-07-15 Online:2020-02-13 Published:2020-02-13

Abstract: Aiming at the problem that the traditional word vector can not effectively represent polysemous words in text summarization, which reduces the accuracy and readability of summarization, this paper proposes an automatic text summarization model construction method based on BERT (Bidirectional Encoder Representations from Transformers)Embedding. This method introduces the BERT pre-training language model to enhance the semantic representation of word vector. The generated word vectors are input into the Seq2Seq model for training to form an automatic text summarization model, which realizes the rapid generation of text summarization. The experimental results show that the model can effectively improve the accuracy and readability of the generated summarization on Gigaword dataset, and can be used for automatic text summarization generation tasks.

Key words: text summarization, BERT model, attention mechanism, Sequence-to-Sequence(Seq2Seq) model

CLC Number: