Computer and Modernization ›› 2024, Vol. 0 ›› Issue (09): 101-106.doi: 10.3969/j.issn.1006-2475.2024.09.017

Previous Articles     Next Articles

Short Text Classification Combining Attention Mechanism and Mengzi Model

  

  1. (1. School of Electrical and Information Engineering, Northeast Petroleum University, Daqing 163318, China;
    2. School of Computer and Information Technology, Northeast Petroleum University, Daqing 163318, China)
  • Online:2024-09-27 Published:2024-09-29

Abstract: How to use short text classification technology to mine useful text information is one of the current hot research directions. To solve the problem of sparse feature information and difficult extraction of short text, a short text classification model named Mengzi-ADCBU is proposed. This model uses Mengzi pre-training model to convert input text information into corresponding text representation. Then, the obtained text vectors are input to the improved deep pyramid convolutional neural network and the bidirectional gated unit integrated with multi-head attention mechanism to extract text feature information, and the extracted feature information is fused and sent to the full connection layer and Softmax function to complete short text classification. Multiple models comparison experiments are carried out on the publicly available THUCNews short text data set and SougouCS short text data set respectively. The experimental results show that the proposed Mengzi-ADCBU model is better than the current mainstream models in the accuracy, precision, recall rate and F1 value of short text classification and has better short text classification ability.

Key words: short text, multi-head attention, deep pyramid convolutional neural netwrks, bidirectional gated unit

CLC Number: