Computer and Modernization ›› 2024, Vol. 0 ›› Issue (04): 66-76.doi: 10.3969/j.issn.1006-2475.2024.04.012

Previous Articles     Next Articles

Short Text Classification Method Based on Improved Adversarial Learning and Fusion Features

  


  1. (1. Yangtze Delta Region (Huzhou) Institute of Intelligent Transportation, Huzhou University, Huzhou 313000, China;
    2. Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou 313000, China;
    3. School of Science and Engineering, Huzhou College, Huzhou 313000, China)
  • Online:2024-04-30 Published:2024-05-13

Abstract: Abstract: Text classification is one of the most important directions in natural language processing research. Short text has the characteristics of less word count, ambiguity, less key information and not easy to capture, but the algorithm models that classify them are often different in training and reasoning, and mainstream classification models basically model key features and ignore non-key feature information, which increases the challenges in accurate classification. In order to solve the above problems, this paper proposes a short text classification framework combining the fusion of multiple adversarial training strategies and improving the self-attention mechanism. At the beginning, the model adds adversarial perturbation to the text vector representation level to strengthen the text representation ability, and adds an adversarial perturbation to improve the model weights after the F1 score reaches a certain threshold to strengthen the generalization ability of the model during training and inference, thereby assisting in improving the feature learning ability of each classifier of the framework. In terms of feature learning network module, this paper uses the combination of multi-scale convolutional module and bidirectional long short-term memory neural network to learn different granular features, in order to learn nonadjacent feature information, introduces hole convolution, increases the convolution receptive field, and designs a gating mechanism to control the learning speed of this layer information. Finally, by adding a new attention mechanism, the key information is modeled and the non-critical information is modeled, and the loss is added for calculation, which enhances the model’s ability to learn feature information and reduces the risk of overfitting. The tests of THUCNews news headline dataset and Toutiao headline dataset of two large-scale public datasets show that the F1 score of this method is increased by up to 4.93 percentage points and 6.14 percentage points compared with the current mainstream model and classical model, and the effectiveness of adding weight disturbance threshold and different modules is also compared and ablation experiments are explored.

Key words: Key words: adversarial training, strategy fusion, void convolution, non-critical information, attention mechanisms

CLC Number: