计算机工程与应用 ›› 2024, Vol. 60 ›› Issue (23): 168-175.DOI: 10.3778/j.issn.1002-8331.2308-0445

• 模式识别与人工智能 • 上一篇    下一篇

结合DistilBERT与标签关联性的多标签文本分类

王旭阳,耿留青,张鑫   

  1. 兰州理工大学 计算机与通信学院,兰州 730050
  • 出版日期:2024-12-01 发布日期:2024-11-29

Multi-Label Text Classification Based on DistilBERT and Label Correlation

WANG Xuyang, GENG Liuqing, ZHANG Xin   

  1. School of Computer and Communication, Lanzhou University of Technology, Lanzhou 730050, China
  • Online:2024-12-01 Published:2024-11-29

摘要: 现有的多标签文本分类方法往往忽视了标签的关联性和语义信息,导致标签特征提取不充分,标签之间的相关性信息难以得到有效的利用,为解决这一问题,提出一个融合DistilBERT和标签关联性信息的模型IDLC。使用DistilBERT获得文本的和标签的词向量表示,同时利用DistilBERT获取到包含文本上下文信息的全局特征,通过CNN提取文本局部特征,再使用图的形式来表示标签之间的相互关系,通过具有多头注意力机制的图注意力网络捕获包含关联信息的标签特征,最后通过特征融合方法将标签特征和文本特征进行融合,构建包含更多特征信息的文本表示,以此来提升模型的分类精度。在基准数据集上的实验结果表明,与基准模型相比,该方法能有效提升模型性能,在多标签文本分类任务中有更好的分类效果。

关键词: 多标签文本分类, DistilBERT, 图注意力网络, 卷积神经网络(CNN)

Abstract: Existing multi-label text classification methods often ignore the relevance and semantic information of the labels, resulting in inadequate extraction of label features, and the correlation information between the labels is difficult to be effectively utilized. In order to solve this problem, a model IDLC is proposed that fuses DistilBERT and label correlation. Firstly, the word vector representations of the text and the labels are obtained by using DistilBERT and simultaneously DistilBERT is used to obtain global features containing text context information. Secondly, local features of the text are extracted by CNN, then the interrelationships between the labels are represented in the form of a graph, and then label features containing correlation information are captured by a graph attention network with a multi-attention mechanism.Thirdly, label features and text features are fused by a feature fusion method to construct a text representation with more feature information. Finally, this paper fuses labeled features and text features by feature fusion method to construct a text representation containing more feature information, so as to improve the classification accuracy of the model. The experimental results on the benchmark dataset show that compared with the benchmark model, this method can effectively improve the performance of the model and has better classification effects in the multi-label text classification task.

Key words: multi-label text classification, DistilBERT, graph attention network, convolutional neural network (CNN)