面向情感分类的非监督异构对比蒸馏模型

UNSUPERVISED CONTRASTIVE DISTILLATION MODEL WITH HETEROGENEOUS SETTINGS FOR SENTIMENT CLASSIFICATION

  • 摘要: 网络评论的数量呈爆炸式增长,对这些评论进行情感分析有着重要的研究价值。自从 BERT模型被提出后,预训练模型成为情感分析任务上的常用方法,但存在模型参数量过大、推理速度缓慢的缺点。在此之前情感分析的方法是一些简单的神经网络模型,训练速度快,可部署性强,但效果一般。因此,结合两类方法的优点,该文提出一种异构设置的非监督对比蒸馏模型,用于网络评论情感分析。在相同数据集和计算资源的情况下,该模型较BERT模型参数量减少146倍,推理时间减少270倍;较DistilBERT蒸馏参数量减少88倍,推理时间减少42.3倍,效果提升1.8个百分点(68.3% vs 70.1%)。

     

    Abstract: With the rapid development of Internet technology, the number of online reviews is increasing. So, sentiment analysis on these reviews is of important research value. Since the BERT model was proposed, pre-trained language models became a common method for sentiment analysis tasks. However, these models had a large number of parameters and took long inference time. The common sentiment analysis methods were some simple neural network models. These models had fast training speed and strong deployability, but mediocre effect. Therefore, combining the advantages of two types of methods, this paper proposes an unsupervised comparative distillation model with heterogeneous settings. With same dataset settings and computing resources, our model reduces the number of parameters of the BERT model by 146 times, and the reasoning time is reduced by 207 times. Compared with DistilBERT, the amount of distillation parameters is reduced by 88 times, the inferencing time is reduced by 42.3 times, and the performance is increased by 1.8% (68.3% vs 70.1%).

     

/

返回文章
返回