基于本体的对话状态跟踪及其知识蒸馏方法

ONTOLOGY-BASED DIALOGUE STATE TRACKING AND ITS KNOWLEDGE DISTILLATION METHOD

  • 摘要: 对话状态跟踪是任务型对话系统中不可缺少的一部分,用于获取和管理用户在对话上下文中的意图。在以往研究中,对话状态跟踪方法使用多槽位学习以捕捉槽位之间的关联,但没有考虑各个任务的难易差异。此外,现有模型是基于预训练模型的大模型,不利于部署,也不符合对话系统对实时性的要求。针对以上问题,根据槽位分类难易程度为槽位损失加权联合优化;使用知识蒸馏的方法,压缩模型并保持精度;在没有大教师模型的情况下,使用小模型从头开始相互蒸馏,也能达到有教师模型蒸馏的精度。实验结果表明,提出的方法在标准WOZ2.0任务型对话数据集上获得了良好的结果。

     

    Abstract: Dialogue state tracking is an indispensable part of task-oriented dialogue systems, which can acquire and manage the user’sintentions during the dialogue process. In previous studies, dialogue state tracking methods use multi - slot learning to capture the associations between slots, but the difference in difficulty of each task is not considered. In addition, most existing models are large models based on pre-trained models, which are not conducive to deployment and do not meet the real-time requirements of dialogue systems. In view of the above problems, the weighted joint optimization of slot loss was performed according to the difficulty of slot classification. The knowledge distillation method was used to compress the model and maintained the accuracy. In the absence of a large teacher model, the small models were used to distill each other from scratch, and it could also achieve the accuracy of the teacher model distillation. Experimental results show that the proposed method achieves good results on the standard WOZ2. 0 task-oriented dialogue dataset.

     

/

返回文章
返回