Back to Search Start Over

Constraint based Knowledge Base Distillation in End-to-End Task Oriented Dialogs

Authors :
Dinesh Raghu
Atishya Jain
null Mausam
Sachindra Joshi
Source :
ACL/IJCNLP (Findings)
Publication Year :
2021
Publisher :
Association for Computational Linguistics, 2021.

Abstract

End-to-End task-oriented dialogue systems generate responses based on dialog history and an accompanying knowledge base (KB). Inferring those KB entities that are most relevant for an utterance is crucial for response generation. Existing state of the art scales to large KBs by softly filtering over irrelevant KB information. In this paper, we propose a novel filtering technique that consists of (1) a pairwise similarity based filter that identifies relevant information by respecting the n-ary structure in a KB record. and, (2) an auxiliary loss that helps in separating contextually unrelated KB information. We also propose a new metric -- multiset entity F1 which fixes a correctness issue in the existing entity F1 metric. Experimental results on three publicly available task-oriented dialog datasets show that our proposed approach outperforms existing state-of-the-art models.<br />D. Raghu and A. Jain contributed equally to this work

Details

Database :
OpenAIRE
Journal :
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
Accession number :
edsair.doi.dedup.....e2108d754838609b60a93c538cff5cce