Back to Search
Start Over
Constraint based Knowledge Base Distillation in End-to-End Task Oriented Dialogs
- Source :
- ACL/IJCNLP (Findings)
- Publication Year :
- 2021
- Publisher :
- Association for Computational Linguistics, 2021.
-
Abstract
- End-to-End task-oriented dialogue systems generate responses based on dialog history and an accompanying knowledge base (KB). Inferring those KB entities that are most relevant for an utterance is crucial for response generation. Existing state of the art scales to large KBs by softly filtering over irrelevant KB information. In this paper, we propose a novel filtering technique that consists of (1) a pairwise similarity based filter that identifies relevant information by respecting the n-ary structure in a KB record. and, (2) an auxiliary loss that helps in separating contextually unrelated KB information. We also propose a new metric -- multiset entity F1 which fixes a correctness issue in the existing entity F1 metric. Experimental results on three publicly available task-oriented dialog datasets show that our proposed approach outperforms existing state-of-the-art models.<br />D. Raghu and A. Jain contributed equally to this work
- Subjects :
- FOS: Computer and information sciences
Structure (mathematical logic)
Computer Science - Machine Learning
Multiset
Computer Science - Computation and Language
Correctness
business.industry
Computer science
Filter (signal processing)
computer.software_genre
Machine Learning (cs.LG)
Constraint (information theory)
Knowledge base
Metric (mathematics)
Artificial intelligence
Dialog box
business
Computation and Language (cs.CL)
computer
Natural language processing
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
- Accession number :
- edsair.doi.dedup.....e2108d754838609b60a93c538cff5cce