1. Joint Turn and Dialogue level User Satisfaction Estimation on Multi-Domain Conversations
- Author
-
Josep Valls-Vargas, Lazaros Polymenakos, Spyros Matsoukas, Aditya Tiwari, and Praveen Kumar Bodigutla
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer science ,Computer Science - Artificial Intelligence ,media_common.quotation_subject ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Machine Learning (cs.LG) ,0202 electrical engineering, electronic engineering, information engineering ,Generalizability theory ,Quality (business) ,0105 earth and related environmental sciences ,media_common ,Computer Science - Computation and Language ,Artificial neural network ,End user ,business.industry ,User satisfaction ,Artificial Intelligence (cs.AI) ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Computation and Language (cs.CL) - Abstract
Dialogue level quality estimation is vital for optimizing data driven dialogue management. Current automated methods to estimate turn and dialogue level user satisfaction employ hand-crafted features and rely on complex annotation schemes, which reduce the generalizability of the trained models. We propose a novel user satisfaction estimation approach which minimizes an adaptive multi-task loss function in order to jointly predict turn-level Response Quality labels provided by experts and explicit dialogue-level ratings provided by end users. The proposed BiLSTM based deep neural net model automatically weighs each turn's contribution towards the estimated dialogue-level rating, implicitly encodes temporal dependencies, and removes the need to hand-craft features. On dialogues sampled from 28 Alexa domains, two dialogue systems and three user groups, the joint dialogue-level satisfaction estimation model achieved up to an absolute 27% (0.43->0.70) and 7% (0.63->0.70) improvement in linear correlation performance over baseline deep neural net and benchmark Gradient boosting regression models, respectively., Findings of EMNLP, 2020
- Published
- 2020