Authors
Marilyn Walker, Rebecca J Passonneau, Julie E Boland
Publication date
2001/7
Conference
Proceedings of the 39th Annual Meeting of the Association for Computational Linguistics
Pages
515-522
Description
This paper describes the application of the PARADISE evaluation framework to the corpus of 662 human-computer dialogues collected in the June 2000 Darpa Communicator data collection. We describe results based on the standard logfile metrics as well as results based on additional qualitative metrics derived using the DATE dialogue act tagging scheme. We show that performance models derived via using the standard metrics can account for 37% of the variance in user satisfaction, and that the addition of DATE metrics improved the models by an absolute 5%.
Total citations
200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320241411015121392815168127471515815971
Scholar articles
M Walker, RJ Passonneau, JE Boland - Proceedings of the 39th Annual Meeting of the …, 2001