Abstract
In evaluating spoken dialogue systems, especially those aimed at actualizing a natural conversation with users, it seems necessary to assess whether the interactions between users and systems are chained appropriately. In this paper, we propose a response evaluation coding scheme for user dialogue actions and evaluate the interactional chains achieved by the users and the systems from the viewpoint of appropriateness of reactions. Two dialogue data sets of interactions between users and sightseeing guidance systems were evaluated by this scheme. The differences between the two systems and the points that should be improved were determined by analyzing the data coded by the scheme. From the result of multiple regression analysis, it was also suggested that the appropriate ratings (AR) for user actions might contribute to predicting the subjective judgment concerning naturalness of the system actions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
N. Beringer, U. Kartal, K. Louka, F. Schiel, and U. Turk. Promise - a procedure for multimodal interactive system evaluation. In 3rd International Conference on Language Resources and Evaluation (LREC), pages 77–80, 2002.
Mark G. Core and James F. Allen. Coding dialogs with the DAMSL annotation scheme. In Working Notes of AAAI Fall Symposium on Communicative Action in Humans and Machines, Boston, 1997.
Klaus-Peter Engelbrecht and Sebastian M¨oller. Sequential classifiers for the prediction of user judgments about spoken dialog systems. Speech Communication, 52, 10:816–833, 2010.
H.P. Grice. Logic and conversation. In P. Cole and J Morgan, editors, Syntax and Semantics volume 3, pages 41–58. Academic Press, New York, 1975.
C. Hori, K. Othake, T. Misu, H. Kashioka, and S. Nakamura. Statistical dialog management applied to wfst-based dialog systems. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 4793–4796, 2009.
Kazunori Komatani, Shinichi Ueno, Tatsuya Kawahara, and Hiroshi G. Okuno. User modeling in spoken dialogue systems to generate flexible guidance. User Modeling and User-Adapted Interaction, 15, 1:169–183, 2005.
Ramón López-Cózar and Michael Mctear Zoraida Callejas. Testing the performance of spoken dialogue systems by means of an artificially simulated user. Artificial Intelligence Review, 26, 4:291–323, 2006.
Teruhisa Misu, Kiyonori Ohtake, Chiori Hori, Hideki Kashioka, and Satoshi Nakamura. Annotating communicative function and semantic content in dialogue act for construction of consulting dialogue systems. In Proceedings of Interspeech 2009, pages 1843–1846, 2009.
S. Möller. Evaluating interactions with spoken dialogue telephone services. In L. Dybkjær, editor, Recent Trends in Discourse and Dialogue, pages 69–100. Springer, 2007.
Sebastian Möller, Klaus-Peter Engelbrecht, and Robert Schleicher. Predicting the quality and usability of spoken dialogue services. Speech Communication, 50, 8–9:730–744, 2008.
Kiyonori Ohtake, Teruhisa Misu, Chiori Hori, Hideki Kashioka, and Satoshi Nakamura. Dialogue acts annotation to construct dialogue systems for consulting. In Spoken Dialogue Systems Technology and Design, pages 231–254. Springer, 2010.
Susan Robinson, Antonio Roque, and David Traum. Dialogues in context: An objective user-oriented evaluation approach for virtual human dialogue. In Proceedings of the Seventh Conference on International Language Resources and Evaluation (LREC’10), 2010.
Susan Robinson, David Trauma, Midhun Ittycheriah, and Joe Henderer. What would you ask a conversational agent? Observations of human-agent dialogues in a museum setting. In Proceedings of the Sixth International Language Resources and Evaluation (LREC’08), pages 64–71, 2008.
Harvey Sacks, Emanuel A. Schegloff, and Gail Jefferson. A simplest systematics for the organization of turn-taking for conversation. Language, 50:696– 735, 1974.
E. Shriberg, R. Dhillon, S. Bhagat, J. Ang, and H. Carvey. The ICSI meeting recorder dialog act (MRDA) corpus. In M. Strube and C. Sidner, editors, Proc. 5th SIGdial Workshop on Discourse and Dialogue, pages 97–100. Cambridge, MA, 2004.
David R. Traum, Susan Robinson, and Jens Stephan. Evaluation of multi-party virtual reality dialogue interaction. In European Language Resources Association (ELRA), editor, Proceedings of the Sixth International Language Resources and Evaluation (LREC’04), Marrakech, Morocco, 2004.
Marilyn Walker, Diane Litman, Candace Kamm, and Alicia Abella. Paradise: A framework for evaluating spoken dialogue agents. In Proceedings of the 35th Annual Meeting of the Association of Computational Linguistics (ACL 97), 1997.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer Science+Business Media, LLC
About this paper
Cite this paper
Mizukami, E., Kashioka, H. (2011). Evaluating User-System Interactional Chains for Naturalness-oriented Spoken Dialogue Systems. In: Delgado, RC., Kobayashi, T. (eds) Proceedings of the Paralinguistic Information and its Integration in Spoken Dialogue Systems Workshop. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-1335-6_30
Download citation
DOI: https://doi.org/10.1007/978-1-4614-1335-6_30
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-1334-9
Online ISBN: 978-1-4614-1335-6
eBook Packages: EngineeringEngineering (R0)