Abstract
Following the increase in use of smart devices, various real-time environmental information becomes available everywhere. To provide more context-aware information, we also need to know emotion and a satisfaction level in a viewpoint of users. In this paper, we define it as “a user satisfaction impact (USI)” and propose a method to estimate USI by combining dialogue features and physical reaction features. As dialogue features, facial expression and acoustic feature are extracted from multimodal dialogue system on a smartphone. As physical reactions, head motion, eye motion, and heartbeat are collected by wearable devices. We conducted the preliminary experiments in the real-world to confirm the feasibility of this study in the tourism domain. Among various features, we confirmed that eye motion correlates with satisfaction level up to 0.36.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alegre J, Garau J (2010) Tourist satisfaction and dissatisfaction. Ann Tour Res 37(1):52–73
AlHanai TW, Ghassemi MM (2017) Predicting latent narrative mood using audio and physiologic data. In: AAAI, pp 948–954
Balandina E, Balandin S, Koucheryavy Y, Mouromtsev D (2015) Iot use cases in healthcare and tourism. In: 2015 IEEE 17th conference on business informatics, vol 2, pp 37–44
Burke JA, Estrin D, Hansen M, Parker A, Ramanathan N, Reddy S, Srivastava MB (2006) Participatory sensing. Center for embedded network sensing
Morishita S, Maenaka S, Daichi N, Tamai M, Yasumoto K, Fukukura T, Sato K (2015) Sakurasensor: quasi-realtime cherry-lined roads detection through participatory video sensing by cars. In: Proceedings of the UBICOMP ’15, pp 695–705
Nakamura Y, Arakawa Y, Kanehira T, Fujiwara M, Yasumoto K (2017) Senstick: comprehensive sensing platform with an ultra tiny all-in-one sensor board for IOT research. J Sens
Quck WY, Huang DY, Lin W, Li H, Dong M (2016) Mobile acoustic emotion recognition. In: 2016 IEEE region 10 conference (TENCON), pp 170–174
Resch B, Summa A, Sagl G, Zeile P, Exner JP (2014) Urban emotions – geo-semantic emotion extraction from technical sensors, human sensors and crowdsourced data. pp 199–212
Ringeval F, Eyben F, Kroupi E, Yuce A, Thiran JP, Ebrahimi T, Lalanne D, Schuller B (2015) Prediction of asynchronous dimensional emotion ratings from audiovisual and physiological data. Pattern Recognit Lett 66:22–30
Soleymani M, Pantic M, Pun T (2012) Multimodal emotion recognition in response to videos. IEEE Trans Affect Comput 3(2):211–223
Acknowledgements
This work was supported by JSPS KAKENHI Grant Number 16J09670, 16H01721.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Matsuda, Y., Fedotov, D., Takahashi, Y., Arakawa, Y., Yasumoto, K., Minker, W. (2019). Estimating User Satisfaction Impact in Cities Using Physical Reaction Sensing and Multimodal Dialogue System. In: D'Haro, L., Banchs, R., Li, H. (eds) 9th International Workshop on Spoken Dialogue System Technology. Lecture Notes in Electrical Engineering, vol 579. Springer, Singapore. https://doi.org/10.1007/978-981-13-9443-0_15
Download citation
DOI: https://doi.org/10.1007/978-981-13-9443-0_15
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-9442-3
Online ISBN: 978-981-13-9443-0
eBook Packages: Literature, Cultural and Media StudiesLiterature, Cultural and Media Studies (R0)