Abstract
In this chapter, we propose methods for learning the model components of intent-based dialog POMDPs from unannotated and noisy dialogs.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Atrash, A., & Pineau, J. (2010). A Bayesian method for learning POMDP observation parameters for robot interaction management systems. In The POMDP Practitioners Workshop.
Bishop, C. M. (2006). Pattern recognition and machine learning. Secaucus, New York: Springer.
Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, 993–1022.
Cassandra, A., Kaelbling, L., & Littman, M. (1995). Acting optimally in partially observable stochastic domains. In Proceedings of the 12th National Conference on Artificial Intelligence (AAAI’95), Seattle, Washington.
Chinaei, H. R., & Chaib-Draa, B. (2011). Learning dialogue POMDP models from data. In Proceedings of the 24th Canadian Conference on Advances in Artificial Intelligence (Canadian AI’11), St. John’s, Newfoundland.
Chinaei, H. R., & Chaib-Draa, B. (2014b). Dialogue POMDP components (part i): Learning states and observations. International Journal of Speech Technology, 17(4), 309–323.
Chinaei, H. R., Chaib-Draa, B., & Lamontagne, L. (2009). Learning user intentions in spoken dialogue systems. In Proceedings of the 1st International Conference on Agents and Artificial Intelligence (ICAART’09), Porto.
Chinaei, H. R., Chaib-draa, B., & Lamontagne, L. (2012). Learning observation models for dialogue POMDPs. In Proceedings of the 24th Canadian Conference on Advances in Artificial Intelligence (Canadian AI’12), Toronto.
Doshi, F., & Roy, N. (2007). Efficient model learning for dialog management. In Proceedings of the 2nd ACM SIGCHI/SIGART Conference on Human-Robot Interaction (HRI’07), Arlington, VA.
Doshi, F., & Roy, N. (2008). Spoken language interaction with model uncertainty: An adaptive human-robot interaction system. Connection Science, 20(4), 299–318.
Griffiths, T., & Steyvers, J. (2004). Finding scientific topics. Proceedings of the National Academy of Science, 101, 5228–5235.
Gruber, A., & Popat, A. (2007). Notes regarding computations in open htmm. http://openhtmm.googlecode.com/files/htmm_computations.pdf
Gruber, A., Rosen-Zvi, M., & Weiss, Y. (2007). Hidden topic Markov models. In Artificial Intelligence and Statistics (AISTATS’07), San Juan, PR.
Kim, D., Kim, J., & Kim, K. (2011). Robust performance evaluation of POMDP-based dialogue systems. IEEE Transactions on Audio, Speech, and Language Processing, 19(4), 1029–1040.
Lison, P. (2013). Model-based bayesian reinforcement learning for dialogue management. In Proceedings of 14th Annual Conference of the International Speech Communication Association (INTERSPEECH’13), Lyon.
Matsubara, S., Kimura, S., Kawaguchi, N., Yamaguchi, Y., & Inagaki, Y. (2002). Example-based speech intention understanding and its application to in-car spoken dialogue system. In Proceedings of the 19th International Conference on Computational linguistics - Volume 1, Taipei.
Ortiz, L. E., & Kaelbling, L. P. (1999). Accelerating EM: An empirical study. In Proceedings of the 15th Conference on Uncertainty in Artificial Intelligence (UAI’99), Stockholm.
Png, S., & Pineau, J. (2011). Bayesian reinforcement learning for POMDP-based dialogue systems. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP’11), Prague.
Rabiner, L. R. (1990). A tutorial on hidden Markov models and selected applications in speech recognition. In Readings in speech recognition (pp. 267–296). San Francisco: Morgan Kaufmann Publishers.
Ross, S., Chaib-draa, B., & Pineau, J. (2007). Bayes-adaptive POMDPs. In Proceedings of the 21st Annual Conference on Neural Information Processing Systems (NIPS’07), Vancouver, BC.
Ross, S., Pineau, J., Chaib-draa, B., & Kreitmann, P. (2011). A Bayesian approach for learning and planning in partially observable Markov decision processes. Journal of Machine Learning Research, 12, 1729–1770.
Roy, N., Pineau, J., & Thrun, S. (2000). Spoken dialogue management using probabilistic reasoning. In Proceedings of the 38th Annual Meeting on Association for Computational Linguistics (ACL’00), Hong Kong.
Sutton, R. S., & Barto, A. G. (1998). Reinforcement learning: An introduction. Cambridge: MIT Press.
Weilhammer, K., Williams, J. D., & Young, S. (2004). The SACTI-2 corpus: Guide for research users. Cambridge University. Technical Report.
Williams, J. D. (2006). Partially observable Markov decision processes for spoken dialogue management. Ph.D. thesis, Department of Engineering, University of Cambridge.
Williams, J. D., & Young, S. (2005). The SACTI-1 corpus: Guide for research users. Department of Engineering, University of Cambridge. Technical Report.
Williams, J. D., & Young, S. (2007). Partially observable Markov decision processes for spoken dialog systems. Computer Speech and Language, 21, 393–422.
Zhang, B., Cai, Q., Mao, J., & Guo, B. (2001b). Planning and acting under uncertainty: A new model for spoken dialogue system. In Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence (UAI’01), Seattle, Washington.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2016 The Authors
About this chapter
Cite this chapter
Chinaei, H., Chaib-draa, B. (2016). Learning the Dialog POMDP Model Components. In: Building Dialogue POMDPs from Expert Dialogues. SpringerBriefs in Electrical and Computer Engineering(). Springer, Cham. https://doi.org/10.1007/978-3-319-26200-0_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-26200-0_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-26198-0
Online ISBN: 978-3-319-26200-0
eBook Packages: EngineeringEngineering (R0)