Skip to main content

Conclusions and Future Work

  • Chapter
  • First Online:
Building Dialogue POMDPs from Expert Dialogues

Part of the book series: SpringerBriefs in Electrical and Computer Engineering ((BRIEFSSPEECHTECH))

  • 475 Accesses

Abstract

Spoken dialog systems (SDSs) are the systems that help the human user to accomplish a task using the spoken language. Dialog management is a difficult problem since automatic speech recognition (ASR) and natural language understanding (NLU) make errors which are the sources of uncertainty in SDSs. Moreover, the human user behavior is not completely predictable. The users may change their intents during the dialog, which makes the SDS environment stochastic. Furthermore, the users may express an intent in several ways which makes dialog management more challenging.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Abbeel, P., & Ng, A. Y. (2004). Apprenticeship learning via inverse reinforcement learning. In Proceedings of the 21st International Conference on Machine Learning (ICML’04), Banff, AB.

    Google Scholar 

  • Atrash, A., & Pineau, J. (2010). A Bayesian method for learning POMDP observation parameters for robot interaction management systems. In The POMDP Practitioners Workshop.

    Google Scholar 

  • Blei, D. (2012). Probabilistic topic models. Communications of the ACM, 55(4), 77–84.

    Article  MathSciNet  Google Scholar 

  • Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, 993–1022.

    MATH  Google Scholar 

  • Boularias, A., Kober, J., & Peters, J. (2011). Relative entropy inverse reinforcement learning. Journal of Machine Learning Research - Proceedings Track, 15, 182–189.

    Google Scholar 

  • Daud, A., Li, J., Zhou, L., & Muhammad, F. (2010). Knowledge discovery through directed probabilistic topic models: A survey. Frontiers of Computer Science in China, 4(2), 280–301.

    Article  Google Scholar 

  • Doshi, F., & Roy, N. (2008). Spoken language interaction with model uncertainty: An adaptive human-robot interaction system. Connection Science, 20(4), 299–318.

    Article  Google Scholar 

  • Jurafsky, D., & Martin, J. H. (2009). Speech and language processing (2nd ed.). Upper Saddle River, NJ: Prentice-Hall.

    Google Scholar 

  • Kim, D., Kim, J., & Kim, K. (2011). Robust performance evaluation of POMDP-based dialogue systems. IEEE Transactions on Audio, Speech, and Language Processing, 19(4), 1029–1040.

    Article  Google Scholar 

  • Manning, C. D., & Schütze, H. (1999). Foundations of statistical natural language processing. Cambridge, MA: MIT Press.

    MATH  Google Scholar 

  • Neu, G., & Szepesvári, C. (2007). Apprenticeship learning using inverse reinforcement learning and gradient methods. In Proceedings of the 23rd Conference on Uncertainty in Artificial Intelligence (UAI’07), Vancouver, BC.

    Google Scholar 

  • Ng, A. Y., & Russell, S. J. (2000). Algorithms for inverse reinforcement learning. In Proceedings of the 17th International Conference on Machine Learning (ICML’00), Stanford, CA.

    Google Scholar 

  • Png, S., & Pineau, J. (2011). Bayesian reinforcement learning for POMDP-based dialogue systems. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP’11), Prague.

    Google Scholar 

  • Png, S., Pineau, J., & Chaib-Draa, B. (2012). Building adaptive dialogue systems via bayes-adaptive pomdps. IEEE Journal of Selected Topics in Signal Processing, 6(8), 917–927.

    Article  Google Scholar 

  • Ramachandran, D., & Amir, E. (2007). Bayesian inverse reinforcement learning. In Proceedings of the 20th International Joint Conference on Artificial Intelligence (IJCAI’07), Hyderabad.

    Google Scholar 

  • Syed, U., & Schapire, R. (2008). A game-theoretic approach to apprenticeship learning. In Proceedings of the Twenty-First Annual Conference on Neural Information Processing Systems, Vancouver, BC.

    Google Scholar 

  • Williams, J. D. (2006). Partially observable Markov decision processes for spoken dialogue management. Ph.D. thesis, Department of Engineering, University of Cambridge.

    Google Scholar 

  • Ziebart, B., Maas, A., Bagnell, J., & Dey, A. (2008). Maximum entropy inverse reinforcement learning. In Proceedings of the 23rd National Conference on Artificial Intelligence (AAAI’08), Chicago, IL.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2016 The Authors

About this chapter

Cite this chapter

Chinaei, H., Chaib-draa, B. (2016). Conclusions and Future Work. In: Building Dialogue POMDPs from Expert Dialogues. SpringerBriefs in Electrical and Computer Engineering(). Springer, Cham. https://doi.org/10.1007/978-3-319-26200-0_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-26200-0_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-26198-0

  • Online ISBN: 978-3-319-26200-0

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics