Abstract
In this paper, we address the modeling of topic and role information in multiparty meetings, via a nonparametric Bayesian model called the hierarchical Dirichlet process. This model provides a powerful solution to topic modeling and a flexible framework for the incorporation of other cues such as speaker role information. We present our modeling framework for topic and role on the AMI Meeting Corpus, and illustrate the effectiveness of the approach in the context of adapting a baseline language model in a large-vocabulary automatic speech recognition system for multiparty meetings. The adapted LM produces significant improvements in terms of both perplexity and word error rate.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. Journal of Machine Learning Research 3, 1137–1155 (2003)
Blitzer, J., Globerson, A., Pereira, F.: Distributed latent variable models of lexical co-occurrences. In: Proceedings of the 10th International Workshop on Artificial Intelligence and Statistics (2005)
Teh, Y.W.: A hierarchical Bayesian language model based on Pitman-Yor processes. In: Proc. of the Annual Meeting of the ACL, vol. 44 (2006)
Huang, S., Renals, S.: Hierarchical Pitman-Yor language models for ASR in meetings. In: Proc. IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU 2007) (2007)
Bilmes, J.A., Kirchhoff, K.: Factored language models and generalized parallel backoff. In: Proceedings of HLT/NACCL, pp. 4–6 (2003)
Xu, P., Emami, A., Jelinek, F.: Training connectionist models for the structured language model. In: Empirical Methods in Natural Language Processing, EMNLP 2003 (2003)
Wallach, H.M.: Topic modeling: Beyond bag-of-words. In: Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA (2006)
Carletta, J.: Unleashing the killer corpus: experiences in creating the multi-everything AMI Meeting Corpus. Language Resources and Evaluation Journal 41(2), 181–190 (2007)
Teh, Y., Jordan, M., Beal, M., Blei, D.: Hierarchical Dirichlet processes. Journal of the American Statistical Association 101(476), 1566–1581 (2006)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. Journal of Machine Learning Research 3 (2003)
Mrva, D., Woodland, P.C.: Unsupervised language model adaptation for Mandarin broadcast conversation transcription. In: Proc. of Interspeech (2006)
Tam, Y.C., Schultz, T.: Unsupervised LM adaptation using latent semantic marginals. In: Proc. of Interspeech (2006)
Hsu, B.J., Glass, J.: Style and topic language model adaptation using HMM-LDA. In: Proc. of EMNLP (2006)
Akita, Y., Nemoto, Y., Kawahara, T.: PLSA-based topic detection in meetings for adaptation of lexicon and language model. In: Proc. of Interspeech (2007)
Teh, Y.W.: Dirichlet processes. Encyclopedia of Machine Learning (Submitted, 2007)
Teh, Y.W., Kurihara, K., Welling, M.: Collapsed variational inference for HDP. Advances in Neural Information Processing Systems 20 (2008)
Ferguson, T.S.: A Bayesian analysis of some nonparametric problems. Annals of Statistics 1(2), 209–230 (1973)
Sethuraman, J.: A constructive definition of Dirichlet priors. Statistica Sinica 4, 639–650 (1994)
Kneser, R., Peters, J., Klakow, D.: Language model adaptation using dynamic marginals. In: Proc. of Eurospeech, Rhodes (1997)
Hain, T., et al.: The AMI system for the transcription of speech in meetings. In: Proc. of ICASSP 2007 (2007)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Huang, S., Renals, S. (2008). Modeling Topic and Role Information in Meetings Using the Hierarchical Dirichlet Process. In: Popescu-Belis, A., Stiefelhagen, R. (eds) Machine Learning for Multimodal Interaction. MLMI 2008. Lecture Notes in Computer Science, vol 5237. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-85853-9_20
Download citation
DOI: https://doi.org/10.1007/978-3-540-85853-9_20
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-85852-2
Online ISBN: 978-3-540-85853-9
eBook Packages: Computer ScienceComputer Science (R0)