Skip to main content

Adjusting Mixture Weights of Gaussian Mixture Model via Regularized Probabilistic Latent Semantic Analysis

  • Conference paper
Advances in Knowledge Discovery and Data Mining (PAKDD 2005)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 3518))

Included in the following conference series:

Abstract

Mixture models, such as Gaussian Mixture Model, have been widely used in many applications for modeling data. Gaussian mixture model (GMM) assumes that data points are generated from a set of Gaussian models with the same set of mixture weights. A natural extension of GMM is the probabilistic latent semantic analysis (PLSA) model, which assigns different mixture weights for each data point. Thus, PLSA is more flexible than the GMM method. However, as a tradeoff, PLSA usually suffers from the overfitting problem. In this paper, we propose a regularized probabilistic latent semantic analysis model (RPLSA), which can properly adjust the amount of model flexibility so that not only the training data can be fit well but also the model is robust to avoid the overfitting problem. We conduct empirical study for the application of speaker identification to show the effectiveness of the new model. The experiment results on the NIST speaker recognition dataset indicate that the RPLSA model outperforms both the GMM and PLSA models substantially. The principle of RPLSA of appropriately adjusting model flexibility can be naturally extended to other applications and other types of mixture models.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Bellegarda, J.R., Nahamoo, D.: Tied mixture continuous parameter modeling for speech recognition. IEEE Trans. Acoustic., Speech, Signal Processing 38 (1990)

    Google Scholar 

  2. Blei, D., Ng, A., Jordan., M.: Latent Dirichlet allocation. Journal of Machine Learning Research, 993–1022 (2003)

    Google Scholar 

  3. Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society B39, 1–38 (1977)

    MathSciNet  Google Scholar 

  4. Furui, S.: An overview of speaker recognition technology. In: Lee, C., Soong, F., Paliwal, K. (eds.) Automatic speech and speaker Recognition. Kluwer Academic Press, Dordrecht (1996)

    Google Scholar 

  5. Hofmann, T.: Probabilistic latent semantic analysis. In: Proceedings of the 15th Conference on Uncertainty in AI (UAI) (1999)

    Google Scholar 

  6. Hofmann, T.: Probabilistic Latent Semantic Indexing. In: Proceedings of the 22nd International Conference on Research and Development in Information Retrieval (SIGIR) (1999)

    Google Scholar 

  7. Permuter, H., Francos, J., Jermyn, I.H.: Gaussian mixture models of texture and colour for image database retrieval. In: Proc. ICASSP, vol. 1, pp. 25–88 (2003)

    Google Scholar 

  8. Povinelli, R.J., Johnson, M.T., Lindgren, A.C., Ye, J.J.: Time Series Classification Using Gaussian Mixture Models of Reconstructed Phase Spaces. IEEE Transactions on Knowledge and Data Engineering 16(6) (2004)

    Google Scholar 

  9. Reynolds, D.A.: Speaker identification and verification using Gaussian mixture speaker models. Speech Communication (17), 91–108 (1998)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Si, L., Jin, R. (2005). Adjusting Mixture Weights of Gaussian Mixture Model via Regularized Probabilistic Latent Semantic Analysis. In: Ho, T.B., Cheung, D., Liu, H. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2005. Lecture Notes in Computer Science(), vol 3518. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11430919_72

Download citation

  • DOI: https://doi.org/10.1007/11430919_72

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-26076-9

  • Online ISBN: 978-3-540-31935-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics