skip to main content
10.1145/1143844.1143887acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
Article

The rate adapting poisson model for information retrieval and object recognition

Published:25 June 2006Publication History

ABSTRACT

Probabilistic modelling of text data in the bag-of-words representation has been dominated by directed graphical models such as pLSI, LDA, NMF, and discrete PCA. Recently, state of the art performance on visual object recognition has also been reported using variants of these models. We introduce an alternative undirected graphical model suitable for modelling count data. This "Rate Adapting Poisson" (RAP) model is shown to generate superior dimensionally reduced representations for subsequent retrieval or classification. Models are trained using contrastive divergence while inference of latent topical representations is efficiently achieved through a simple matrix multiplication.

References

  1. Airoldi, E., Cohen, W., & Fienberg, S. (2005). Bayesian methods for frequent terms in text. Proc. of the CSNA & INTERFACE Annual Meetings.]]Google ScholarGoogle Scholar
  2. Blei, D. M., & Jordan, M. I. (2004). Variational inference for dirichlet process mixtures. Bayesian Analysis, 1, 121--144.]]Google ScholarGoogle ScholarCross RefCross Ref
  3. Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent Dirichlet allocation. Journal of Machine Learning Research, 3, 993--1022.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Buntine, W. (Ed.). (2002). Variational extensions to em and multinomial pca, vol. 2430 of Lecture Notes in Computer Science. Helsinki, Finland: Springer.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Buntine, W., & Jakulin, A. (2004). Applying discrete pca in data analysis. Proceedings of the 20th conference on Uncertainty in artificial intelligence (pp. 59--66). Banff, Canada.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Carreira-Perpinan, M., & Hinton, G. (2005). On contrastive divergence learning. Tenth International Workshop on Artificial Intelligence and Statistics. Barbados.]]Google ScholarGoogle Scholar
  7. Casella, G., & Robert, C. (1996). Rao-blackwellisation of sampling schemes. Biometrika, 83(1), 81--94.]]Google ScholarGoogle ScholarCross RefCross Ref
  8. Deerwester, S., Dumais, S., Landauer, T., Furnas, G., & Harshman, R. (1990). Indexing by latent semantic analysis. Journal of the American Society of Information Science, 41, 391--407.]]Google ScholarGoogle ScholarCross RefCross Ref
  9. Fergus, R., Fei-Fei, L., Perona, P., & Zisserman, A. (2005). Learning object categories from google's image search. Proceedings of the International Conference on Computer Vision.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Girolami, M., & Kaban, A. (2003). On an equivalence between PLSI and LDA. Proceedings of SIGIR 2003.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Griffiths, T., & Steyvers, M. (2002). A probabilistic approach to semantic representation. Proceedings of the 24th Annual Conference of the Cognitive Science Society.]]Google ScholarGoogle Scholar
  12. Hinton, G. (2002). Training products of experts by minimizing contrastive divergence. Neural Computation, 14, 1771--1800.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Hinton, G., Osindero, S., & Teh, Y. (2006). A fast learning algorithm for deep belief networks. Neural Computation. to appear.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Hofmann, T. (1999). Probabilistic latent semantic analysis. Proc. of Uncertainty in Artificial Intelligence, UAI '99. Stockholm.]]Google ScholarGoogle Scholar
  15. Kadir, T., & Brady, M. (2001). Saliency, scale and image description. Int. J. Comput. Vision, 45, 83--105.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Lee, D., & Seung, H. (1999). Learning the parts of objects by non-negative matrix factorization. Nature, 401, 788--791.]]Google ScholarGoogle ScholarCross RefCross Ref
  17. Li, F., & Perona, P. (2005). A bayesian hierarchical model for learning natural scene categories. Proceedings of the Conference on Computer Vision and Pattern Recognition.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. McCallum, A. (1996). Bow: A toolkit for statistical language modeling, text retrieval, classification and clustering. http://www.cs.cmu.edu/~mccallum/bow.]]Google ScholarGoogle Scholar
  19. Minka, T., & Lafferty, J. (2002). Expectation-propogation for the generative aspect model. Proc. of the 18th Annual Conference on Uncertainty in Artificial Intelligence (pp. 352--359).]]Google ScholarGoogle Scholar
  20. Olshausen, A., & Field, D. (1997). Sparse coding with over-complete basis set: A strategy employed by v1? Vision Research, 37, 3311--3325.]]Google ScholarGoogle ScholarCross RefCross Ref
  21. Roweis, S. (1997). Em algorithms for pca and spca. Neural Information Processing Systems (pp. 626--632).]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Welling, M., & Hinton, G. (2001). A new learning algorithm for mean field Boltzmann machines. Proc. of the Int'l Conf. on Artificial Neural Networks. Madrid, Spain.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Welling, M., Hinton, G., & Osindero, S. (2002). Learning sparse topographic representations with products of student-t distributions. Neural Information Processing Systems.]]Google ScholarGoogle Scholar
  24. Welling, M., Rosen-Zvi, M., & Hinton, G. (2004). Exponential family harmoniums with an application to information retrieval. Neural Information Processing Systems.]]Google ScholarGoogle Scholar
  25. Xing, E., Yan, R., & Hauptman, A. (2005). Mining associated text and images with dual-wing harmoniums. Proc. of the Conf. on Uncertainty in Artificial Intelligence.]]Google ScholarGoogle Scholar

Index Terms

  1. The rate adapting poisson model for information retrieval and object recognition

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in
            • Published in

              cover image ACM Other conferences
              ICML '06: Proceedings of the 23rd international conference on Machine learning
              June 2006
              1154 pages
              ISBN:1595933832
              DOI:10.1145/1143844

              Copyright © 2006 ACM

              Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

              Publisher

              Association for Computing Machinery

              New York, NY, United States

              Publication History

              • Published: 25 June 2006

              Permissions

              Request permissions about this article.

              Request Permissions

              Check for updates

              Qualifiers

              • Article

              Acceptance Rates

              ICML '06 Paper Acceptance Rate140of548submissions,26%Overall Acceptance Rate140of548submissions,26%

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader