skip to main content
10.1145/130385.130424acmconferencesArticle/Chapter ViewAbstractPublication PagescoltConference Proceedingsconference-collections
Article
Free Access

Toward efficient agnostic learning

Published:01 July 1992Publication History

ABSTRACT

In this paper we initiate an investigation of generalizations of the Probably Approximately Correct (PAC) learning model that attempt to significantly weaken the target function assumptions. The ultimate goal in this direction is informally termed agnostic learning, in which we make virtually no assumptions on the target function. The name derives from the fact that as designers of learning algorithms, we give up the belief that Nature (as represented by the target function) has a simple or succinct explanation.

We give a number of both positive and negative results that provide an initial outline of the possibilities for agnostic learning. Our results include hardness results for the most obvious generalization of the PAC model to an agnostic setting, an efficient and general agnostic learning method based on dynamic programming, relationships between loss functions for agnostic learning, and an algorithm for learning in a model for problems involving hidden variables.

References

  1. 1.David Aldous and Umesh Vazirani. A Markovian extension of Valiant's learning model. In 31st Annual Symposium on Foundations of Computer Science, pages 392-404, October 1990.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. 2.Anselm Blumer, Andrzej Ehrenfeucht, David Haussler, and Manfred K. Warmuth. Learnability and the Vapnik-Chervonenkis dimension. Journal of the Association for Computing Machinery, 36(4):929-965, October 1989. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. 3.Richard O. Duda and Peter E. Hart. Pattern Classification and Scene Analysis. Wiley, 1973.Google ScholarGoogle Scholar
  4. 4.R. M. Dudley. Central limit theorems for empirical measures. The Annals of Probability, 6(6):899-929, 1978.Google ScholarGoogle ScholarCross RefCross Ref
  5. 5.M. Garey and D. Johnson. Computers and Intractability: A Guide to the Theory o/NP-Completeness. W. H. Freeman, San Francisco, 1979. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. 6.David Haussler. Decision theoretic generalizations of the PAC model for neural net and other learning applications. Information and Computation, To appear. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. 7.David P. Helmbold and Philip M. Long. Tracking drifting concepts using random examples. In Proceedings o/the Fourth Annual Workshop on Computational Learning Theory, pages 13-23, August 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. 8.Michael Kearns and Ming Li. Learning in the presence of malicious errors. In Proceedings o/ the Twentieth Annual ACM Symposium on Theory of Computing, pages 267-280, May 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. 9.Michael Kearns, Ming Li, Leonard Pitt, and Leslie Valiant. On the learnability of Boolean formulae. In Proceedings of the Nineteenth Annual A CM Symposium on Theory of Computing, pages 285-295, May 1987. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. 10.Michael Kearns and Leslie G. Valiant. Cryptographic limitations on learning Boolean formulae and finite automata. In Proceedings of the Twenty First Annual ACM Symposium on Theory o/Computing, pages 433-444, May 1989. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. 11.Michael J. Kearns and Robert E. Schapire. Efficient distribution-free learning of probabilistic concepts. In 31st Annual Symposium on Foundations of Computer Science, pages 382-391, October 1990. To appear, Journal of Computer and System Sciences. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. 12.Nathan Linial, Yishay Mansour, and Noam Nisan. Constant depth circuits, Fourier transform, and learnability. In 30th Annual Symposium on Foundations of Computer Science, pages 574-579, October 1989.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. 13.Leonard Pitt and Leslie G. Valiant. Computational limitations on learning from examples. Journal of the Association for Computing Machinery, 35(4):965-984, October 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. 14.David Pollard. Convergence of Stochastic Processes. Springer-Verlag, 1984.Google ScholarGoogle ScholarCross RefCross Ref
  15. 15.Robert E. Schapire. The strength of weak learnability. Machine Learning, 5(2):197-227, 1990. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. 16.L. G. Valiant. A theory of the learnable. Communications of the ACM, 27(11):1134-1142, November 1984. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. 17.L. G. Valiant. Learning disjunctions of conjunctions. In Proceedings of the 9th International Joint Conference on Artificial Intelligence, pages 560- 566, August 1985.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. 18.V. N. Vapnik. Estimation of Dependences Based on Empirical Data. Springer-Verlag, 1982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. 19.Halbert White. Learning in artificial neural networks: A statistical perspective. Neural Computation, 1(4):425-464, 1989.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. 20.Kenji Yamanishi. A learning criterion for stochastic rules. In Proceedings of the Third Annual Workshop on Computational Learning Theory, August 1990. To appear, Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Toward efficient agnostic learning

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          COLT '92: Proceedings of the fifth annual workshop on Computational learning theory
          July 1992
          452 pages
          ISBN:089791497X
          DOI:10.1145/130385

          Copyright © 1992 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 1 July 1992

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • Article

          Acceptance Rates

          Overall Acceptance Rate35of71submissions,49%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader