skip to main content
10.1145/279943.279965acmconferencesArticle/Chapter ViewAbstractPublication PagescoltConference Proceedingsconference-collections
Article
Free Access

Improved lower bounds for learning from noisy examples: an information-theoretic approach

Authors Info & Claims
Published:24 July 1998Publication History
First page image

References

  1. 1.D. Angluin, "Learning regular sets from queries and counterexamples", Information and Computation, vol. 75, n. 2, pp. 87-106, 1987. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. 2.D. Angluin, "Queries and concept learning", Machine Learning, vol. 2, n. 4, pp 319-342, 1988. Google ScholarGoogle ScholarCross RefCross Ref
  3. 3.D. Angluin, M. Krikis, R. Sloan, G. Turfin, "Malicious Omissions and Errors in Answers to Membership Queries", Machine Learning, vol. 28, no. 2/3, pp. 211- 255, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. 4.D. Angluin, P.D. Laird, "Learning from Noisy Examples'', Machine Learning, vol. 2, no. 2, pp. 343-370, 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. 5.B. Apolloni, C. Gentile, "Sample Size Lower Bounds in PAC Learning by Algorithmic Complexity Theory", Theor. Comp. Sc., to appear. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. 6.J.A. Aslam, S.E. Decatur, "On the sample complexity of noise-tolerant learning", Inf. Proc. Lett., vol. 57, pp. 189-195, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. 7.P. Bartlett, "Learning with a slowly changing distribution'', in Proc. of the 5th Workshop on Comput. Learn. Th., 1992, pp. 243-252. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. 8.P. Bartlett, D. Helmbold, manuscript, 1996.Google ScholarGoogle Scholar
  9. 9.R.D. Barve, P.M. Long, "On the complexity of learning from drifting distributions", Information and Computation, vol. 138, no. 2, pp. 170-193, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. 10.G. Benedek, A. Itai, "Leamability by Fixed Distributions'', Theor. Comp. Sc., vol. 86, no. 2, pp. 377-389, 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. 11.A. Blumer, A. Ehrenfeucht, D. Haussler, M. Warmuth, "Leamability and the Vapnik-Chervonenkis Dimension'', J. ofACM, vol. 36, pp. 929-965, 1989. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. 12.N. Cesa-Bianchi, E. Dichterman, P. Fischer, E. Shamir, H.U. Simon, "Sample-efficient Strategies for Learning in the Presence of Noise", eCOLT Tech. Rep. 97- 003, WWW.' http://ecolt.informatik, uni-dortmund, de/. Preliminary versions in 28th STOC, 1996 and 3rd EuroCOLT, 1997Google ScholarGoogle Scholar
  13. 13.T.M. Cover, J.A. Thomas, Elements of information theory. NY: John Wiley & Sons, Inc., 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. 14.R.L. Dobrushin, "General formulation of Shannon's main theorem in information theory", Uspelchi Mat. Nauk, vol. 14, pp. 3-104, 1959; translated in Amer. Math. Soc. Translations, Ser. 2, vol. 33, pp. 323-438, 1963.Google ScholarGoogle Scholar
  15. 15.R.M. Dudley, A Course on Empirical Processes. Lecture Notes in Mathematics, vol. 1097, Spfinger-Vefiag, Berlin/New York, 1984.Google ScholarGoogle Scholar
  16. 16.A. Ehrenfeucht, D. Haussler, M. Kearns, L. Valiant, "A General Lower Bound on the Number of Examples Needed for Learning", Information and Computation, vol. 82, no. 3, pp. 247-261, 1989. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. 17.B. Eisenberg, R.L. Rivest, "On the Sample Complexity of Pac-Leaming using Random and Chosen Examples'', in Proc. of the 3th Workshop on Comput. Learn. Th., 1990, pp. 154~162. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. 18.C. Gentile, "A note on sample size lower bounds for PAC-leaming", manuscript, 1997.Google ScholarGoogle Scholar
  19. 19.D. Haussler, A. Barron, "How well do Bayes methods work for on-line prediction of {-1 ,+ 1 } values'?.", in Proc. of the 3rd NEC Symposium on Computation and Cognition, 1992, pp. 74-100.Google ScholarGoogle Scholar
  20. 20.D. Haussler, M. Keams, R. Schapire, "Bounds on the Sample Complexity of Bayesian Learning Using Information Theory and the VC Dimension", Machine Learning, vol. 14, pp. 84-114, 1994. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. 21.D. Haussler, N. Littlestone, and M. K. Warmuth. Predicting {0, 1 } functions on randomly drawn points. Information and Computation, 115(2):284-293, 1994. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. 22.D. Haussler, M. Opper, "Mutual Information, Metric Entropy, and Cumulative Relative Entropy Risk", Annals of Statistics, 1997. To appear.Google ScholarGoogle Scholar
  23. 23.D. Helmbold, N. Littlestone, P. Long, "Apple Tasting'', manuscript 1997. An extended abstract appeared in Proc. of the 33rd Symposium on the Foundations of Comp. Sci., 1992, pp.493-502.Google ScholarGoogle Scholar
  24. 24.S. Ihara, Information theory for continuous systems. River Edge, NJ: World Scientific, 1993.Google ScholarGoogle Scholar
  25. 25.M. Keams, M. Li, "Learning in the presence of malicious errors", SIAM J. Comput, vol. 22, pp. 807-837, 1993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. 26.A.N. Kolmogorov, V.M. Tihomirov, "e-entropy and ecapacity of sets in functional spaces", Amer. Math. Soc. Translations (Ser. 2), vol. 17, pp. 277-364, 1961.Google ScholarGoogle Scholar
  27. 27.P. Laird, Learning from Good and Bad Data. Kluwer International Series in Engineering and Computer Science, Kluwer Academic Publishers, Boston, MA, 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. 28.W. Maass, G. Turfin, "On the complexity of learning from counterexamples and membership queries", in Proc. of the 31th Symposium on the Foundations of Comp. Sci., 1990, pp. 203-210.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. 29.K. Sakakibara, "On learning from queries and counterexamples in the presence of noise", Inf, Proc. Lett., vol. 37, no. 5, pp. 279-284, 1991. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. 30.H.U. Simon, "General Bounds on the Number of Examples Needed for Learning Probabilistic Concepts", Journal of Comp. System Sci., vol. 52, no. 2, pp. 239- 254, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. 31.R. Sloan, "Four types of noise in data for PAC learning'', Inf. Proc. Lett., vol. 54, pp. 157-162, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. 32.G. Shackelford, D. Volper, "Leaming k-DNF with noise in the attributes", in Proc. of the 1988 Workshop on Comput. Learn. Th., 1988, pp. 97-103. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. 33.G. Turhn, "Lower bounds for PAC learning with queries", in Proc. of the 6th Workshop on Comput. Learn. Th., 1993, pp. 384-391. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. 34.L. Valiant, "A theory of the learnable", Communication ofACM, vol. 27, no. ll,pp. 1134-1142, 1984. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. 35.V.N. Vapnik, Estimation of dependencies based on empirical data. NY: Springer Verlag, 1982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. 36.V.N. Vapnik, The Nature of Statistical Learning Theory. NY: Springer Verlag, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. 37.B. Yu, "Lower Bounds on Expected Redundancy for Nonparametric Classes", IEEE Trans. on Inf. Th., vol. 42, no. 1, pp. 272-275, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Improved lower bounds for learning from noisy examples: an information-theoretic approach

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            COLT' 98: Proceedings of the eleventh annual conference on Computational learning theory
            July 1998
            304 pages
            ISBN:1581130570
            DOI:10.1145/279943

            Copyright © 1998 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 24 July 1998

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • Article

            Acceptance Rates

            Overall Acceptance Rate35of71submissions,49%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader