skip to main content
10.1145/1273496.1273621acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
Article

On learning with dissimilarity functions

Published:20 June 2007Publication History

ABSTRACT

We study the problem of learning a classification task in which only a dissimilarity function of the objects is accessible. That is, data are not represented by feature vectors but in terms of their pairwise dissimilarities. We investigate the sufficient conditions for dissimilarity functions to allow building accurate classifiers. Our results have the advantages that they apply to unbounded dissimilarities and are invariant to order-preserving transformations. The theory immediately suggests a learning paradigm: construct an ensemble of decision stumps each depends on a pair of examples, then find a convex combination of them to achieve a large margin. We next develop a practical algorithm called Dissimilarity based Boosting (DBoost) for learning with dissimilarity functions under the theoretical guidance. Experimental results demonstrate that DBoost compares favorably with several existing approaches on a variety of databases and under different conditions.

References

  1. Balcan, M.-F., & Blum, A. (2006). On a theory of learning with similarity functions. International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Balcan, M.-F., Blum, A., & Vempala, S. (2004). On kernels, margins, and low-dimensional mappings. International Workshop on Algorithmic Learning Theory.Google ScholarGoogle ScholarCross RefCross Ref
  3. Balcan, M.-F., Blum, A., & Vempala, S. (2006). Kernels as features: On kernels, margins, and low-dimensional mappings. Machine Learning, 65, 79--94. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Breiman, L., Friedman, J. H., Olshen, R. A., & Stone, C. J. (1984). Classification and regression trees. Wadsworth.Google ScholarGoogle Scholar
  5. Chang, C. C., & Lin, C. J. (2001). A library for support vector machines. Available at http://www.csie.ntu.edu.tw/cjlin/libsvm.Google ScholarGoogle Scholar
  6. Freund, Y., & Schapire, R. E. (1996). Experiments with a new boosting algorithm. International Conference on Machine Learning.Google ScholarGoogle Scholar
  7. Goldfarb, L. (1985). A new approach to pattern recognition. In L. N. Kannal and A. Rosenfeld (Ed.), Progress in Pattern Recognition, 2, 241--402.Google ScholarGoogle Scholar
  8. Graepel, T., Herbrich, R., Bollmann-sdorra, P., & Obermayer, K. (1999). Classification on pairwise proximity data. Advances in Neural Information Processing Systems. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Hinton, G. E., & Revow, M. (1996). Using pairs of data points to define splits for decision trees. Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  10. Jacobs, D. W., Weinshall, D., & Gdalyahu, Y. (2000). Classification with nonmetric distances: Image retrieval and class representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22, 583--560. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Jain, A. K., & Zongker, D. E. (1997). Representation and recognition of handwritten digits using deformable templates. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19, 1386--1391. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Li, J., Chen, G., & Chi, Z. (2002). A fuzzy image metric with application to fractal coding. IEEE Transactions on Image Processing, 11, 636--643. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Maltoni, D., Maio, D., Jain, A. K., & Prabhakar, S. (2003). Handbook of fingerprint recognition. New York: Springer. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Pekalska, E., & Duin, R. P. W. (2002). Dissimilarity representations allow for building good classifiers. Pattern Recognition Letters, 23, 943--956. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Pekalska, E., & Duin, R. P. W. (2005). The dissimilarity representation for pattern recognition. World Scientific. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Pekalska, E., Paclík, P., & Duin, R. P. W. (2002). A generalized kernel approach to dissimilarity-based classification. Journal of Machine Learning Research, 2, 175--211. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Ripley, B. (1996). Pattern recognition and neural networks. Cambridge: Cambridge University Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Schapire, R. E., & Singer, Y. (1999). Improved boosting algorithms using confidence-rated predictions. Machine Learning, 37, 297--336. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Simard, P., Cun, Y. L., & Denker, J. (1993). Efficient pattern recognition using a new transformation distance. Advances in Neural Information Processing Systems. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Zhao, W., Chellappa, R., Phillips, P. J., & Rosenfeld, A. (2003). Face recognition: A literature survey. ACM Computing Surveys, 35, 399--458. Google ScholarGoogle ScholarDigital LibraryDigital Library
  1. On learning with dissimilarity functions

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        ICML '07: Proceedings of the 24th international conference on Machine learning
        June 2007
        1233 pages
        ISBN:9781595937933
        DOI:10.1145/1273496

        Copyright © 2007 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 20 June 2007

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • Article

        Acceptance Rates

        Overall Acceptance Rate140of548submissions,26%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader