Skip to main content
Log in

Adjusted support vector machines based on a new loss function

  • Published:
Annals of Operations Research Aims and scope Submit manuscript

Abstract

Support vector machine (SVM) has attracted considerable attentions recently due to its successful applications in various domains. However, by maximizing the margin of separation between the two classes in a binary classification problem, the SVM solutions often suffer two serious drawbacks. First, SVM separating hyperplane is usually very sensitive to training samples since it strongly depends on support vectors which are only a few points located on the wrong side of the corresponding margin boundaries. Second, the separating hyperplane is equidistant to the two classes which are considered equally important when optimizing the separating hyperplane location regardless the number of training data and their dispersions in each class. In this paper, we propose a new SVM solution, adjusted support vector machine (ASVM), based on a new loss function to adjust the SVM solution taking into account the sample sizes and dispersions of the two classes. Numerical experiments show that the ASVM outperforms conventional SVM, especially when the two classes have large differences in sample size and dispersion.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Batista, G., Prati, R. C., & Monard, M. C. (2004). A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explorations, 6(1), 20–29.

    Article  Google Scholar 

  • Boser, B. E., Guyon, I. M., & Vapnik, V. N. (1992). A training algorithm for optimal margin classifiers. In 5th annual ACM workshop on COLT. Pittsburgh, PA.

  • Burges, C. J. C. (1998). A tutorial on support vector machines for pattern recognition. Knowledge Discovery and Data Mining, 2(2), 121–167.

    Article  Google Scholar 

  • Chawla, N. V., Japkowicz, N., & Kolcz, A. (2004). Editorial: special issue on learning from imbalanced data sets. SIGKDD Explorations, 6(1), 1–6.

    Article  Google Scholar 

  • Cortes, C., & Vapnik, V. (1995). Support vector networks. Machine Learning, 20, 273–297.

    Google Scholar 

  • Eitrich, T., & Lang, B. (2006). Efficient optimization of support vector machine learning parameters for unbalanced datasets. Journal of Computational and Applied Mathematics, 196, 425–436.

    Article  Google Scholar 

  • Elkan, C. (2001). The foundations of cost-sensitive learning. In Proceedings of the seventeenth international conference on machine learning (pp. 239–246).

  • Guyon, I., Boser, B., & Vapnik, V. (1993). Automatic capacity tuning of very large VC-dimension classifiers. Advances in Neural Information Processing Systems, 5, 147–155.

    Google Scholar 

  • Guyon, I., Weston, J., Barnhill, S., & Vapnik, V. (2002). Gene selection for cancer classification using support vector machines. Machine Learning, 46, 389–422.

    Article  Google Scholar 

  • Hall, P., Marron, J., & Neeman, A. (2005). Geometric representation of high dimension, low sample size data. Journal of the Royal Statistical Society Series B, 67, 427–444.

    Article  Google Scholar 

  • Hastie, T., Tibshirani, R., & Friedman, J. (2001). The elements of statistical learning: data mining, inference, and prediction. Berlin: Springer.

    Google Scholar 

  • Lee, Y., & Mangasarian, O. L. (2001). SSVM: smooth support vector machine for classification. Computational Optimization and Applications, 20(1), 5–22.

    Article  Google Scholar 

  • Lin, Y. (2002a). Support vector machines and the Bayes rule in classification. Data Mining and Knowledge Discovery, 6(3), 259–275.

    Article  Google Scholar 

  • Lin, Y. (2002b). A note on margin-based loss function in classification (Technical Report). Department of Statistics, University of Wisconsin.

  • Lin, Y., Lee, Y., & Wahba, G. (2002). Support vector machines for classification in nonstandard situations. Machine Learning, 46, 191–202.

    Article  Google Scholar 

  • Luenberger, D. G. (1984). Linear and nonlinear programming (2nd ed.). Reading: Addison-Wesley.

    Google Scholar 

  • Osuna, E., Freund, R., & Girosi, F. (1997). Support vector machines: training and applications (AI Memo 1602). MIT.

  • Shen, X., Tseng, G. C., Zhang, X., & Wong, W. H. (2003). On ψ-learning. Journal of American Statistical Association, 98, 724–734.

    Article  Google Scholar 

  • Statnikov, A., Aliferis, C., Tsamardinos, I., Hardin, D., & Levy, S. (2005). A comprehensive evaluation of multicategory classification methods for microarray gene expression cancer diagnosis. Bioinformatics, 21(5), 631–643.

    Article  Google Scholar 

  • Vapnik, V. (1998). Statistical learning theory. New York: Wiley.

    Google Scholar 

  • Weiss, G. M. (2004). Mining with rarity: a unifying framework. SIGKDD Explorations, 6(1), 7–19.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wei Jiang.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Wang, S., Jiang, W. & Tsui, KL. Adjusted support vector machines based on a new loss function. Ann Oper Res 174, 83–101 (2010). https://doi.org/10.1007/s10479-008-0495-y

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10479-008-0495-y

Keywords

Navigation