Risk Minimization in the Presence of Label Noise

Authors

  • Wei Gao Nanjing University and Collaborative Innovation Center of Novel Software Technology and Industrialization
  • Lu Wang Nanjing University and Collaborative Innovation Center of Novel Software Technology and Industrialization
  • Yu-Feng li Nanjing University and Collaborative Innovation Center of Novel Software Technology and Industrialization
  • Zhi-Hua Zhou Nanjing University and Collaborative Innovation Center of Novel Software Technology and Industrialization

DOI:

https://doi.org/10.1609/aaai.v30i1.10293

Abstract

Matrix concentration inequalities have attracted much attention in diverse applications such as linear algebra, statistical estimation, combinatorial optimization, etc. In this paper, we present new Bernstein concentration inequalities depending only on the first moments of random matrices, whereas previous Bernstein inequalities are heavily relevant to the first and second moments. Based on those results, we analyze the empirical risk minimization in the presence of label noise. We find that many popular losses used in risk minimization can be decomposed into two parts, where the first part won't be affected and only the second part will be affected by noisy labels. We show that the influence of noisy labels on the second part can be reduced by our proposed LICS (Labeled Instance Centroid Smoothing) approach. The effectiveness of the LICS algorithm is justified both theoretically and empirically.

Downloads

Published

2016-02-21

How to Cite

Gao, W., Wang, L., li, Y.-F., & Zhou, Z.-H. (2016). Risk Minimization in the Presence of Label Noise. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.10293

Issue

Section

Technical Papers: Machine Learning Methods