skip to main content
research-article

Noise-aware Local Model Training Mechanism for Federated Learning

Published:15 June 2023Publication History
Skip Abstract Section

Abstract

As a new paradigm in training intelligent models, federated learning is widely used to train a global model without requiring local data to be uploaded from end devices. However, there are often mislabeled samples (i.e., noisy samples) in the dataset, which will cause the model update to deviate from the correct direction during the training process, thus reducing the convergence accuracy of the global model. Existing works employ noisy label correction techniques to reduce the impact of noisy samples on model updates by correcting labels; however, such methods necessitate the use of prior knowledge and additional communication costs, which cannot be directly applied to federated learning due to data privacy concerns and limited communication resources. Therefore, this paper proposes a noise-aware local model training method that corrects the noisy labels directly at the end device under the constraints of federated learning. By constructing a label correction model, a joint optimization problem is formally defined for optimizing both the label correction model and the client-side local training model (e.g., classification model). As a solution to this optimization problem, we propose a robustness training algorithm using label correction, along with a cross-validation data sampling algorithm that updates both models simultaneously. It is verified through experiments that the mechanism can effectively improve the model convergence accuracy on noisy datasets in federated learning scenarios.

REFERENCES

  1. [1] Chen Pengfei, Liao Ben Ben, Chen Guangyong, and Zhang Shengyu. 2019. Understanding and utilizing deep neural networks trained with noisy labels. In International Conference on Machine Learning. PMLR, 10621070.Google ScholarGoogle Scholar
  2. [2] Chen Yiqiang, Yang Xiaodong, Qin Xin, Yu Han, Chan Piu, and Shen Zhiqi. 2020. Dealing with label quality disparity in federated learning. Federated Learning: Privacy and Incentive (2020), 108121.Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Chen Ziqi, Jiang Liangxiao, and Li Chaoqun. 2022. Label distribution-based noise correction for multiclass crowdsourcing. International Journal of Intelligent Systems 37, 9 (2022), 57525767.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. [4] Finn Chelsea, Abbeel Pieter, and Levine Sergey. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In International Conference on Machine Learning. PMLR, 11261135.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Frénay Benoît and Verleysen Michel. 2013. Classification in the presence of label noise: A survey. IEEE Transactions on Neural Networks and Learning Systems 25, 5 (2013), 845869.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Han Bo, Yao Jiangchao, Niu Gang, Zhou Mingyuan, Tsang Ivor, Zhang Ya, and Sugiyama Masashi. 2018. Masking: A new perspective of noisy supervision. Advances in Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  7. [7] Han Bo, Yao Quanming, Yu Xingrui, Niu Gang, Xu Miao, Hu Weihua, Tsang Ivor, and Sugiyama Masashi. 2018. Co-teaching: Robust training of deep neural networks with extremely noisy labels. Advances in Neural Information Processing Systems 31 (2018).Google ScholarGoogle Scholar
  8. [8] Hu Peng, Peng Xi, Zhu Hongyuan, Zhen Liangli, and Lin Jie. 2021. Learning cross-modal retrieval with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 54035413.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Jiang Lu, Zhou Zhengyuan, Leung Thomas, Li Li-Jia, and Fei-Fei Li. 2018. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In International Conference on Machine Learning. PMLR, 23042313.Google ScholarGoogle Scholar
  10. [10] Jiang Nan, Gao Li, Duan Fuxian, Wen Jie, Wan Tao, and Chen Honglong. 2022. SAN: Attention-based social aggregation neural networks for recommendation system. International Journal of Intelligent Systems 37, 6 (2022), 33733393.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. [11] Krause Jonathan, Sapp Benjamin, Howard Andrew, Zhou Howard, Toshev Alexander, Duerig Tom, Philbin James, and Fei-Fei Li. 2016. The unreasonable effectiveness of noisy data for fine-grained recognition. In European Conference on Computer Vision. Springer, 301320.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Krizhevsky Alex et al. 2009. Learning multiple layers of features from tiny images. Citeseer (2009).Google ScholarGoogle Scholar
  13. [13] Kwon Soonil. 2021. Optimal feature selection based speech emotion recognition using two-stream deep convolutional neural network. International Journal of Intelligent Systems 36, 9 (2021), 51165135.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Lee Kimin, Yun Sukmin, Lee Kibok, Lee Honglak, Li Bo, and Shin Jinwoo. 2019. Robust inference via generative classifiers for handling noisy labels. In International Conference on Machine Learning. PMLR, 37633772.Google ScholarGoogle Scholar
  15. [15] Lee Kuang-Huei, He Xiaodong, Zhang Lei, and Yang Linjun. 2018. CleanNet: Transfer learning for scalable image classifier training with label noise. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 54475456.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Liu Yang and Guo Hongyi. 2020. Peer loss functions: Learning from noisy labels without knowing noise rates. In International Conference on Machine Learning. PMLR, 62266236.Google ScholarGoogle Scholar
  17. [17] Liu Zhiwei, Yang Liangwei, Fan Ziwei, Peng Hao, and Yu Philip S.. 2022. Federated social recommendation with graph neural network. ACM Trans. Intell. Syst. Technol. 13, 4, Article 55 (Aug. 2022), 24 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Lu Yang. 2019. Artificial intelligence: A survey on evolution, models, applications and future trends. Journal of Management Analytics 6, 1 (2019), 129.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] McMahan Brendan, Moore Eider, Ramage Daniel, Hampson Seth, and Arcas Blaise Aguera y. 2017. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics. PMLR, 12731282.Google ScholarGoogle Scholar
  20. [20] Paolacci Gabriele, Chandler Jesse, and Ipeirotis Panagiotis G.. 2010. Running experiments on Amazon Mechanical Turk. Judgment and Decision Making 5, 5 (2010), 411419.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Patrini Giorgio, Rozza Alessandro, Menon Aditya Krishna, Nock Richard, and Qu Lizhen. 2017. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 19441952.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Reed Scott, Lee Honglak, Anguelov Dragomir, Szegedy Christian, Erhan Dumitru, and Rabinovich Andrew. 2014. Training deep neural networks on noisy labels with bootstrapping. arXiv preprint arXiv:1412.6596 (2014).Google ScholarGoogle Scholar
  23. [23] Shu Jun, Xie Qi, Yi Lixuan, Zhao Qian, Zhou Sanping, Xu Zongben, and Meng Deyu. 2019. Meta-weight-net: Learning an explicit mapping for sample weighting. Advances in Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  24. [24] Song Hwanjun, Kim Minseok, and Lee Jae-Gil. 2019. Selfie: Refurbishing unclean samples for robust deep learning. In International Conference on Machine Learning. PMLR, 59075915.Google ScholarGoogle Scholar
  25. [25] Tu Xiaoguang, Ma Zheng, Zhao Jian, Du Guodong, Xie Mei, and Feng Jiashi. 2020. Learning generalizable and identity-discriminative representations for face anti-spoofing. ACM Trans. Intell. Syst. Technol. 11, 5, Article 60 (Jul. 2020), 19 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Vahdat Arash. 2017. Toward robustness against label noise in training deep discriminative neural networks. Advances in Neural Information Processing Systems 30 (2017).Google ScholarGoogle Scholar
  27. [27] Wang Qizhou, Han Bo, Liu Tongliang, Niu Gang, Yang Jian, and Gong Chen. 2021. Tackling instance-dependent label noise via a universal probabilistic model. arXiv preprint arXiv:2101.05467 (2021).Google ScholarGoogle Scholar
  28. [28] Wei Hongxin, Tao Lue, Xie Renchunzi, and An Bo. 2021. Open-set label noise can improve robustness against inherent label noise. Advances in Neural Information Processing Systems 34 (2021).Google ScholarGoogle Scholar
  29. [29] Xia Xiaobo, Liu Tongliang, Han Bo, Gong Chen, Wang Nannan, Ge Zongyuan, and Chang Yi. 2020. Robust early-learning: Hindering the memorization of noisy labels. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  30. [30] Xia Xiaobo, Liu Tongliang, Wang Nannan, Han Bo, Gong Chen, Niu Gang, and Sugiyama Masashi. 2019. Are anchor points really indispensable in label-noise learning? Advances in Neural Information Processing Systems 32 (2019).Google ScholarGoogle Scholar
  31. [31] Xiao Han, Rasul Kashif, and Vollgraf Roland. 2017. Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017).Google ScholarGoogle Scholar
  32. [32] Xiao Tong, Xia Tian, Yang Yi, Huang Chang, and Wang Xiaogang. 2015. Learning from massive noisy labeled data for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 26912699.Google ScholarGoogle Scholar
  33. [33] Yang Seunghan, Park Hyoungseob, Byun Junyoung, and Kim Changick. 2022. Robust federated learning with noisy labels. IEEE Intelligent Systems 37, 2 (2022), 3543.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Zeng Bixiao, Yang Xiaodong, Chen Yiqiang, Yu Hanchao, and Zhang Yingwei. 2022. CLC: A consensus-based label correction approach in federated learning. ACM Transactions on Intelligent Systems and Technology (TIST) 13, 5 (2022), 123.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. [35] Zhang Zixing, Geiger Jürgen, Pohjalainen Jouni, Mousa Amr El-Desoky, Jin Wenyu, and Schuller Björn. 2018. Deep learning for environmentally robust speech recognition: An overview of recent developments. ACM Trans. Intell. Syst. Technol. 9, 5, Article 49 (Apr. 2018), 28 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Zheng Guoqing, Awadallah Ahmed Hassan, and Dumais Susan. 2021. Meta label correction for noisy label learning. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI’21), Vol. 35. 1105311061.Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] Zhu Qing, Gao Lijian, Song Heping, and Mao Qirong. 2021. Learning to disentangle emotion factors for facial expression recognition in the wild. International Journal of Intelligent Systems 36, 6 (2021), 25112527.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Noise-aware Local Model Training Mechanism for Federated Learning

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Intelligent Systems and Technology
        ACM Transactions on Intelligent Systems and Technology  Volume 14, Issue 4
        August 2023
        481 pages
        ISSN:2157-6904
        EISSN:2157-6912
        DOI:10.1145/3596215
        • Editor:
        • Huan Liu
        Issue’s Table of Contents

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 15 June 2023
        • Online AM: 2 May 2023
        • Accepted: 31 March 2023
        • Revised: 3 March 2023
        • Received: 7 October 2022
        Published in tist Volume 14, Issue 4

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      View Full Text