Abstract
Federated learning, a new distributed learning paradigm, has the advantage of sharing model information without revealing data privacy. However, considering the selfishness of organizations, they will not participate in federated learning without compensation. To address this problem, in this paper, we design a feature importance-aware vertical federated learning incentive mechanism. We first synthesize a small amount of data locally using the interpolation method at the organization and send it to the coordinator for evaluating the contribution of each feature to the learning task. Then, the coordinator calculates the importance value of each feature in the dataset for the current task using the Shapley value method according to the synthetic data. Next, we formulate the process of organization participation in the federation as a feature importance maximization problem based on reverse auction which is a knapsack auction problem. Finally, we design an approximate algorithm to solve the proposed optimization problem and the solution of the approximation algorithm is shown to be \(\frac{1}{2}\)-approximate to the optimal solution. Furthermore, we prove that the proposed mechanism is truthfulness, individual rationality, and computational efficiency. The superiority of our proposed mechanism is verified through experiments on real-world datasets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Cai, S., Zhao, Y., Liu, Z., Qiu, C., Wang, X., Hu, Q.: Mgfl: multi-granularity federated learning in edge computing systems. In: Algorithms and Architectures for Parallel Processing (ICA3PP), pp. 549–563. Springer International Publishing, Cham (2022). https://doi.org/10.1007/978-3-030-95384-3_34
Chen, F., Chen, N., Mao, H., Hu, H.: Assessing four neural networks on handwritten digit recognition dataset (MNIST). arXiv preprint arXiv:1811.08278 (2018). https://doi.org/10.48550/ARXIV.1811.08278
Covert, I., Lundberg, S.M., Lee, S.I.: Understanding global feature contributions with additive importance measures. Adv. Neural Inform. Process. Syst. (NeurIPS) 33, 17212–17223 (2020)
Deng, Y., et al.: Fair: Quality-aware federated learning with precise user incentive and model aggregation. In: IEEE Conference on Computer Communications (INFOCOM), pp. 1–10. IEEE (2021). https://doi.org/10.1109/INFOCOM42981.2021.9488743
Fu, F., et al.: VF2Boost: very fast vertical federated gradient boosting for cross-enterprise learning. In: Proceedings of the 2021 International Conference on Management of Data, pp. 563–576 (2021)
Gul, F.: Bargaining foundations of shapley value. Econometrica: J. Econom. Society 81–95 (1989)
Horowitz, E., Sahni, S.: Computing partitions with applications to the knapsack problem. J. ACM 21(2), 277–292 (1974)
Kairouz, P., et al.: Advances and open problems in federated learning. Found. Trends® in Mach. Learn. 14(1–2), 1–210 (2021)
Kang, J., Xiong, Z., Niyato, D., Xie, S., Zhang, J.: Incentive mechanism for reliable federated learning: a joint optimization approach to combining reputation and contract theory. IEEE Internet Things J. 6(6), 10700–10714 (2019). https://doi.org/10.1109/JIOT.2019.2940820
Kolesnikov, V., Kumaresan, R., Rosulek, M., Trieu, N.: Efficient batched oblivious prf with applications to private set intersection. In: Proceedings of the 2016 ACM Conference on Computer and Communications Security (CCS), pp. 818–829 (2016)
Konečnỳ, J., McMahan, H.B., Yu, F.X., Richtárik, P., Suresh, A.T., Bacon, D.: Federated learning: Strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492 (2016)
Le, T.H.T., et al.: An incentive mechanism for federated learning in wireless cellular networks: an auction approach. IEEE Trans. Wireless Commun. 20(8), 4874–4887 (2021). https://doi.org/10.1109/TWC.2021.3062708
Li, X., Chen, X., Wang, S., Ding, Y., Li, K.: Multi-initial-center federated learning with data distribution similarity-aware constraint. In: Algorithms and Architectures for Parallel Processing (ICA3PP), pp. 752–772. Springer Nature Switzerland, Cham (2023). https://doi.org/10.1007/978-3-031-22677-9_41
Lim, W.Y.B., et al.: Hierarchical incentive mechanism design for federated machine learning in mobile networks. IEEE Internet Things J. 7(10), 9575–9588 (2020). https://doi.org/10.1109/JIOT.2020.2985694
Liu, T., Di, B., An, P., Song, L.: Privacy-preserving incentive mechanism design for federated cloud-edge learning. IEEE Trans. Network Sci. Eng. 8(3), 2588–2600 (2021). https://doi.org/10.1109/TNSE.2021.3100096
Liu, Y., Wu, G., Zhang, W., Li, J.: Federated learning-based intrusion detection on non-iid data. In: Algorithms and Architectures for Parallel Processing (ICA3PP), pp. 313–329. Springer Nature Switzerland, Cham (2023). https://doi.org/10.1007/978-3-031-22677-9_17
Lu, J., Pan, B., Seid, A.M., Li, B., Hu, G., Wan, S.: Truthful incentive mechanism design via internalizing externalities and lp relaxation for vertical federated learning. IEEE Transactions on Computational Social Systems, pp. 1–15 (2022). https://doi.org/10.1109/TCSS.2022.3227270
Lu, S., Wang, Y., Wu, Y.: Novel high-precision simulation technology for high-dynamics signal simulators based on piecewise hermite cubic interpolation. IEEE Trans. Aerosp. Electron. Syst. 54(5), 2304–2317 (2018). https://doi.org/10.1109/TAES.2018.2814278
Luo, X., Wu, Y., Xiao, X., Ooi, B.C.: Feature inference attack on model predictions in vertical federated learning. In: 2021 IEEE 37th International Conference on Data Engineering (ICDE), pp. 181–192. IEEE (2021)
McMahan, B., Moore, E., Ramage, D., Hampson, S., y Arcas, B.A.: Communication-efficient learning of deep networks from decentralized data. In: Artificial intelligence and statistics (AISTATS), pp. 1273–1282. PMLR (2017)
Myerson, R.B.: Optimal auction design. Math. Oper. Res. 6(1), 58–73 (1981)
Ng, J.S., Lim, W.Y.B., Xiong, Z., Cao, X., Niyato, D., Leung, C., Kim, D.I.: A hierarchical incentive design toward motivating participation in coded federated learning. IEEE J. Sel. Areas Commun. 40(1), 359–375 (2022). https://doi.org/10.1109/JSAC.2021.3126057
Shin, M., Hwang, C., Kim, J., Park, J., Bennis, M., Kim, S.L.: XOR Mixup: privacy-preserving data augmentation for one-shot federated learning. In: International Workshop on Federated Learning for User Privacy and Data Confidentiality in Conjunction with ICML (FL-ICML) (2020)
Štrumbelj, E., Kononenko, I.: Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst. 41(3), 647–665 (2014)
Tang, M., Wong, V.W.: An incentive mechanism for cross-silo federated learning: a public goods perspective. In: IEEE Conference on Computer Communications (INFOCOM), pp. 1–10. IEEE (2021). https://doi.org/10.1109/INFOCOM42981.2021.9488705
Tseng, C.C., Lee, S.L.: Design of fractional delay filter using hermite interpolation method. IEEE Trans. Circuits Syst. I Regul. Pap. 59(7), 1458–1471 (2012). https://doi.org/10.1109/TCSI.2011.2177136
Wang, Y., Lin, L., Chen, J.: Communication-efficient adaptive federated learning. In: Proceedings of the 39th International Conference on Machine Learning (ICML). Proceedings of Machine Learning Research, vol. 162, pp. 22802–22838 (17–23 Jul 2022)
Weng, J., Weng, J., Huang, H., Cai, C., Wang, C.: Fedserving: a federated prediction serving framework based on incentive mechanism. In: IEEE Conference on Computer Communications (INFOCOM), pp. 1–10. IEEE (2021)
Wu, Y., Cai, S., Xiao, X., Chen, G., Ooi, B.C.: Privacy preserving vertical federated learning for tree-based models. Proceedings of the VLDB Endowment 13(11) (2020)
Xiao, H., Rasul, K., Vollgraf, R.: Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv: 1708.07747 (2017). https://doi.org/10.48550/arXiv.1708.07747
Yang, Q., Liu, Y., Chen, T., Tong, Y.: Federated machine learning: concept and applications. ACM Trans. Intell. Syst. Technol. 10(2), 1–19 (2019)
Yeh, I.C., Lien, C.: The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients. Expert Syst. Appl. 36(2), 2473–2480 (2009)
Yoon, T., Shin, S., Hwang, S.J., Yang, E.: Fedmix: approximation of mixup under mean augmented federated learning. In: International Conference on Learning Representations (ICLR) (2021)
Yu, Y., Chen, D., Tang, X., Song, T., Hong, C.S., Han, Z.: Incentive framework for cross-device federated learning and analytics with multiple tasks based on a multi-leader-follower game. IEEE Trans. Netw. Sci. Eng. 9(5), 3749–3761 (2022). https://doi.org/10.1109/TNSE.2022.3190377
Zhan, Y., Li, P., Qu, Z., Zeng, D., Guo, S.: A learning-based incentive mechanism for federated learning. IEEE Internet Things J. 7(7), 6360–6368 (2020). https://doi.org/10.1109/JIOT.2020.2967772
Zhang, J., Wu, Y., Pan, R.: Incentive mechanism for horizontal federated learning based on reputation and reverse auction. In: Proceedings of the Web Conference 2021, pp. 947–956 (2021)
Zhang, Q., et al.: AsySQN: faster vertical federated learning algorithms with better computation resource utilization. In: Proceedings of the 27th ACM Conference on Knowledge Discovery & Data Mining (SIGKDD), pp. 3917–3927 (2021)
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant U1911201, U2001209, the Natural Science Foundation of Guangdong under Grant 2021A1515011369, and the Science and Technology Program of Guangzhou under Grant 2023A04J2029.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Tan, L., Yang, Y., Hu, M., Zhou, Y., Wu, D. (2024). FRAIM: A Feature Importance-Aware Incentive Mechanism for Vertical Federated Learning. In: Tari, Z., Li, K., Wu, H. (eds) Algorithms and Architectures for Parallel Processing. ICA3PP 2023. Lecture Notes in Computer Science, vol 14491. Springer, Singapore. https://doi.org/10.1007/978-981-97-0808-6_8
Download citation
DOI: https://doi.org/10.1007/978-981-97-0808-6_8
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-0807-9
Online ISBN: 978-981-97-0808-6
eBook Packages: Computer ScienceComputer Science (R0)