Abstract
The classification analysis of imbalanced data remains a challenging task since the base classifier usually focuses on the majority class and ignores the minority class. This paper proposes a reliability-based imbalanced data classification approach (RIC) with Dempster-Shafer theory to address this issue. First, based on the minority class, multiple under-sampling for the majority one are implemented to obtain the corresponding balanced training sets, which results in multiple globally optimal trained classifiers. Then, the neighbors are employed to evaluate the local reliability of different classifiers in classifying each test sample, making each global optimal classifier focus on the sample locally. Finally, the revised classification results based on various local reliability are fused by the Dempster-Shafer (DS) fusion rule. Doing so, the test sample can be directly classified if more than one classifier has high local reliability. Otherwise, the neighbors belonging to different classes are employed again as the additional knowledge to revise the fusion result. The effectiveness has been verified on synthetic and several real imbalanced datasets by comparison with other related approaches.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In applications, users can employ other appropriate under-sampling approaches according to the request of practice.
- 2.
http://www.keel.es/.
References
Zhu, Z., Wang, Z., Li, D., Du, W.: Globalized multiple balanced subsets with collaborative learning for imbalanced data. IEEE Trans. Cybern. 52(4), 2407–2417 (2022)
Chaovalitwongse, W.A., Fan, Y.J., Sachdeo, R.C.: On the time series \( k \)-nearest neighbor classification of abnormal brain activity. IEEE Trans. Syst. Man Cybern. A Syst. Hum. 37(6), 1005–1016 (2007)
Kashef, R.: A boosted SVM classifier trained by incremental learning and decremental unlearning approach. Expert Syst. Appl. 167, 114154 (2021)
He, H., Garcia, E.A.: Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 21(9), 1263–1284 (2009)
Wang, N., Liang, R., Zhao, X., Gao, Y.: Cost-sensitive hypergraph learning with f-measure optimization, IEEE Trans. Cybern. (2021)
Seiffert, C., Khoshgoftaar, T.M., van Hulse, J., et al.: RUSBoost: a hybrid approach to alleviating class imbalance. IEEE Trans. Syst. Man Cybern. A Syst. Hum. 40(1), 185–197 (2009)
Dempster, A.P.: Upper and lower probabilities induced by a multivalued mapping. Ann. Statist. 83, 325–339 (1967)
Shafer, G.: A Mathematical Theory of Evidence. Princeton Univ, Press (1976)
Liu, Z., Pan, Q., Dezert, J., Han, J.W., He, Y.: Classifier fusion with contextual reliability evaluation. IEEE Trans. Cybern. 48(5), 1605–1618 (2018)
Zhang, Z.W., Tian, H.P., Yan, L.Z., Martin, A., Zhou, K.: Learning a credal classifier with optimized and adaptive multiestimation for missing data imputation, IEEE Trans. Syst. Man Cybern. (2021)
Niu, J., Liu, Z., Lu, Y., Wen, Z.: Evidential combination of classifiers for imbalanced data. IEEE Trans. Syst. Man Cybern. (2022)
Zhang, Z.W., Ye, S.T., Zhang, Y.R., Ding, W.P., Wang, H.: Belief combination of classifiers for incomplete data. IEEE-CAA J. Automatica Sin. 9(4), 652–667 (2022)
Grina, Fares, Elouedi, Zied, Lefevre, Eric: A preprocessing approach for class-imbalanced data using SMOTE and belief function theory. In: Analide, Cesar, Novais, Paulo, Camacho, David, Yin, Hujun (eds.) IDEAL 2020. LNCS, vol. 12490, pp. 3–11. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-62365-4_1
Niu, Jiawei, Liu, Zhunga: Imbalance data classification based on belief function theory. In: Denœux, Thierry, Lefèvre, Eric, Liu, Zhunga, Pichon, Frédéric. (eds.) BELIEF 2021. LNCS (LNAI), vol. 12915, pp. 96–104. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-88601-1_10
Smets, P.: Constructing the pignistic probability function in a context of uncertainty. Uncertainty Artif. Intell. 5, 29–39 (1990)
Zhang, Y., Liu, G., Luan, W.: An approach to class imbalance problem based on stacking and inverse random under sampling methods, In: 2018 IEEE 15th International Conference on Networking, Sensing and Control (ICNSC), pp. 1–6. IEEE (2018)
Chawla, N.V., Bowyer, K.W., Hall, L.O., et al.: SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)
Lin, W.C., Tsai, C.F., Hu, Y.H., et al.: Clustering-based undersampling in class-imbalanced data. Inf. Sci. 409, 17–26 (2017)
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant U20B2067, Grant 61790552, and Grant 61790554; the Aeronautical Science Foundation of China under Grant 201920007001.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Tian, H., Zhang, Z., Martin, A., Liu, Z. (2022). Reliability-Based Imbalanced Data Classification with Dempster-Shafer Theory. In: Le Hégarat-Mascle, S., Bloch, I., Aldea, E. (eds) Belief Functions: Theory and Applications. BELIEF 2022. Lecture Notes in Computer Science(), vol 13506. Springer, Cham. https://doi.org/10.1007/978-3-031-17801-6_8
Download citation
DOI: https://doi.org/10.1007/978-3-031-17801-6_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-17800-9
Online ISBN: 978-3-031-17801-6
eBook Packages: Computer ScienceComputer Science (R0)