Skip to main content
Log in

Active Learning by Extreme Learning Machine with Considering Exploration and Exploitation Simultaneously

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

As an important machine learning paradigm, active learning has been widely applied to scenarios in which it is easy to acquire a large number of instances but labeling them is expensive and/or time-consuming. In such scenario, active learning can significantly reduce the cost of labeling the instances. Extreme learning machine (ELM) is a popular supervised learning model that has the structure of a single-hidden-layer feed-forward network, and has such merits as low computational cost, high training speed, and high generalization ability. Previous studies have shown that the integration of active learning with the ELM can yield effective and efficient results. However, the currently used method of integration considers only the capability for exploitation neglecting that for exploration, further increasing the risk of the results falling into local optima in context of a cold start. To address this problem, we propose an improved algorithm called the AL-SNN-ELM in this paper. It contains two sub-procedures for a sequential query: The exploration strategy, which uses the shared nearest neighbor (SNN) clustering algorithm, takes charge of exploring the sample space to query representative instances, and the exploitation strategy is responsible for transforming the actual outputs of the ELM into posterior probabilities to query uncertain instances. That is to say, the exploration sub-procedure helps roughly locate the decision boundary for sound classification by observing the global distribution of the data, while the exploitation sub-procedure subtly tunes this decision boundary by observing the distribution of local instances surrounding it. In addition, to reduce the time-complexity of active learning, online-sequential extreme learning machine is also adopted to replace the traditional ELM. The results of experiments on 20 UCI benchmark datasets and two real-world datasets show that the proposed AL-SNN-ELM algorithm can yield a significant improvement in performance in comparison with the traditional AL-ELM algorithm, indicating that it is useful to consider the exploration and exploitation simultaneously in the framework of active learning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Data Availability

The data that support the findings of this study are available from the openly UCI machine learning repository (http://archive.ics.uci.edu/ml/datasets.php), and Kaggle platform (https://www.kaggle.com/datasets/brjapon/gearbox-fault-diagnosis-stdev-of-accelerations; https://www.kaggle.com/datasets/subhajournal/credit-card-fraud-dataset). The codes of the proposed algorithm can be downloaded from https://github.com/ML-YanGu/AL-SNN-ELM.git.

References

  1. Chakraborty S, Balasubramanian V, Panchanathan S (2015) Adaptive batch mode active learning. IEEE Trans Neural Netw Learn Syst 26(8):1747–1760

    Article  MathSciNet  Google Scholar 

  2. Hazarika BB, Gupta D (2021) Density weighted twin support vector machines for binary class imbalance learning. Neural Process Lett 54(2):1091–1130

    Article  Google Scholar 

  3. Hazarika BB, Gupta D (2020) Density-weighted support vector machines for binary class imbalance learning. Neural Comput Appl 33(9):4243–4261

    Article  Google Scholar 

  4. Du B, Wang Z, Zhang L et al (2017) Exploring representativeness and informativeness for active learning. IEEE Trans Cybern 47(1):14–26

    Article  Google Scholar 

  5. Settles B (2011) From theories to queries: active learning in practice. In: JMLR workshop and conference proceedings, vol 16, pp 1–18

  6. Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370

    Article  Google Scholar 

  7. Konyushkova K, Sznitman R, Fua P (2015) Introducing geometry in active learning for image segmentation. In: 2015 IEEE international conference on computer vision (ICCV), Santiago, Chile, pp 2974–2982

  8. Liu B, Ferrari V (2017) Active learning for human pose estimation. In: 2017 IEEE international conference on computer vision (ICCV), Venice, Italy, pp 4363–4372

  9. She Q, Chen K, Luo Z et al (2020) Double-criteria active learning for multiclass brain–computer interfaces. Comput Intell Neurosci 2020:1–13

    Article  Google Scholar 

  10. Malhotra K, Bansal S, Ganapathy S (2019) Active learning methods for low resource end-to-end speech recognition. In: Interspeech, Graz, Austria, pp 2215–2219

  11. Han X, Kwoh CK, Kim J (2016) Clustering based active learning for biomedical named entity recognition. In: 2016 International joint conference on neural networks (IJCNN), Vancouver, BC, Canada, pp 1253–1260

  12. Flores CA, Figueroa RL, Pezoa JE (2021) Active learning for biomedical text classification based on automatically generated regular expressions. IEEE Access 9:38767–38777

    Article  Google Scholar 

  13. Sharma M, Bilgic M (2016) Evidence-based uncertainty sampling for active learning. Data Min Knowl Disc 31:164–202

    Article  MathSciNet  MATH  Google Scholar 

  14. Lughofer E, Pratama M (2018) Online active learning in data stream regression using uncertainty sampling based on evolving generalized fuzzy models. IEEE Trans Fuzzy Syst 26(1):292–309

    Article  Google Scholar 

  15. Wang G, Hwang JN, Rose C, Wallace F (2019) Uncertainty-based active learning via sparse modeling for image classification. IEEE Trans Image Process 28(1):316–329

    Article  MathSciNet  MATH  Google Scholar 

  16. Wang R, Kwong S, Chen D (2012) Inconsistency-based active learning for support vector machines. Pattern Recognit 45(10):3751–3767

    Article  Google Scholar 

  17. Yu G, Yang Y, Wang X et al (2020) Adversarial active learning for the identification of medical concepts and annotation inconsistency. J Biomed Inform 108:103481

    Article  Google Scholar 

  18. Smith JS, Nebgen B, Lubbers N et al (2018) Less is more: sampling chemical space with active learning. J Chem Phys 148(24):241733

    Article  Google Scholar 

  19. Settles B, Craven M (2008) An analysis of active learning strategies for sequence labeling tasks. In: Proceedings of the conference on empirical methods in natural language processing (EMNLP), Honolulu, Hawaii, USA, pp 1070–1079

  20. Mingkun Li, Sethi IK (2006) Confidence-based active learning. IEEE Trans Pattern Anal Mach Intell 28(8):1251–1261

    Article  Google Scholar 

  21. Roy N, McCallum A (2001) Toward optimal active learning through monte carlo estimation of error reduction. In: Proceedings of the international conference on machine learning (ICML), Williamstown, MA, USA, vol 2, pp 441–448

  22. Ling C, Lu Z, Zhu X (2019) Efficient methods by active learning kriging coupled with variance reduction based sampling methods for time-dependent failure probability. Reliab Eng Syst Saf 188:23–35

    Article  Google Scholar 

  23. Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370

    Article  Google Scholar 

  24. Huang SJ, Jin R, Zhou ZH (2014) Active learning by querying informative and representative examples. IEEE Trans Pattern Anal Mach Intell 36(10):1936–1949

    Article  Google Scholar 

  25. Yang Y, Ma Z, Nie F et al (2014) Multi-class active learning by uncertainty sampling with diversity maximization. Int J Comput Vis 113(2):113–127

    Article  MathSciNet  Google Scholar 

  26. Wang M, Min F, Zhang ZH, Wu YX (2017) Active learning through density clustering. Expert Syst Appl 85:305–317

    Article  Google Scholar 

  27. He D, Yu H, Wang G, Li J (2021) A two-stage clustering-based cold-start method for active learning. Intell Data Anal 25(5):1169–1185

    Article  Google Scholar 

  28. Reitmaier T, Calma A, Sick B (2015) Transductive active learning: a new semi-supervised learning approach based on iteratively refined generative models to capture structure in data. Inf Sci 293:275–298

    Article  Google Scholar 

  29. Yu K, Bi J, Tresp V (2006) Active learning via transductive experimental design. In: Proceedings of the 23rd international conference on machine learning (ICML), Pittsburgh, Pennsylvania, pp 1081–1088

  30. Yang Y, Yin X, Zhao Y et al (2021) Batch mode active learning based on multi-set clustering. IEEE Access 9:51452–51463

    Article  Google Scholar 

  31. Chen DW, Jin YH (2020) An active learning algorithm based on Shannon entropy for constraint-based clustering. IEEE Access 8:171447–171456

    Article  Google Scholar 

  32. Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501

    Article  Google Scholar 

  33. Huang GB, Wang DH, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2(2):107–122

    Article  Google Scholar 

  34. Huang GB, Zhou HM, Ding XJ, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst Man Cybern B 42(2):513–529

    Article  Google Scholar 

  35. Borah P, Gupta D (2020) Unconstrained convex minimization based implicit lagrangian twin extreme learning machine for classification (ULTELMC). Appl Intell 50(4):1327–1344

    Article  Google Scholar 

  36. Hazarika BB, Gupta D, Berlin M (2020) A Coiflet LDMR and coiflet OB-elm for river suspended sediment load prediction. Int J Environ Sci Technol 18(9):2675–2692

    Article  Google Scholar 

  37. Yu H, Sun C, Yang W et al (2015) AL-ELM: one uncertainty-based active learning algorithm using extreme learning machine. Neurocomputing 166:140–150

    Article  Google Scholar 

  38. Yu H, Yang X, Zheng S, Sun C (2019) Active learning from imbalanced data: a solution of online weighted extreme learning machine. IEEE Trans Neural Netw Learn Syst 30(4):1088–1103

    Article  Google Scholar 

  39. Qin J, Wang C, Zou Q et al (2021) Active learning with extreme learning machine for online imbalanced multiclass classification. Knowl Based Syst 231:107385

    Article  Google Scholar 

  40. Yoon J, Hwang SJ (2017) Combined group and exclusive sparsity for deep neural networks. In: Proceedings of international conference on machine learning (ICML), Sydney, NSW, Australia, vol 70, pp 3958–3966

  41. Kumar V, Pujari AK, Padmanabhan V, Kagita VR (2019) Group preserving label embedding for multi-label classification. Pattern Recognit 90:23–34

    Article  Google Scholar 

  42. Ertöz L, Steinbach M, Kumar V (2003) Finding clusters of different sizes, shapes, and densities in noisy, high dimensional data. In: Proceedings of the 2003 SIAM international conference on data mining (SDM), San Francisco, CA, USA, pp 47–58

  43. Liang NY, Huang GB, Saratchandran P, Sundararajan N (2006) A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans Neural Netw Learn Syst 17(6):1411–1423

    Article  Google Scholar 

  44. Wang Z, Du B, Tu W et al (2021) Incorporating distribution matching into uncertainty for multiple kernel active learning. IEEE Trans Knowl Data Eng 33(1):128–142

    Article  Google Scholar 

  45. Jarvis RA, Patrick EA (1973) Clustering using a similarity measure based on shared near neighbors. IEEE Trans Comput C –22(11):1025–1034

    Article  Google Scholar 

  46. Wang WT, Wu YL, Tang CY, Hor MK (2015) Adaptive density-based spatial clustering of applications with noise (DBSCAN) according to Data. In: 2015 International conference on machine learning and cybernetics (ICMLC), GuangDong, China, vol 1, pp 445–451

  47. Sawant K (2014) Adaptive methods for determining dbscan parameters. Int J Innov Sci Eng Technol 1(4):329–334

    Google Scholar 

  48. Blake C, Keogh E, Merz CJ (1998) UCI repository of machine learning databases, Department of Information and Computer Science, University of California, Technical Report 213, Irvine, CA

  49. https://www.kaggle.com/datasets/brjapon/gearbox-fault-diagnosis-stdev-of-accelerations

  50. https://www.kaggle.com/datasets/subhajournal/credit-card-fraud-dataset

  51. Xu Z, Yu K, Tresp V et al (2003) Representative sampling for text classification using support vector machines. In: European conference on information retrieval (ECIR), Berlin, Heidelberg, pp 393–407

  52. Zhang X, Delpha C, Diallo D (2020) Incipient fault detection and estimation based on Jensen–Shannon divergence in a data-driven approach. Signal Process 169:107410

    Article  Google Scholar 

  53. Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30

    MathSciNet  MATH  Google Scholar 

  54. Garcia S, Herrera F (2008) An extension on "statistical comparisons of classifiers over multiple data sets" for all pairwise comparisons. J Mach Learn Res 9:12

    MATH  Google Scholar 

  55. García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf Sci 180(10):2044–2064

    Article  Google Scholar 

Download references

Acknowledgements

The work was supported in part by Natural Science Foundation of Jiangsu Province of China under Grant No. BK20191457, National Natural Science Foundation of China under Grants No. 62176107, No. 62076111 and No. 62076215, Postgraduate Research & Practice Innovation Program of Jiangsu Province, China No. SJCX22_1901.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hualong Yu.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Gu, Y., Yu, H., Yang, X. et al. Active Learning by Extreme Learning Machine with Considering Exploration and Exploitation Simultaneously. Neural Process Lett 55, 5245–5267 (2023). https://doi.org/10.1007/s11063-022-11089-w

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-022-11089-w

Keywords

Navigation