Adaboost Ensemble Data Classification Based on Diversity of Classifiers

Article Preview

Abstract:

In this research we propose an ensemble classification technique based on decision tree, artificial neural network, and support vector machine models weighting classifier by adaboost in order to increase classification accuracy. we used a total of 30 classifiers. The technique generated random data used Bootstrap. Testing Diabites Data from UCI, classification accuracy tests on Diabites Data found that the proposed ensemble classification models weighting classifier by Adaboost yields better performance than that of a single model with the same type of classifier. The result as follows, Diabites Data achieved the best performance with 75.21%. we can conclude that there are two essential requirements in the model. The first is that the ensemble members or learning agents must be diverse or complementary, i.e., agents must exhibit different properties. Another condition is that an optimal ensemble strategy is also required to fuse a set of diverse by AdaBoost.

You might also be interested in these eBooks

Info:

Periodical:

Advanced Materials Research (Volumes 403-408)

Pages:

3682-3687

Citation:

Online since:

November 2011

Export:

Price:

[1] K. Zeitouni and N. Chelghoum, Spatial Decision Tree-Application to Traffic Risk Analysis, in Proceedings of the ACS/IEEE International Conference on Computer Systems and Applications: IEEE Computer Society, (2001).

DOI: 10.1109/aiccsa.2001.933978

Google Scholar

[2] D. Thammasiri N. Phothi V. Nuipain P. Sangsiri P. Aummarukkarat and P. Meesad, Credit scoring with a data mining approach based on Back-Propagation Neural Network, 2rd National Conference on Information Technology, (2008) pp.476-483.

Google Scholar

[3] D. Thammasiri and P. Meesad, Ensemble Data Classification based on Support Vector Machine Optimized parameters by Genetic Algorithm also Feature Selection by Correlation Coefficient", The 6th National Conference on Computing and Information Technology (NCCIT, 10), (2010).

Google Scholar

[4] S. j. Wang, A. Mathew, Y. Chen, L. f. Xi, L. Ma, and J. Lee, Empirical analysis of support vector machine ensemble classifiers, Expert Syst. Appl., vol. 36, pp.6466-6476, (2009).

DOI: 10.1016/j.eswa.2008.07.041

Google Scholar

[5] Y. Yao, Z. Fu, X. Zhao and W. Cheng, Combining Classifier Based on Decision Tree, icie, vol. 2, pp.37-40, 2009 WASE International Conference on Information Engineering, (2009).

Google Scholar

[6] J. Wang, J. Yang, S. Li, Q. Dai, and J. Xie, Number Image Recognition Based on Neural Network Ensemble, in Proceedings of the Third International Conference on Natural Computation - Volume 01: IEEE Computer Society, (2007).

DOI: 10.1109/icnc.2007.506

Google Scholar

[7] S. Chen, W. Wang, and H. V. Zuylen, Construct support vector machine ensemble to detect traffic incident, Expert Systems with Applications, vol. 36, pp.10976-10986, (2009).

DOI: 10.1016/j.eswa.2009.02.039

Google Scholar

[8] K. C. Ying, S. -W. Lin, Z. J. Lee, and Y. T. Lin, An ensemble approach applied to classify spam e-mails, Expert Syst. Appl., vol. 37, pp.2197-2201, (2010).

DOI: 10.1016/j.eswa.2009.07.080

Google Scholar

[9] X. Pei-Yong, D. X. Qian, and J. B. Ning, A GA-based feature selection and ensemble learning for high-dimensional datasets, in Machine Learning and Cybernetics, 2009 International Conference on, (2009), pp.7-12.

DOI: 10.1109/icmlc.2009.5212542

Google Scholar

[10] L. Zhou, K. K. Lai, and L. Yu, Least squares support vector machines ensemble models for credit scoring, Expert Systems with Applications, vol. 37, pp.127-133, (2010).

DOI: 10.1016/j.eswa.2009.05.024

Google Scholar

[11] W. Yu and L. Cheng De, Learning by Bagging and Adaboost based on Support Vector Machine, in Industrial Informatics, 2007 5th IEEE International Conference on, (2007), pp.663-668.

DOI: 10.1109/indin.2007.4384852

Google Scholar

[12] S. Khaleghian, H. R. Rabiee, and M. H. Rohban, Face recognition across large pose variations via Boosted Tied Factor Analysis, in Applications of Computer Vision (WACV), 2011 IEEE Workshop on, (2011), pp.190-195.

DOI: 10.1109/wacv.2011.5711502

Google Scholar

[13] M. T. Mitchell, Machine Learning. New York: McGraw-Hill, (1997).

Google Scholar

[14] Quinlan, J. R. Induction of decision trees. Machine Learning, 81–106, (1986).

Google Scholar

[15] Defu Zhang, Qingshan Jiang, and Xin Li. A Hybrid Mining Model Based on Neural Network and Kernel Smoothing Technique. Lecture Notes in Computer Science. (2005), 3516: 801–805.

DOI: 10.1007/11428862_110

Google Scholar

[16] V. Vapnik, The nature of statistical learning theory, New York: Wiley, (1998).

Google Scholar

[17] B. SchP olkopf, A. Smola, Learning with Kernels, MIT Press, Cambridge, MA, (2002).

Google Scholar

[18] Y. Freund and R.E. Schapire, A decision-theoretic generalization of on-line learning and an application to Boosting, Journal of Computer and System Sciences 55(1) (1997) 119–139.

DOI: 10.1006/jcss.1997.1504

Google Scholar

[19] Schapire, R.E., Singer, Y., Improved boosting algorithms using confidence-rated predictions. Machine Learning 37 (3), 297–336, (1999).

Google Scholar

[20] C.L. Blake, C.J. Merz, UCI repository of machine learning databases, Website: http: /www. ics. uci. edu/~mlearn/MLRepository. html, (1998).

Google Scholar

[21] Rayner Alfred, Knowledge Discovery: Enhancing Data Mining and Decision Support Integration, (2005).

Google Scholar

[22] L. Breiman, Bagging predictors, Machine Learning., vol. 24, pp.123-140, (1996).

Google Scholar