Abstract
Sparse models are preferred in machine learning problems because of their computational interpretability and it is seen in many applications such as in Google Page Rank, classification and regression problems, in the method of Principal Component Analysis (PCA) that finds the most important features and further applications in graphical models. In this study, we derive optimality conditions for the quadratic problem which has cardinality constraint imposing sparse solution. Our Quadratic model is a special application of ensemble pruning model in ensemble learning algorithms. Here, we refer to our previous study on this application in ensemble selection for clustering problems. The quadratic model proposed in this study optimizes trade-off between accuracy and diversity of discriminant functions (classifiers) simultaneously so that the best candidates of ensemble are selected for prediction step. The selection of the best classifiers in the ensemble is crucial for the overall performance of ensemble learning algorithms since redundant/outlier solutions in the ensemble library will decrease the overall prediction accuracy. In order to eliminate such candidates, both accuracy and diversity are taken into account when selecting the best subset of the ensemble. The cardinality constraint is further relaxed by considering various approximations such as \(l_1\)-norm regularization and student t-log likelihood approximations. Under these considerations and approximations, we build optimality criteria for our quadratic optimization problem with a cardinality constraint.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Akyüz, S., Otar, B.: Ensemble clustering selection by optimization of accuracy-diversity trade off. In: 2017 25th Signal Processing and Communications Applications Conference (SIU), pp. 1–4 (2017)
Azimi, J., Fern, X.: Adaptive cluster ensemble selection. In: IJCAI International Joint Conference on Artificial Intelligence, pp. 992–997 (2009)
Bache, K., Lichman, M.: UCI Machine Learning Repository (2013)
Bradley, P.S., Mangasarian, O.L., Rosen, J.B.: Parsimonious least norm approximation. Comput. Optim. Appl. (1998)
Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)
Caruana, R., Niculescu-Mizil, A., Crew, G., Ksikes, A.: Ensemble selection from libraries of models. In: Twenty-First International Conference on Machine Learning - ICML 2004, p. 18 (2004)
Daszykowski, M., Walczak, B.: Density-based clustering methods. Compr. Chemom. 2, 635–654 (2010). https://doi.org/10.1016/B978-044452701-1.00067-3
Fern, X.Z., Lin, W.: Cluster ensemble selection. Stat. Anal. Data Min. 1(3), 128–141 (2008). https://doi.org/10.1002/sam.10008
Ferrari, D.G., De Castro, L.N.: Clustering algorithm selection by meta-learning systems: a new distance-based problem characterization and ranking combination methods. Inf. Sci. 301, 181–194 (2015)
Forgy, E.W.: Cluster analysis of multivariate data: efficiency versus interpretability of classifications. Biometrics 21(3), 768–769 (1965)
Ghosh, J., Acharya, A.: Wiley interdisciplinary reviews: data mining and knowledge discovery. Clust. Ensembles 1(4), 305–315 (2011)
Horst, R., Thoai, N.V.: Dc programming: overview. J. Optim. Theor. Appl. 103(1), 1–43 (1999)
Jia, J., Xiao, X., Liu, B., Jiao, L.: Bagging-based spectral clustering ensemble selection. Pattern Recognit. Lett. 32(Fskd), 1456–1467 (2011)
Le Thi, H.A., Pham Dinh, T., Le, H.M., Vo, X.T.: DC approximation approaches for sparse optimization. Eur. J. Oper. Res. 244, 26–46 (2015). https://doi.org/10.1016/j.ejor.2014.11.031
Mangasarian, O.: Machine learning via polyhedral concave minimization. In: Applied Mathematics and Parallel Computing, pp. 175–188. Springer, Heidelberg (1996)
Moon, T.: The expectation-maximization algorithm. IEEE Sig. Process. Mag. 13(6), 47–60 (1996)
Muñoz, A., Diego, I.: From indefinite to positive semi-definite matrices. Struct. Syntactic Stat. Pattern Recognit. 4109, 764–772 (2006)
Pham Dinh, T., Le Thi, H.A.: Recent advances in DC programming and DCA. In: Nguyen, N.T., Le-Thi, H.A. (eds.) Trans. Comput. Intell. XIII, pp. 1–37. Springer, Heidelberg (2014)
Rinaldi, F., Schoen, F., Sciandrone, M.: Concave programming for minimizing the zero-norm over polyhedral sets. Comput. Optim. Appl. (2010)
Rockafeller, R.T.: Convex analysis (1970)
Schapire, R.: The boosting approach to machine learning: an overview (2001)
Shen, X., Diamond, S., Gu, Y., Boyd, S.: Disciplined convex-concave programming. In: 2016 IEEE 55th Conference on Decision and Control (CDC), pp. 1009–1014. IEEE (2016)
Sriperumbudur, B.K., Torres, D.A., Lanckriet, G.R.: A majorization-minimization approach to the sparse generalized eigenvalue problem. Mach. Learn. (2011)
Strehl, A., Ghosh, J.: Cluster ensembles - a knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 3, 583–617 (2002)
Tao, P.D., et al.: Algorithms for solving a class of nonconvex optimization problems. Methods of subgradients. In: North-Holland Mathematics Studies, vol. 129, pp. 249–271. Elsevier (1986)
Thiao, M., Dinh, T.P., Le Thi, H.A.: Dc programming approach for a class of nonconvex programs involving l 0 norm. In: Modelling, Computation and Optimization in Information Systems and Management Sciences, pp. 348–357. Springer (2008)
Tibshirani, R.: Regression selection and shrinkage via the lasso (1996). https://doi.org/10.2307/2346178
Weston, J., Elisseeff, A., Scholkopf, B., Tipping, M.: Use of the zero-norm with linear models and kernel methods. J. Mach. Learn. Res. (2003)
Wolpert, D.H.: Stacked generalization. Neural Netw. 5(2), 241–259 (1992)
Yu, Z., Li, L., Gao, Y., You, J., Liu, J., Wong, H.S., Han, G.: Hybrid clustering solution selection strategy. Pattern Recognit. 47(10), 3362–3375 (2014)
Zhang, Y., Burer, S., Street, W.N.: Ensemble pruning via semi-definite programming. J. Mach. Learn. Res. 7, 1315–1338 (2006)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Üçüncü, D., Akyüz, S., Gül, E., Wilhelm-Weber, G. (2019). Optimality Conditions for Sparse Quadratic Optimization Problem. In: Rodrigues, H., et al. EngOpt 2018 Proceedings of the 6th International Conference on Engineering Optimization. EngOpt 2018. Springer, Cham. https://doi.org/10.1007/978-3-319-97773-7_67
Download citation
DOI: https://doi.org/10.1007/978-3-319-97773-7_67
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-97772-0
Online ISBN: 978-3-319-97773-7
eBook Packages: EngineeringEngineering (R0)