Abstract
Prediction performance and optimization attributes of the transparent open-box learning networks (TOB) applied to a large database of US coals are further explored to complement recently published base-case analysis. Nine sensitivity cases configured with the 6339 data records allocated in different ways to the TOB’s tuning, training, and testing subsets are developed. These cases demonstrate for this data set that the TOB algorithm provides robust, reliable, and repeatable predictions provided that the tuning subset contains 40 or so data records. On the other hand, increasing the number of records in the testing subset to numbers much greater than 100 does not lead to improved prediction accuracy. A comparison of the prediction performance of three optimizers applied with the TOB for each of the sensitivity cases reveals that a memetic firefly optimizer matches the optimized solutions found by Excel’s GRG Solver optimizer. The functionality of the memetic firefly optimizer enables it to be used effectively in a fully coded version of the TOB optimizer (not involving Excel cell formula for the optimizer to operate). This is an advantage when evaluating larger data sets with larger tuning subset requirements. The memetic firefly optimizer also introduces further transparency, flexibility, and control to the TOB optimization process by facilitating metaheuristic profiling that aids the tuning and customization of the optimizer for deployment with a range of data sets involving complex, non-linear feasible-solution spaces.
Similar content being viewed by others
References
Arora S, Singh S (2013) The firefly optimization algorithm: convergence analysis and parameter selection. Int J Comput Appl 69(3):48–52
Arora S, Singh S (2014) Performance research on firefly optimization algorithm with mutation. In: International conference on communication, computing & systems, pp 168–172
Atkeson CG, Moore AW, Schaal S (1997) Locally weighted learning. Artif Intell Rev 11(1–5):11–73
Birattari M, Bontempi G, Bersini H (1999) Lazy learning meets the recursive least squares algorithm. In: Kearns MJ, Solla SA, Cohn DA (eds) Proceedings of the 1998 conference on advances in neural information processing systems, vol 11. MIT Press, Cambridge, pp 375–381
Bontempi G, Birattari M, Bersini H (1999) Lazy learning for local modeling and control design. Int J Control 72(7/8):643–658
Bragg LJ, Oman JK, Tewalt SJ, Oman CJ, Rega NH, Washington PM, Finkelman RB (1997) U.S. geological survey coal quality (COALQUAL) database: version 2.0. U.S. geological survey open-file report 97-134. https://doi.org/10.3133/ofr97134. Accessed 19 Jan 2019
Chen GH, Shah D (2018) Explaining the success of nearest neighbor methods in prediction. Found Trends R Mach Learn 10(5–6):337–588
Cover TM, Hart PE (1967) Nearest neighbor pattern classification. IEEE Trans Inf Theory 13(1):21–27
Feng Q, Zhang J, Zhang X, Wen S (2015) Proximate analysis-based prediction of gross calorific value of coals: a comparison of support vector machine, alternating conditional expectation and artificial neural network. Fuel Process Technol 129:120–129
Fix E, Hodges JL Jr (1951) Discriminatory analysis, nonparametric discrimination: consistency properties. Technical report, USAF School of Aviation Medicine
Fix E, Hodges JL Jr (1989) Discriminatory analysis. nonparametric discrimination: consistency properties. Int Stat Rev 57(3):238–247. https://doi.org/10.2307/1403797
Frontline Solvers (2019) Standard excel solver—limitations of nonlinear optimization. https://www.solver.com/standard-excel-solver-limitations-nonlinear-optimization. Accessed 19 Jan 2019
Garcia S, Derrac J, Cano J, Herrera F (2012) Prototype selection for nearest neighbor classification: taxonomy and empirical study. IEEE Trans Pattern Anal Mach Intell 34(3):417–435
Gul A, Perperoglou A, Khan Z, Mahmoud O, Miftahuddin M, Adler W, Lausen B (2018) Ensemble of a subset of kNN classifiers. Adv Data Anal Classif 12(4):827–840. https://doi.org/10.1007/s11634-015-0227-5
Lever J, Krywinski M, Altman N (2016) Model selection and overfitting. Nat Methods 13:703–704. https://doi.org/10.1038/nmeth.3968
Li Z, Wang W, Yan Y, Li Z (2015) PS–ABC: a hybrid algorithm based on particle swarm and artificial bee colony for high-dimensional optimization problems. Expert Syst Appl 42:8881–8895
Martínez-Muñoz G, Suárez A (2010) Out-of-bag estimation of the optimal sample size in bagging. Pattern Recogn 43:143–152
Matin SS, Chelgani SC (2016) Estimation of coal gross calorific value based on various analyses by random forest method. Fuel 177:274–278
Pal SK, Raj CS, Singh AP (2012) Comparative study of firefly algorithm and particle swarm optimization for noisy non-linear optimization problems. Int J Intell Syst Appl 10:50–57
Sabzevari M, Martınez-Munoz G, Suarez A (2014) Improving the robustness of bagging with reduced sampling size. In: Proceedings of the European symposium on artificial neural networks, computational intelligence and machine learning. Bruges (Belgium), 23–25 April 2014. ISBN 978-287419095-7. https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2014-185.pdf. Accessed 19 Jan 2019.
Samworth R (2012) Optimal weighted nearest neighbour classifiers. Ann Stat 40(5):2733–2763
Shakhnarovich G, Darrell T, Indyk P (2006) Nearest-neighbor methods in learning and vision: theory and practice (neural information processing). The MIT Press, Cambridge, MA. ISBN 026219547X
Tan P, Zhang C, Xia J, Fang Q-Y, Chen G (2015) Estimation of higher heating value of coal based on proximate analysis using support vector regression. Fuel Process Technol 138:298–304
Wood DA (2016a) Metaheuristic profiling to assess performance of hybrid evolutionary optimization algorithms applied to complex wellbore trajectories. J Nat Gas Sci Eng 33:751–768. https://doi.org/10.1016/j.jngse.2016.05.041 2016
Wood DA (2016b) Evolutionary memetic algorithms supported by metaheuristic profiling effectively applied to the optimization of discrete routing problems. J Nat Gas Sci Eng 35:997–1014. https://doi.org/10.1016/j.jngse..jngse.2016.09.031
Wood DA (2018a) A transparent open-box learning network provides insight to complex systems and a performance benchmark for more-opaque machine learning algorithms. Adv Geo Energy Res 2(2):148–162
Wood DA (2018b) Transparent open-box learning network provides auditable predictions for coal gross calorific value. Model Earth Syst Environ. https://doi.org/10.1007/s40808-018-0543-9
Wood DA (2018c) Thermal maturity and burial history modelling of shale is enhanced by use of Arrhenius time-temperature index and memetic optimizer. Petroleum 4:25–42. https://doi.org/10.1016/j.petlm.2017.10.004
Yang XS (2009) Firefly algorithms for multimodal optimization. In: Watanabe O, Zeugmann T (eds) Stochastic algorithms: foundations and applications. SAGA 2009. Lecture notes in computer science, vol 5792. Springer, Berlin, Heidelberg, pp 169–178
Yang X-S, He X (2013) Firefly algorithm: recent advances and applications. Int J Swarm Intell 1(1):36–50
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author declares no conflicts of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wood, D.A. Sensitivity analysis and optimization capabilities of the transparent open-box learning network in predicting coal gross calorific value from underlying compositional variables. Model. Earth Syst. Environ. 5, 753–766 (2019). https://doi.org/10.1007/s40808-019-00583-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40808-019-00583-1