Abstract
For the past ten years, Gaussian process has become increasingly popular for modeling numerous inferences and reasoning solutions due to the robustness and dynamic features. Particularly concerning regression and classification data, the combination of Gaussian process and Bayesian learning is considered to be one of the most appropriate supervised learning approaches in terms of accuracy and tractability. However, due to the high complexity in computation and data storage, Gaussian process performs poorly when processing large input dataset. Because of the limitation, this method is ill-equipped to deal with the large-scale system that requires reasonable precision and fast reaction rate. To improve the drawback, our research focuses on a comprehensive analysis of Gaussian process performance issues, highlighting ways to drastically reduce the complexity of hyper-parameter learning and training phases, which could be applicable in predicting the CPU utilization in the demonstrated application. In fact, the purpose of this application is to save the energy by distributively engaging the Gaussian process regression to monitor and predict the status of each computing node. Subsequently, a migration mechanism is applied to migrate the system-level processes between multi-core and turn off the idle one in order to reduce the power consumption while still maintaining the overall performance.
Similar content being viewed by others
References
Lawrence N D (2004) Gaussian process latent variable models for visualisation of high dimensional data. Advances in neural information processing systems 16:329–336
Rasmussen CE (1997) Evaluation of gaussian processes and other methods for non-linear regression, Ph.D. dissertation, Toronto, Ont., Canada, Canada, aAINQ28300
Chalupka K, Williams C K I, Murray I (2012) A framework for evaluating approximation methods for gaussian process regression, CoRR. arXiv: http://arxiv.org/abs/1205.6326
Rasmussen C, Williams C (2005) Gaussian processes for machine learning, ser. adaptive computation and machine learning, MIT Press. http://www.gaussianprocess.org/gpml/chapters/
Brahim-Belhouari S, Vesin J (2001) Bayesian learning using gaussian process for time series prediction. In: Statistical Signal Processing, 2001. Proceedings of the 11th IEEE Signal Processing Workshop on, pp 433–436
Roberts S, Osborne M, Ebden M, Reece S, Gibson N, Aigrain S (2012) Gaussian processes for time-series modeling. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 371(1984)
Petelin D, Filipič B, Kocijan J Optimization of gaussian process models with evolutionary algorithms. In: Proceedings of the 10th International Conference on Adaptive and Natural Computing Algorithms - Volume Part I, ser. ICANNGA’11. Berlin, Heidelberg: Springer-Verlag, 2011, pp. 420–429. [Online]. Available:, http://dl.acm.org/citation.cfm?id=1997052.1997098
Petelin D, Kocijan J. (2014) Evolving gaussian process models for predicting chaotic time-series. In: Evolving and Adaptive Intelligent Systems (EAIS), 2014 IEEE Conference on. IEEE, pp. 1–8.
Shewchuk J R (1994) An introduction to the conjugate gradient method without the agonizing pain, Pittsburgh, PA, USA, Tech. Rep
Grande R, Chowdhary G, How J (2013) Nonparametric adaptive control using gaussian processes with online hyperparameter estimation. In: 2013 IEEE 52nd annual conference on decision and control (CDC), , pp. 861867
Banerjee A, Dunson D, Tokdar S (2011) Efficient gaussian process regression for large data sets. arXiv: http://arxiv.org/abs/e-prints
Hensman J, Fusi N, Lawrence N D (2013) Gaussian processes for big data CoRR. arXiv: http://arxiv.org/abs/1309.6835
Shen Y, Ng A Y, Seeger M (2005) Fast gaussian process regression using kd-trees. In: NIPS
Yang C, Duraiswami R, Davis L S (2004) Efficient kernel machines using the improved fast gauss transform. In: NIPS
Beatson R, Greengard L (1997) A short course on fast multipole methods. Wavelets, multilevel methods and elliptic PDEs 1:1–37
Kress R, Maz’ya V, Kozlov V (1989) Linear integral equations, vol 82. Springer
Cunningham J, Ghahramani Z, Rasmussen C E (2012) Gaussian processes for time-marked time-series data. In: Lawrence N D, Girolami M (eds) AISTATS, ser. JMLR Proceedings. JMLR.org, vol 22, pp 255–263
Hastie T, Tibshirani R, Friedman J (2001) The Elements of Statistical Learning, ser. Springer Series in Statistics. New York, NY, USA: Springer New York Inc
Tipping M E (2003) Bayesian inference: An introduction to principles and practice in machine learning. In: Bousquet O, von Luxburg U, Rtsch G (eds) Advanced lectures on machine learning, ser. lecture notes in computer science. http://dblp.uni-trier.de/db/conf/ac/ml2003.html, vol 3176. Springer, pp 41–62
Chowdhary G, Kingravi H, How J, Vela P (2014) Bayesian nonparametric adaptive control using gaussian processes. IEEE Transactions on Neural Networks and Learning Systems PP(99):1–1
Álvarez M A, Lawrence N D (2011) Computationally efficient convolved multiple output gaussian processes. J Mach Learn Res 12:1459–1500. http://dl.acm.org/citation.cfm?id=1953048.2021048
Gallager R (2013) Stochastic Processes: Theory for Applications. Cambridge University Press. http://books.google.co.kr/books?id=CGFbAgAAQBAJ
Muller K, Mika S, Ratsch G, Tsuda K, Scholkopf B (2001) An introduction to kernel-based learning algorithms. IEEE Transactions on Neural Networks 12(2):181–201
Tsionas E G (2012) Maximum likelihood estimation of stochastic frontier models by the fourier transform. J Econ 170 (1):234–248
Bergstra J, Bardenet R, Bengio Y, Kgl B (2011) Algorithms for hyper-parameter optimization. In: Shawe-Taylor J, Zemel R S, Bartlett PL, Pereira F C N, Weinberger K Q (eds) NIPS, pp 2546–2554
Rodner E, Freytag A, Bodesheim P, Denzler J (2012) Large-scale gaussian process classification with flexible adaptive histogram kernels. In: Fitzgibbon A W, Lazebnik S, Perona P, Sato Y, Schmid C (eds) ECCV (4), ser. Lecture Notes in Computer Science, vol 7575. Springer, pp 85–98
Pace R, LeSage J P (2004) Chebyshev approximation of log-determinants of spatial weight matrices. Comput Stat Data Anal 45(2):179–196
Cai T T, Liang T, Zhou HH (2013) Law of log determinant of sample covariance matrix and optimal estimation of differential entropy for high-dimensional gaussian distributions. CoRR. arXiv: http://arxiv.org/abs/1309.0482
Quinonero-Candela J, Rasmussen C E (2005) A unifying view of sparse approximate gaussian process regression. J Mach Learn Res 6:1939–1959
de Baar J, Dwight R, Bijl H (2013) Speeding up kriging through fast estimation of the hyperparameters in the frequency-domain. Comput Geosci 54(0):99–106
Sollich P, Williams C K I (2004). In: Winkler J, Niranjan M, N. D. Lawrence (eds) Understanding gaussian process regression using the equivalent kernel. in Deterministic and Statistical Methods in Machine Learning, ser. Lecture Notes in Computer Science, vol 3635. Springer, pp 211–228. http://dblp.uni-trier.de/db/conf/dsmml/dsmml2004.html
Boyd S, Vandenberghe L (2004) Convex optimization. Cambridge University Press. http://books.google.co.kr/books?id=mYm0bLd3fcoC
Robbins H, Monro S (1951) A stochastic approximation method, The annals of mathematical statistics:400–407
Robbins H, Siegmund D (1985) A convergence theorem for non negative almost supermartingales and some applications, in Herbert Robbins Selected Papers. Springer, pp 111–135
Yang C, Duraiswami R, Gumerov N, Davis L (2003) Improved fast gauss transform and efficient kernel density estimation. In: Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on, pp.664–671 vol.1
Alecu T I, Voloshynovskiy S, Pun T (2005) The gaussian transform. In: EUSIPCO2005, 13th European Signal Processing Conference, pp.4–8
Greengard L, Strain J (1991) The fast gauss transform. SIAM J Sci Stat Comput 12(1):79–94
Yamamoto Y (2006) Efficient parallel implementation of a weather derivatives pricing algorithm based on the fast gauss transform. In: Parallel and Distributed Processing Symposium, 2006. IPDPS 2006. 20th International, pp. 8
Spivak M, Veerapaneni S K, Greengard L (2010) The fast generalized gauss transform. SIAM J Sci Comput 32(5):3092–3107. 10.1137/100790744
Sampath R S, Sundar H, Veerapaneni S K (2010) Parallel fast gauss transform. In: Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, ser. SC ’10. Washington, DC, USA: IEEE Computer Society, pp. 1–10. 10.1109/SC.2010.39
Simoncini V, Szyld D B (2003) Theory of inexact krylov subspace methods and applications to scientific computing. SIAM J Sci Comput 25(2):454–477
Greengard L, Rokhlin V (1987) A fast algorithm for particle simulations. J Comp Physiol 73(2):325–348
Morariu V I, Srinivasan B V, Raykar V C, Duraiswami R, Davis L S (2008) Automatic online tuning for fast gaussian summation. In: Koller D, Schuurmans D, Bengio Y, Bottou L (eds) NIPS. Curran Associates, Inc, pp 1113–1120
Fangohr H (2004) A comparison of c, matlab, and python as teaching languages in engineering. In: Computational Science-ICCS 2004. Springer, pp 1210–1217
Wu X (1999) Performance Evaluation, Prediction and Visualization of Parallel Systems, ser. The International Series on Asian Studies in Computer and Information Science. Springer, US. http://books.google.co.kr/books?id=IJZt5H6R8OIC
Acknowledgments
This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) NRF-2014R1A2A2A01003914.
This work was supported by the IT R&D program of MSIP/IITP. [14-000-09-001, Development of General-Purpose OS and Virtualization Technology to Reduce 30 % of Energy for High-density Servers based on Low-power Processors].
Author information
Authors and Affiliations
Corresponding author
Additional information
Conflict of interests
The authors declare that they have no conflict of interests.
Compliance with Ethical Standards
Research involving Human Participants and/or Animals. The authors declare that this research does not involve any Human Participants and/or Animals.
Rights and permissions
About this article
Cite this article
Bui, DM., Nguyen, HQ., Yoon, Y. et al. Gaussian process for predicting CPU utilization and its application to energy efficiency. Appl Intell 43, 874–891 (2015). https://doi.org/10.1007/s10489-015-0688-4
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-015-0688-4