Skip to main content

Advertisement

Log in

Gaussian process for predicting CPU utilization and its application to energy efficiency

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

For the past ten years, Gaussian process has become increasingly popular for modeling numerous inferences and reasoning solutions due to the robustness and dynamic features. Particularly concerning regression and classification data, the combination of Gaussian process and Bayesian learning is considered to be one of the most appropriate supervised learning approaches in terms of accuracy and tractability. However, due to the high complexity in computation and data storage, Gaussian process performs poorly when processing large input dataset. Because of the limitation, this method is ill-equipped to deal with the large-scale system that requires reasonable precision and fast reaction rate. To improve the drawback, our research focuses on a comprehensive analysis of Gaussian process performance issues, highlighting ways to drastically reduce the complexity of hyper-parameter learning and training phases, which could be applicable in predicting the CPU utilization in the demonstrated application. In fact, the purpose of this application is to save the energy by distributively engaging the Gaussian process regression to monitor and predict the status of each computing node. Subsequently, a migration mechanism is applied to migrate the system-level processes between multi-core and turn off the idle one in order to reduce the power consumption while still maintaining the overall performance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Lawrence N D (2004) Gaussian process latent variable models for visualisation of high dimensional data. Advances in neural information processing systems 16:329–336

    MathSciNet  Google Scholar 

  2. Rasmussen CE (1997) Evaluation of gaussian processes and other methods for non-linear regression, Ph.D. dissertation, Toronto, Ont., Canada, Canada, aAINQ28300

  3. Chalupka K, Williams C K I, Murray I (2012) A framework for evaluating approximation methods for gaussian process regression, CoRR. arXiv: http://arxiv.org/abs/1205.6326

  4. Rasmussen C, Williams C (2005) Gaussian processes for machine learning, ser. adaptive computation and machine learning, MIT Press. http://www.gaussianprocess.org/gpml/chapters/

  5. Brahim-Belhouari S, Vesin J (2001) Bayesian learning using gaussian process for time series prediction. In: Statistical Signal Processing, 2001. Proceedings of the 11th IEEE Signal Processing Workshop on, pp 433–436

  6. Roberts S, Osborne M, Ebden M, Reece S, Gibson N, Aigrain S (2012) Gaussian processes for time-series modeling. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 371(1984)

  7. Petelin D, Filipič B, Kocijan J Optimization of gaussian process models with evolutionary algorithms. In: Proceedings of the 10th International Conference on Adaptive and Natural Computing Algorithms - Volume Part I, ser. ICANNGA’11. Berlin, Heidelberg: Springer-Verlag, 2011, pp. 420–429. [Online]. Available:, http://dl.acm.org/citation.cfm?id=1997052.1997098

  8. Petelin D, Kocijan J. (2014) Evolving gaussian process models for predicting chaotic time-series. In: Evolving and Adaptive Intelligent Systems (EAIS), 2014 IEEE Conference on. IEEE, pp. 1–8.

  9. Shewchuk J R (1994) An introduction to the conjugate gradient method without the agonizing pain, Pittsburgh, PA, USA, Tech. Rep

  10. Grande R, Chowdhary G, How J (2013) Nonparametric adaptive control using gaussian processes with online hyperparameter estimation. In: 2013 IEEE 52nd annual conference on decision and control (CDC), , pp. 861867

  11. Banerjee A, Dunson D, Tokdar S (2011) Efficient gaussian process regression for large data sets. arXiv: http://arxiv.org/abs/e-prints

  12. Hensman J, Fusi N, Lawrence N D (2013) Gaussian processes for big data CoRR. arXiv: http://arxiv.org/abs/1309.6835

  13. Shen Y, Ng A Y, Seeger M (2005) Fast gaussian process regression using kd-trees. In: NIPS

  14. Yang C, Duraiswami R, Davis L S (2004) Efficient kernel machines using the improved fast gauss transform. In: NIPS

  15. Beatson R, Greengard L (1997) A short course on fast multipole methods. Wavelets, multilevel methods and elliptic PDEs 1:1–37

    MathSciNet  Google Scholar 

  16. Kress R, Maz’ya V, Kozlov V (1989) Linear integral equations, vol 82. Springer

  17. Cunningham J, Ghahramani Z, Rasmussen C E (2012) Gaussian processes for time-marked time-series data. In: Lawrence N D, Girolami M (eds) AISTATS, ser. JMLR Proceedings. JMLR.org, vol 22, pp 255–263

  18. Hastie T, Tibshirani R, Friedman J (2001) The Elements of Statistical Learning, ser. Springer Series in Statistics. New York, NY, USA: Springer New York Inc

  19. Tipping M E (2003) Bayesian inference: An introduction to principles and practice in machine learning. In: Bousquet O, von Luxburg U, Rtsch G (eds) Advanced lectures on machine learning, ser. lecture notes in computer science. http://dblp.uni-trier.de/db/conf/ac/ml2003.html, vol 3176. Springer, pp 41–62

  20. Chowdhary G, Kingravi H, How J, Vela P (2014) Bayesian nonparametric adaptive control using gaussian processes. IEEE Transactions on Neural Networks and Learning Systems PP(99):1–1

  21. Álvarez M A, Lawrence N D (2011) Computationally efficient convolved multiple output gaussian processes. J Mach Learn Res 12:1459–1500. http://dl.acm.org/citation.cfm?id=1953048.2021048

    MATH  MathSciNet  Google Scholar 

  22. Gallager R (2013) Stochastic Processes: Theory for Applications. Cambridge University Press. http://books.google.co.kr/books?id=CGFbAgAAQBAJ

  23. Muller K, Mika S, Ratsch G, Tsuda K, Scholkopf B (2001) An introduction to kernel-based learning algorithms. IEEE Transactions on Neural Networks 12(2):181–201

    Article  Google Scholar 

  24. Tsionas E G (2012) Maximum likelihood estimation of stochastic frontier models by the fourier transform. J Econ 170 (1):234–248

    Article  MathSciNet  Google Scholar 

  25. Bergstra J, Bardenet R, Bengio Y, Kgl B (2011) Algorithms for hyper-parameter optimization. In: Shawe-Taylor J, Zemel R S, Bartlett PL, Pereira F C N, Weinberger K Q (eds) NIPS, pp 2546–2554

  26. Rodner E, Freytag A, Bodesheim P, Denzler J (2012) Large-scale gaussian process classification with flexible adaptive histogram kernels. In: Fitzgibbon A W, Lazebnik S, Perona P, Sato Y, Schmid C (eds) ECCV (4), ser. Lecture Notes in Computer Science, vol 7575. Springer, pp 85–98

  27. Pace R, LeSage J P (2004) Chebyshev approximation of log-determinants of spatial weight matrices. Comput Stat Data Anal 45(2):179–196

    Article  MATH  MathSciNet  Google Scholar 

  28. Cai T T, Liang T, Zhou HH (2013) Law of log determinant of sample covariance matrix and optimal estimation of differential entropy for high-dimensional gaussian distributions. CoRR. arXiv: http://arxiv.org/abs/1309.0482

  29. Quinonero-Candela J, Rasmussen C E (2005) A unifying view of sparse approximate gaussian process regression. J Mach Learn Res 6:1939–1959

    MATH  MathSciNet  Google Scholar 

  30. de Baar J, Dwight R, Bijl H (2013) Speeding up kriging through fast estimation of the hyperparameters in the frequency-domain. Comput Geosci 54(0):99–106

    Article  Google Scholar 

  31. Sollich P, Williams C K I (2004). In: Winkler J, Niranjan M, N. D. Lawrence (eds) Understanding gaussian process regression using the equivalent kernel. in Deterministic and Statistical Methods in Machine Learning, ser. Lecture Notes in Computer Science, vol 3635. Springer, pp 211–228. http://dblp.uni-trier.de/db/conf/dsmml/dsmml2004.html

  32. Boyd S, Vandenberghe L (2004) Convex optimization. Cambridge University Press. http://books.google.co.kr/books?id=mYm0bLd3fcoC

  33. Robbins H, Monro S (1951) A stochastic approximation method, The annals of mathematical statistics:400–407

  34. Robbins H, Siegmund D (1985) A convergence theorem for non negative almost supermartingales and some applications, in Herbert Robbins Selected Papers. Springer, pp 111–135

  35. Yang C, Duraiswami R, Gumerov N, Davis L (2003) Improved fast gauss transform and efficient kernel density estimation. In: Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on, pp.664–671 vol.1

  36. Alecu T I, Voloshynovskiy S, Pun T (2005) The gaussian transform. In: EUSIPCO2005, 13th European Signal Processing Conference, pp.4–8

  37. Greengard L, Strain J (1991) The fast gauss transform. SIAM J Sci Stat Comput 12(1):79–94

    Article  MATH  MathSciNet  Google Scholar 

  38. Yamamoto Y (2006) Efficient parallel implementation of a weather derivatives pricing algorithm based on the fast gauss transform. In: Parallel and Distributed Processing Symposium, 2006. IPDPS 2006. 20th International, pp. 8

  39. Spivak M, Veerapaneni S K, Greengard L (2010) The fast generalized gauss transform. SIAM J Sci Comput 32(5):3092–3107. 10.1137/100790744

  40. Sampath R S, Sundar H, Veerapaneni S K (2010) Parallel fast gauss transform. In: Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, ser. SC ’10. Washington, DC, USA: IEEE Computer Society, pp. 1–10. 10.1109/SC.2010.39

  41. Simoncini V, Szyld D B (2003) Theory of inexact krylov subspace methods and applications to scientific computing. SIAM J Sci Comput 25(2):454–477

    Article  MATH  MathSciNet  Google Scholar 

  42. Greengard L, Rokhlin V (1987) A fast algorithm for particle simulations. J Comp Physiol 73(2):325–348

    Article  MATH  MathSciNet  Google Scholar 

  43. Morariu V I, Srinivasan B V, Raykar V C, Duraiswami R, Davis L S (2008) Automatic online tuning for fast gaussian summation. In: Koller D, Schuurmans D, Bengio Y, Bottou L (eds) NIPS. Curran Associates, Inc, pp 1113–1120

  44. Fangohr H (2004) A comparison of c, matlab, and python as teaching languages in engineering. In: Computational Science-ICCS 2004. Springer, pp 1210–1217

  45. Wu X (1999) Performance Evaluation, Prediction and Visualization of Parallel Systems, ser. The International Series on Asian Studies in Computer and Information Science. Springer, US. http://books.google.co.kr/books?id=IJZt5H6R8OIC

Download references

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) NRF-2014R1A2A2A01003914.

This work was supported by the IT R&D program of MSIP/IITP. [14-000-09-001, Development of General-Purpose OS and Virtualization Technology to Reduce 30 % of Energy for High-density Servers based on Low-power Processors].

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dinh-Mao Bui.

Additional information

Conflict of interests

The authors declare that they have no conflict of interests.

Compliance with Ethical Standards

Research involving Human Participants and/or Animals. The authors declare that this research does not involve any Human Participants and/or Animals.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bui, DM., Nguyen, HQ., Yoon, Y. et al. Gaussian process for predicting CPU utilization and its application to energy efficiency. Appl Intell 43, 874–891 (2015). https://doi.org/10.1007/s10489-015-0688-4

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-015-0688-4

Keywords

Navigation