Abstract
Twenty-five machine learning (ML) methods and ordinary least squares regression (OLS) are trained to detect in-sample U.S. annual inflation rates up to a year in advance. The FRED-MD monthly dataset with 134 economic and financial variables from 1959 to April 2022 is used for training, validation, and forecasting. Out of these twenty-five ML methods, top-ten (by root mean square error or RMSE) are chosen to forecast the out-of-sample annual inflation rate. The ML methods are more accurate than the OLS in forecasting the annual inflation rate. The OLS does not appear in the top-10 list in any forecasting period. The ML methods robustly classify the labor market as the top factor in forecasting inflation. The labor market has a significantly higher impact on inflation than the housing or stock market.
Similar content being viewed by others
Notes
All 134 variables: https://s3.amazonaws.com/real.stlouisfed.org/wp/2015/2015-012.pdf.
Definition and details of CPIAUCSL from the FRED website: https://fred.stlouisfed.org/series/CPIAUCSL.
List of ML methods at scikit-learn: https://scikit-learn.org/stable/index.html.
Ensemble methods: https://en.wikipedia.org/wiki/Ensemble_learning.
25 ML regression methods: https://www.mathworks.com/help/stats/choose-regression-model-options.html.
References
Abiodun, O. I., Jantan, A., Omolara, A. E., Dada, K. V., Mohamed, N. A., & Arshad, H. (2018). State-of-the-art in artificial neural network applications: A survey. Heliyon, 4(11), e00938. https://doi.org/10.1016/j.heliyon.2018.e00938
Abramowitz, M., & Stegun, I. A. (1965). Handbook of mathematical functions (Vol. 361). Dover Publications.
Agrawal, A., Gans, J., & Goldfarb, A. (2019). The economics of artificial intelligence: An agenda. University of Chicago Press. https://doi.org/10.7208/chicago/9780226613475.001.0001
Alvarez, F., Lucas, R. E., & Weber, W. E. (2001). Interest rates and inflation. American Economic Review, 91(2), 219–225. https://doi.org/10.1257/aer.91.2.219
Athey, S. (2018). The impact of machine learning on economics (No. c14009). National Bureau of Economic Research. Retrieved from, https://www.nber.org/books-and-chapters/economics-artificial-intelligence-agenda/impact-machine-learning-economics
Athey, S., & Imbens, G. W. (2019). Machine learning methods that economists should know about. Annual Review of Economics, 11(1), 685–725. https://doi.org/10.1146/annurev-economics-080217-053433
Bai, J., & Ng, S. (2008). Forecasting economic time series using targeted predictors. Journal of Econometrics, 146(2), 304–317. https://doi.org/10.1016/j.jeconom.2008.08.010
Bennett, J., & Lanning, S. (2007). The Netflix prize. In Proceedings of KDD cup and workshop, 2007, 35
Bergmeir, C., & Benítez, J. M. (2012). On the use of cross-validation for time series predictor evaluation. Information Sciences, 191, 192–213. https://doi.org/10.1016/j.ins.2011.12.028
Bishop, C. M., & Nasrabadi, N. M. (2006). Pattern recognition and machine learning (Vol. 4). Springer.
Breiman, L. (1996). Bagging predictors. Machine Learning, 24(2), 123–140.
Breiman, L. (2001). Random forests. Machine Learning, 45, 5–32.
Breiman, L., Friedman, J. H., Olshen, R. A., & Stone, C. J. (2017). Classification and regression trees. Routledge. https://doi.org/10.1201/9781315139470
Bresnahan, T. F., & Trajtenberg, M. (1995). General purpose technologies ‘engines of growth’? Journal of Econometrics, 65(1), 83–108. https://doi.org/10.1016/0304-4076(94)01598-T
Bruno, M., & Easterly, W. (1998). Inflation crises and long-run growth. Journal of Monetary Economics, 41(1), 3–26. https://doi.org/10.1016/S0304-3932(97)00063-9
Chollet, F., & Allaire, J. J. (2018). Deep learning with R. Manning Publications Co.
Clark, D., Schreter, Z., & Adams, A. (1996). A quantitative comparison of dystal and backpropagation. In Australian conference on neural networks.
de Prado, M. (2020). Elements in quantitative finance. Cambridge University Press. https://doi.org/10.1017/9781108883658
Domingos, P. (2012). A few useful things to know about machine learning. Communications of the ACM, 55(10), 78–87. https://doi.org/10.1145/2347736.2347755
Dornbusch, R. (1985). Inflation, exchange rates and stabilization (Working Paper No. 1739; working paper series). National Bureau of Economic Research. https://doi.org/10.3386/w1739
Dumouchel, W., & O’Brien, F. (1989). Integrating a robust option into a multiple regression computing environment. In Computer science and statistics: Proceedings of the 21st symposium on the interface (pp. 297–302).
Edwards, S. (2006). The relationship between exchange rates and inflation targeting revisited (Working paper no. 12163; working paper series). National Bureau of Economic Research. https://doi.org/10.3386/w12163
Efron, B., & Morris, C. (1977). Stein’s paradox in statistics. Scientific American, 236(5), 119–127.
Elliott, G., Gargano, A., & Timmermann, A. (2013). Complete subset regressions. Journal of Econometrics, 177(2), 357–373. https://doi.org/10.1016/j.jeconom.2013.04.017
Elliott, G., & Timmermann, A. (2008). Economic forecasting. Journal of Economic Literature, 46(1), 3–56. https://doi.org/10.1257/jel.46.1.3
Fan, R.-E., Chen, P.-H., Lin, C.-J., & Joachims, T. (2005). Working set selection using second order information for training support vector machines. Journal of Machine Learning Research, 6(12), 1889.
Freund, Y., & Schapire, R. E. (1997). A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences, 55(1), 119–139.
Friedman, J. H. (2001). Greedy function approximation: A gradient boosting machine. Annals of Statistics, 29, 1189–1232.
Friedman, M. (1963). Inflation: Causes and consequences. Asia Publishing House.
Friedman, M. (1977). Nobel lecture: Inflation and unemployment. Journal of Political Economy, 85(3), 451–472. https://doi.org/10.1086/260579
Fuentes, M., & Raftery, A. E. (2005). Model evaluation and spatial interpolation by Bayesian combination of observations with outputs from numerical models. Biometrics, 61(1), 36–45. https://doi.org/10.1111/j.0006-341X.2005.030821.x
Glorot, X., & Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics (pp. 249–256).
Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning. MIT press.
Görtler, J., Kehlbeck, R., & Deussen, O. (2019). A visual exploration of Gaussian processes. Distill, 4(4), e17. https://doi.org/10.23915/distill.00017
Guyon, I., & Elisseeff, A. (2003). An introduction to variable and feature selection. Journal of Machine Learning Research, 3(Mar), 1157–1182.
Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements of statistical learning: Data mining, inference, and prediction (2nd ed.). Springer.
He, K., Zhang, X., Ren, S., & Sun, J. (2015). Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision (pp. 1026–1034).
Holland, P. W., & Welsch, R. E. (1977). Robust regression using iteratively reweighted least-squares. Communications in Statistics-Theory and Methods, 6(9), 813–827.
Huang, P.-S., Avron, H., Sainath, T. N., Sindhwani, V., & Ramabhadran, B. (2014). Kernel methods match deep neural networks on timit. In 2014 IEEE international conference on acoustics, speech and signal processing (ICASSP) (pp. 205–209).
Huber, P. J. (1981). Robust statistics. Wiley.
Hull, I. (2021). Machine learning for economics and finance in TensorFlow 2. Apress.
Inoue, A., & Kilian, L. (2008). How useful is bagging in forecasting economic time series? A case study of U.S. consumer price inflation. Journal of the American Statistical Association, 103(482), 511–522. https://doi.org/10.1198/016214507000000473
In-sample vs. out-of-sample forecasts. (2015, February 9). Retrieved from, https://ec.europa.eu/eurostat/statistics-explained/index.php?title=Glossary:In-sample_vs._out-of-sample_forecasts
James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning (Vol. 112). Springer.
Kecman, V., Huang, T.-M., & Vogt, M. (2005). Iterative single data algorithm for training kernel machines from huge data sets: Theory and performance. In L. Wang (Ed.), Support vector machines: Theory and applications (pp. 255–274). Springer. https://doi.org/10.1007/10984697_12
Keynes, J. M. (1936). The general theory of employment, interest and money. Macmillan and Company. https://www.jstor.org/stable/2278703
Kleinberg, J., Ludwig, J., Mullainathan, S., & Obermeyer, Z. (2015). Prediction policy problems. The American Economic Review, 105(5), 491–495. https://doi.org/10.1257/aer.p20151023
Lagarias, J. C., Reeds, J. A., Wright, M. H., & Wright, P. E. (1998). Convergence properties of the Nelder—Mead simplex method in low dimensions. SIAM Journal on Optimization, 9(1), 112–147. https://doi.org/10.1137/S1052623496303470
Le, Q., Sarlós, T., & Smola, A. (2013). Fastfood-approximating kernel expansions in loglinear time. In Proceedings of the international conference on machine learning, 85, 8.
Lichman, M. (2013). UCI machine learning repository. Retrieved from, http://Archive.Ics.Uci.Edu/Ml
Loh, W.-Y. (2002). Regression tress with unbiased variable selection and interaction detection. Statistica Sinica, 12(2), 361–386.
Loh, W.-Y., & Shih, Y.-S. (1997). Split selection methods for classification trees. Statistica Sinica, 7(4), 815–840.
Malladi, R. K. (2022). Application of supervised machine learning techniques to forecast the COVID-19 U.S. recession and stock market crash. Computational Economics. https://doi.org/10.1007/s10614-022-10333-8
Mankiw, N. G., & Reis, R. (2002). Sticky information versus sticky prices: A proposal to replace the new Keynesian Phillips curve*. The Quarterly Journal of Economics, 117(4), 1295–1328. https://doi.org/10.1162/003355302320935034
Mankiw, N. G., Reis, R., & Wolfers, J. (2003). Disagreement about Inflation Expectations. NBER Macroeconomics Annual, 18, 209–248. https://doi.org/10.1086/ma.18.3585256
McCallum, B. T. (1990). Chapter 18 Inflation: Theory and evidence. Handbook of monetary economics (Vol. 2, pp. 963–1012). Elsevier. https://doi.org/10.1016/S1573-4498(05)80025-6
McCracken, M. W., & Ng, S. (2016). FRED-MD: A monthly database for macroeconomic research. Journal of Business & Economic Statistics, 34(4), 574–589. https://doi.org/10.1080/07350015.2015.1086655
Medeiros, M. C., Vasconcelos, G. F. R., Veiga, Á., & Zilberman, E. (2021). Forecasting inflation in a data-rich environment: The benefits of machine learning methods. Journal of Business & Economic Statistics, 39(1), 98–119. https://doi.org/10.1080/07350015.2019.1637745
Mitchel, T. (1997). Machine learning. McGraw-Hill.
Mullainathan, S., & Spiess, J. (2017). Machine learning: An applied econometric approach. Journal of Economic Perspectives, 31(2), 87–106. https://doi.org/10.1257/jep.31.2.87
Murphy, K. (2022). Probabilistic machine learning: Advanced topics. MIT Press.
Nash, W. J., Sellers, T. L., Talbot, S. R., Cawthorn, A. J., & Ford, W. B. (1994). The population biology of abalone (Haliotis species) in tasmania. I. Blacklip abalone (H. Rubra) from the north coast and islands of bass strait. Sea Fisheries Division, Technical Report, 48 (p 411.).
Nocedal, J., & Wright, S. J. (1999). Numerical optimization. Springer.
Opitz, D., & Maclin, R. (1999). Popular ensemble methods: An empirical study. Journal of Artificial Intelligence Research, 11, 169–198. https://doi.org/10.1613/jair.614
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., & Dubourg, V. (2011). Scikit-learn: Machine learning in Python. The Journal of Machine Learning Research, 12, 2825–2830.
Phelps, E. S. (1967). Phillips curves, expectations of inflation and optimal unemployment over Time. Economica, 34(135), 254–281. https://doi.org/10.2307/2552025
Phillips, A. W. (1958). The relation between unemployment and the rate of change of money wage rates in the United Kingdom, 1861–1957. Economica, 25(100), 283–299. https://doi.org/10.2307/2550759
Polikar, R. (2006). Ensemble based systems in decision making. IEEE Circuits and Systems Magazine, 6(3), 21–45. https://doi.org/10.1109/MCAS.2006.1688199
Rabinowicz, A., & Rosset, S. (2020). Cross-validation for correlated data. Journal of the American Statistical Association. https://doi.org/10.1080/01621459.2020.1801451
Racine, J. (2000). Consistent cross-validatory model-selection for dependent data: Hv-block cross-validation. Journal of Econometrics, 99(1), 39–61. https://doi.org/10.1016/S0304-4076(00)00030-0
Rahimi, A., & Recht, B. (2007). Random features for large-scale kernel machines. In Advances in neural information processing systems, 20.
Rasmussen, C., & Williams, C. (2006). Gaussian processes for machine learning. MIT Press.
Rokach, L. (2010). Ensemble-based classifiers. Artificial Intelligence Review, 33(1), 1–39. https://doi.org/10.1007/s10462-009-9124-7
Stein, C. (1956). Inadmissibility of the usual estimator for the mean of a multivariate normal distribution. In Proceedings of the third Berkeley symposium on mathematical statistics and probability, volume 1: Contributions to the theory of statistics. Retrieved from, https://projecteuclid.org/euclid.bsmsp/1200501656
Stigler, S. M. (1981). Gauss and the invention of least squares. The Annals of Statistics, 9(3), 465–474. https://doi.org/10.1214/aos/1176345451
Stock, J. H., & Watson, M. W. (1999). Forecasting inflation. Journal of Monetary Economics, 44(2), 293–335. https://doi.org/10.1016/S0304-3932(99)00027-6
Stock, J. H., & Watson, M. W. (2006a). Chapter 10 Forecasting with many predictors. In G. Elliott, C. W. J. Granger, & A. Timmermann (Eds.), Handbook of economic forecasting (Vol. 1, pp. 515–554). Elsevier. https://doi.org/10.1016/S1574-0706(05)01010-4
Street, J. O., Carroll, R. J., & Ruppert, D. (1988). A note on computing robust regression estimates via iteratively reweighted least squares. The American Statistician, 42(2), 152–154. https://doi.org/10.1080/00031305.1988.10475548
Taylor, J. B. (1993). Discretion versus policy rules in practice. Carnegie-Rochester Conference Series on Public Policy, 39, 195–214. https://doi.org/10.1016/0167-2231(93)90009-L
Varian, H. R. (2014a). Beyond big data. Business Economics, 49(1), 27–31. https://doi.org/10.1057/be.2014.1
Varian, H. R. (2014b). Big data: New tricks for econometrics. Journal of Economic Perspectives, 28(2), 3–28. https://doi.org/10.1257/jep.28.2.3
Wai, U. T. (1959). The relation between inflation and economic development: A statistical inductive study. Staff Papers (international Monetary Fund), 7(2), 302–317. https://doi.org/10.2307/3866244
Waugh, S. G. (1995). Extending and benchmarking cascade-correlation: Extensions to the cascade-correlation architecture and benchmarking of feed-forward supervised artificial neural networks. Ph.D., University of Tasmania. https://eprints.utas.edu.au/21965/
Zhang, N., Xiong, J., Zhong, J., & Leatham, K. (2018). Gaussian process regression method for classification for high-dimensional data with limited samples. In 2018 eighth international conference on information science and technology (ICIST) (pp. 358–363).
Funding
This study is not funded by any entity.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Authors do not have any conflicts of interest associated with this article.
Ethical Approval
This article does not contain any studies with human participants or animals.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Malladi, R.K. Benchmark Analysis of Machine Learning Methods to Forecast the U.S. Annual Inflation Rate During a High-Decile Inflation Period. Comput Econ (2023). https://doi.org/10.1007/s10614-023-10436-w
Accepted:
Published:
DOI: https://doi.org/10.1007/s10614-023-10436-w