Skip to main content

Survey-Based Forecasting: To Average or Not to Average

  • Chapter
  • First Online:
Behavioral Predictive Modeling in Economics

Part of the book series: Studies in Computational Intelligence ((SCI,volume 897))

Abstract

Forecasting inflation rate is of tremendous importance for firms, consumers, as well as monetary policy makers. Besides macroeconomic indicators, professional surveys deliver experts’ expectation and perception of the future movements of the price level. This research studies survey-based inflation forecast in an extended recent sample covering the Great Recession and its aftermath. Traditional methods extract the central tendency in mean or median and use it as a predictor in a simple linear model. Among the three widely cited surveys, we confirm the superior forecasting capability of the Survey of Professional Forecasters (SPF). While each survey consists of many individual experts, we utilize machine learning methods to aggregate the individual information. In addition to the off-the-shelf machine leaning algorithms such as the Lasso, the random forest and the gradient boosting machine (GBM), we tailor the standard Lasso by differentiating the penalty level according to an expert’s experience, in order to handle for participants’ frequent entries and exits in surveys. The tailored Lasso delivers strong empirical results in the SPF and beats all other methods except for the overall best performer, GBM. Combining forecasts of the tailored Lasso model and GBM further achieves the most accurate inflation forecast in both the SPF and the Livingston Survey, which beyonds the reach of a single machine learning algorithm. We conclude that combination of machine learning forecasts is a useful technique to predict inflation, and averaging should be exercised in a new generation of algorithms capable of digesting disaggregated information.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 149.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 199.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 199.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Note that sample periods mentioned in this section would be defined as the periods of data which we retrieved from corresponding sources. However, since data from surveys are denoting one-year-ahead forecasts of the inflation rates, to provide clearer interpretation, we define sample periods as the data’s forecasting periods in implementation. That is, the sample periods of surveys is lagged by one year in all other sections.

  2. 2.

    Available at FRED St. Louis: https://fred.stlouisfed.org/series/CPIAUCSL/.

  3. 3.

    Available at FRED Philadelphia https://www.philadelphiafed.org/research-and-data/real-time-center/survey-of-professional-forecasters/data-files/cpi/.

  4. 4.

    Available at FRED Philadelphia https://www.philadelphiafed.org/research-and-data/real-time-center/livingston-survey/historical-data/.

  5. 5.

    The Regents of the University of Michigan. (2018). Expected Change in Prices During the Next Year. Available at https://data.sca.isr.umich.edu/data-archive/mine.php.

References

  1. Ang, A., Bekaert, G., Wei, M.: Do macro variables, asset markets, or surveys forecast inflation better? J. Monetary Econ. 54, 1163–1212 (2007)

    Article  Google Scholar 

  2. Atkeson, A., Ohanian, L.E., et al.: Are phillips curves useful for forecasting inflation? Fed. Reserve Bank Minneap. Q. Rev. 25, 2–11 (2001)

    Google Scholar 

  3. Belloni, A., Chen, D., Chernozhukov, V., Hansen, C.: Sparse models and methods for optimal instruments with an application to eminent domain. Econometrica 80, 2369–2429 (2012)

    Article  MathSciNet  Google Scholar 

  4. Breiman, L.: Random forests. Mach. Learn. 45, 5–32 (2001)

    Article  Google Scholar 

  5. Caner, M.: Lasso-type GMM estimator. Econom. Theory 25, 270–290 (2009)

    Article  MathSciNet  Google Scholar 

  6. Capistrán, C., Timmermann, A.: Forecast combination with entry and exit of experts. J. Bus. Econ. Stat. 27, 428–440 (2009)

    Article  MathSciNet  Google Scholar 

  7. Carlson, J.A.: A study of price forecasts. In: Annals of Economic and Social Measurement, vol. 6, no. 1, pp. 27–56. NBER (1977)

    Google Scholar 

  8. Diebold, F.X., Shin, M.: Machine learning for regularized survey forecast combination: partially-egalitarian lasso and its derivatives. Int. J. Forecast. 35(4), 1679–1691 (2018)

    Article  Google Scholar 

  9. Friedman, J.H.: Greedy function approximation: a gradient boosting machine. Ann. Stat. 29, 1189–1232 (2001)

    Article  MathSciNet  Google Scholar 

  10. Granger, C.W., Ramanathan, R.: Improved methods of combining forecasts. J. Forecast. 3, 197–204 (1984)

    Article  Google Scholar 

  11. Grant, A.P., Thomas, L.B.: Inflationary expectations and rationality revisited. Econom. Lett. 62, 331–338 (1999)

    Article  Google Scholar 

  12. Lee, J.H., Shi, Z., Gao, Z.: On lasso for predictive regression. arXiv:1810.03140 (2018)

  13. Mehra, Y.P.: Survey measures of expected inflation: revisiting the issues of predictive content and rationality. FRB Richmond Econ. Q. 88, 17–36 (2002)

    Google Scholar 

  14. Nowotarski, J., Raviv, E., Trück, S., Weron, R.: An empirical comparison of alternative schemes for combining electricity spot price forecasts. Energy Econ. 46, 395–412 (2014)

    Article  Google Scholar 

  15. Phillips, P.C., Shi, Z.: Boosting the Hodrick-Prescott filter. arXiv:1905.00175 (2019)

  16. Shi, Z.: Econometric estimation with high-dimensional moment equalities. J. Econom. 195, 104–119 (2016)

    Article  MathSciNet  Google Scholar 

  17. Shi, Z.: Estimation of sparse structural parameters with many endogenous variables. Econom. Rev. 35, 1582–1608 (2016)

    Article  MathSciNet  Google Scholar 

  18. Shi, Z., Huang, J.: Forward-selected panel data approach for program evaluation. arXiv:1908.05894 (2019)

  19. Stock, J.H., Watson, M.W.: Combination forecasts of output growth in a seven-country data set. J. Forecast. 23, 405–430 (2004)

    Article  Google Scholar 

  20. Stock, J.H., Watson, M.W.: Why has us inflation become harder to forecast? J. Money Credit Bank. 39, 3–33 (2007)

    Article  Google Scholar 

  21. Su, L., Shi, Z., Phillips, P.C.: Identifying latent structures in panel data. Econometrica 84, 2215–2264 (2016)

    Article  MathSciNet  Google Scholar 

  22. Thomas, L.B.: Survey measures of expected us inflation. J. Econ. Perspect. 13, 125–144 (1999)

    Article  Google Scholar 

  23. Tibshirani, R.: Regression shrinkage and selection via the lasso. J. Roy. Stat. Soc.: Ser. B (Methodol.) 58, 267–288 (1996)

    MathSciNet  MATH  Google Scholar 

  24. Varma, S., Simon, R.: Bias in error estimation when using cross-validation for model selection. BMC Bioinform. 7, 91 (2006)

    Article  Google Scholar 

  25. Wager, S., Athey, S.: Estimation and inference of heterogeneous treatment effects using random forests. J. Am. Stat. Assoc. 113, 1228–1242 (2018)

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

Zhentao Shi acknowledges financial support from the Research Grants Council (RGC) No. 14500118.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhentao Shi .

Editor information

Editors and Affiliations

Appendix: Cross Validation and Tuning of the Hyperparameters

Appendix: Cross Validation and Tuning of the Hyperparameters

Hyperparameters in models of Sect. 2.1 were tuned within sets of designed grids through a series of nested cross-validation procedure. The nested cross-validation are described in Fig. 7 while the tuning grids of the hyperparameters are displayed in Table 6.

Fig. 7
figure 7

The Nested Cross Validation Procedure \(\bullet \) Notes This graph is for illustration purpose only and the number of data points involved are different from the real data

The nested cross-validation is conducted by both the inner loops and the outer loops, as shown Fig. 7. Given that the outer loop is built by a series of rolling windows to train for optimal parameters, each of the training set is then divided into training subsets for tuning hyperparameters. Optimal model is then selected according to the RMSE. In this study, lengths of training subsets were set to be 80% of the training length. According to Varma and Simon [24] (2006), this nested cross-validation approach estimates the true error with nearly zero bias.

Table 6 lists the grids of the tuning hyperparameters on which we evaluate the machine learning methods in Sect. 2.1. For Lasso, we took exponential for the lambda values \(\lambda \) and \(\lambda _{2}\), in order to be compatible with the standard way to compare the forecasting errors with values of \(\ln \left( \lambda \right) \), instead of \(\lambda \).

Table 6 Tuning Grids of Hyperparameters in models of Sect. 2.1

For our tailored Lasso, the adaptive penalty level \(\lambda _{1}\) is designed to be \(\ln \left( x\right) \). Based on our design, the penalty factor of the model, \(\lambda _{1}\exp (\lambda _{2}/s_{it})\), is more sensitive to an increase in the accumulated number of forecasts \(s_{it}\) when \(s_{it}\) is at a relatively low level. In other words, while forecasters who have accumulated 2 forecasts are penalized much lighter than those who have made merely 1 forecast, the difference between the penalties imposed on forecasters who have accumulated 20 forecasts and 21 forecasts are much smaller. It is intuitive that professional skills improve with experience in a decreasing rate. On the other hand, as SPF consist of quarterly data and the Livingston Survey is of half-yearly data, we set \(\lambda _{1}=\ln (10)\) for the SPF while \(\lambda _{1}=\ln (5)\) for the Livingston Survey. Optimal design of the adaptive factor \(\lambda _{1}\) requires further tuning in the future.

Furthermore, since more trees is in general beneficial to the performance of the random forest algorithm, we skipped its tuning procedure and take the default value of 500. On the other hand, to avoid potential overfitting caused by applying an excessive number of trees on GBM algorithm, we tuned for the number of trees with values from 100 to 500 for a fair comparison between GBM and the random forest.

Since data on individual response is not available for the Survey of Consumers, models in Sect. 2.1 would be applied on the datasets of the Survey of Profession Forecasters and the Livingston Survey, but not the Survey of Consumers. Besides, parameters in models of Sect. 2.1 were tuned through a series of nested cross-validation procedure. This nested cross-validation approach could estimate the true error with nearly zero bias (Varma and Simon [24]).

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Cheng, K., Huang, N., Shi, Z. (2021). Survey-Based Forecasting: To Average or Not to Average. In: Sriboonchitta, S., Kreinovich, V., Yamaka, W. (eds) Behavioral Predictive Modeling in Economics. Studies in Computational Intelligence, vol 897. Springer, Cham. https://doi.org/10.1007/978-3-030-49728-6_5

Download citation

Publish with us

Policies and ethics