Abstract
Modern regularization techniques are able to select the relevant variables and features in prediction problems where much more predictors than observations are available. We investigate how regularization methods can be used to select the influential predictors of an autoregressive model with a very large number of potentially informative predictors. The methods are used to forecast the quarterly gross added value in the manufacturing sector by use of the business survey data collected by the ifo Institute. Also ensemble methods, which combine several forecasting methods are exemplarily evaluated.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Armstrong, J. (2001). Combining forecasts. International Series in Operations Research and Management Science, 30, 417–440.
Armstrong, J. S., & Collopy, F. (1992). Error measures for generalizing about forecasting methods: Empirical comparisons. International Journal of Forecasting, 8, 69–80.
Bai, J., & Ng, S. (2002). Determining the number of factors in approximate factor models. Econometrica, 70(1), 191–221.
Bai, J., & Ng, S. (2008). Forecasting economic time series using targeted predictors. Journal of Econometrics, 146(2), 304–317.
Bai, J., & Ng, S. (2009). Boosting diffusion indices. Journal of Applied Econometrics, 24(4), 607–629.
Becker, S. O., & Wohlrabe, K. (2007). Micro Data at the Ifo Institute for Economic Research - The “Ifo Business Survey”, Usage and Access. Ifo Working Paper Series Ifo Working Paper No. 47, Ifo Institute for Economic Research at the University of Munich.
Buchen, T., & Wohlrabe, K. (2011). Forecasting with many predictors: Is boosting a viable alternative? Economics Letters, 113(1), 16–18.
Bühlmann, P., & Hothorn, T. (2007). Boosting algorithms: Regularization, prediction and model fitting (with discussion). Statistical Science, 22, 477–505.
Bühlmann, P., & Yu, B. (2003). Boosting with the L2 loss: Regression and classification. Journal of the American Statistical Association, 98, 324–339.
Clemen, R. T. (1989). Combining forecasts: A review and annotated bibliography. International Journal of Forecasting, 5(4), 559–583.
Clements, M. P., & Harvey, D. I. (2009). Forecast combination and encompassing. In Palgrave handbook of econometrics. Volume 2: Applied econometrics (pp. 169–198). London: Palgrave Macmillan.
De Mol, C., Giannone, D., & Reichlin, L. (2008). Forecasting using a large number of predictors: Is bayesian shrinkage a valid alternative to principal components? Journal of Econometrics, 146(2), 318–328.
Diebold, F. X., & Mariano, R. S. (1995). Comparing predictive accuracy. Journal of Business & Economic Statistics, 13, 253–263.
Efron, B., Hastie, T., Johnstone, I., & Tibshirani, R. (2004). Least angle regression. Annals of Statistics, 32, 407–499.
Feng, Y., & Heiler, S. (1998). Locally weighted autoregression. In Econometrics in theory and practice (pp. 101–117). Heidelberg: Springer.
Freund, Y., & Schapire, R. E. (1996). Experiments with a new boosting algorithm. In Proceedings of the Thirteenth International Conference on Medicine Learning (pp. 148–156). San Francisco: Morgan Kaufmann.
Friedman, J. H. (2012). Fast sparse regression and classification. International Journal of Forecasting, 28(3), 722–738.
Friedman, J. H., Hastie, T., & Tibshirani, R. (2010). Regularization paths for generalized linear models via coordinate descent. Journal of Statistical Software, 33(1), 1–22.
Goeman, J. J. (2010). L1 penalized estimation in the Cox proportional hazards model. Biometrical Journal, 52, 70–84.
Harvey, D. I., Leybourne, S. J., & Newbold, P. (1998). Tests for forecast encompassing. Journal of Business & Economic Statistics, 16, 254–259.
Hastie, T., Tibshirani, R., & Friedman, J. H. (2001). The elements of statistical learning. New York: Springer.
Hastie, T., Tibshirani, R., & Friedman, J. H. (2009). The elements of statistical learning (2nd ed.). New York: Springer.
Heiler, S., & Feng, Y. (2000). Data-driven decomposition of seasonal time series. Journal of Statistical Planning and Inference, 91(2), 351–363.
Hoerl, A. E., & Kennard, R. W. (1970). Ridge regression: Bias estimation for nonorthogonal problems. Technometrics, 12, 55–67.
Kendall, M. G. (1957). A course in multivariate analysis. New York: Hafner Pub. Co.
Kisinbay, T. (2010). The use of encompassing tests for forecast combinations. Journal of Forecasting, 29, 715–727.
Meier, L. (2009). grplasso: Fitting user specified models with Group Lasso penalty. R package version 0.4-2.
Meier, L., van de Geer, S., & Bühlmann, P. (2008). The group lasso for logistic regression. Journal of the Royal Statistical Society, Series B, 70, 53–71.
Mevik, B.-H., Wehrens, R., & Liland, K. H. (2011). pls: Partial Least Squares and Principal Component regression. R package version 2.3-0.
Robinzonov, N., Tutz, G., & Hothorn, T. (2012). Boosting techniques for nonlinear time series models. AStA Advances in Statistical Analysis, 96, 99–122.
Shafik, N., & Tutz, G. (2009). Boosting nonlinear additive autoregressive time series. Computational Statistics and Data Analysis, 53, 2453–2464.
Stock, J. H., & Watson, M. W. (2004). Combination forecasts of output growth in a seven-country data set. Journal of Forecasting, 23, 405–430.
Stock, J. H., & Watson, M. W. (2006). Forecasting with many predictors. In C. G. G. Elliott & A. Timmermann (Eds.), Handbook of economic forecasting (Vol. 1, pp. 515–554). Amsterdam: Elsevier.
Stock, J. H., & Watson, M. W. (2011, February). Generalized shrinkage methods for forecasting using many predictors generalized shrinkage methods for forecasting using many predictors. Manuscript, Harvard University, 0-62.
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society B, 58, 267–288.
Wold, H. (1975). Soft Modeling by latent variables; The nonlinear iterative partial least squares approach. Perspectives in probability and statistics. Papers in Honour of M. S. Bartlett. London: Academic Press.
Yuan, M., & Lin, Y. (2006). Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society B, 68, 49–67.
Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society B, 67, 301–320.
Acknowledgements
We thank the Munich ifo Institute for providing the data from the surveys of the ifo Business Climate Index. Especially, we thank Kai Carstensen, Johannes Mayr, Klaus Wohlrabe, Teresa Buchen and Steffen Henzel from the ifo Institute for helpful discussions.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Schauberger, G., Tutz, G. (2015). Regularization Methods in Economic Forecasting. In: Beran, J., Feng, Y., Hebbel, H. (eds) Empirical Economic and Financial Research. Advanced Studies in Theoretical and Applied Econometrics, vol 48. Springer, Cham. https://doi.org/10.1007/978-3-319-03122-4_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-03122-4_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-03121-7
Online ISBN: 978-3-319-03122-4
eBook Packages: Business and EconomicsEconomics and Finance (R0)