Abstract
Forecasting is one of the main goals in time series analysis and it has had a great development in the last decades. In forecasting, the prediction intervals provide additional assessment of the uncertainty compared with a point forecast, which can better guide risk management decisions. The construction of prediction intervals requires fitting a model and the knowledge of the distribution of the observed data, which is typically unknown. Hence, data are usually assumed to follow some hypothetical distribution, and the resulting prediction interval can be adversely affected by departures from that assumption (Thombs and Schucany, J Am Stat Assoc 85:486–492, 1990). For this reason, in the last two decades several works based on free distributions have been proposed as an alternative for the construction of prediction intervals. Some alternatives consist in the sieve bootstrap approach, which assumes that the linear process admits typically an autoregressive AR representation, and it generates “new” realizations from the same model but with the resampled innovations (Alonso et al., J Stat Plan Inference 100:1–11, 2002; Chen et al., J Forecast 30:51–71, 2011). The linear nature of the models has not limited the implementation of the sieve bootstrap methodology in nonlinear models such as GARCH, since the squared returns can also be represented as linear ARMA process (Shumway and Stoffer, Time series analysis and its applications with R examples (2nd ed.). New York: Springer, 2006; Francq and Zakoian, GARCH Models: Structure, statistical inference and financial applications. Chichester: Wiley, 2010; Chen et al., J Forecast 30:51–71, 2011).
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Alonso, A., Peña, D., & Romo, J. (2002). Forecasting time series with sieve bootstrap. Journal of Statistical Planning and Inference, 100, 1–11.
Alonso, A., Peña, D., & Romo, J. (2002). Una revisión de los métodos de remuestreo en series temporales. Estadística Española, 44(150), 133–159.
Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal of Econometrics, 31, 307–327.
Bühlmann, P. (1997). Sieve bootstrap for time series. Bernoulli, 2(3), 123–148.
Bühlmann, P. (2002). Bootstrap for time series. Statistical Science, 17(1), 52–72.
Cao, R., Febrero-Bande, M., Gonzalez-Manteiga, W., & Garcia-Jurado, I. (1997). Saving computer time in constructing consistent bootstrap prediction intervals for autoregressive processes. Communications in Statistics—Simulation and Computation, 26, 961–978.
Chen, B., Gel, Y., Balakrishna, N., & Abraham, B. (2011). Computationally efficient bootstrap prediction intervals for returns and volatilities in arch and garch processes. Journal of Forecasting, 30, 51–71.
Davison, A., & Hinkley, D. (1997). Bootstrap methods and their application. Cambridge University Press.
Efron, B., & Tibshirani, R. (1995). An introduction to the bootstrap. Chapman and Hall.
Fox, A. J. (1972). Outliers in time series. Journal of the Royal Statistical Society, 34(B), 350–363.
Francq, C., & Zakoian, J. (2010). GARCH Models: Structure, statistical inference and financial applications. Wiley.
Hall, P. (1992). The bootstrap and Edgeworth expansion. Springer.
Härdle, W. (2003). Bootstrap methods for time series. International Statistical Review, 71, 435–459.
Künsch, H. (1989). The jackknife and the bootstrap for general stationary observations. Statistic, 17, 1217–1241.
Pascual, L., Romo, J., & Ruiz, E. (2004). Bootstrap predictive inference for arima processes. Journal of Time Series Analysis, 25, 449–465.
Pascual, L., Romo, J., & Ruiz, E. (2006). Bootstrap prediction for returns and volatilities in garch models. Computational Statistics and Data Analysis, 50, 2293–2312.
Politis, D. N. (2003). The impact of bootstrap methods on time series analysis. Statistical Science, 28, 219–230.
Shumway, R., & Stoffer, D. (2006). Time series analysis and its applications with R examples (2nd ed.). Springer.
Stine, R. A. (1987). Estimating properties of autoregressive forecasts. Journal of the American Statistical Association, 82(400), 1072–1078.
Thombs, L. A., & Schucany, W. R. (1990). Bootstrap prediction intervals for autoregression. Journal of the American Statistical Association, 85, 486–492.
Tsay, R. (2005). Analysis of financial time series (2nd ed.). Wiley.
Acknowledgements
This work was supported by the following Research Grants: Fondecyt 1110854 and FB0821 Centro Científico Tecnológico de Valparaíso. Partial support was also received from Mecesup FSM 0707. We also want to thank Dr. Ricardo Ñanculef for his important contribution in the realization of this chapter.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Allende, H., Ulloa, G., Allende-Cid, H. (2015). Prediction Intervals in Linear and Nonlinear Time Series with Sieve Bootstrap Methodology. In: Beran, J., Feng, Y., Hebbel, H. (eds) Empirical Economic and Financial Research. Advanced Studies in Theoretical and Applied Econometrics, vol 48. Springer, Cham. https://doi.org/10.1007/978-3-319-03122-4_16
Download citation
DOI: https://doi.org/10.1007/978-3-319-03122-4_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-03121-7
Online ISBN: 978-3-319-03122-4
eBook Packages: Business and EconomicsEconomics and Finance (R0)