Skip to main content

Linking PISA Competencies over Three Cycles – Results from Germany

  • Chapter
  • First Online:

Abstract

Since the publication of the PISA 2006 study results the question of reporting trends over the PISA cycles has received a lot of interest. This chapter discusses the possibilities and limitations of trend analyses based on data from this international comparative study and using complex test designs. The chapter succeeds trend analyses which were carried out with the German data from the first three PISA studies in 2000, 2003 and 2006 (Carstensen CH, Prenzel M, Baumert J, Trendanalysen in PISA: Wie haben sich die Kompetenzen in Deutschland zwischen PISA 2000 und PISA 2006 entwickelt? [Trend analyses in PISA: how did competencies in Germany develop between PISA 2000 and PISA 2006?] Zeitschrift für Erziehungswissenschaften, Sonderheft 10:11–34, 2008; Prenzel M, Artelt C, Baumert J, Blum W, Hammann M, Klieme E et al (eds), PISA 2006. Die Ergebnisse der dritten internationalen Vergleichsstudie [PISA 2006. Results of the third international comparison]. Waxmann, Münster, 2007).

The choice of a scaling and trend analysis model depends on the focus of the analysis and on the assessment design. With respect to international comparisons, very strict assumptions on the uni-dimensionality of the test instruments used have to be made to allow for trend analyses. What if these conditions are not met across all participating countries for all assessment cycles? This paper presents an alternative model for trend analyses, assuming uni-dimensionality only within a particular country but not across all participating countries. Trend results with this model can only be interpreted within the particular country and are not intended for use in international comparisons.

To establish the validity of the presented trend model, an empirical analysis of the different tests and subscales used in different assessment cycles was performed. As far as different versions of the instruments were administered within cycles, the correlations of these test forms give an empirical indication of the uni-dimensionality of the underlying constructs. Monte Carlo simulations were performed to analyse whether the correlations of these test forms indicate a uni-dimensional construct being measured over time. Having analyzed the correlations of the tests, a fit analysis at the item level followed. Further assumptions refer to the stability of item difficulties over time. This is addressed by estimating item by time interaction parameters, allowing for a descriptive analysis of items changing their difficulty over time and a model fit comparison to check whether item drift has an impact on their difficulties.

Results show that trends might be reported for the German data, using the short test for reading and using all pair wise link items in Mathematics and Science. In the conclusion, the results and some implications for the design of future PISA assessments will be discussed.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    Due to deletion of one reading item for the German data set, the short test in the following analysis includes 27 items.

  2. 2.

    Earlier trend reports were restricted to two subscales which were assessed with sufficiently large item numbers.

References

  • Adams, R. J., & Carstensen, C. H. (2002). Scaling outcomes. In R. J. Adams & M. Wu (Eds.), PISA 2000 technical report. Paris: OECD.

    Google Scholar 

  • Adams, R. J., & Wu, M. L. (2002). PISA 2000 technical report. Paris: OECD.

    Google Scholar 

  • Bozdogan, H. (1987). Model selection and Akaike’s information criterion (AIC): The general theory and its analytical extensions. Psychometrika, 52(3), 345–370.

    Article  Google Scholar 

  • Carstensen, C. H, Prenzel, M., & Baumert, J. (2008). Trendanalysen in PISA: Wie haben sich die Kompetenzen in Deutschland zwischen PISA 2000 und PISA 2006 entwickelt? [Trend analyses in PISA: How did competencies in Germany develop between PISA 2000 and PISA 2006?] Zeitschrift für Erziehungswissenschaften, Sonderheft 10/2008, pp. 11–34.

    Google Scholar 

  • Gebhardt, E., & Adams, R. J. (2007). The influence of equating methodology on reported trends in PISA. Journal of Applied Measurement, 8(3), 305–322.

    Google Scholar 

  • Kolen, M. J., & Brennan, R. L. (2004). Test equating, scaling and linking. New York: Springer.

    Google Scholar 

  • Mazzeo, J., & von Davier, M. (2008). Review of the Programme for International Student Assessment (PISA) test design: Recommendations for fostering stability in assessment results. Princeton, NJ: ETS.

    Google Scholar 

  • Masters, G. N. (1982). A Rasch model for partial credit scoring. Psychometrika, 47, 149–174.

    Google Scholar 

  • Monseur, C., & Berezner, A. (2007). The computation of equating errors in international surveys in education. Journal of Applied Measurement, 8(3), 323–335.

    Google Scholar 

  • OECD. (2003). The PISA 2003 assessment framework – Mathematics, reading, science and problem solving knowledge and skills. Paris: OECD.

    Google Scholar 

  • OECD. (2004). Learning for tomorrow’s world: First results from PISA 2003. Paris: OECD.

    Google Scholar 

  • OECD. (2005). PISA 2003 technical report. Paris: OECD.

    Book  Google Scholar 

  • OECD. (2006). Assessing scientific, reading and mathematical literacy. A framework for PISA 2006. Paris: OECD.

    Book  Google Scholar 

  • OECD. (2007). PISA 2006: Science competencies for tomorrow’s world (Analysis, Vol. 1). Paris: OECD.

    Google Scholar 

  • OECD. (2009). PISA 2006 technical report. Paris: OECD.

    Book  Google Scholar 

  • Prenzel, M., Artelt, C., Baumert, J., Blum, W., Hammann, M., Klieme, E., et al. (Eds.). (2007). PISA 2006. Die Ergebnisse der dritten internationalen Vergleichsstudie [PISA 2006. Results of the third international comparison]. Münster, Germany: Waxmann.

    Google Scholar 

  • Prenzel, M., Artelt, C., Baumert, J., Blum, W., Hammann, M., Klieme, E., et al. (Eds.). (2008). PISA 2006. Die Kompetenzen der Jugendlichen im dritten Ländervergleich. Münster, Germany: Waxmann.

    Google Scholar 

  • Wright, B. D., & Masters, G. N. (1982). Rating scale analysis: Rasch measurement. Chicago: University Press.

    Google Scholar 

  • Wu, M., Adams, R. J., Wilson, M., & Haldane, S. (2007). Conquest 2.0. Camberwell, Australia: ACER Press.

    Google Scholar 

  • Xu, X., & von Davier, M. (2008). Linking for the general diagnostic model. In IERI monograph series: Issues and methodologies in large-scale assessments (Vol. 1, pp. 97–112). Princeton, NJ: IEA-ETS Research Institute.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Claus H. Carstensen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Carstensen, C.H. (2013). Linking PISA Competencies over Three Cycles – Results from Germany. In: Prenzel, M., Kobarg, M., Schöps, K., Rönnebeck, S. (eds) Research on PISA. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-4458-5_12

Download citation

Publish with us

Policies and ethics