Skip to main content

Using Item Response Theory as a Tool in Educational Measurement

  • Chapter
  • First Online:
  • 2141 Accesses

Part of the book series: Education in the Asia-Pacific Region: Issues, Concerns and Prospects ((EDAP,volume 18))

Abstract

Item response theory (IRT) and classical test theory (CTT) are invaluable tools for the construction of assessment instruments and the measurement of student proficiencies in educational settings. However, the advantages of IRT over CTT are not always clear. This chapter uses an example item analysis to contrast IRT and CTT. It is hoped that the readers can gain a deeper understanding of IRT through comparisons of similarities and differences between IRT and CTT statistics. In particular, this chapter discusses item properties such as the difficulty and discrimination power of items, as well as person ability measures contrasting the weighted likelihood estimates and plausible values in non-technical ways. The main advantage of IRT over CTT is outlined through a discussion on the construction of a developmental scale on which individual students are located. Further, some limitations of both IRT and CTT are brought to light to guide the valid use of IRT and CTT results. Lastly, the IRT software program, ConQuest (Wu et al. ACERConQuest version 2: Generalised item response modelling software. Australian Council for Educational Research, Camberwell, 2007), is used to run the item analysis to illustrate some of the program’s functionalities.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  • Crocker, L., & Algina, J. (1986). Introduction to classical and modern test theory. New York: CBS College Publishing.

    Google Scholar 

  • Lord, F. M. (1952). A theory of test scores. Psychometrika Monograph, No. 7, 17 (4, Pt. 2).

    Google Scholar 

  • Lord, F. M., & Novick, M. R. (1968). Statistical theories of mental test scores. Reading: Addison-Wesley.

    Google Scholar 

  • Rasch, G. (1960). Probabilistic models for some intelligence and attainment tests. Copenhagen: Danish Institute for Educational Research.

    Google Scholar 

  • Wu, M. L. (2005). The role of plausible values in large-scale surveys. In Postlethwaite (Ed.), Special issue of studies in educational evaluation (SEE) in memory of R M Wolf. 31 (2005) (pp. 114–128).

    Google Scholar 

  • Wu, M. L., & Adams, R. J. (2008). Properties of Rasch residual fit statistics. Unpublished paper.

    Google Scholar 

  • Wu, M. L., Adams, R. J., Wilson, M. R., & Haldane, S. A. (2007). ACERConQuest version 2: Generalised item response modelling software. Camberwell: Australian Council for Educational Research.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Margaret Wu .

Editor information

Editors and Affiliations

Appendices

Appendices

1.1 Appendix A.1 Mathematics Test

figure b
figure c
figure d
figure e

1.2 Appendix A.2 Recoding of the Item Responses

Question

New code  =  student response

1. Place value

1 = 6 ones

2 = 6 tens

3 = 6 hundreds

4 = 6 thousands

9 = missing response

2. Time

1 = 25 (minutes)

2 = 30 (minutes)

3 = 35 (minutes)

4 = 40 (minutes)

0 = all other responses

9 = missing response

3. Stamp

1 = 1 cm

2 = 2 cm

3 = 3 cm

4 = 4 cm

9 = missing response

4. Map

1 = 12 (km)

0 = all other responses

9 = missing response

5. Multiplication

1 = 1720

0 = all other responses

9 = missing response

6. Floor plan

1 = 50 (m)

2 = 54 (m)

3 = 56 (m)

0 = all other responses

9 = missing response

7. Sports graph

1 = netball

2 = football

3 = cricket

4 = athletics

8. Transport graph

1 = 4 (students)

2 = 5 (students)

0 = all other responses

9 = missing response

9. Lollies

1,2,3,4 according to the order of the 4 response options.

9 = missing response

10. Spinner

1 = 1

2 = 2

3 = 3

4 = 4

5 = 5

9 = missing response

11. Shape fraction

1 = 1/2

2 = 1/3

3 = 1/4

4 = 1/5

9 = missing response

12. Number sentence

1 = 2

2 = 3

3 = 36

0 = all other responses

9 = missing response

13. Gingerbread man

1 = ($) 5

2 = ($) 7.5

3 = ($) 8.5

4 = ($) 10

0 = all other responses

9 = missing response

14. Party pies

1 = 3 (boxes)

2 = 4 (boxes)

3 = 5 (boxes)

0 = all other responses

9 = missing response

15. Cubes

1 = 14

2 = 15

3 = 16

4 = 17

0 = all other responses

9 = missing response

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Wu, M. (2012). Using Item Response Theory as a Tool in Educational Measurement. In: Mok, M. (eds) Self-directed Learning Oriented Assessments in the Asia-Pacific. Education in the Asia-Pacific Region: Issues, Concerns and Prospects, vol 18. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-4507-0_9

Download citation

Publish with us

Policies and ethics