Skip to main content

Assessing Student Learning Online

Overcoming Reliability Issues

  • Chapter
  • First Online:
Digital Systems for Open Access to Formal and Informal Learning

Abstract

Assessing students in online university courses poses challenges to instructors in maintaining reliability of the measurement instruments being utilized and their delivery processes. Some programs have the latitude to incorporate proctored assessments, but this is not always practical in asynchronously structured courses reaching out across a broad geographic region. Adding challenge to the process, it is difficult to know with 100 % certainty who is actually completing the assignments and various other checks on learning in online courses. Reliability in the assessment sphere is associated with a continuum, higher and lower, rather than absolute. A great deal has been written about constructing reliable tests, but less focus has been placed upon the testing process. Given the substantial expansion of online education, an equitable amount of attention must be contributed to the online assessment process. Rather than focusing on test construct reliability, this chapter will explore the process of assessment in the online environment, strategies for increasing reliability via the delivery processes, and alternative student outputs as viable representations of learning.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Airasian, P. W., & Russell, M. K. (2007). Classroom assessment: Concepts and application (6th ed.). New York: McGraw-Hill.

    Google Scholar 

  • Arnold, S. D. (2010). Student e-cooperative: Computer mediated group projects. In the International Technology, Education, and Development Conference Proceedings CD (ISBN: 978-84-613-5538-9). Valencia, Spain: INTED.

    Google Scholar 

  • Arnold, S. D., & Moshchenko, M. (2009). Technology input versus input and output: Does it result in learning differences among elementary school students? In C. D. Maddux (Ed.), Research highlights in technology and teacher education 2009 (pp. 1–9). Chesapeake: Society for Information Technology & Teacher Education.

    Google Scholar 

  • Arnold, S. D. (2012, October). Assessing online learning with digital audio and video. In Proceedings of World Conference on E-Learning in Corporate, Government, Healthcare, and Higher Education 2012 (pp. 473–479). Chesapeake, VA: AACE.

    Google Scholar 

  • Brimi, H. M. (2011). Reliability of grading high school work in English. Practical assessment Research & Evaluation, 16(17), 1–12.

    Google Scholar 

  • Calandra, B., Brantley-Dias, L., Lee, J. K., & Fox, D. L. (2009). Using video editing to cultivate novice teachers’ practice. Journal of Research on Technology in Education, 42(1), 73–94.

    Article  Google Scholar 

  • Craik, F., & Lockhart, R. (1972). Levels of processing: A framework for memory research. Journal of Verbal Learning and Verbal Behavior, 11(6), 671–684.

    Article  Google Scholar 

  • Gronlund, N. E., & Waugh, C. K. (2009). Assessment of student achievement (9th ed.). Upper Saddle River: Pearson Education Inc.

    Google Scholar 

  • Hunter, R. (2004). Madeline Hunter’s mastery teaching: Increasing instructional effectiveness in elementary and secondary schools. Thousand Oaks: Corwin Press.

    Google Scholar 

  • Kolowich, S. (22 March 2013a). The minds behind the MOOCs. The chronicle of higher education. http://chronicle.com. Retrieved October 1, 2013.

  • Kolowich, S. (8 April 2013b). Coursera takes a nuanced view of MOOC dropout rates. The chronicle of higher education. http://chronicle.com. Retrieved October 1, 2013.

  • McKenzie, G. R. (1979). Effects of questions and test-like events on achievement and on-task behavior in a classroom concept learning presentation. Journal of Educational Research, 72(6), 348–351.

    Google Scholar 

  • McMillan, J. H. (2011). Classroom assessment: Principles and practices for effective standards-based instruction (5th ed.). Boston: Pearson Education Inc.

    Google Scholar 

  • McTighe, J. (1996). What happens between assessments? Educational Leadership, 54(4), 6–12.

    Google Scholar 

  • Montgomery, K. (2002). Authentic tasks and rubrics: Going beyond traditional assessments in college teaching. College Teaching, 50(1), 34–39.

    Article  Google Scholar 

  • National Survey of Student Engagement [NSSE]. (2010). Major differences: Examining student engagement by field of study-annual results 2010. Bloomington: Indiana University Center for Postsecondary Research.

    Google Scholar 

  • Popham, J. W. (2011). Classroom assessment (6th ed.). Boston: Pearson Education Inc.

    Google Scholar 

  • Roblyer, M. D., & Doering, A. H. (2012). Integrating educational technology into teaching (6th ed.). Boston: Pearson Education Inc.

    Google Scholar 

  • Roediger, H. L. III., Putnam, A. L., & Smith, M. A. (2011). Ten benefits of testing and their applications to educational practice. In J. P. Mestre & B. H. Ross (Eds.), Psychology of learning and motivation Academic Press, 2011, Volume 55, pp. 1–36, ISSN 0079-7421, ISBN 9780123876911, http://dx.doi.org/10.1016/B978-0-12-387691-1.00001-6.

    Google Scholar 

  • Slavin, R. E. (2012). Educational psychology: Theory and practice (10th ed.). Boston: Pearson Education Inc.

    Google Scholar 

  • Struyven, K., Dochy, F., Janssens, S., Schelfhout, W., & Gielen, S. (2006). The overall effects of end-of-course assessment on student performance: A comparison between multiple choice testing, peer assessment, case-based assessment and portfolio assessment. Studies in Educational Evaluation, 32, 202–222.

    Article  Google Scholar 

  • Swan, K., Schenker, J., Arnold, S., & Kuo, C.-L. (2007). Shaping online discussion: Assessment matters. E-mentor, 1 (18). http://www.e-mentor.edu.pl/_xml/wydania/18/390.pdf. Retrieved October 1, 2014.

  • Thorpe, M. (1998). Assessment and third generation distance education. Distance Education, 19(2), 265–286.

    Article  Google Scholar 

  • Tuckman, B. W. (1999). Conducting educational research (5th Ed.). Belmont: Wadsworth Group.

    Google Scholar 

  • Watson, G., & Sottile, J. (2010). Cheating in the digital age: Do students cheat more in online courses? Online Journal of Distance Administration, 13 (1). http://www.westga.edu/~distance/ojdla/spring131/watson131.html. Retrieved October 1, 2013.

  • Williams, B. L., & Suen, H. K. (1998). Formal versus informal assessment methods. American Journal of Health Behavior, 22(4), 308–313.

    Google Scholar 

  • Young, C. B., & Henquinet, J. A. (2000). A conceptual framework for designing group projects. Journal of Education for Business, 76(1), 56–60.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Stephen D. Arnold PhD .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Arnold, S. (2014). Assessing Student Learning Online. In: Sampson, D., Ifenthaler, D., Spector, J., Isaias, P. (eds) Digital Systems for Open Access to Formal and Informal Learning. Springer, Cham. https://doi.org/10.1007/978-3-319-02264-2_7

Download citation

Publish with us

Policies and ethics