Skip to main content

Methodology III: Empirical Evaluation

  • Chapter
  • First Online:
Foundations for Designing User-Centered Systems

Abstract

Evaluation is an essential part of development. There are several good reasons for carrying out user testing in particular. A successful evaluation requires careful planning. Here we describe the issues that you need to take into account and discuss several effective methods that can be used to collect data. User testing reduces the risk that you will deliver a system to your users that is unusable and is therefore ineffective. We also touch briefly on the need make sure that any evaluation that you carry out is conducted according to appropriate ethical guidelines.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 34.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 44.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Berg, M. (1997). Rationalizing medical work: Decision support techniques and medical practices. Cambridge, MA: MIT Press.

    Google Scholar 

  • Brooke, J. (1996). SUS: A ‘quick and dirty’ usability scale. In P. W. Jordan, B. Thomas, B. A. Weerdmeester, & I. L. McClelland (Eds.), Usability evaluation in industry (pp. 189–194). London: Taylor & Francis.

    Google Scholar 

  • Brown, C. M. L. (1988). Human-computer interface design guidelines. Norwood, NJ: Ablex.

    Google Scholar 

  • Calfee, R. C. (1985). Experimental methods in psychology. New York, NY: Holt, Rinehart and Winston.

    Google Scholar 

  • Campbell, D. T., & Stanley, J. C. (1963). Experimental and quasi-experimental designs for research. Boston, MA: Houghton Mifflin.

    Google Scholar 

  • Card, S. K., Moran, T., & Newell, A. (1983). The psychology of human-computer interaction. Hillsdale, NJ: Erlbaum.

    Google Scholar 

  • Carroll, J. M. (2000). Making use: Scenario-based design of human-computer interactions. Cambridge, MA: MIT Press.

    Google Scholar 

  • Carroll, J. M., & Rosson, M. B. (1992). Getting around the task-artifact cycle: How to make claims and design by scenario. ACM Transactions on Information Systems, 10, 181–212.

    Article  Google Scholar 

  • Churchill, E. F., Nelson, L., Denoue, L., Murphy, P., & Helfman, J. I. (2003). The Plasma poster network: Social hypermedia on public display. In K. O’Hara, M. Perry, E. Churchill, & D. Russell (Eds.), Social and interactional aspects of shared display technologies. London: Kluwer Academic Publishers.

    Google Scholar 

  • Endsley, M. R. (1995). Toward a theory of situation awareness in dynamic systems. Human Factors, 37(1), 32–64.

    Article  Google Scholar 

  • Ericsson, K. A., & Simon, H. A. (1980). Protocol analysis: Verbal reports as data. Psychological Review, 87, 215–251.

    Article  Google Scholar 

  • Ericsson, K. A., & Simon, H. A. (1993). Protocol analysis: Verbal reports as data (2nd ed.). Cambridge, MA: MIT Press.

    Google Scholar 

  • Feurzeig, W., & Ritter, F. (1988). Understanding reflective problem solving. In J. Psotka, L. D. Massey, & S. A. Mutter (Eds.), Intelligent tutoring systems: Lessons learned. Hillsdale, NJ: Erlbaum.

    Google Scholar 

  • Goodman, E., Kuniavsky, M., & Moed, A. (2012). Observing the user experience: A practitioner’s guide to user research (2nd ed.). Waltham, MA: Morgan Kaufmann.

    Google Scholar 

  • Hart, S. G., & Staveland, L. E. (1988). Development of the NASA-TLX (Task Load Index): Results of empirical and theoretical research. In P. A. Hancock & N. Meshkati (Eds.), Human mental workload (pp. 139–185). Amsterdam: North Holland.

    Chapter  Google Scholar 

  • Hock, R. R. (2002). Forty studies that changed psychology. Upper Saddle River, NJ: Prentice Hall.

    Google Scholar 

  • Howell, D. C. (2010). Statistical methods for psychology (7th ed.). Belmont, CA: Wadsworth.

    Google Scholar 

  • Kukreja, U., Stevenson, W. E., & Ritter, F. E. (2006). RUI—recording user input from interfaces under Windows and Mac OS X. Behavior Research Methods, 38(4), 656–659.

    Article  Google Scholar 

  • LaPiere, R. T. (1934). Attitude versus action. Social Forces, 13, 230–237.

    Article  Google Scholar 

  • Lazar, J., Feng, J. H., & Hochheiser, H. (2010). Research methods in human-computer interaction. New York, NY: Wiley.

    Google Scholar 

  • Miles, M. B., Huberman, A. M., & Saldaña, J. (2013). Qualitative data analysis: A methods sourcebook. Thousand Oaks, CA: Sage.

    Google Scholar 

  • Molich, R., & Nielsen, J. (1990). Improving a human-computer dialogue. Communications of the ACM, 33(3), 338–348.

    Google Scholar 

  • Monk, A., Wright, P., Haber, J., & Davenport, L. (1993). Apendix 1—cooperative evaluation: A run-time guide. In Improving your human-computer interface: A practical technique. New York: Prentice-Hall.

    Google Scholar 

  • Monk, A. F. (1998). Lightweight techniques to encourage innovative user interface design. In L. Wood (Ed.), User interface design: Bridging the gap between user requirements and design (pp. 109–129). Boca Raton, FL: CRC Press.

    Google Scholar 

  • Mosier, J. N., & Smith, S. L. (1986). Application of guidelines for designing user interface software. Behaviour and Information Technology, 5, 39–46.

    Article  Google Scholar 

  • Müller, M. J., Haslwanter, J. H., & Dayton, T. (1997). Participatory practices in the software lifecycle. In M. G. Helander, T. K. Landauer, & P. V. Prabhu (Eds.), Handbook of human-computer interaction (2nd ed., pp. 255–297). Amsterdam, NL: Elsevier.

    Google Scholar 

  • NASA. (1987). NASA Task Load Index (TLX) V 1.0. Users Manual. Retrieved 10 March 2014, from http://humansystems.arc.nasa.gov/groups/TLX/downloads/TLX_comp_manual.pdf.

  • Navalpakkam, V., & Churchill, E. F. (in press). Eyetracking: A brief introduction. In J. S. Olson & W. Kellogg (Eds.), Ways of knowing, HCI methods. Heidelberg, Germany: Springer.

    Google Scholar 

  • Nielsen, J. (1993). Usability engineering. Chestnut Hill, MA: AP Professional Press.

    MATH  Google Scholar 

  • Nielsen, J., & Molich, R. (1990). Heuristic evaluation of user interfaces. In Proceedings of CHI 90 (pp. 249–256). New York: ACM.

    Google Scholar 

  • Nielsen, J., & Pernice, K. (2010). Eyetracking web usability. Berkeley, CA: New Riders.

    Google Scholar 

  • Ray, W. J. (2008). Methods toward a science of behavior and experience (9th ed.). Belmont, CA: Wadsworth Publishing.

    Google Scholar 

  • Reeves, B., & Nass, C. (1996). The media equation: How people treat computers, television, and new media like real people and places. New York: NY Cambridge University Press.

    Google Scholar 

  • Ritter, F. E., Kim, J. W., & Morgan, J. H. (2009). Running behavioral experiments with human participants: A practical guide (Tech. Report No. 2009-1). Applied Cognitive Science Lab: The Pennsylvania State University, College of Information Sciences and Technology.

    Google Scholar 

  • Ritter, F. E., Kim, J. W., Morgan, J. H., & Carlson, R. A. (2013). Running behavioral studies with human participants: A practical guide. Thousand Oaks, CA: Sage.

    Google Scholar 

  • Smallman, H. S., & St. John, M. (2005). Naïve realism: Misplaced faith in the utility of realistic displays. Ergonomics in Design, 13(Summer), 6–13.

    Article  Google Scholar 

  • Todd, Z. (Ed.). (2004). Mixing methods in psychology: The integration of qualitative and quantitative methods in theory and practice. Abingdon, UK: Psychology Press.

    Google Scholar 

  • Tullis, T., & Albert, B. (2008). Measuring the user experience. Burlington, MA: Morgan Kaufmann.

    Google Scholar 

  • Whiteside, J., Jones, S., Levy, P. S., & Wixon, D. (1985). User performance with command, menu, and iconic interfaces. In Proceedings of CHI’85 Human Factors in Computing Systems (185–191). New York: ACM.

    Google Scholar 

  • Woods, D. D., & Dekker, S. W. A. (2000). Anticipating the effects of technological change: A new era of dynamics for human factors. Theoretical Issues in Ergonomic Science, 1(3), 272–282.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Frank E. Ritter .

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer-Verlag London

About this chapter

Cite this chapter

Ritter, F.E., Baxter, G.D., Churchill, E.F. (2014). Methodology III: Empirical Evaluation . In: Foundations for Designing User-Centered Systems. Springer, London. https://doi.org/10.1007/978-1-4471-5134-0_13

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-5134-0_13

  • Published:

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-4471-5133-3

  • Online ISBN: 978-1-4471-5134-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics