Abstract
Performance assessments that have the capability of measuring a broad range of reasoning and thinking skills are considered to be fundamental tools in the educational reform movement (Linn, 1993). They are being used to document the need for educational reform as well as to monitor the impact of reform at the national, state, and local levels. Many advocates of performance assessments are optimistic that these assessments will be sensitive to measuring the impact of educational reform and that the differences in performance on these assessments among ethnic, linguistic, and gender subgroups will be narrower than those observed on multiple-choice tests. Thus, performance assessments are considered by many to be fairer and more equitable assessments of achievement for various subgroups than multiple-choice tests. It is important, however, to ensure that empirical evidence is obtained to support the fairness of the assessment for various subgroups of students.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Angoff, W.H. (1993). Perspectives on differential item functioning methodology. In P. W. Holland & H. Wainer (Eds.), Differential item functioning. Hillsdale, NJ: Lawrence Erlbaum.
Baker, E.L., O’Neil, H.F., & Linn, R. L. (1991). Policy and validity prospects for performance-based assessment. Paper presented at the annual meeting of the American Psychological Association.
Barr, R., & Dreeben, R. (1983). How schools work. Chicago: University of Chicago Press.
Bond, L. (1995). Unintended consequences of performance assessment: Issues of bias and fairness. Educational Measurement: Issues and Practice, 14(4), 21–24.
Cai, J., Magone, M. E., Wang, N., & Lane, S. (1996). A cognitive analysis of QUASAR’s mathematics performance assessment tasks and their sensitivity to measuring changes in middle-school students thinking and reasoning Research in Middle Level Education, 19(3), 63–94.
College Entrance Examination Board (1985). Equality and excellence: The educational status of Black Americans. New York: Author.
Cronbach, L.J., Gleser, G.C., Nanda, H., & Rajaratnam, N. (1972). The dependability of behavioral measurement. New York: Wiley.
Darling-Hammond, L., & Snyder, J. (1991). Traditions of curriculum inquiry: The scientific tradition. In P. Jackson (Ed.), Handbook of research on curriculum. New York: Macmillan.
Dossey, J.A., Mullis, I.V.A., & Jones, C.A. (1993) Can students do mathematical problem solving?: Results from constructed-response questions in NAEP’s 1992 mathematics assessment. Washington, D.C.: National Center for Educational Statistics.
Doolittle, A.E. & Cleary, T.A. (1987). Gender-based differential item performance in mathematics achievement items. Journal of Educational Measurement, 24(2), 157–166.
Dunbar, S.B. (1987, April). Comparability of indirect measures of writing as predictors of writing performance across demographic groups. Paper presented at the annual meeting of the American Educational Research Association, Washington, DC.
Dunbar, S.B., Koretz, & Hoover, H.D. (1991). Quality control in the development and use of performance assessments. Applied Measurement in Education, 4, 289–304.
Feinberg, L. (1990, Fall). Multiple choice and its critics. The College Board Review, No. 157.
Fenema, E., & Tartre, L. (1985). The use of spatial visualization in mathematics by girls and boys. Journal for Research in Mathematics Education, 16. 184–206.
Harris, A.M., & Carlton, S.T. (1993). Patterns of gender differences on mathematics items on the Scholastic Aptitude Test. Applied Measurement in Education, 6(2), 137–151.
Holland, P.W., & Thayer, D.T. (1986). Differential item performance and the Mantel-Haenszel procedure. (Research Report No. 86-31) Princeton: Educational Testing Service.
Lane, S. (1993). The conceptual framework for the development of a mathematics performance assessment, Educational Measurement: Issues and Practice, 12(2), 16–23.
Lane, S., Liu, M., Ankenmann, R. D., & Stone, C. A., (1996). Generalizability and validity of a mathematics performance assessment. Journal of Educational Measurement, 33(1), 71–92.
Lane, S., & Parke, C., (1996). Consequences of a mathematics performance assessment and the relationship between the consequences and student learning. Paper presented at the annual meeting of the National Council of Measurement in Education, N.Y.
Lane, S., Parke, C., & Moskal, B. (1992). Principles for developing performance assessments. Papa-presented at the annual meeting of the American Educational research Association, San Francisco, CA.
Lane, S., & Silver, E.A. (1995). Equity and validity considerations in the design and implementation of a mathematics performance assessment: The experience of the QUASAR project. In M. T. Nettles and A.L. Nettles (Eds.) Equity and Excellence in Educational Testing and Assessment, pp. 185–220.
Lane, S., & Silver, E.A. (April 1994). Examining students’ capacities for mathematical thinking and reasoning in the QUASAR project. Paper presented at the annual meeting of the American Educational Research Association, New Orleans.
Lane, S., Silver, E.A., & Wang, N. (April 1995). An examination of the performance of culturally and linguistically diverse students on a mathematics performance assessment within the QUASAR project. Paper presented at the annual meeting of the American Educational Research Association, San Francisco.
Lane, S., Stone, C.A., Ankenmann, R.D., & Liu, M. (1995). Examination of the assumptions and properties of the graded item response model: An example using a mathematics performance assessment. Applied Measurement in Education, 8(4), 313–340.
Lane, S., Stone, C.A., Ankenmann, R.D., & Liu, M. (1994). Reliability and validity of a mathematics performance assessment. International Journal of Educational Research, 21(3), p. 247–266.
Lane, S., Wang, N., & Magone, M. (in press). Gender related differential item functioning on a middle school mathematics performance assessment: The use of an analytical analysis of student performance to complement a statistical DIF procedure. Educational Measurement: Issues and Practice.
Linn, R. L. (1993). Educational assessment: Expanded expectations and challenges. Educational Evaluation and Policy Analysis, 15(1), 1–16.
Linn, R. L., Baker, E. L., & Dunbar, S. B. (1991). Complex, performance-based assessment: Expectations and validation criteria. Educational Researcher, 20(8), 15–21.
Magone, M. E., Cai, J., Silver, E. A., & Wang, N. (1994). Validity evidence for cognitive complexity of performance assessments: An analysis of selected QUASAR tasks. International Journal of Educational Research, 21(3), 317–340.
Messick, S. (1989). Validity. In R. L. Linn (Ed.), Educational Measurement (3rd ed.) (pp. 13–104). New York: American Council on Education.
Miller, T.R., & Spray, J.A. (1993). Logistic discriminant function analysis for DIF identification of polytomously scored items. Journal of Educational Measurement, 30(2), 107–122.
National Council of Teachers of Mathematics (1989). Curriculum and evaluation standards for school mathematics. Reston, VA: Author.
Oakes, J. (1990). Multiplying inequalities: The effects of race, social class, and tracking on opportunities to learn mathematics and science. Santa Monica: RAND Corporation.
O’Neill, K. A., & McPeek, W.M. (1993). Item and test characteristics that are associated with differential item functioning. In P.W. Holland & H. Wainer (Eds.), Differential item functioning, Hillsdale, NJ: Lawrence Erlbaum Associates.
Parke, C., Lane, S., & Guo, F. (April 1995). The consequential validity of a performance assessments in a mathematics education reform project. Paper presented at the annual meeting of the National Council of Measurement in Education, San Francisco.
Resnick, L. B. (1990, October). Assessment and educational standards. Paper presented at the Office of Educational Research and Improvement Conference, The Promise and Peril of Alternative Assessment, Washington, DC.
Silver, E.A., & Lane, S. (1993). Assessment in the context of mathematics instruction reform: The design of assessment in the QUASAR project. In M. Niss (Ed.), Assessment in mathematics education and its effects (pp. 59–70). London: Kluwer Academic.
Silver, E. A., & Stein, M. K. (1996). The QUASAR project: The “revolution of the possible” in mathematics instructional reform in urban middle schools. Urban Education, 30(4), 476–521.
Smith, M. (October 1994). Personal communication.
Stein, M. K., Grover, B. W., & Henningsen, M. (1996). Building student capacity for mathematical thinking and reasoning: An analysis of mathematical tasks used in reform classrooms. American Educational Research Journal, 33(2), 454–488.
Stein, M.K., & Lane, S., (1996). Instructional tasks and the development of student capacity to think and reason: An analysis of the relationship between teaching and learning in a reform mathematics project. Educational Research and Evaluation, 2(1), 50–80.
Tartre, L. (1990). Spatial skills, gender and mathematics. In E. Fennema, & G. Leder (Eds.), Mathematics and gender, (pp. 27–59). New York: Teacher’s College Press.
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer Science+Business Media New York
About this chapter
Cite this chapter
Lane, S., Silver, E.A. (1999). Fairness and Equity in Measuring Student Learning Using a Mathematics Performance Assessment: Results from the Quasar Project. In: Nettles, A.L., Nettles, M.T. (eds) Measuring Up. Evaluation in Education and Human Services, vol 48. Springer, Dordrecht. https://doi.org/10.1007/978-94-011-4399-8_6
Download citation
DOI: https://doi.org/10.1007/978-94-011-4399-8_6
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-5890-2
Online ISBN: 978-94-011-4399-8
eBook Packages: Springer Book Archive