Predicting Group Membership using National Assessment of Educational Progress (NAEP) Mathematics Data
- 1.2k Downloads
Since 1969 in the United States, the federally mandated National Assessment of Educational Progress (NAEP) has been used to assess the condition of student learning at the state and national levels, in particular subject areas, and in specific grades and/or ages (US National Center for Educational Statistics [NCES], n.d.-a). In the 1990s, academic achievement in various disciplines derived from NAEP scores were tracked by state and measured via the percentage of students at or above the levels established in a three-tiered scoring model: basic, proficient, or advanced (Hombo, 2003).
In concert with the No Child Left Behind Act of 2001 (NCLB, 2002), the US Congress decreed NAEP as the Nation's Report Card, to be used to indicate student-achievement score trends in academic areas such as mathematics, reading, science, and history among states in nationally representative samples (US NCES, n.d.-a). Results derived from NAEP data aggregated at the state level pertaining to student achievement—which in the latter years of its nearly 40-year existence have been employed as a means toward ascertaining accountability affiliated with high-stakes testing, such as NCLB—have yielded mixed interpretations and perceptions (see the American Educational Research Association, 2000, definition for a standard interpretation of the term high-stakes testing). For example, Hanushek and Raymond (2006) found that NAEP mathematics scores provided some positive evidence for accountability, whereas Amrein and Berliner (2002) determined the contrary with various state-level NAEP scores in certain academic areas. Darling-Hammond (2007) contended that NAEP, in areas such as mathematics, did not measure higher-order cognitive domains but instead “measures less complex application of knowledge” (p. 319). This lack of NAEP's application of knowledge via measuring—for instance, problem-solving skills—is linked to multiplechoice- type tests used for accountability purposes under NCLB that do not allow for the assessment of student achievement with the following question: “What can students do with what they have learned?” (p. 319).
Keywords
Student Achievement Limited English Proficiency American Educational Research Association Gain Score Predict Group MembershipPreview
Unable to display preview. Download preview PDF.
References
- American Educational Research Association. (2000). AERA position statement on high-stakes testing in pre-K — 12 education. Retrieved October 7, 2007, from policyandprograms/?id = 378&terms = AERA + position + statement + concerning&searchtype = 1&fragment = FalseGoogle Scholar
- Amrein, A. L., & Berliner, D. C. (2002). High-stakes testing, uncertainty, and student learning. Educational Policy Analysis Archives, 10(18). Retrieved from http://epaa.asu.edu/epaa/v10n18
- Ananth, C. V., & Kleinbaum, D. G. (1997). Regression models for ordinal responses: A review of methods and applications. International Journal of Epidemiology, 26(6), 1323–1333.CrossRefGoogle Scholar
- Borman, G. D., & D'Agostino, J. V. (2001). Title I and student achievement: A quantitative synthesis. In G. D. Borman, S. C. Stringfield, & R. E. Slavin (Eds.), Title I: Compensatory education at the crossroads (pp. 25–57). Mahwah, NJ: Lawrence Erlbaum.Google Scholar
- Brennan, R. L. (1995). The conventional wisdom about group mean scores. Journal of Educational Measurement, 32(4), 385–396.CrossRefGoogle Scholar
- Choi, K., Seltzer, M., Herman, J., & Yamashiro, K. (2007). Children left behind in AYP and non-AYP schools: Using student progress and the distribution of student gains to validate AYP. Educational Measurement: Issues and Practice, 26(3), 21–32CrossRefGoogle Scholar
- Chubb, J. E., & Hanushek, E. A. (1990). Reforming educational reform. In H. J. Aaron (Ed.), Setting national priorities: Policy for the nineties (pp. 213–248). Washington, DC: Brookings Institution Press.Google Scholar
- Cliff, N. (1987). Analyzing multivariate data. San Diego, CA: Harcourt Brace Jovanovich.Google Scholar
- Coleman, J. S. (1987). Families and schools. Educational Researcher, 16(6), 32–38.Google Scholar
- Coley, R. J. (2003). Growth in school revisited: Achievement gains from the fourth to the eighth grade. Princeton, NJ: Educational Testing Service.Google Scholar
- Cooper, B. S., Fusarelli, L. D., & Randall, E. V. (2004). Better policies, better schools: Theories and applications. Boston: Pearson/Allyn & Bacon.Google Scholar
- Cronbach, L. J. (1976). Research on classrooms and schools: Formulation of questions, design, and analysis. Stanford, CA: Stanford Evaluation Consortium.Google Scholar
- Darling-Hammond, L. (2007). The flat earth and education: How America's commitment to equity will determine our future [Third Annual Brown Lecture in Education Research]. Educational Researcher, 36(6), 318–334.CrossRefGoogle Scholar
- Dorn, S. (2006). No more aggregate NAEP studies? Education Policy Analysis Archives, 14(31). Retrieved from http://epaa.asu.edu/epaa/v14n31
- Education Policy Studies Laboratory (2005, September 14). Study predicts at least 85 percent of Great Lakes schools will be labeled ‘failing’ by 2014. [Press release]. Retrieved from http:// www.asu.edu/educ/epsl/EPRU/documents/EPSL-0509-109-EPRU-press.pdf Elementary and Secondary Education Act of 1965. Pub. L. No. 89–10. (1965).Google Scholar
- Goals 2000: Educate America Act of 1994. Title III. 20 U.S.C. 5801 § 302. (1994).Google Scholar
- Hanushek, E. A. (1996). School resources and student performance. In G. Burtless (Ed.), Does money matter? The effect of school resources on student achievement and adult success (pp. 43–73). Washington, DC: Brookings Institution Press.Google Scholar
- Hanushek, E. A., & Raymond, M. E. (2006). Early returns from school accountability. In P. E. Peterson (Ed.), Generational change: Closing the test score gap (pp. 143–166). Lanham, MD: Rowman & Littlefield.Google Scholar
- Hedges, L. V., Laine, R. D., & Greenwald, R. (1994). An exchange: Part I: Does money matter? A meta-analysis of studies of the effects of differential school inputs on student outcomes. Educational Researcher, 23(3), 5–14.Google Scholar
- Hombo, C. M. (2003). NAEP and No Child Left Behind: Technical challenges and practical solutions. Theory into Practice, 42(1), 59–65.CrossRefGoogle Scholar
- Huberty, C. J. (1994). Applied discriminant analysis. New York: Wiley.Google Scholar
- Huberty, C. J., & Lowman, L. L. (1998). Discriminant analysis in higher education research. In J. C. Smart (Ed.), Higher education: Handbook of theory and research (pp. 181–234). New York: Agathon Press.Google Scholar
- Huberty, C. J., & Lowman, L. L. (2000). Group overlap as a basis for effect size. Educational and Psychological Measurement, 60(4), 543–563.Google Scholar
- Huberty, C. J., Wisenbaker, J. M., & Smith, J. C. (1987). Assessing predictive accuracy in discriminant analysis. Multivariate Behavioral Research, 22(3), 307–329.CrossRefGoogle Scholar
- Improving America's Schools Act of 1994. Pub. L. No. 103–382, 108 Stat. 3518. (1994).Google Scholar
- Jennings, J. F. (2001). Title I: Its legislative history and its promise. In G. D. Borman, S. C. Stringfield, & R. E. Slavin (Eds.), Title I: Compensatory education at the crossroads (pp. 1–24). Mahwah, NJ: Lawrence Erlbaum.Google Scholar
- Klein, S., Benjamin, R., Shavelson, R. J., & Bolus, R. (2007). The Collegiate Learning Assessment: Facts and fantasies [White paper]. Retrieved October 15, 2007, from http://www.cae.org/ content/pdf/CLA.Facts.n.Fantasies.pdf
- Knapp, T. R. (1977). The unit-of-analysis problem in applications of simple correlation analysis to educational research. Journal of Educational Statistics, 2(3), 171–186.CrossRefGoogle Scholar
- Lachenbruch, P. A. (1967). An almost unbiased method of obtaining confidence intervals for the probability of misclassification in discriminant analysis. Biometrics, 23(4), 639–645.CrossRefGoogle Scholar
- Lachenbruch, P. A., & Mickey, M. R. (1968). Estimation of error rates in discriminant analysis. Technometrics, 10, 1–11.CrossRefGoogle Scholar
- Lee, J. (2004). How feasible is adequate yearly progress (AYP)? Simulations of school AYP “uniform averaging” and “safe harbor” under the No Child Left Behind Act. Education Policy Analysis Archives, 12(14). Retrieved from http://epaa.asu.edu/epaa/v12n14
- Lieberman, M. G., & Morris, J. D. (2004, April). Selecting predictor variables in logistic regression. Paper presented at the annual meeting of the American Educational Research Association, San Diego, CA.Google Scholar
- Linn, R. L. (2001). Assessment and accountability (condensed version). Practical Assessment, Research and Evaluation, 7(11). Retrieved from http://PAREonline.net/getvn.asp?v = 7 &n = 11
- Linn, R. L. (2003). Performance standards: Utility for different uses of assessments. Education Policy Analysis Archives, 11(31). Retrieved from http://epaa.asu.edu/epaa/v11n31
- Linn, R. L. (2005). Conflicting demands of No Child Left Behind and state systems: Mixed messages about school performance. Education Policy Analysis Archives, 12(33). Retrieved from http://epaa.asu.edu/epaa/v13n33/
- Lyons, J. E., & Algozzine, B. (2006). Perceptions of the impact of accountability on the role of principals. Education Policy Analysis Archives, 14(16). Retrieved from http://epaa.asu.edu/ epaa/v14n16/
- Morris, J. D., & Meshbane, A. (1995). Selecting predictor variables in two-group classification problems. Educational and Psychological Measurement, 55(3), 438–441.CrossRefGoogle Scholar
- No Child Left Behind Act of 2001. Pub. L. No. 107–110, 115 Stat. 1425. (2002).Google Scholar
- Nichols, S. L., Glass, G. V., & Berliner, D. C. (2006). High-stakes testing and student achievement: Does accountability pressure increase student learning? Educational Policy Analysis Archives, 14(1). Retrieved from http://epaa.asu.edu/epaa/v14n1Google Scholar
- Osborne, J. W. (2000). Advantages of hierarchical linear modeling. Practical Assessment, Research and Evaluation, 7(1). Retrieved from http://www.pareonline.net/getvn.asp?v = 7&n = 1
- Schafer, W. D., Liu, M., & Wang, H.-F. (2007). Content and grade trends in state assessments and NAEP. Practical Assessment, Research and Evaluation, 12(9). Retrieved from http://pareon-line.net/pdf/v12n9.pdf
- Schumacker, R. E., Mount, R. E., & Monahan, M. P. (2002). Factors affecting multiple regression and discriminant analysis with a dichotomous dependent variable: Prediction, explanation, and classification. Multiple Linear Regression Viewpoints, 28(2), 32–39.Google Scholar
- Slavin, R. E. (2001). How Title I can become an engine of reform in America's schools. In G. D. Borman, S. C. Stringfield, & R. E. Slavin (Eds.), Title I: Compensatory education at the crossroads (pp. 235–260). Mahwah, NJ: Lawrence Erlbaum.Google Scholar
- United States Department of Education. (2001). The longitudinal evaluation of school change and performance in Title I schools (Policy Studies Associates Contract No. EA 96008001). Rockville, MD: Author.Google Scholar
- United States Department of Education. (2003). No Child Left Behind, Accountability and adequate yearly progress (AYP) [National Title I Directors' Conference 2003]. Retrieved June 5, 2008, from http://www.ed.gov/admins/lead/account/ayp203/edlite-index.html
- United States National Center for Education Statistics. (n.d.-a). NAEP overview. Retrieved June 27, 2008, from http://nces.ed.gov/nationsreportcard/about/
- United States National Center for Education Statistics. (n.d.-b). State education reforms: Standards, assessment, and accountability. Retrieved June 27, 2008, from http://nces.ed.gov/ programs/statereform/saa.asp