Pathway to Proficiency: Linking the STAR Reading
Transcription
Pathway to Proficiency: Linking the STAR Reading
March 30, 2016 Pathway to Proficiency: Linking the STAR Reading™ and STAR Math™ Scales with Performance Levels on the Georgia Milestones Quick reference guide to the STAR Assessments™ STAR Reading™—used for screening and progress-monitoring assessment—is a reliable, valid, and efficient computer-adaptive assessment of general reading achievement and comprehension for grades 1–12. STAR Reading provides nationally norm-referenced reading scores and criterion-referenced scores. A STAR Reading assessment can be completed without teacher assistance in about 10 minutes and repeated as often as weekly for progress monitoring. STAR Math™—used for screening, progress-monitoring, and diagnostic assessment—is a reliable, valid, and efficient computer-adaptive assessment of general math achievement for grades 1–12. STAR Math provides nationally norm-referenced math scores and criterionreferenced evaluations of skill levels. A STAR Math assessment can be completed without teacher assistance in less than 15 minutes and repeated as often as weekly for progress monitoring. STAR Reading and STAR Math received the highest possible ratings for screening and progress monitoring by the National Center on Response to Intervention, are highly rated for progress monitoring by the National Center on Intensive Intervention, and meet all criteria for scientifically based progress-monitoring tools set by the National Center on Student Progress Monitoring. © 2016 by Renaissance Learning, Inc. All rights reserved. Printed in the United States of America. All logos, designs, and brand names for Renaissance Learning’s products and services, including but not limited to Renaissance Learning, Renaissance Place, STAR, STAR Assessments, STAR Math, and STAR Reading are trademarks of Renaissance Learning, Inc., and its subsidiaries, registered, common law, or pending registration in the United States and other countries. All other product and company names should be considered the property of their respective companies and organizations. This publication is protected by U.S. and international copyright laws. It is unlawful to duplicate or reproduce any copyrighted material without authorization from the copyright holder. For more information, contact: RENAISSANCE LEARNING P.O. Box 8036 Wisconsin Rapids, WI 54495-8036 (800) 338-4204 www.renaissance.com Page 2 of 22 Introduction Educators face many challenges; chief among them is making decisions regarding how to allocate limited resources to best serve diverse student needs. A good assessment system supports teachers by providing timely, relevant information that can help address key questions such as which students are on track to meet important performance standards? And which students are not on track and thus need additional help? Different educational assessments serve different purposes, but those that can identify students early in the school year as being at-risk to miss academic standards can be especially useful because they can help inform instructional decisions that can improve student performance and reduce gaps in achievement. Assessments that can do that while taking little time away from instruction are particularly valuable. Indicating which students are on track to meet later expectations is one of the potential capabilities of a category of educational assessments called “interim” (Perie, Marian, Gong, & Wurtzel, 2007). They are one of three broad categories of assessment: 1. Summative – typically annual tests that evaluate the extent to which students have met a set of standards. Most common are state-mandated tests such as the Georgia Milestones assessments. 2. Formative – short and frequent processes embedded in the instructional program that support learning by providing feedback on student performance and identifying specific things students know and can do as well as gaps in their knowledge. 3. Interim – assessments that fall in between formative and summative in terms of their duration and frequency. Some interim tests can serve one or more purposes, including informing instruction, evaluating curriculum and student responsiveness to intervention, and forecasting likely performance on a high-stakes summative test later in the year. This study focuses on the application of interim test results, notably their power to inform educators about which students are on track to succeed on the year-end summative state test and which students might need additional assistance to reach proficiency. Specifically, it involves linking the Georgia Milestones performance levels for ELA and Mathematics with scales from two interim tests, STAR Reading and STAR Math. The Georgia Milestones assessments are used statewide as summative assessments in Georgia. The STAR tests are the most widely used assessments in the U.S., are computer adaptive and use item response theory, require very little time (on average, less than 10–15 minutes group administration time), and may be given repeatedly throughout the school year.1 An outcome of this study is that educators using the STAR Reading Enterprise or STAR Math Enterprise assessments can access the Reading Dashboard and STAR Performance Reports (see Sample Reports, pp. 12–16) that indicate whether individual students or groups of students (by class, grade, or demographic characteristics) are on track to meet proficiency as measured by the Georgia Milestones. The dashboard and reports allow instructors to evaluate student progress toward proficiency and make instructional decisions based on data well in advance of annual state tests. They also help educators screen for later difficulties and progress monitor students’ responsiveness to interventions. 1 For an overview of the STAR tests and how they work, please see the References section for a link to download The research foundation for STAR Assessments report. For additional information, full technical manuals are available for each STAR assessment by contacting Renaissance Learning at research@renaissance.com Page 3 of 22 Sources of data STAR Reading and STAR Math data were gathered from schools that use those assessments on the Renaissance Place hosted platform2. Performance-level distributions from the Georgia Milestones English Language Arts (ELA) End-ofGrade assessments (GA Milestones ELA) and the GA Milestones Mathematics End-of-Grade assessments (GA Milestones Mathematics) were retrieved from Georgia Milestones school summary data available on the Georgia Department of Education website. The Georgia Milestones use four performance levels: Level 1: Beginning Learning, Level 2: Developing Learner, Level 3: Proficient Learner, Level 4: Distinguished Learner. Students scoring in the Level 3 or Level 4 categories would be counted as meeting proficiency standards for state and federal performance-level reporting. This study uses STAR Reading, STAR Math, and Georgia Milestones end-of-grade data from the 2014–15 school year. Georgia Milestones Performance Levels: Level 1: Beginning Learner Level 2: Developing Learner Level 3: Proficient Learner Level 4: Distinguished Learner Methodology Many of the ways to link scores between two tests require that the scores from each test be available at a student level. Obtaining a sufficient sample of student-level data can be a lengthy and difficult process. However, there is an alternative technique that produces similar results without requiring us to know each individual student’s Georgia Milestones score and STAR scaled score. The alternative involves using school-level data to determine the STAR scaled scores that correspond to each Georgia Milestones performance level cutscore. School level Georgia Milestones end-ofgrade data are publically available, allowing us to streamline the linking process and complete linking studies more rapidly. The STAR scores used in this analysis were “projected” scaled scores using STAR Reading and STAR Math decile-based growth norms. The growth norms are both grade- and subject-specific and are based on the growth patterns of more than one million students using STAR Assessments over a three-year period. They provide typical growth rates for students based on their starting STAR test score, making predictions much more accurate than a “one-size-fits-all” growth rate. For each observed score, the number of weeks between the STAR test administration date and the middle of the Georgia Milestones window was calculated. Then, the number of weeks between the two dates was multiplied by the student’s expected weekly scaled score growth (from our decile-based growth norms, which take into account grade and starting observed score). Expected growth was added to the observed scaled score to determine each student’s projected STAR score. For students with multiple STAR tests within a school year, the average of their projected scores was used. This method used to link our STAR scale to Georgia Milestones proficiency levels is equivalent groups equipercentile equating. This method looks at the distribution of Georgia Milestones performance levels in the sample and compares that to the distribution of projected STAR scores for the sample; the STAR scaled score that cuts off the same percentage 2 Renaissance Place is a service that involves “hosting” schools’ data from the STAR tests and other products. For more information about Renaissance Place, see http://www.renaissance.com/products/renaissance-place Page 4 of 22 of students as each Georgia Milestones performance level is taken to be the cutscore for each respective proficiency level. For several different states, we compared the results from the equivalent groups equipercentile equating to results from student-level data and found the accuracy of the two methods to be nearly identical (Renaissance Learning, 2015a, 2015b). McLaughlin and Bandeira de Mello (2002) employed a similar method in their comparison of NAEP scores and state assessment results, and this method has been used multiple times since 2002 (Bandeira de Mello, Blankenship, & McLaughlin, 2009; McLaughlin & Bandeira de Mello, 2003; McLaughlin & Bandeira de Mello, 2006; McLaughlin, Bandeira de Mello, Blankenship, Chaney, Esra, Hikawa, Rojas, William, & Wolman, 2008). Additionally, Cronin et al. (2007) found this method was able to determine performance level cutscore estimates very similar to the cutscores generated by statistical methods requiring student-level data. Sample selection To find a sample of students who were assessed by both the Georgia Milestones and STAR, we began by gathering all hosted STAR Reading and STAR Math test records for the year 2014-2015. Then, each school’s STAR Reading and STAR Math data were aggregated by grade and subject area. The next step was to match STAR data with the Georgia Milestones end-of-grade data. To do this, performance level distribution data from the Georgia Milestones was obtained from the Georgia Department of Education website. The file included the total number of students tested in each grade and the percentage of students in each performance level. The STAR Reading and STAR Math data were matched to the Georgia Milestones end-of-grade ELA and mathematics data by system and school name. Once we determined how many students in each grade at a school were tested on the Georgia Milestones for English Language Arts and took a STAR Reading assessment, we calculated the percentage of enrolled students assessed on both tests. Then we repeated this exercise for the mathematics assessments. In each grade at each school, if between 95% and 105% of the students who tested on the Georgia Milestones had taken a STAR assessment, that grade was included in the sample. The process was conducted separately for the English Language Arts and mathematics assessments. This method of sample selection ensured that our sample consisted of schools in which all or nearly all of the enrolled students who took the Georgia Milestones also took STAR within the specified window of time. If a total of approximately 1,000 or more students per grade met the sample criteria, that grade’s sample was considered sufficiently large for analysis. Demographic information for the schools in our sample was available from the student enrollment by grade level report available on the Georgia Department of Education website. We aggregated this data to the grade level to create Table 1 on the following page and Table 3 on page 7. Page 5 of 22 Sample description: Reading A total of 354 unique schools across grades 3 through 8 met the sample requirements (explained previously in Sample Selection). Racial/ethnic characteristics for each grade of the sample are presented along with statewide averages in Table 1. White students were slightly over-represented and Black students were slightly under-represented in the reading sample. Table 2 displays by-grade test summaries for the reading sample. It includes counts of students taking STAR Reading and the Georgia Milestones ELA. It also includes percentages of students in each performance level, both for the sample and statewide. Students in the reading sample had similar performance to students statewide. Table 1. Characteristics of reading sample: Racial/ethnic statistics Percent of students by racial/ethnic category Grade Number of Schools 3 Native American Asian & Pac. Islander Hispanic Black White Multiple Races 114 0.2% 3.1% 13.4% 30.2% 49.6% 3.5% 4 142 0.2% 3.4% 13.6% 31.1% 48.4% 3.4% 5 175 0.2% 3.4% 12.8% 30.9% 49.4% 3.4% 6 50 0.2% 3.5% 10.6% 33.2% 49.2% 3.3% 7 69 0.2% 3.3% 10.6% 33.8% 48.9% 3.3% 8 53 0.2% 4.2% 9.6% 34.2% 48.9% 3.0% Statewide 0.0% 4.0% 14.0% 37.0% 42.0% 3.0% Table 2. Characteristics of reading sample: Performance on STAR Reading ™ and Georgia Milestones ELA Level 1: Level 2: Level 3: Level 4: Students Students Beginning Developing Proficient Distinguished taking GA Grade taking STAR Learner Learner Learner Learner Milestones Reading™ ELA Sample State Sample State Sample State Sample State 3 12,658 12,374 25% 33% 29% 30% 31% 27% 15% 10% 4 14,885 14,494 23% 29% 33% 34% 32% 28% 12% 9% 5 18,222 17,804 22% 27% 31% 34% 36% 31% 11% 8% 6 10,161 9,964 26% 31% 29% 30% 35% 31% 10% 8% 7 15,796 15,481 27% 31% 32% 33% 33% 30% 8% 6% 8 12,446 12,267 22% 24% 33% 37% 34% 31% 11% 8% Page 6 of 22 Sample description: Math Georgia Milestones mathematics grade counts were matched to counts of students who took STAR Math the same year. A total of 206 unique schools across grades 3 through 8 met the sample requirements for math (explained in Sample selection). Racial/ethnic characteristics for each of the math samples were calculated in the same manner as for reading and are presented along with statewide averages in Table 3. White students were slightly over-represented and Black students were slightly under-represented in the math sample. Table 4 displays by-grade test summaries for the math sample. It includes counts of students taking STAR Math and a Georgia Milestones Mathematics assessment. It also includes percentages of students in each performance level, both for the sample and statewide. Students in the math sample had similar Georgia Milestones Mathematics performance to the statewide population. Table 3. Characteristics of math sample: Racial/ethnic statistics Percent of students by racial/ethnic category Grade Number of Schools Native American Asian & Pac. Islander Hispanic Black White Multiple Races 3 85 0.2% 3.4% 13.8% 30.6% 48.5% 3.5% 4 77 0.2% 3.6% 14.3% 31.4% 47.2% 3.3% 5 103 0.2% 3.7% 12.8% 30.4% 49.5% 3.4% 6 33 0.2% 5.6% 13.1% 29.8% 48.1% 3.2% 7 35 0.2% 4.2% 10.6% 33.0% 49.0% 3.1% 8 34 0.2% 4.6% 10.9% 33.1% 48.4% 2.9% Statewide 0.0% 4.0% 14.0% 37.0% 42.0% 3.0% Table 4. Characteristics of math sample: Performance on STAR Math™ and Georgia Milestones Mathematics Level 1: Level 2: Level 3: Level 4: Students Students taking Beginning Developing Proficient Distinguished taking Grade GA Milestones Learner Learner Learner Learner STAR Mathematics Math™ Sample State Sample State Sample State Sample State 3 9,466 9,271 16% 21% 38% 41% 34% 30% 12% 8% 4 7,828 7,641 17% 20% 35% 40% 34% 31% 14% 9% 5 11,179 10,911 20% 25% 34% 37% 30% 27% 16% 11% 6 7,288 7,110 21% 26% 33% 39% 30% 26% 16% 9% 7 8,999 8,789 19% 25% 34% 38% 28% 25% 19% 12% 8 8,227 8,080 19% 25% 35% 38% 27% 25% 19% 12% Page 7 of 22 Analysis First, we aggregated the sample of schools for each subject to grade level. Next, we calculated the percentage of students scoring in each Georgia Milestones performance level for each grade. Finally, we ordered STAR scores and analyzed the distribution to determine the scaled score at the same percentile as the Georgia Milestones proficiency level. For example, in our third grade reading sample, 25% of students were at Level 1: Beginning Learner, 29% at Level 2: Developing Learner, 31% at Level 3: Proficient Learner, and 15% at Level 4: Distinguished Learner. Therefore, the cutscores for achievement levels in the third grade are at the 25th percentile for Level 2: Developing Learner, the 54th percentile for Level 3: Proficient Learner, and the 85th percentile for Level 4: Distinguished Learner. Figure 1. Illustration of linked STAR Reading™ and Georgia Milestones ELA scales for 3rd grade. STAR Reading 1400 15% GA Milestones ELA 830 15% Level 4 581 525 475 618 31% Level 3 29% Level 2 31% 469 29% 25% Level 1 180 349 25% 0 Page 8 of 22 Results and reporting Table 5 presents estimates of equivalent scores on STAR Reading and Georgia Milestones ELA. Table 6 presents estimates of equivalent scores on STAR Math and Georgia Milestones Mathematics. These results will be incorporated into the Renaissance Learning’s Reading Dashboard and STAR Performance Reports (see Sample dashboard and reports, pp. 12–16) that can be used to help educators determine early and periodically which students are on track to reach the Proficient Learner status or higher in order to make instructional decisions accordingly. Table 5. Estimated STAR Reading™ cut scores for Georgia Milestones ELA performance levels Grade Level 1: Beginning Learner Level 2: Developing Learner Level 3: Proficient Learner Level 4: Distinguished Learner 3 < 349 349 469 618 4 < 416 416 564 797 5 < 479 479 667 960 6 < 544 544 743 1087 7 < 582 582 843 1231 8 < 615 615 908 1274 Table 6. Estimated STAR Math™ cut scores for Georgia Milestones Mathematics performance levels Grade Level 1: Beginning Learner Level 2: Developing Learner Level 3: Proficient Learner Level 4: Distinguished Learner 3 < 542 542 634 707 4 < 604 604 701 779 5 < 675 675 768 832 6 < 706 706 800 867 7 < 716 716 820 880 8 < 743 743 843 897 Page 9 of 22 Accuracy analyses Following the linking study summarized earlier in this report, 3 Georgia school districts shared student level 2015 Georgia Milestones scores in order to explore the accuracy of using STAR Reading and STAR Math for forecasting Georgia Milestones performance. Classification diagnostics were derived from counts of correct and incorrect classifications that could be made when using STAR scores to predict whether or not a student would achieve proficiency on the GA Milestones.3 The results are summarized in the following tables and figures, and indicate that the statistical linkages provide an effective means of estimating end-of-year achievement on the GA Milestones based on STAR scores obtained earlier in the year. Table 7. Fourfold schema of proficiency classification diagnostic data GA Milestones Result STAR Estimate Proficient Not Total Total Proficient Not True Positive (TP) False Negative (FN) Observed Proficient (TP + FN) False Positive (FP) True Negative (TN) Observed Not (FP + TN) Projected Proficient (TP + FP) Projected Not (FN + TN) N = TP+FP+FN+TN Table 8a. Proficiency classification frequencies: ELA Classification Grade 3 4 5 6 7 8 True Positive (TP) 170 201 132 90 80 99 True Negative (TN) 448 424 500 318 325 293 False Positive (FP) 83 62 71 21 28 38 False Negative (FN) 33 31 44 37 22 31 Total (N) 734 718 747 466 455 461 Table 8b. Proficiency classification frequencies: Mathematics Classification Grade 3 4 5 6 7 8 True Positive (TP) 250 220 235 80 90 89 True Negative (TN) 444 406 491 338 309 309 False Positive (FP) 77 68 65 32 37 26 False Negative (FN) 43 27 54 21 17 39 Total (N) 814 721 845 471 453 463 3 Classifications were based upon STAR scores projected to April 15, 2015, the midpoint of the 2014-2015 GA Milestones end-of-grade main administration window (March 30, 2015 – May 1, 2015). Page 10 of 22 Table 9a. Accuracy of projected STAR™ scores for predicting GA Milestones proficiency: ELA Measure Formula Interpretation Grade 3 4 5 6 7 8 Percentage of correct classifications 84% 87% 85% 88% 89% 85% Percentage of proficient students identified as such using STAR 84% 87% 75% 71% 78% 76% Percentage of not proficient students identified as such using STAR 84% 87% 88% 94% 92% 89% Percentage of students STAR finds proficient who actually are proficient 67% 76% 65% 81% 74% 72% Percentage of students STAR finds not proficient who actually are not proficient 93% 93% 92% 90% 94% 90% Percentage of students who achieve proficient 28% 32% 24% 27% 22% 28% Percentage of students STAR finds proficient 34% 37% 27% 24% 24% 30% Difference between projected and observed proficiency rates 6% 5% 3% -3% 2% 2% Area Under the ROC Curve Summary measure of diagnostic accuracy 0.93 0.95 0.91 0.93 0.94 0.92 Correlation (r) Relationship between STAR and GA Milestones ELA scores 0.84 0.85 0.81 0.83 0.84 0.84 Overall classification accuracy Sensitivity Specificity Positive predictive value (PPV) Negative predictive value (NPV) Observed proficiency rate (OPR) Projected proficiency rate (PPR) Proficiency status projection error TP + TN N TP TP + FN TN TN + FP TP TP + FP TN FN + TN TP + FN N TP + FP N PPR - OPR For reading, STAR correctly classified students as proficient or not between 84% and 89% of the time depending on grade. Sensitivity (i.e., the percentage of students correctly forecasted as proficient) ranged from 71% to 87%. Specificity (i.e., the percentage of students correctly forecasted as not proficient) values ranged from 84% to 94%. Positive predictive values ranged from 65% to 81%, meaning that when STAR forecasted students to be proficient, they actually were 65%-81% of the time depending on grade. Negative predictive values ranged from 90% to 94%. When STAR forecasted that students were not proficient, they actually were not 90%-94% of the time depending on grade. Proficiency status projection errors ranged from -3% to 6% across grades. The area under the ROC curve (AUC) ranged from 0.91 to 0.95, providing a strong indication that STAR did a good job of discriminating between which students reached the GA Milestones ELA proficiency benchmarks and which did not. Finally, the correlations between STAR and GA Milestones ELA were between 0.81 and 0.85, reflecting a strong positive relationship between projected STAR Reading scores and GA Milestones ELA scores. Page 11 of 22 Table 9b. Accuracy of projected STAR™ scores for predicting GA Milestones proficiency: Mathematics Measure Formula Interpretation Grade 3 4 5 6 7 8 Percentage of correct classifications 85% 87% 86% 89% 88% 86% Percentage of proficient students identified as such using STAR 85% 89% 81% 79% 84% 70% Percentage of not proficient students identified as such using STAR 85% 86% 88% 91% 89% 92% Percentage of students STAR finds proficient who actually are proficient 76% 76% 78% 71% 71% 77% Percentage of students STAR finds not proficient who actually are not proficient 91% 94% 90% 94% 95% 89% Percentage of students who achieve proficient 36% 34% 34% 21% 24% 28% Percentage of students STAR finds proficient 40% 40% 36% 24% 28% 25% Difference between projected and observed proficiency rates 4% 6% 2% 3% 4% -3% Area Under the ROC Curve Summary measure of diagnostic accuracy 0.94 0.94 0.93 0.96 0.95 0.93 Correlation (r) Relationship between STAR and GA Milestones Mathematics scores 0.86 0.84 0.83 0.80 0.80 0.77 Overall classification accuracy Sensitivity Specificity Positive predictive value (PPV) Negative predictive value (NPV) Observed proficiency rate (OPR) Projected proficiency rate (PPR) Proficiency status projection error TP + TN N TP TP + FN TN TN + FP TP TP + FP TN FN + TN TP + FN N TP + FP N PPR - OPR For mathematics, STAR correctly classified students as proficient or not 85% to 89% of the time depending on grade. Sensitivity ranged from 70% to 89% and specificity ranged from 85% to 92%. Positive predictive values ranged from 71% to 78%, meaning that when STAR forecasted students to be proficient, they actually were 71%-78% of the time depending on grade. Negative predictive values ranged from 89% to 95%. When STAR forecasted that students were not proficient, they actually were not 89%-95% of the time depending on grade. Proficiency status projection errors ranged from -3% to 6% across grades and the area under the ROC curve (AUC) ranged from 0.93 to 0.96, providing a strong indication that STAR did a good job of discriminating between which students reached the GA Milestones mathematics proficiency benchmarks and which did not. Finally, the correlations between STAR and GA Milestones mathematics were between 0.77 and 0.86, reflecting a positive relationship between projected STAR Math scores and GA Milestones Mathematics scores. Page 12 of 22 Scatter plots: ELA 800 800 700 700 600 600 GA Milestones ELA GA Milestones ELA For each student, projected STAR Reading™ scores and GA Milestones ELA scores were plotted across grades. 500 400 300 200 500 400 300 200 0 200 400 600 800 1000 1200 1400 0 200 400 800 800 700 700 600 600 500 400 300 200 1000 1200 1400 500 400 300 200 0 200 400 600 800 1000 1200 1400 0 200 400 STAR Reading (Grade 5) 600 800 1000 1200 1400 STAR Reading (Grade 6) 800 800 700 700 600 600 GA Milestones ELA GA Milestones ELA 800 STAR Reading (Grade 4) GA Milestones ELA GA Milestones ELA STAR Reading (Grade 3) 600 500 400 300 200 500 400 300 200 0 200 400 600 800 1000 1200 STAR Reading (Grade 7) 1400 0 200 400 600 800 1000 1200 1400 STAR Reading (Grade 8) Page 13 of 22 Scatter plots: Mathematics 800 800 700 700 GA MilestonesMathematics GA Milestones Mathematics For each student, projected STAR Math™ scores and GA Milestones Mathematics scores were plotted across grades. 600 500 400 300 200 600 500 400 300 200 0 200 400 600 800 1000 1200 1400 0 200 400 STAR Math (Grade 3) 1000 1200 1400 1200 1400 800 GA MilestonesMathematics GA Milestones Mathematics 800 STAR Math (Grade 4) 800 700 600 500 400 300 200 700 600 500 400 300 200 0 200 400 600 800 1000 1200 1400 0 200 400 STAR Math (Grade 5) 600 800 1000 STAR Math (Grade 6) 800 800 700 700 GA Milestones Mathematics GA Milestones Mathematics 600 600 500 400 300 200 600 500 400 300 200 0 200 400 600 800 1000 STAR Math (Grade 7) 1200 1400 0 200 400 600 800 1000 1200 1400 STAR Math (Grade 8) Page 14 of 22 References and additional information Bandeira de Mello, V., Blankenship, C., & McLaughlin, D. H. (2009). Mapping state proficiency standards onto NAEP scales: 2005–2007 (NCES 2010-456). Washington, DC: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics. Retrieved from http://www.schooldata.org/Portals/0/uploads/Reports/LSAC_2003_handout.ppt Cronin, J., Kingsbury, G. G., Dahlin, M., & Bowe, B. (2007, April). Alternate methodologies for estimating state standards on a widely used computer adaptive test. Paper presented at the American Educational Research Association, Chicago, IL. McGlaughlin, D. H., & V. Bandeira de Mello. (2002, April). Comparison of state elementary school mathematics achievement standards using NAEP 2000. Paper presented at the annual meeting of the American Educational Research Association, New Orleans, LA. McLaughlin, D., & Bandeira de Mello, V. (2003, June). Comparing state reading and math performance standards using NAEP. Paper presented at the National Conference on Large-Scale Assessment, San Antonio, TX. McLaughlin, D., & Bandeira de Mello, V. (2006). How to compare NAEP and state assessment results. Paper presented at the annual National Conference on Large-Scale Assessment. Retrieved from http://www.naepreports.org/task 1.1/LSAC_20050618.ppt McLaughlin, D. H., Bandeira de Mello, V., Blankenship, C., Chaney, K., Esra, P., Hikawa, H., et al. (2008a). Comparison between NAEP and state reading assessment results: 2003 (NCES 2008-474). Washington, DC: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics. McLaughlin, D.H., Bandeira de Mello, V., Blankenship, C., Chaney, K., Esra, P., Hikawa, H., et al. (2008b). Comparison between NAEP and state mathematics assessment results: 2003 (NCES 2008-475). Washington, DC: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics. Perie, M., Marion, S., Gong, B., & Wurtzel, J. (2007). The role of interim assessments in a comprehensive assessment system. Aspen, CO: Aspen Institute. Renaissance Learning. (2014). The research foundation for STAR Assessments: The science of STAR. Wisconsin Rapids, WI: Author. Available online from http://doc.renlearn.com/KMNet/R003957507GG2170.pdf Renaissance Learning. (2015a). STAR Math: Technical manual. Wisconsin Rapids, WI: Author. Available from Renaissance Learning by request to research@renaissance.com Renaissance Learning. (2015b). STAR Reading: Technical manual. Wisconsin Rapids, WI: Author. Available from Renaissance Learning by request toresearch@renaissance.com Page 15 of 22 Independent technical reviews of STAR Reading™ and STAR Math™ U.S. Department of Education: National Center on Intensive Intervention. (2015a). Review of progress monitoring tools [Review of STAR Math]. Washington, DC: Author. Available online from http://www.intensiveintervention.org/chart/progress-monitoring U.S. Department of Education: National Center on Intensive Intervention. (2015b). Review of progress monitoring tools [Review of STAR Reading]. Washington, DC: Author. Available online from http://www.intensiveintervention.org/chart/progress-monitoring U.S. Department of Education: National Center on Response to Intervention. (2010a). Review of progress monitoring tools [Review of STAR Math]. Washington, DC: Author. Available online from https://web.archive.org/web/20120813035500/http://www.rti4success.org/pdf/progressMonitoringGOM.pdf U.S. Department of Education: National Center on Response to Intervention. (2010b). Review of progress monitoring tools [Review of STAR Reading]. Washington, DC: Author. Available online from https://web.archive.org/web/20120813035500/http://www.rti4success.org/pdf/progressMonitoringGOM.pdf U.S. Department of Education: National Center on Response to Intervention. (2011a). Review of screening tools [Review of STAR Math]. Washington, DC: Author. Available online from http://www.rti4success.org/resources/toolscharts/screening-tools-chart U.S. Department of Education: National Center on Response to Intervention. (2011b). Review of screening tools [Review of STAR Reading]. Washington, DC: Author. Available online from http://www.rti4success.org/resources/toolscharts/screening-tools-chart Page 16 of 22 Sample dashboard and reports4 Reading Dashboard. The Reading Dashboard combines key data from across Renaissance’s products to provide a 360 degree view of student performance and progress toward goals. At-a-glance data paired with actionable insight helps teachers drive student growth to meet standards and ensure college and career readiness. Educators can review data at the student, group, or class level and various timeframes to determine what’s working, what isn’t, and most importantly, where to focus attention to drive growth. 4 Reports are regularly reviewed and may vary from those shown as enhancements are made. Page 17 of 22 Sample STAR™ Performance Reports focusing on the Pathway to Proficiency. This report will be available to schools using STAR Reading Enterprise or STAR Math Enterprise. The report graphs the student’s STAR Reading or STAR Math scores and trend line (projected growth) for easy comparison with the pathway to proficiency on the PAWS. Page 18 of 22 Sample Group Performance Report. For the groups and for the STAR test date ranges identified by educators, the Group Performance Report compares your students' performance on the STAR Assessments to the pathway to proficiency for your annual state tests and summarizes the results. It helps you see how groups of your students (whole class, for example) are progressing toward proficiency. The report displays the most current data as well as historical data as bar charts so that you can see patterns in the percentages of students on the pathway to proficiency and below the pathway—at a glance. Page 19 of 22 Sample Growth Proficiency Chart. Using the classroom GPC, school administrators and teachers can better identify best practices that are having a significant educational impact on student growth. Displayed on an interactive, web-based growth proficiency chart, STAR Assessments’ Student Growth Percentiles and expected State Assessment performance are viewable by district, school, grade, or class. In addition to Student Growth Percentiles, the Growth Report displays other key growth indicators such as grade equivalency, percentile rank, and instructional reading level. Page 20 of 22 Sample Performance Reports. This report is for administrators using STAR Reading and STAR Math assessments. It provides users with periodic, high level forecasts of student performance on your state's reading and math tests. It includes a performance outlook for each performance level of your annual state tests. The report includes options for how to group and list information. These reports are adapted to each state to indicate the appropriate number and names of performance levels. Page 21 of 22 Acknowledgments The following experts have advised Renaissance Learning in the development of the STAR Assessments. Thomas P. Hogan, Ph.D., is a professor of psychology and a Distinguished University Fellow at the University of Scranton. He has more than 40 years of experience conducting reviews of mathematics curricular content, principally in connection with the preparation of a wide variety of educational tests, including the Stanford Diagnostic Mathematics Test, Stanford Modern Mathematics Test, and the Metropolitan Achievement Test. Hogan has published articles in the Journal for Research in Mathematics Education and Mathematical Thinking and Learning, and he has authored two textbooks and more than 100 scholarly publications in the areas of measurement and evaluation. He has also served as consultant to a wide variety of school systems, states, and other organizations on matters of educational assessment, program evaluation, and research design. James R. McBride, Ph.D., is vice president and chief psychometrician for Renaissance Learning. He was a leader of the pioneering work related to computerized adaptive testing (CAT) conducted by the Department of Defense. McBride has been instrumental in the practical application of item response theory (IRT) and since 1976 has conducted test development and personnel research for a variety of organizations. At Renaissance Learning, he has contributed to the psychometric research and development of STAR Math, STAR Reading, and STAR Early Literacy. McBride is co-editor of a leading book on the development of CAT and has authored numerous journal articles, professional papers, book chapters, and technical reports. Michael Milone, Ph.D., is a research psychologist and award-winning educational writer and consultant to publishers and school districts. He earned a Ph.D. in 1978 from The Ohio State University and has served in an adjunct capacity at Ohio State, the University of Arizona, Gallaudet University, and Minnesota State University. He has taught in regular and special education programs at all levels, holds a Master of Arts degree from Gallaudet University, and is fluent in American Sign Language. Milone served on the board of directors of the Association of Educational Publishers and was a member of the Literacy Assessment Committee and a past chair of the Technology and Literacy Committee of the International Reading Association. He has contributed to both readingonline.org and Technology & Learning magazine on a regular basis. Over the past 30 years, he has been involved in a broad range of publishing projects, including the SRA reading series, assessments developed for Academic Therapy Publications, and software published by The Learning Company and LeapFrog. He has completed 34 marathons and 2 Ironman races. Page 22 of 22 R45876.160330