NCES assesses student performance in reading by administering NAEP assessments to samples of students who are representative of the nation's students. The content of the NAEP reading assessment is determined by a framework developed with the help of researchers, policymakers, and the interested public as well as experts in reading and its measurement. Read more about what the assessment measures, how it was developed, and how the assessment was administered.
Since 2003, all states and jurisdictions have participated in the grades 4 and 8 NAEP reading assessments. For grades 4 and 8, the NAEP national sample was obtained by aggregating the samples of public school students from each state and jurisdiction, and then supplementing the aggregate sample with a nationally representative sample of students from nonpublic schools, rather than by obtaining an independently selected national sample. As a consequence, the national sample size increased, and smaller differences between years or between groups of students were found to be statistically significant than would have been detected in previous assessments.
At grade 12, the national sample of schools and students is drawn from across the country, and results from the assessed students are combined to provide accurate estimates of the overall performance of twelfth-graders in the nation.
In 2024, the national results from the reading assessment are compared to 15 previous assessment years at grade 4 and 14 previous years at grade 8 (the 2000 reading assessment was administered at grade 4 only). In 2019, national results are compared to 8 previous assessment years at grade 12. State results for 2024 are compared to 14 previous assessment years at grade 4 and 12 previous years at grade 8. The same framework that has guided assessment development since 2009 was used to guide development of the 2024 digitally based assessments. In 2009 it was determined that, despite the changes to the framework, the results from the 2009 and subsequent assessments could be compared to those results from earlier assessment years. Read a summary of the analysis that was conducted to inform that decision.
The national results from 1998 or later are based on administration procedures in which testing accommodations were permitted for students with disabilities (SD) and English learners (EL). Accommodations were not permitted prior to 1998. Read more about NAEP's policy of inclusion. A split sample design was used in 1998 and NAEP reports show results for both the accommodations permitted sample and the accommodations not permitted sample for that year. For subsequent assessment years, only results from the sample that included accommodations are shown. Any comparisons to 1998 are based on the sample that included accommodations even though significant differences in results when accommodations were not permitted may be noted in figures and tables. Changes in student performance across years or differences between groups of students in 2024 are discussed only if they have been determined to be statistically significant.
The results of student performance on the NAEP reading assessment are presented in two ways: as average scores on the NAEP reading scale and as the percentages of students attaining NAEP reading achievement levels. The average scale scores represent how students performed on the assessment. The NAEP achievement levels represent how that performance measured up against set expectations for achievement. Thus, the average scale scores represent what students know and can do, while the NAEP achievement-level results indicate the degree to which student performance meets expectations of what they should know and be able to do.
Average reading scale score results are based on the NAEP reading scale, which ranges from 0 to 500. The NAEP reading assessment scale is a composite combining separately estimated scales for each type of reading (literary and informational) specified by the reading framework. When the reading framework was changed in 2009, the way the composite reading scale was defined was changed; however, special analyses determined that the 2009 and subsequent results could be compared to those from previous years.
Average scale scores are computed for groups of students; NAEP does not produce scores for individual students. The results for all grades assessed are placed together on one reporting scale. In the base year of the trend line, the assessed grades are analyzed together to create a cross-grade scale. In subsequent years, the data from each grade level are analyzed separately and then linked to the original cross-grade scale established in the base year. Comparisons of overall national performance across grade levels on a cross-grade scale are acceptable; however, other types of comparisons or inferences may not be supported by the available information. Note that while the scale is cross-grade, the skills tested and the material on the test increase in complexity and difficulty at each higher grade level.
Adopted by the National Assessment Governing Board, NAEP achievement levels are performance standards that describe expectations for what students should know and be able to do. For each grade tested, the Governing Board has adopted three NAEP achievement levels: NAEP Basic, NAEP Proficient, and NAEP Advanced. For reporting purposes, the NAEP achievement-level cut scores are placed on the reading scales, resulting in four ranges: below NAEP Basic, NAEP Basic, NAEP Proficient, and NAEP Advanced.
The Governing Board established the cut points for the reading NAEP achievement levels in 1992 based upon the reading content framework and revised the NAEP achievement level descriptions in 2009 to account for the differences between the 1992–2007 framework and the new and current reading framework. A cross section of educators and interested citizens from across the nation were asked to judge what students should know and be able to do relative to the content reflected in the NAEP reading framework. NAEP achievement-level setting is based on the collective judgments of a broadly representative panel of teachers, education specialists, and members of the general public. The authorizing legislation for the National Assessment of Educational Progress (NAEP) requires that the NAEP achievement levels be used on a trial basis until the Commissioner of the National Center for Education Statistics (NCES) determines that the NAEP achievement levels are reasonable, valid, and informative to the public (20 USC § 9622(e)(2)(C)). The NCES Commissioner's determination is to be based on a congressionally mandated, rigorous, and independent evaluation. The latest evaluation of the NAEP achievement levels was conducted by a committee convened by the National Academies of Sciences, Engineering and Medicine 2016. The evaluation concluded that further evidence should be gathered to determine whether the NAEP achievement levels are reasonable, valid, and informative. Accordingly, the NCES commissioner determined that the trial status of the NAEP achievement levels should be maintained at this time.
Item maps illustrate the knowledge and skills demonstrated by students performing at different scale points on the NAEP reading assessment. In order to provide additional context, the cut points for the three NAEP achievement levels are marked on the item maps. The map location for each question represents the probability that, for a given score point, 65 percent probability of obtaining credit at a specific level of constructed-response questions or polytomously scored selected-response questions, a 74 percent probability of correctly answering a four-option single-selection multiple choice type of selected-response question, or a 72 percent probability of correctly answering a five-option single-selection multiple choice type of selected-response question in certain subjects. For dichotomously scored, non-single-selection multiple-choice type of selected-response questions, the position reflects the scale score for students with a 65 percent probability of obtaining credit, unless evidence indicates a high rate of guessing, in which case the probability is recalculated as for a single-selection multiple-choice question. Selected-response question includes question types such as single-selection multiple choice, matching, grid, zone, and in-line choice.
Approximately 40-50 reading questions per grade have been selected and placed on each item map.
The average scores and percentages presented are estimates because they are based on representative samples of students rather than on the entire population of students. Moreover, the collection of subject-area questions used at each grade level is but a sample of the many questions that could have been asked. As such, NAEP results are subject to a measure of uncertainty, reflected in the standard error of the estimates. The standard errors for the estimated scale scores and percentages in the figures and tables presented on this website are available through the NAEP Data Explorer.
Results are provided for groups of students defined by shared characteristics—race/ethnicity, gender, economically disadvantaged status, NAEP SES (socioeconomic status) index, highest level of parental education, type of school, charter school, type of school location, region of the country, status as students with disabilities, and status as students identified as English learners. Based on participation rate criteria, results are reported for various student populations only when sufficient numbers of students and adequate school representation are present. The minimum requirement is at least 62 students in a particular group from at least five
primary sampling units (PSU). However, the data for all students, regardless of whether their group was reported separately, were included in computing overall results. Explanations of the reporting groups are presented below.
Prior to 2011, student race/ethnicity was obtained from school records and reported for the six mutually exclusive categories shown below:
Students who identified with more than one of the other five categories were classified as “other” and were included as part of the "unclassified" category along with students who had a background other than the ones listed or whose race/ethnicity could not be determined.
In compliance with new standards from the U.S. Office of Management and Budget for collecting and reporting data on race/ethnicity, additional information was collected beginning in 2011 so that results could be reported separately for Asian students, Native Hawaiian/Other Pacific Islander students, and students identifying with two or more races. Beginning in 2011, all of the students participating in NAEP were identified by school reports as one of the seven racial/ethnic categories listed below:
Students identified as Hispanic were classified as Hispanic in subsequent years even if they were also identified with another racial/ethnic group. Students who identified with two or more of the other racial/ethnic groups (e.g., White and Black) would have been classified as “other” and reported as part of the "unclassified" category prior to 2011, but from 2011 on were classified as “Two or More Races.” Results for these students are presented under the "Two or More Races" category in the graphics and tables in the report.
When comparing the results for racial/ethnic groups from 2011, 2013, 2015, 2017, 2019, 2022, and 2024 to earlier assessment years, the data for Asian and Native Hawaiian/Other Pacific Islander students were combined into a single Asian/Pacific Islander category. Information based on student self-reported race/ethnicity will continue to be reported in the NAEP Data Explorer.
Results are reported separately for male and female students.
Economically disadvantaged status variable was formerly labeled as eligibility for the National School Lunch Program in the assessments prior to 2024. NAEP first began collecting data in 1996 on student eligibility for the National School Lunch Program (NSLP) as an indicator of poverty for students' socioeconomic status (SES). Based on available school records, students were classified as either currently eligible for the free/reduced-price school lunch or not eligible. Eligibility for free and reduced-price lunches is determined by students' family income in relation to the federally established poverty level. Students from families with incomes at or below 130 percent of the poverty level qualify to receive free lunches and those from families with incomes between 130 and 185 percent of the poverty level qualify to receive reduced-price lunch. If school records were not available, the student was classified as "Information not available." If the school did not participate in the program, all students in that school were classified as information not available. As a result of the passage of the Healthy, Hunger-Free Kids Act of 2010, schools can use a universal meal service option, the "Community Eligibility Provision" (CEP). Through CEP, eligible schools can provide meal service to all students at no charge, regardless of economic status and without the need to collect eligibility data through household applications. CEP became available nationwide in the 2014–2015 school year. Beginning in 2015, NAEP switched from using the free meal application-based NSLP eligibility criteria (as used in the NAEP assessments prior to the 2015) to state-defined economically disadvantaged status, since the free meal application criteria were no longer administered in CEP schools. Changes in the operation of NSLP led the NAEP program to revise its approach to data collection for NSLP eligibility in order to account for the changes due to the introduction of CEP. Beginning with the 2024 assessment, the previously reported NSLP eligibility variable is relabeled as economically disadvantaged status. The NSLP eligible (students eligible for free or reduced-price lunch) category is relabeled as economically disadvantaged, and the NSLP not eligible category is relabeled as not economically disadvantaged. The information not available label remains unchanged.
To better understand SES and its relationship to academic performance, the NAEP program embarked on a series of research and development steps to produce an SES measure
that accounts for more than state-reported proxies of SES (as with economically disadvantaged status or NSLP). An expert panel convened for this purpose
recommended creating
a composite measure of SES
Click to open pdf. using several variables as opposed to using a single variable.
NCES-sponsored research found that using student eligibility for NSLP,
the overall percentage of students eligible for NSLP at the school the student attends, the number of books in the student’s home, and the highest education level of either
parent (grade 8 only) explained more variance in performance than using any one of these variables as the sole measure. Beginning with the 2024 assessment, NAEP will report
results for two SES indices described in the table below. Both 3-component and 4-component NAEP SES indices are available at grade 8, and only the 3-component version is
available at grade 4. The table below shows the component variables with scoring information for each variable category that formed the SES index.
Index type | Component variables | Index categories – complete version | Variable categories – collapsed version |
---|---|---|---|
NAEP SES index (3 components) for grades 4 and 8 | Component 1: Economically disadvantaged status from school records:
Component 2: Percentage of students in school classified as economically disadvantaged, based on NAEP grade sample:
Component 3: About how many books are there in your home? (student-reported)
| 0, 1, 2, 3 ,4, 5, 6, 7, 8, 9 | Low SES (0-2) Middle SES (3-6) High SES (7-9) |
NAEP SES index (4 components) for grade 8 | Component 1: Economically disadvantaged status from school records:
Component 2: Percentage of students in school classified as economically disadvantaged, based on NAEP grade sample:
Component 3: About how many books are there in your home? (student-reported)
Component 4: Parental education: Highest level achieved by either parent (based on student responses to two background questions)
Note: Students who indicated “Unknown” to both questions are treated as missing. | 0, 1, 2, 3 ,4, 5, 6, 7, 8, 9, 10, 11, 12 | Low SES (0-4) Middle SES (5-9) High SES (10-12) |
Results for the SES indices and component variables are available in the NAEP Data Explorer (NDE).
Parents' highest level of education is defined by the highest level reported by eighth-graders and twelfth-graders for either parent. Fourth-graders were not asked to indicate their parents' highest level of education because their responses in previous studies were highly variable, and a large percentage of them chose the "I don't know" option. Parental education attainment is one component used to measure student's socioeconomic status (SES).
The national results are based on a representative sample of students in both public schools and nonpublic schools. Nonpublic schools include private schools, Bureau of Indian Affairs schools, and Department of Defense schools. Private schools include Catholic, Conservative Christian, Lutheran, and other private schools. Results are reported for private schools overall, as well as disaggregated by Catholic and other private schools. Sample size was insufficient to permit a reliable estimate for other private schools in 2005, 2007, 2009, 2011, 2013, 2015, 2017, 2019, 2022, and 2024. The state results are based on public school students only. The participation rate for private schools overall did not meet the NAEP reporting standards in 2024; therefore, their results are not shown in the "type of school" charts on the website. The participation rate for Catholic schools, a subcategory of private schools, did not meet the reporting standards in 2005, 2009, 2013, and 2019 at grade 12; therefore, their results are not available for these assessment years.
A pilot study of America's charter schools and their students was conducted as part of the 2003 NAEP assessments in reading and mathematics at grade 4. Results are available for charter schools starting in 2003 at grade 4, 2005 at grade 8, and 2009 at grade 12. Results for this variable are reported for public school students.
NAEP results are reported for four mutually exclusive categories of school location: city, suburb, town, and rural. The categories are based on standard definitions established by the Federal Office of Management and Budget using population and geographic information from the U.S. Census Bureau. Schools are assigned to these categories in the NCES Common Core of Data based on their physical address. In 2007, the classification system was revised; therefore, results for school location are available for 2007 and later assessment years.
In 2007, the classification system was revised; therefore, trend comparisons to previous years are not available. The new locale codes are based on an address's proximity to an urbanized area (a densely settled core with densely settled surrounding areas). This is a change from the original system based on metropolitan statistical areas. To distinguish the two systems, the new system is referred to as "urban-centric locale codes." The urban-centric locale code system classifies territory into four major types: city, suburban, town, and rural. Each type has three subcategories. For city and suburb, these are gradations of size—large, midsize, and small. Towns and rural areas are further distinguished by their distance from an urbanized area. They can be characterized as fringe, distant, or remote.
Prior to 2003, NAEP results were reported for four NAEP-defined regions of the nation: Northeast, Southeast, Central, and West. As of 2003, to align NAEP with other federal data collections, NAEP analysis and reports have used the U.S. Census Bureau's definition of "region." The four regions defined by the U.S. Census Bureau are Northeast, South, Midwest, and West. The Central region used by NAEP before 2003 contained the same states as the Midwest region defined by the U.S. Census. The former Southeast region consisted of the states in the Census-defined South minus Delaware, the District of Columbia, Maryland, Oklahoma, Texas, and the section of Virginia in the District of Columbia metropolitan area. The former West region consisted of Oklahoma, Texas, and the states in the Census-defined West. The former Northeast region consisted of the states in the Census-defined Northeast plus Delaware, the District of Columbia, Maryland, and the section of Virginia in the District of Columbia metropolitan area. Therefore trend data by region are not provided for the 2005 science assessment. The list below shows how states are subdivided into these Census regions. All 50 states and the District of Columbia are listed. Other jurisdictions, including the Department of Defense Educational Activity schools, are not assigned to any region. In 2009, the results were not available for this variable due to insufficient sample sizes to permit reliable estimates at grade 12.
SOURCE: U.S. Department of Commerce Economics and Statistics Administration.
Results are reported for students who were identified by school records as having a disability. A student with a disability may need specially designed instruction to meet his or her learning goals. A student with a disability will usually have an
Individualized Education Program (IEP) which guides his or her special education instruction. Students with disabilities (SD) are often referred to as special education students and may be classified by their school as learning disabled (LD) or emotionally disturbed (ED). The goal of NAEP is that students who are capable of participating meaningfully in the assessment are assessed, but some students with disabilities selected by NAEP may not be able to participate, even with the accommodations provided. Beginning in 2009, NAEP disaggregated students with disabilities from students who were identified under
section 504 of the Rehabilitation Act of 1973; however, trend results dating back to 1998 are available in reading for the SD variable that includes section 504 students. The results for SD are based on students who were assessed and could not be generalized to the total population of such students.
Results are reported for students who were identified by school records as being English learners (EL). (Note that English learners were previously referred to as limited English proficient (LEP). The results for EL are based on students who were assessed and could not be generalized to the total population of such students.
All 50 states and 2 jurisdictions (District of Columbia and the Department of Defense Education Activity (DoDEA)) participated in the 2017 reading assessment at grades 4 and 8. To ensure that the samples in each state are representative, NAEP has established policies and procedures to maximize the inclusion of all students in the assessment. Every effort is made to ensure that all selected students who are capable of participating meaningfully in the assessment are assessed. While some students with disabilities (SD) and/or English learners (EL) can be assessed without any special procedures, others require accommodations to participate in NAEP. Still other SD and/or EL students selected by NAEP may not be able to participate. Local school authorities determine whether SD/EL students require accommodations or should be excluded because they cannot be assessed. The percentage of SD and/or EL students who are excluded from NAEP assessments varies from one jurisdiction to another and within a jurisdiction over time. Read more about the potential effects of exclusion rates on assessment results.
Use the "Data Quick View" to see additional information about the percentages of students with disabilities and English learners identified, excluded, and assessed at the national, state, and district level in 2024 at
grades 4 and 8.
View this information at the national level for grade 12 in 2019.
Use the "Data Quick View" to see the types of accommodations permitted for students with disabilities and/or English learners at the national level in 2024 at grades 4 and 8, and in 2019 at grade 12.
Exclusion rates for other subjects, as well as rates of use of specific accommodations, are available.
Differences between scale scores and between percentages that are discussed in the results on this website take into account the standard errors associated with the estimates. Comparisons are based on statistical tests that consider both the magnitude of the difference between the group average scores or percentages and the standard errors of those statistics. Throughout the results, differences between scores or between percentages are discussed only when they are significant from a statistical perspective.
All differences reported are significant at the 0.05 level with appropriate adjustments for multiple comparisons. The term "significant" is not intended to imply a judgment about the absolute magnitude or the educational relevance of the differences. It is intended to identify statistically dependable population differences to help inform dialogue among policymakers, educators, and the public.
Comparisons across states use a t-test (the method most commonly used to evaluate the differences in means between two groups) to detect whether a difference is statistically significant or not. There are four possible outcomes when comparing the average scores of jurisdictions A and B:
When comparing all jurisdictions to each other, the testing procedures are based on all pairwise combinations of the jurisdictions in a particular year or pair of years. It may be possible that a given state or jurisdiction has a numerically higher average scale score than the nation or another state but that the difference is not statistically significant, while another state with the same average score may show a statistical significance compared to the nation or the other state. These situations may arise due to the fact that standard errors vary across states/jurisdictions and estimates.
Users are cautioned against interpreting NAEP results as implying causal relations. Inferences related to student group performance or to the effectiveness of public and nonpublic schools, for example, should take into consideration the many socioeconomic and educational factors that may also have an impact on performance.
The NAEP reading scale makes it possible to examine relationships between students' performance and various background factors measured by NAEP. However, a relationship that exists between achievement and another variable does not reveal its underlying cause, which may be influenced by a number of other variables. Similarly, the assessments do not reflect the influence of unmeasured variables. The results are most useful when they are considered in combination with other knowledge about the student population and the educational system, such as trends in instruction, changes in the school-age population, and societal demands and expectations.
Return to the reading subject information.