Natalie Wexler challenges a fundamental assumption in American education: that reading is merely a set of transferable skills like finding the main idea, rather than a product of specific knowledge. Her analysis of a quietly abandoned experiment in Louisiana offers a rare, data-backed glimpse into what happens when tests finally measure what students actually know, rather than their ability to guess at unfamiliar topics.
The Knowledge Gap
Wexler begins by dismantling the standard model of reading assessment. She notes that while states claim to test what is taught, the reality is that standardized tests often rely on passages about obscure topics—like rugby—that assume a background knowledge many students, particularly those from less educated families, simply do not possess. "To be able to make an inference, for example, you need a certain threshold of relevant background knowledge," Wexler explains, highlighting how the current system penalizes students for gaps in their world knowledge rather than their reading ability.
This framing is crucial because it shifts the blame from the student's lack of skill to the test's lack of context. The author argues that the traditional approach creates a "superficial" instructional environment where teachers drill isolated skills instead of building robust understanding. As she puts it, "The problem, as then-state superintendent of Louisiana John White explained... is that those skills don't transfer from one context to another." This observation lands with particular force for busy professionals who understand that expertise in any field comes from deep domain knowledge, not just generic problem-solving techniques.
The Louisiana Experiment
The core of Wexler's piece is the story of Louisiana's "Innovative Assessment Pilot," a bold attempt to align testing with a content-rich curriculum. Instead of generic reading passages, the pilot tested students on specific units of history and literature they had just studied, including "warm" reads that connected thematically to the core text. The results were striking. A white paper from NWEA found that students felt less anxious and more engaged. More importantly, the achievement gap between economically disadvantaged students and their affluent peers was "significantly smaller than on the LEAP."
Wexler writes, "The new test design may be 'leveling the playing field' by providing students a more equitable opportunity to show what they know." This is the piece's most compelling evidence: when the test rewards knowledge acquisition, it stops being a gatekeeper for the privileged and starts being a tool for equity. The experiment also transformed teacher behavior. Educators stopped wasting time on generic test prep and began focusing on the actual content. "We don't do that anymore," one teacher told White. "We devote our time to diving into the unit and making sure that students have a strong understanding, as much background knowledge as we can possibly give them."
Critics might note that the pilot's success relied heavily on the fact that 80% of Louisiana schools used the same curriculum, a level of standardization that is nearly impossible to replicate in other states with fragmented systems. However, Wexler acknowledges this limitation while arguing that the principle remains valid: if the test changes, the teaching changes.
"What gets tested gets taught," according to a timeworn but clearly evidence-based adage. If we continue to test illusory skills, that's what teachers will continue to focus on, to the continued detriment of many students.
The Political and Practical Hurdles
Despite the promising data, the experiment was quietly discontinued in 2024. Wexler attributes this to a mix of administrative turnover, the disruption of the pandemic, and the high cost of implementation. She points out a deeper structural issue: the disconnect between curriculum experts and the psychometricians who design tests. The latter prioritize statistical reliability over educational substance. "There's not great curiosity, in that technocratic worldview, of getting under the messy hood of, well, was it worth it?" Wexler quotes White asking. "Did kids learn the content that people need to be productive?"
The author suggests that the federal government could help by funding research into new assessment models, noting a potentially surprising political dynamic. She observes that the current Republican administration might be more open to such experimentation than previous Democratic ones, which often viewed testing reforms as threats to civil rights protections. This nuance is vital; it suggests that the path forward may not be ideological but rather pragmatic, requiring a coalition that values outcomes over process.
Bottom Line
Wexler's strongest argument is that the current testing regime is not just ineffective, but actively harmful, forcing teachers to ignore the very content that would help disadvantaged students catch up. The piece's vulnerability lies in the immense logistical and political difficulty of scaling a content-based test across a fragmented national system. The reader should watch for whether any state other than Louisiana is willing to take the risk of aligning their assessments with what students actually learn, rather than what they can guess.