For decades, the American education system has compromised its own promise by measuring what is easy, rather than what matters most. Because assessments act as powerful signals, they shape what is taught and learned. Yet, by relying on “drop-in-from-the-sky” standardized tests that reward isolated fact fluency and procedural mimicry, we calcify our educational objectives in the past. This approach disconnects schooling from the preparation students desperately need to navigate a future defined by rapid technological acceleration and profound societal uncertainty.
Thriving in the twenty-first century requires far more than basic factual recall. Reading, writing, and arithmetic must be joined by vital skills such as: creativity, collaboration, and computer literacy. The mark of an educated person is no longer the sheer quantity of knowledge retained, but rather their knowledgeability and intellective competence—the ability to actively use knowledge, strategies, and values to solve novel problems. Yet, traditional tests serve merely as retrospective audits—static snapshots that ignore how learning actually happens. To deliver on the promise of human potential, we must rebalance from the “assessment OF education” to “assessment FOR education,” transforming measurement into a dynamic pedagogical transaction that actively improves teaching and learning.
Embracing Broader, Holistic Definitions of Learning and Context
Student growth unites academic achievement with social, emotional, and cultural development. Learning does not unfold in a vacuum, and neither should assessment. As Pamela Cantor M.D. & Kate Felsen argue, unlocking human potential requires a paradigm shift that debunks the myth of fixed intelligence. We must treat human variance not as “noise,” but as an asset leveraged for learning and justice.
Grounded in the RISE Principles—which recognize that learning is Rooted in biology, Integrated with identity, Shaped by culture, and Experienced in the body (Pea, Lee, Nasir, & McKinney de Royston)—our tools must capture the entire learning ecosystem. By crediting the varied, durable strengths students develop within their communities, asset-based assessments can measure a holistic profile of human agency.
Measuring the Processes of Learning
The tyranny of the “final correct answer” obscures human cognition. To understand learning, we should measure the underlying processes, including: self-regulation, motivation, attention, and engagement. Useful web-based tools that center student voice—such as the School Climate Walkthrough developed by Christina Cipriano and her colleagues—can provide real-time, actionable feedback. By making these invisible processes visible, a student’s response to a setback becomes just as illuminating as their success.
Tracking metacognition and what researchers term “productive failure” (Foster & Piacentini) reveals exactly how students adapt when they lack immediate answers. Educators are adopting practical measurement for improvement (LeMahieu & Cobb). Instead of waiting for end-of-year educational autopsies, teachers embed low-burden leading indicators directly into daily Plan-Do-Study-Act (PDSA) workflows. Measuring these real-time shifts ensures immediate “validity-in-use” to drive continuous instructional growth.
Measuring Complex, Durable Skills for the Future
Cultivating student agency requires measuring the durable skills that multiple-choice bubbles inherently miss: collaboration, communication, critical thinking, perseverance, and AI literacy. Above all, we must prioritize intellective competence. A modern, skills-based vision, like the one championed by Liu, Liu, Sherer, & LeMahieu and the Skills for the Future (SFF) team, values affective and behavioral skills alongside cognitive domains. Crucially, this approach validates competencies acquired across multiple pathways, acknowledging that durable skills forged in out-of-school, workforce, and community settings matter just as much as those learned at a desk.
To capture this comprehensive profile, the “Observation” vertex of the Assessment Triangle (Pellegrino) must evolve. We are shifting from static exams to immersive, AI-driven digital simulations. As Jack Buckley & Erica Snow demonstrate, Game-Based Assessment within environments like Roblox enables the “stealth assessment” of complex traits like systems thinking. By analyzing continuous telemetry—strategic choices, chat logs, clickstreams, and physical gestures—in multimodal, collaborative environments, we might seamlessly measure the dynamic application of teamwork without interrupting the flow of learning.
Reimagining Core Subjects in Authentic Contexts
This transformation must also extend to our core subjects. Next-generation assessments in math and language arts are abandoning decontextualized abstractions and framing challenges in authentic, real-world scenarios that measure deep “knowledge-in-use.”
To rebuild this cognitive foundation, developers are utilizing Principled Assessment Design (PAD) (Huff). Instead of writing multiple-choice items and attaching standards post-hoc, PAD maps the learner’s cognitive journey from novice to expert before a task is ever designed. This proactive blueprint yields “low-floor, high-ceiling” performance tasks that authentically measure a student’s capacity for cross-disciplinary literacy and conceptual transfer.
Changing What is Possible
By redefining what we measure, we transform assessment from a retrospective audit into a dynamic engine for human agency. Powered by multimodal AI, the learning science and technological infrastructure required to scale this vision are already maturing. Now, system leaders, policymakers, and philanthropists must aggressively redirect investments away from twentieth-century compliance and toward modern, dynamic assessment ecosystems. We cannot build tomorrow’s intellective competence on yesterday’s testing infrastructure. When we innovate what we measure, we change what is possible. It is time to commit to measuring what matters most.
