If we are serious about using assessment to support learning, then we have to be equally serious about how those systems are experienced by the people inside them.
For most of my career, I’ve worked at the intersection of classroom practice, large-scale assessment systems, and public policy. I began as a classroom teacher, moved into district and state leadership, and eventually served as Director of Assessment at the Oregon Department of Education. Over time, I noticed I was hearing from colleagues more and more often, “You talk about assessment differently.”
At first, I honestly didn’t know what they meant.
However, I came to realize something important: Most conversations about assessment are dominated by technical quality, compliance, and systems management. Those things matter, they matter deeply, but they are not the whole story.
Where This Work Comes From
There was a moment that shifted how I understood the work of large-scale assessments. It involved a parent, a “good mom.” The kind of parent we say we want engaged in their child’s learning. Yet, through our systems, processes, and communication choices, we created an experience that left her confused, frustrated, and questioning both herself and her child. You can read the story here.
Nothing about that situation was caused by a lack of technical quality. In fact, from a measurement perspective, everything was working as designed.
And that’s the problem.
Because if a system is technically sound but produces harm, confusion, or disengagement for the people it is meant to serve, then we have to ask a harder question:
What exactly are we optimizing for?
The Core Belief
This question sits at the center of my work and the work of Metimur:
Assessment systems should be designed and implemented in ways that are both technically sound and human-centered.
Not one or the other. Both.
We have spent decades building increasingly sophisticated models, leveraging online testing to capture more nuanced evidence of student learning, and expanding reporting systems to tell richer stories about that learning from shorter tests. We have worked to scale these systems across states, often in pursuit of comparability, even when that comparability is more assumed than realized.
Yet, too often, the lived experience of those systems by students, families, and educators remains an afterthought. Compassionate Assessment is not about lowering standards or abandoning rigor. It is about aligning rigor with responsibility.
What Is Compassionate Assessment?
Compassionate Assessment is a framework for designing and implementing assessment systems that center the people they are intended to serve, without compromising technical quality.
It recognizes that assessment is not just a technical exercise. It is also a human experience.
Those two dimensions are inseparable.
The Four Pillars of Compassionate Assessment
The framework is built on four overlapping and interdependent pillars:
1. Technical Quality
This is the foundation.
Validity, reliability, fairness, alignment, and accessibility are non-negotiable. Without technical quality, assessment results cannot be trusted or used responsibly.
However, technical quality alone does not ensure that a system is effective in practice.
2. Assessment Environment
This includes the conditions under which assessment occurs.
It is not limited to the physical or digital testing environment, but also encompasses the local implementation practices that shape those environments. Decisions about scheduling, messaging, technology access ,and functionality all influence how assessments are experienced.
These factors directly affect how students show up to the assessment and what they are actually able to demonstrate. A technically sound assessment administered under inconsistent or unfair conditions will not yield meaningful results.
3. Adult Attitudes & Beliefs
Educators, leaders, and policymakers do not enter assessment systems as neutral actors.
They bring with them a set of beliefs about what assessment is for, what it measures, and how results should be interpreted and used. Those beliefs show up in visible ways. Through the decisions they make, the messages they communicate, and the expectations they set. In other words, attitudes are the outward expression of underlying beliefs.
If we focus only on shifting attitudes (what people say or do) without addressing the beliefs driving those actions, change will be surface-level and short-lived. The way adults frame assessment, intentionally or not, shapes how it is experienced by students and families.
If assessment is treated primarily as compliance or accountability, that belief will influence everything from implementation decisions to how results are communicated and used, regardless of the technical design of the system.
4. Student Attitudes & Beliefs
Students are not passive participants in assessment systems.
They are constantly making sense of what tests are asking of them, what their performance means, and whether the system is fair. These interpretations are rooted in their beliefs, which then show up in their attitudes and how they approach the assessment, the effort they invest, and the degree to which they engage with the experience.
As with adults, focusing only on student behaviors without understanding the power of the beliefs beneath them limits what we can meaningfully influence.
A student who disengages may not be responding to the content of the assessment itself, but to what they believe it represents about their ability or their future. When those beliefs go unexamined (or worse assumed) we risk interpreting results as a reflection of learning alone, when in reality they may also reflect confusion, mistrust, or a lack of connection to the purpose of the assessment.
Why This Matters
In practice, assessment systems are often designed and discussed as if these four pillars operate independently, if we even address all four at all. Technical quality is addressed in one conversation. Test administration logistics in another. Communication and use of results somewhere else entirely.
But that is not how these systems are actually experienced.
For students, families, and educators, these elements are not separate. They show up all at once, interacting in ways that either reinforce clarity and trust or create confusion and doubt.
A technically strong assessment administered in an inconsistent environment produces results that misrepresent what students know. Clear reporting can still fall flat if the adults interpreting it carry beliefs that limit how those results are understood or used, and even the most carefully designed system can yield distorted outcomes when students disengage because the experience itself doesn’t make sense to them.
What we often treat as isolated design decisions are, in reality, deeply interconnected conditions shaping what assessment data actually represents.
When we focus narrowly, most often on technical quality, we don’t eliminate problems. We relocate them. The system may function as intended on paper, while producing unintended consequences in practice, and those consequences are not abstract. They show up in how students perform, how educators respond, and how families interpret what they are seeing.
A Shift in Responsibility
Recognizing this interconnectedness requires us to rethink what we mean by success.
For a long time, success in assessment systems has been defined by precision and efficiency: accurate scores, clean data, and reports delivered on time. Those are important outcomes. They reflect a system that is functioning as designed.
However, they are not sufficient.
A system can meet every technical benchmark and still fail to serve the people it was built for. Students may leave the experience unsure of what was expected of them. Families may struggle to make sense of results that are technically precise but practically inaccessible. Educators may receive data that is difficult to translate into meaningful instructional decisions. Over time, these gaps erode trust, not because the system is inaccurate, but because it is incomplete.
Compassionate Assessment expands the definition of success to include these realities. It asks whether the system supports understanding, not just measurement. Whether it enables use, not just reporting. Whether it builds trust, not just compliance.
This is not about lowering expectations or softening rigor. It is about recognizing that rigor, on its own, is not enough. A system that is technically sound but humanly disconnected will always fall short of its intended purpose.
Where We Go From Here
This overview is the starting point.
Each pillar is explored in more depth through dedicated resources that translate this framework into real-world practice across classrooms, schools, districts, and state systems.
Perhaps more importantly, this work is already happening, whether intentionally or not.
Assessment systems are shaping student experiences every day. The question is whether we are willing to examine those systems more closely, learn from them, and begin designing them with the same level of care we bring to their technical quality.
I’ll also return to the story of that parent because it illustrated to me more clearly than any framework, what happens when we overlook the human side of implementation.


Leave a Reply