Every organisation wants a clearer view of how its people are growing. As work becomes more complex and interdependent, we talk about capability, learning agility, systems thinking and relational intelligence as if they are things we can measure with precision. Self-assessments often become the tool of choice. They are easy to administer, simple to scale, and they give us a sense of movement over time.
But a self-assessment is not a mirror. It is a conversation between how I see myself, how I think I am supposed to see myself, and how safe it feels to admit uncertainty. If we rely on that conversation as data, we need to understand both its limits and its quiet strengths.
This piece explores the validity of self-assessments for tracking development, the distortions introduced by unconscious incompetence, and the implications of using this data to shape workforce decisions. Underneath all of it sits a more systemic question: what happens when we try to measure human change inside a living, shifting organisational environment?
The nature of self-assessment
A self-assessment captures a person's current sensemaking. When someone rates their creativity or leadership or ability to work with ambiguity, they are expressing not only a judgment but a worldview. They are telling us what "good" looks like and how close they feel to it. The validity of any given score is fragile. But the pattern of change over time can be meaningful, because it reveals how someone's understanding is evolving.
The difference between a snapshot and a trajectory matters. A single rating tells us very little. A sequence of ratings, especially when paired with narrative reflection, can show a person becoming more aware of nuance, more honest about their limits, or more able to name the specific behaviours that represent progress.
Self-assessments are most accurate for skills that carry little identity weight. People tend to overestimate in areas that feel core to who they are. They also tend to describe what they intend rather than what they do under pressure. So if we treat these scores as precise indicators of capability, we're guaranteed to misread them. If we treat them as windows into a person's meaning-making, the picture becomes more useful.
Unconscious incompetence and the distortion built into growth
The four stages of competence offer a simple reminder: early in learning, people do not know what they do not know. This creates a pattern that shows up again and again in self-reported data.
In the first stage, unconscious incompetence, people often give high ratings because they lack the criteria to judge themselves. They compare against local norms, which might be weak. They may feel confident because they have not yet encountered the complexity of the domain. This is the territory explored by Dunning and Kruger: low skill paired with high confidence.
When organisations interpret these early ratings at face value, workforce planning becomes distorted. People appear more capable than they are. Development resources get allocated to the wrong places. High-stakes decisions lean on data that cannot support the weight placed on it.
The second stage, conscious incompetence, creates the opposite distortion. Awareness grows faster than skill. People start to see how much they do not know, and their scores often drop. On paper, it looks like regression. In reality, this is the beginning of real growth. The person is recalibrating. Their mental model is becoming more accurate.
If we expect upward-only trajectories, we will misinterpret this dip as failure. But when we view self-assessment through a developmental lens, this dip becomes a positive signal.
A holistic view: individuals nested in systems
Emergent holism reminds us that capability is not stored inside a person like a resource. It arises from interactions between the individual and their surroundings. Collaboration emerges from team dynamics, not individual preference. Systems thinking emerges when the environment rewards long-term, cross-boundary reasoning. Psychological safety emerges between people, not within one person's mind.
From this perspective, a self-assessment is not just a statement about the person. It is also a quiet statement about the system around them. Consistently low ratings in influence may signal more about power structures than personal weakness. High ratings in learning agility may reflect aspiration rather than lived experience if the environment punishes experimentation.
When organisations read self-assessments as solely individual data points, they miss these systemic signals. A more useful reading sees patterns across teams and functions as clues about the conditions in which development is or is not emerging.
How valid are self-assessments for tracking growth?
Self-assessments can be valid enough when we ask the right question. Rather than "Is this score accurate?" we can ask "Does the movement of this score make sense in the context of a person's developmental journey?"
Three principles help make this possible.
- Look for change rather than level. We often learn more from the direction and shape of a trajectory than from the number itself. A shift from confident overestimation to a more grounded self-view may signal real maturation.
- Combine numbers with narrative. Small amounts of reflective writing reveal how someone is constructing meaning. They show whether an increase in score reflects genuine growth or simply a shift in interpretation.
- Triangulate. Self-assessment becomes more powerful when paired with observed behaviour, outcomes, and peer perspectives. None of these sources is definitive. Together they form a more coherent picture.
What happens when we use this data for workforce planning?
This is where things get risky. Once self-assessments feed into promotions, succession decisions, or talent pipelines, they stop being purely developmental. They become a steering mechanism.
The first risk is false precision. Dashboards invite us to treat 3.7 as different from 4.1 in a meaningful way. But human development does not move in linear increments. It shifts through experiences, relationships, shocks, and opportunities. Self-ratings capture only a faint outline of that process.
The second risk is gaming. When people know ratings influence career outcomes, they adjust their behaviour. Some inflate. Others, especially those shaped by cultural or gender norms that discourage self-promotion, deflate. Scores then reflect power dynamics rather than capability.
A third risk involves equity. Without careful interpretation, self-assessment data can amplify existing inequities. Groups that tend to under-rate themselves appear less capable on paper. Groups that over-rate themselves appear readier for advancement. The organisation interprets this as individual difference rather than systemic pattern.
To use this data responsibly, workforce planning needs to treat self-assessment as soft evidence. It should spark conversation, not close it. It should inform hypotheses about where growth may be emerging, not serve as the basis for high-stakes decisions on its own.
Designing for wiser use
If we want self-assessments to support genuine human development, we can design for it.
We can name the distortions of early learning so people recognise them in themselves. We can build calibration exercises that help people align their sense of skill with shared examples of practice. We can protect a developmental space where honesty is not punished. And we can read patterns across groups to understand not only individuals but the ecosystem shaping them.
Used this way, self-assessments become less about judging capability and more about supporting growth. They help individuals see themselves more clearly. They help organisations notice the conditions that enable development. And they remind us that human change is not a simple metric but an emergent property of people learning together in complex systems.