The Skills-First Trap

Most hiring processes are built around a simple premise: find the person who knows the most. Screen for certifications, years of experience, and technical knowledge. Give them a test, score it, rank the candidates, hire the top result.

It makes sense on paper. In practice, it selects for a very specific kind of person — someone who is good at having answers. Not someone who is good at finding them. And in any environment that changes faster than a textbook gets updated (which is every environment), those are very different qualities.

I've watched teams hire the most technically qualified candidate on paper and end up with someone who can't adapt when the requirements shift, won't ask for help when they're stuck, and quietly builds the wrong thing for three weeks rather than admit they didn't understand the brief. The knowledge was there. The behaviour wasn't.

One case sticks with me. A team I was advising hired a senior developer who aced every technical question — scored higher than any candidate they'd seen. Within two months, they'd introduced a critical architectural decision without consulting the team, dismissed feedback from a junior who'd flagged a legitimate concern, and when the system broke in production, their first response was to explain why their approach was theoretically correct. The team spent a week cleaning up the damage. The hire didn't last three months. On paper, they were the best candidate in the room. In practice, they were the most expensive mistake that quarter.

What I Actually Test For

I've written the evaluation tests for new hires at multiple organisations. The structure looks like a technical assessment, but the design is more deliberate than most candidates — or hiring managers — realise.

Every test I build has three categories of questions, though the candidate never sees that structure:

  • Knowledge questions — These are straightforward. They test whether the candidate has the baseline technical understanding the role requires. If they can't answer these, the conversation is probably over.
  • Logic questions — These test problem-solving approach. There's a correct answer, but getting there requires reasoning through something unfamiliar. I'm watching how they work, not just what they produce.
  • Behaviour questions — These are the ones that matter most. They're intentionally difficult — sometimes beyond what we'd expect anyone at this level to know. We don't expect them to get the answer. We want to see what they do when they can't. These reveal how someone operates under uncertainty — whether they adapt, engage, or shut down. That's team-operating fit, and no résumé will tell you that.

The Impossible Question

This is where the real signal is. When a candidate hits a question they genuinely don't know the answer to, one of several things happens:

  • They guess confidently and get it wrong
  • They leave it blank and move on
  • They ask questions — probing the problem, trying to narrow it down
  • They engage with the person available in the room and try to work through it collaboratively
  • They reason through it out loud, showing their thought process even without certainty

Each of these responses tells me something the rest of the test can't. The candidate who guesses confidently is showing me how they'll behave when they're stuck on a real project — they'll push forward with assumptions rather than verify. The candidate who asks questions is showing me they'll use the resources around them. The candidate who engages and collaborates is showing me they'll function well on a team.

The question isn't designed to test what they know. It's designed to reveal what they do when they don't know.

If they end up getting the answer right — even if someone guided them there — that's a positive signal, not a failure. It means they can learn in real time with support. That's exactly the behaviour you want on a team.

Why "Confidently Wrong" Is the Biggest Red Flag

Given a choice between two candidates — one who answered every question but one correctly and got that last question confidently wrong, and one who missed a few but asked smart questions and collaborated to reach better answers — I'll take the second candidate every time.

Here's why: the confidently wrong candidate has a blind spot they don't know about, and a pattern of pushing through without checking. On a test, that costs one question. On a production system, that costs an outage. In a client relationship, that costs trust. And because they're confident, they won't flag it. You'll find out when something breaks.

The Real Risk

A knowledge gap can be closed in a week. A behavioural pattern — especially one the person doesn't recognise in themselves — takes months to surface and longer to change. Hire for the pattern, train for the gap.

The candidate who asks questions, who says "I'm not sure about this — can I walk through my thinking?", who engages with the team to get to a better answer — that person will do the same thing on day one of the job. And day one hundred. That's not a weakness. That's exactly how high-performing teams operate.

Culture Isn't a Vibe — It's an Operating Model

When people hear "hire for culture," they often think of personality fit — does this person seem like someone we'd enjoy working with? That's part of it, but it's the least important part.

Culture, in the context that actually affects outcomes, is an operating model. It's the set of behaviours a team defaults to under pressure:

  • When someone is stuck, do they ask for help or go silent?
  • When someone makes a mistake, does the team learn from it or assign blame?
  • When priorities shift, does the team adapt or resist?
  • When someone doesn't know something, is that safe to say out loud?

These behaviours aren't trained in onboarding. They're selected for — or against — in the hiring process. If your evaluation only measures knowledge, you're only selecting for knowledge. Everything else is a coin flip.

Skills Can Be Taught. Behaviour Is Harder.

A junior developer who asks great questions, learns from every review, and collaborates naturally will outperform a senior developer who works in isolation and doesn't adapt — within a year, sometimes less. I've seen it happen repeatedly.

Technical skills have a well-defined learning path. Languages, frameworks, tools, platforms — these are teachable. They have documentation, courses, mentors, and practice environments. A motivated person with the right behaviour will acquire them.

But the willingness to learn, the instinct to collaborate, the humility to say "I don't know" — those are much harder to develop in someone who doesn't already have them. They're not skills. They're dispositions. And they're what separate a team that can grow from a team that stays stuck.

I'd rather hire someone who doesn't have all the answers but knows how to find them than someone who has all the answers but one — and got that one confidently wrong.

Designing Better Evaluations

If this resonates, here's where to start:

  • Include at least one question in every evaluation that's beyond expected skill level — and watch what the candidate does with it
  • Make a knowledgeable team member available during the evaluation so candidates have the option to ask questions
  • Score process and approach separately from correct answers — weight them equally or higher
  • Debrief as a team after evaluations, discussing behaviour patterns, not just scores
  • Treat "I don't know, but here's how I'd find out" as a stronger answer than a confident guess

None of this means you lower the bar. You're still testing for competence — the knowledge questions handle that. But you're also testing for something most evaluations miss entirely: how this person will actually function on your team.

The Bottom Line

Hiring is the highest-leverage decision a team makes. Every person you add either strengthens the operating culture or dilutes it. A test that only measures what someone already knows will only tell you what they've already learned — not whether they can learn what's next.

If your hiring process is selecting for knowledge but leaving behaviour to chance, the evaluation itself is the blind spot. I help teams redesign their technical evaluations and interview loops to test for adaptability, learning behaviour, and team-operating fit — not just what someone can recite from memory. Let's talk about what that looks like for your team.