The Debrief
The interview is over. Now comes the part that actually determines whether someone gets hired: the debrief. This is where all the signal from every interviewer gets combined into a single decision. A great debrief turns good interviews into good hiring decisions. A bad debrief turns them into groupthink, gut feelings, and wasted effort.
The debrief is where decisions happen
You can run perfect interviews, ask brilliant questions, and take flawless notes. None of it matters if the debrief falls apart. The debrief is the mechanism that converts individual signal into a collective hiring decision — and it has more failure modes than most people realize.
The critical rule: independent write-ups first
Before the debrief meeting starts, every interviewer must submit their evaluation independently. This is non-negotiable. Here's why:
If the senior director speaks first and says "I thought this candidate was outstanding," every other interviewer in the room will unconsciously shift their assessment upward. This is anchoring bias, and it is devastatingly effective at destroying the value of your interview loop.
The fix is simple: everyone writes up their evaluation and submits it before seeing anyone else's feedback. No peeking. No hallway conversations. No "hey, what did you think?" between interviews.
| Without independent write-ups | With independent write-ups |
|---|---|
| First opinion anchors the group | Each interviewer's signal is preserved |
| Dissenting voices stay quiet | Disagreements surface naturally |
| One strong personality drives the outcome | Evidence drives the outcome |
| You get one opinion multiplied by five | You get five independent data points |
Structure of an effective debrief
A good debrief follows a predictable structure. Predictability is the point — it prevents the meeting from devolving into an unstructured conversation where the loudest voice wins.
Step 1: Confirm independent submissions
Before anyone speaks, confirm that every interviewer has submitted their written evaluation. If someone hasn't, they write it now — before hearing from anyone else. No exceptions.
Step 2: Share evidence by principle
Go principle by principle, not interviewer by interviewer. For each principle being evaluated:
- Each interviewer shares the evidence they collected (what the candidate said and did)
- Each interviewer shares their rating for that principle
- The group identifies where assessments agree and where they diverge
| Approach | What happens |
|---|---|
| Interviewer-by-interviewer | "Let me tell you everything about my interview." Others zone out. No structured comparison. |
| Principle-by-principle | "Here's what we all heard about Ownership." Direct comparison. Gaps become obvious. |
Step 3: Discuss gaps and disagreements
When interviewers disagree, that's valuable information. Don't smooth it over — dig into it:
- Did different interviewers see different behaviors from the candidate?
- Did one interviewer probe deeper and uncover something others missed?
- Is the disagreement about what happened, or about what the bar should be?
Step 4: Identify missing data
Sometimes the honest answer is: "We don't have enough data on this principle." That's a legitimate outcome. More on this below.
Step 5: Make the decision
After reviewing all the evidence, the group makes a hire / no-hire decision. This decision should be based on evidence, not vibes.
The hiring bar: "Would I hire this person on MY team?"
There are two ways to think about the bar:
- "Is this person good enough to hire somewhere?" — This is the wrong question. It's too vague and leads to "sure, someone could use them" hires.
- "Would I want this person on MY team?" — This is the right question. It forces you to imagine working with this person daily. It raises the bar because you're accountable for the outcome.
Every interviewer should answer the second question honestly. If you wouldn't want them on your team, that's signal. If three out of five interviewers feel that way, that's a clear pattern.
The Bar Raiser role
Many organizations designate a Bar Raiser (or equivalent role) in every interview loop. This person has a specific job: protect the hiring bar.
| Regular interviewer | Bar Raiser |
|---|---|
| Evaluates the candidate for their team | Evaluates the candidate against the company-wide bar |
| May feel pressure to fill a headcount | Has no stake in this particular role being filled |
| Focuses on the principles assigned to them | Looks across all principles for the overall pattern |
| Can be overruled by the hiring manager | Has veto power on the hire decision |
The Bar Raiser's independence is what makes the role effective. They're not trying to fill a req. They're not the candidate's future manager. They're asking one question: "Does this candidate meet our standard?"
If your organization doesn't have a formal Bar Raiser program, you can still apply the principle: designate one person in every debrief whose explicit job is to hold the line on quality. Give them permission to say no.
Common debrief anti-patterns
These are the ways debriefs go wrong. Learn to spot them.
| Anti-pattern | What it looks like | Why it's harmful |
|---|---|---|
| Loudest voice wins | A senior leader states their opinion first and forcefully. Others fall in line. | You lose independent signal from every other interviewer. The debrief becomes one person's decision. |
| Anchoring | The first person to speak sets the frame. "I thought they were strong" means everyone evaluates relative to "strong." | Later opinions are contaminated. Even honest disagreement gets softened. |
| "I liked them" | An interviewer's entire contribution is "I liked them" or "they seemed smart" with no supporting evidence. | Liking someone is not signal. Without evidence, this is a gut reaction that introduces bias. |
| Recency bias | The last interviewer's assessment carries disproportionate weight because it's freshest in everyone's mind. | Earlier interviews had equally valid signal that gets discounted. |
| Halo effect | The candidate nailed one interview, so everyone assumes they're strong across the board. | Excellence in one area doesn't guarantee competence in others. Evaluate each principle independently. |
| Urgency override | "We really need to fill this role" becomes the deciding factor. | Hiring the wrong person costs far more than leaving a role open another month. |
| Consensus without evidence | Everyone agrees to hire but nobody can articulate why. | Agreement isn't the goal — an evidence-based decision is. |
Resolving split assessments
It's common for two interviewers to see the same principle differently. One rates Ownership as "strong hire," the other as "no hire." This doesn't mean someone is wrong — it means you need to dig in.
Step 1: Compare the evidence, not the ratings.
Put the specific behavioral evidence side by side. Often you'll find:
- They saw different behaviors — the candidate showed strong ownership in one interview and weak ownership in another. Both observations are valid.
- They interpreted the same behavior differently — one interviewer saw "took on a project outside their scope" as ownership; another saw it as "poor boundary management." This is a calibration gap, not a data gap.
- One interviewer probed deeper — they uncovered something the other missed. The deeper data should carry more weight.
Step 2: Ask the group for corroborating evidence.
When you're on the fence or two interviewers disagree, the right move is to ask: "Did anyone else see evidence of this principle — one way or the other?"
Sometimes a third interviewer has a data point that wasn't in their assigned principles but is relevant. A candidate's story about collaboration might have included an ownership moment that the Collaboration interviewer noted but didn't flag. The debrief is where these cross-cutting observations surface.
| What to say | When to say it |
|---|---|
| "I'm split on Ownership. I saw X but also Y. Did anyone else see evidence either way?" | You have conflicting signal from your own interview |
| "Interviewer 2 and I rated Collaboration differently. Can we compare the specific evidence?" | Two interviewers disagree on the same principle |
| "I didn't evaluate Customer Focus, but I noticed something in my interview that might be relevant..." | You have incidental evidence on another interviewer's principle |
Step 3: Decide whether the split is about evidence or about the bar.
If both interviewers have valid evidence and still disagree, the question becomes: "Are we disagreeing about what happened, or about what the bar should be?"
- If it's about what happened → compare evidence, ask for more data points
- If it's about the bar → this is a calibration conversation. What does "strong Ownership" look like for this level? Align on the standard, then re-evaluate the evidence.
Story depth and seniority: how many examples is enough?
During the debrief, a pattern sometimes emerges: the candidate had one or two strong stories that they used to answer multiple questions. Is that a problem?
It depends on the level.
| Level | Expected story depth | What "too few stories" signals |
|---|---|---|
| Entry level / new grad | 1-2 strong stories is normal. They may have limited professional experience. Academic projects, internships, and personal projects all count. | Not a concern unless they can't give ANY specific examples. |
| Mid-level (3-5 years) | 3-4 distinct stories expected. Should be able to cover different principles with different examples. | If they're recycling the same story for every principle, they may lack breadth of experience — but probe before concluding. |
| Senior (5-10 years) | 5+ distinct stories with depth. Should have examples from different roles, projects, and contexts. | Recycling stories at this level is a yellow flag. A senior candidate with only one or two examples may not have the range of experience the role requires. |
| Staff / Principal (10+ years) | Multiple deep stories showing increasing scope and impact. Stories should span leadership, technical depth, cross-org influence. | Limited stories at this level is a significant concern. Flag it in the debrief. |
How to handle it in the debrief:
- Note the pattern, don't assume the cause. "Candidate used the same project for Ownership and Collaboration questions" is an observation. "Candidate lacks experience" is a conclusion that may not follow.
- Check if the interviewer probed for a different example. If the candidate recycled a story and the interviewer didn't ask "Can you give me a different example?", the gap might be in the interview, not the candidate.
- Calibrate for the level. A junior candidate with two strong stories is showing exactly what you'd expect. A senior candidate with two stories needs further evaluation.
- Consider an additional interview if the data is insufficient for the level. One more conversation focused on breadth can resolve the question.
When "not enough data" is the right answer
Sometimes, after reviewing all the evidence, you genuinely don't have enough information to assess a principle. This happens. The right response depends on the situation:
| Scenario | Right response |
|---|---|
| One principle wasn't covered at all due to interview logistics | Schedule an additional focused interview on that principle |
| The candidate gave vague answers and probing didn't help | That itself is data — a strong candidate should be able to provide specific examples |
| Two interviewers got conflicting signal on the same principle | Bring in a third interviewer to break the tie with a targeted interview |
| The candidate is strong everywhere except one area where signal is ambiguous | Decide whether that area is critical for the role. If yes, get more data. If no, weigh the rest. |
"Not enough data" is always better than "I'll just guess." The cost of an extra interview is trivial compared to the cost of a bad hire.
Structuring your notes for the debrief
Your written evaluation should be structured so the debrief panel can quickly understand your evidence. Use this format for each principle you assessed:
Format: Evidence + Principle + Rating
| Component | What to write | Example |
|---|---|---|
| Principle | Which principle you were evaluating | Ownership |
| Evidence | What the candidate said and did — specific facts from your notes | "Noticed payment failures going undetected, analyzed logs independently, identified 12 failure modes, built alerting system over 2 weeks outside sprint commitments. MTTD dropped from 45 min to under 5 min." |
| Assessment | Your rating with a one-sentence rationale | Strong hire. Candidate identified a problem outside their scope, took initiative without being asked, and delivered measurable results. |
| Gaps | What you couldn't evaluate or where signal was weak | "Didn't probe how sprint work was affected — unclear if ownership came at the expense of existing commitments." |
When every interviewer follows this format, the debrief becomes a structured comparison of evidence rather than a free-form discussion of feelings.
Exercise: Write a debrief evaluation
You interviewed a candidate for a senior engineer role. Based on the scenario below, write your debrief evaluation organized by principle, with evidence and a rating.
You interviewed Ravi, a senior backend engineer, for a role on your platform team. You asked him about a time he had to make a difficult technical decision with incomplete information. Here's what he told you: 'About six months ago, our main database was hitting capacity limits and we needed to either scale vertically — bigger machine — or migrate to a sharded architecture. The CTO wanted sharding because it's the long-term solution, but we had a product launch in 8 weeks and sharding would take at least 12 weeks to do safely. I did an analysis of our growth projections and realized vertical scaling would buy us about 14 months before we hit limits again. So I proposed we do vertical scaling now to hit the launch, then start the sharding project right after. I wrote up a one-page decision doc with the cost comparison — vertical scaling was about $2K/month more in server costs, sharding would save us that plus give us 10x headroom. The CTO pushed back because he didn't want to accumulate tech debt. I walked him through the risk analysis — if we tried to shard before launch, we had a 60% chance of slipping the launch date, which would cost us the partnership deal worth about $400K. He agreed to the phased approach. We did the vertical scaling in one week, launched on time, and started the sharding project two weeks later. We finished the migration about four months after that, and we've had zero capacity issues since.' Based on this answer, write your debrief evaluation. Assess the following principles: (1) Sound Judgment / Decision-Making, (2) Ownership, and (3) Communication / Influence. For each principle, include the evidence from Ravi's answer, your rating (strong hire / hire / no hire / strong no hire), and any gaps in signal.
A good debrief evaluation separates evidence from assessment. For each principle, capture what the candidate actually said and did, then give your rating with a brief rationale. Flag any areas where you didn't get enough signal.
This exercise supports AI-powered coaching via Claude. Enter your access code to enable it, or use the offline feedback below.
Exercise 2: Resolve a split assessment
Two interviewers evaluated the same candidate on Collaboration. Read their evidence and decide how to resolve the disagreement.
Interviewer A (Rating: Strong Hire):
"Candidate described leading a cross-team design review for the new API. Set up the meeting, created the agenda, facilitated the discussion, and documented the decisions. Three teams attended and all agreed on the contract."
Interviewer B (Rating: No Hire):
"When I probed about a disagreement with a colleague, the candidate said 'I just escalated it to my manager and let them sort it out.' No evidence of working through the conflict directly. When I asked for another example, they described the same API design review from Interviewer A's question."
Based on these two assessments: (1) Is the split about evidence or about the bar? (2) What does the recycled story tell you? (3) What is your recommendation — hire, no hire, or get more data? Support your answer with reasoning.
Both interviewers have valid evidence. Your job is to reconcile what they saw and determine whether it's a calibration issue, a data gap, or a genuine concern.
This exercise supports AI-powered coaching via Claude. Enter your access code to enable it, or use the offline feedback below.
Exercise 3: Story depth by level
Review these two debrief summaries. For each, assess whether the number of stories is appropriate for the candidate's level.
Candidate 1: Maya — Entry-level software engineer (new grad)
The panel collected evidence from two stories: a capstone project where Maya built a web scraper and an internship where she optimized a database query. Both stories had clear STAR elements with metrics. She recycled the internship story when asked about collaboration, framing it from a teamwork angle.
Candidate 2: David — Senior engineering manager (12 years experience)
The panel collected evidence primarily from one story: a platform migration David led 3 years ago. He used this story to answer questions about Ownership, Collaboration, Dive Deep, and Bias for Action. When interviewers asked for different examples, he gave shorter, less detailed stories that lacked specific metrics or outcomes.
For each candidate, assess: (1) Is the number/depth of stories appropriate for their level? (2) Should the panel be concerned? (3) What would you recommend?
Calibrate your expectations by level. What's normal for a new grad is a red flag for a senior leader.
This exercise supports AI-powered coaching via Claude. Enter your access code to enable it, or use the offline feedback below.
During a debrief, the hiring manager speaks first and says: 'I thought this candidate was exceptional — one of the best I've interviewed this year.' Two other interviewers had concerns but now seem hesitant to share them. What went wrong?
In the debrief, four interviewers rate the candidate as 'hire' but one interviewer — the Bar Raiser — says 'no hire' based on weak evidence of Sound Judgment. The hiring manager says: 'Four out of five is a strong signal, let's move forward.' What should happen?
You're writing your debrief evaluation for a candidate. Which of the following is the strongest write-up for the Ownership principle?