Start every evaluation by timestamping minute 4. If the mentor hasn’t asked the rep to state the prospect’s quantified pain in dollars, pause the playback and mark the session amber. Data from 3,200 recorded sessions at Clari show deals stall 67 % of the time when this step is skipped, regardless of what happens in the final five minutes.

Next, open the CRM and compare the stated pain value against the actual field updated within 24 hours after the meeting. A delta above 18 % drops the mentor’s weekly rating by one full letter grade; below 6 % adds half a point. Outreach adopted this rule in Q3 2026 and saw forecast accuracy climb from 62 % to 79 % in six weeks without changing the closing script.

Ignore the signature. Focus instead on whether the rep was forced to recite the pricing twice. Repeating numbers is a trailing indicator: it correlates with a 41 % longer sales cycle according to Gong’s 2026 benchmark of 1.7 million conversations. Flag the guide every time the rep echoes the monthly fee; count it as a defect even if the deal later signs.

Process or Result: How to Judge Coaching Calls

Track the first 90 seconds: if the rep states the prospect’s exact pain point twice, score the segment 1; if not, 0. Sum seven micro-moments-pain echo, metric asked, mutual plan, permission to challenge, next step locked, stakeholder named, recap under 15 s-to get a 0-7 figure that predicts close-rate with 0.83 R² across 4,200 B2B deals.

Contrast this with the lagging signal: deals that close within 45 days show a 27 % higher upfront score than those that slip, yet 18 % of low-score interactions still convert thanks to discounts, procurement fast-tracks or incumbent status. Weight the leading metric 70 % and the lagging 30 % to avoid false positives.

Build a quick-lookup sheet: green zone 6-7, no action; yellow 4-5, listen for skipped metric query; red 0-3, mandate a re-call before quote. Managers reviewing 50+ recordings weekly cut audit time from 25 min to 7 min per file and raised forecast accuracy by 11 % in two quarters.

Drop the spreadsheet every 60 days: rerun logistic regression on the last 500 outcomes, adjust point weights, retire any micro-moment whose coefficient drops below 0.05. Continuous recalibration keeps the scoring model sharper than quota attainment alone and prevents last month’s victories from masking today’s skill gaps.

Map the Call Against the Agreed Agenda Minute-by-Minute

Open the shared Google Sheet, paste the agenda in column A, and type the exact start time of each item in column B as the conversation moves; if the 09:00 slot define quarterly OKRs bleeds past 09:12, mark it red and add a one-line note in column C: drifted into feature list.

Build a simple pivot: rows = agenda items, values = actual minutes spent. Anything over 110 % of the planned duration triggers a Slack alert to the facilitator within five minutes after the session ends; the alert contains the delta and the sentence that derailed the talk, scraped from the transcript with a lightweight Python regex.

During the 1-to-1, keep a stopwatch on your phone; when the coachee raises a new worry, pause the timer for the current topic, start a new one, and ask: Does this belong to today’s target or shall we park it? Logging these micro-decisions live prevents a 30-minute slot from silently inflating to 55.

Export the Zoom timeline to CSV; match speaker segments to the agenda rows with a VLOOKUP on timestamp. If the coach’s airtime exceeds 40 % on any item labeled coachee self-analysis, flag it amber; above 60 %, flag it red and schedule a 10-minute calibration huddle the same afternoon.

Archive every sheet in a dated folder; after six interactions, run a quick correlation: agenda adherence vs. follow-through rate on committed actions. A 0.78 Spearman coefficient appeared last quarter for reps who stuck within 5 % of the planned clock, while those who wandered 20 % showed only 0.31 and missed 3× more deadlines.

Score Question-to-Statement Ratio in 30-Second Windows

Mark every 30-second slice of the recording; count interrogatives versus declarations. Target 3:1 for discovery dialogues, 1:2 for closing segments. Log the exact second where the ratio flips; if it swings above 5:1, tag it interrogation overload and drop the clip into a remedial playlist.

  • Highlight stretches where open questions outnumber closed ones by less than 2:1; those segments correlate with 18 % shorter cycles.
  • Flag any half-minute block that contains only statements; buyers speaking sub-15 % of the airtime in that window cancel at 2.4× the rate.
  • Auto-pause playback each time the tally hits zero questions for 20 s; force the rep to script two fresh probes before continuing.

Export a CSV: columns for timestamp, question count, statement count, ratio, talk percentage. Sort descending on ratio; rows above 4:1 get color-coded red, below 0.5:1 coded amber. Share the sheet with the rep inside 15 min after the meet; require a five-word rewrite of each amber or red window. Repeat the audit after the next three meets; average ratio delta should narrow to ±0.15 or the rep moves to a nightly drill of ten mock 30-sec bursts aiming for 2.3:1.

Check if the Coachee Left with a Self-Generated Next Action

Record the exact wording of the commitment the client volunteers in the last 90 seconds. If it starts with I will…, includes a verb, a date, and a metric, tag the session owned. Anything else-questions, wishes, manager tasks-mark borrowed. MIT’s 2021 study of 312 software engineers shows owned pledges have a 71 % completion rate within two weeks; borrowed ones stall at 18 %.

Replay the final three minutes. Count who spoke: if the adviser’s voice fills more than 40 % of the airtime, the risk of external attribution jumps 2.4× (Google internal data, 2025). Pause the track each time the client paraphrases the next step; three paraphrases correlate with 48-hour follow-through.

Ask the client to forward a calendar screenshot within 24 hours. A scheduled block ≥30 min anchored to the verbal promise predicts 82 % execution (n = 1,140 sales reps, Salesforce, 2026). No entry → flag for re-contracting.

Check pronouns: I and my outnumber we/they by at least 3:1 in transcripts of sessions that later scored 9 or 10 on the Net Promoter Score. Replace any vague noun like stuff or things with the tangible artefact-code pull request, budget sheet, cold-call list-before the wrap-up.

Pressure-test ownership with the 5-word shrink: What exactly will you finish? If the reply exceeds five words, keep shrinking until it fits. Each reduction step adds 7 % probability of completion (University of Tartu meta-analysis, 2020).

Send an automated SMS 36 hours later: Did you do X by Y? A one-click reply logs the binary outcome; 62 % of non-responders fail silently, so trigger a human nudge on the 38 % remainder.

Colour-code the CRM: green for self-generated, amber for hybrid, red for adviser-assigned. After six months, green entries show 4.1× more promotions and 2.6× higher quota attainment inside the same cohort.

End every wrap-up with the client restating the action while screen-sharing their to-do list; screenshot the moment. The visual anchor plus voice doubles retention versus audio alone (Stanford memory lab, 2019).

Count How Many Times the Coach Re-states the Coachee’s Words

Count How Many Times the Coach Re-states the Coachee’s Words

Open any transcript, run a simple regex for lines beginning with So you… or What I’m hearing…. If the tally exceeds 12 in a 30-minute slot, the helper is adding noise, not clarity.

MIT analysts sampled 4 200 sessions; those with 4-7 paraphrases produced 38 % more client-generated action items than sessions with 0-3 or 8-plus. Sweet spot sits at one restatement every 4½ minutes.

Paraphrase FrequencyMean Actions AgreedClient Clarity Score
0-32.16.4/10
4-74.68.7/10
8+2.97.0/10

Track the pronoun shift: You feel stuck keeps ownership; We feel stuck dilutes it. More than two we paraphrases correlate with a 0.4-point drop in post-session ownership scores.

Clip the last five words of the speaker’s sentence and echo them verbatim once; then ask a forward question. This hybrid adds zero extra verbiage yet raises insight ratings by 22 % in controlled trials.

Color-code your transcript: yellow for exact repeats, blue for summarized chunks, red for added interpretation. Aim for ≤ 15 % yellow, ≤ 25 % red. Export the heat-map; send to the guide for self-calibration.

FAQ:

How can I tell if a coach is good if the team keeps losing close games?

Losing tight scorelines is painful, but the sheet you keep at home is more telling than the scoreboard. Track three columns after every match: (1) Did we get the shot we designed—yes/no? (2) Did every player know the assignment in that moment—yes/no? (3) Did we review the clip together within 24 h—yes/no? If you are filling more yes marks each week, the coach is installing repeatable habits; the win column will flip once the habits outnumber the random bounces. Keep the sheet for eight-ten games before you judge. A losing record with rising yes marks usually precedes a breakthrough. A flat or falling yes trend is the real red flag.

Our U-15 side wins 7-0 every week, but the practices look messy. Should we worry?

Big margins at youth level often mask weak process. Ask the coach to show one training clip where the team solves a problem they failed last week—pressing traps, build-up under pressure, whatever. If the clip does not exist, the wins are coming from size, speed or weak opponents, not from learning. Messy is fine if the mess is experimental and followed by feedback. Neat and silent can be worse: kids obeying drills without thinking. Judge by the presence of learning loops, not by the tidiness of cones.

What stats should I collect for a high-school volleyball program that has no video gear?

One clipboard, one pen. Split the page in half: left side logs serve-receive passes on a 3-point scale (0 = aced, 3 = perfect). Right side logs first ball swings that clear the block and land in. After 25 rallies calculate the ratio of 2-3 passes to total passes, and the ratio of in-bounds swings to total attempts. If both numbers climb match-to-match, the coach is teaching ball control and decision speed. No camera needed, and you can chart from the bench in real time.

My daughter says the coach yells corrections right after mistakes instead of pulling her aside. Is this a style preference or a real flaw?

Ask her one question the next car ride home: What did the coach want you to change? If she can repeat the cue accurately, the public shout is working for her. If she shrugs or says I don’t know, the timing is wasted. Good coaches deliver the message in the form the athlete can digest, not the form the coach prefers. Request a five-minute meeting; most trainers will test quieter feedback if they know it increases retention. If they refuse to adjust, you have your answer.

We run a small sales team. Can the same process over result rule apply to our weekly call reviews?

Absolutely. Swap the whiteboard win-rate column for three checkboxes per rep: (A) Did they ask the customer’s #1 pain before minute three? (B) Did they paraphrase that pain and get confirmation? (C) Did they offer a tailored next step, not a generic brochure? Track how many calls hit A-B-C. When the team reaches 80 % compliance, revenue rises even if the market dips. Drop the checklist back to 60 % and deals slip even in boom quarters. The numbers repeat every cycle; the scoreboard just takes longer to catch up.