Sales managers who implement coaching programs without a tracking system face a common problem six weeks later: they cannot tell whether the coaching worked. Reps seem to be performing better, or worse, or the same, but there is no data connecting a specific coaching session to a measurable change in call behavior. This guide walks through a six-step process for building a coaching log system that produces that data, using QA scoring to isolate improvement at the criterion level.
Step 1: Define the Metrics That Coaching Targets
Before logging a single session, identify which QA criteria you are trying to move. Not all metrics improve from coaching. Activity metrics like call volume are driven by scheduling, not skill. The metrics that respond to coaching are behavioral: objection handling score, discovery question frequency, value framing before pricing, compliance phrase usage.
Build a list of 5-8 specific QA criteria that are both measurable in call recordings and influenced by rep behavior in real time. These become the official coaching targets. Every session in your log will reference at least one criterion from this list. Without this step, coaching logs become notes about conversations rather than records of targeted behavior change.
Insight7's weighted criteria system supports this setup: managers configure main criteria, sub-criteria, and descriptions of what good and poor performance look like for each. These definitions become the shared language between manager and rep during coaching sessions.
What Makes a Coaching Metric Trackable
A trackable coaching metric has three properties: it appears in call recordings (so it can be scored before and after coaching), it is specific enough to define as present or absent (not "communication quality" but "stated the refund policy within the first 60 seconds"), and it is influenced by rep choice rather than by factors outside their control. Metrics that pass all three are worth tracking. Those that fail any one of these tests create noise in your improvement data.
Step 2: Log Each Session with the Criterion Targeted
A coaching log entry should take under two minutes to create. The minimum required fields are: rep name, session date, criterion targeted, baseline score on that criterion, coaching method used (roleplay, call review, script walkthrough), and a one-sentence description of what the rep agreed to do differently.
The criterion field is the most important. Teams that log "discussed objection handling" instead of "targeted criterion: handles price objection before closing" cannot aggregate data later. When every log entry references a specific criterion from your defined list, you can run queries like "which criterion did we coach most this quarter, and what was the average score change after coaching?"
Avoid this common mistake: logging the coaching topic as a broad theme rather than the specific QA criterion. "Communication" and "customer empathy" are themes. "Used empathy statement within 30 seconds of customer frustration" is a criterion. Only criteria produce trackable improvement data.
Step 3: Record Baseline Scores Before the Session
Before the coaching session occurs, pull the rep's current average score on the targeted criterion from the last 10 calls. This is the baseline. Record it in the log entry at the time of session scheduling, not after.
Recording baseline after the fact introduces bias. Managers who know coaching went well tend to remember lower starting scores. Managers who feel the coaching did not land tend to remember higher starting scores. The baseline must be recorded from the scoring system before any subjective judgment about the session outcome enters the picture.
Insight7 generates per-rep, per-criterion scorecards automatically. Before a coaching session, a manager can pull the rep's score on the targeted criterion across their last 10 calls and record the average in under a minute. This removes the manual effort of reviewing individual call recordings to establish a baseline.
Step 4: Re-Score the Same Criterion in the 10 Calls After Coaching
Two to three weeks after the coaching session, collect scores for the targeted criterion from the rep's next 10 calls. This is the post-coaching window. Use the same scoring tool and criteria configuration that generated the baseline, so the two numbers are directly comparable.
Do not use the first call after coaching as the measurement. Reps often overcorrect immediately after a session and then revert. The 10-call window smooths out that overcorrection and captures whether the behavior change held. Platforms that score 100% of calls automatically, like Insight7, generate this data without requiring the manager to manually select and review calls.
What a Valid Pre/Post Comparison Requires
For a pre/post comparison to be valid, four conditions should hold: the same criterion was scored in both periods, the scoring criteria definition did not change between baseline and post window, the call types in both windows are comparable (inbound vs. inbound, not inbound vs. outbound), and the post window represents at least 10 calls. If any of these conditions are not met, the comparison may reflect criterion drift or call type variation rather than actual rep improvement.
Step 5: Compare Pre/Post Scores to Isolate Improvement
Calculate the score change: post-coaching average minus pre-coaching average on the targeted criterion. A positive number indicates improvement. A flat or negative number indicates the coaching approach did not produce the intended change.
Isolating improvement at the criterion level is the key advantage of this system over general performance tracking. If a rep's overall QA score improves, you cannot tell whether the coaching caused it or whether unrelated factors shifted. When you track a specific criterion from before a specific coaching session to after, you have isolated the variable.
SQM Group research on contact center performance improvement shows that criterion-level tracking doubles the likelihood that managers identify which interventions worked versus which did not. This is the difference between knowing coaching happened and knowing coaching worked.
Step 6: Aggregate Across the Team to Identify Which Coaching Approaches Work
After three months of logging, you have enough data to answer questions your coaching program has probably never answered before: which criteria show the largest average improvement after coaching, which coaching methods (roleplay vs. call review vs. script walkthrough) produce better score outcomes, and which reps consistently improve versus which plateau regardless of intervention type.
Aggregate the pre/post deltas by coaching method to find your most effective approach. Aggregate by criterion to find which behaviors are coachable versus which require hiring or process changes. Aggregate by rep to identify who needs a different intervention strategy.
Insight7 generates team-level dashboards showing score trends across reps and criteria. Managers can see improvement trajectories, identify which reps improved the most after coaching, and drill into the specific calls that drove the change. This aggregated view is what transforms a coaching log from an administrative record into a management tool.
FAQ
What are the 5 C's of coaching?
In sales coaching contexts, the 5 C's are typically: context (understanding the rep's situation and call environment), conversation (the coaching dialogue itself), commitment (the rep's agreement to a specific behavior change), consistency (repeated practice until the behavior becomes habitual), and calibration (re-scoring calls to measure whether the change occurred). This six-step system is built around the calibration step, which most coaching programs skip.
How do you track sales productivity?
Sales productivity is typically tracked through a combination of activity metrics (calls made, emails sent, pipeline created) and outcome metrics (close rate, average deal size, quota attainment). For coaching programs, the most useful productivity signal is criterion-level call quality: tracking specific behaviors on calls over time shows whether productivity improvements are sustainable or coincidental. Insight7's scoring system separates behavioral quality from activity volume, giving managers a cleaner view of what coaching is actually changing.
How often should coaching log data be reviewed?
Weekly for active coaching cases, monthly for team-level trend analysis. During the 10-call post-coaching window, managers should check in at the 5-call mark to see whether improvement is emerging or whether a second coaching session is needed before the window closes. Monthly aggregation across the full team reveals systemic patterns that individual session logs obscure.
