Your QA team reviews 3% of calls. We review 100%. See how →
VivrittVivritt
QA & Coaching

From 3% sample to full coverage: a QA lead's practical playbook

James Tan··9 min read

When I started my first QA lead role at a 180-seat contact centre in Quezon City, we evaluated six calls per agent per month. It took a team of four to manage. We thought we had a reasonable programme — until a client audit surfaced 47 compliance gaps in a single week that our sample had never flagged.

That audit was the moment I realised the fundamental problem with sampling-based QA: it creates the illusion of oversight without the substance. This playbook is what I wish I'd had at the start of that journey — a practical, phased approach to getting to full coverage without hiring an army or throwing out what already works.

Why the "industry standard" is not actually a standard

The widely cited benchmark of 4 to 6 calls per agent per month comes from a time when QA meant a supervisor with a headset listening to recordings. It was never a statistical recommendation. It was a practical ceiling determined by how much time a human QA analyst could spend per agent.

The math is not on your side. If an agent handles 400 calls per month and you evaluate 6, you are reviewing 1.5% of their output. At that coverage rate, an agent who delivers a required disclosure incorrectly on 1 in 8 calls has approximately a 45% chance of going undetected in any given month. Over a team of 60 agents, you are statistically certain that systemic compliance problems are invisible to you.

The SQM Group's 2024 research confirms that 60% of contact centres evaluate 5 or fewer calls per agent per month. That is the industry norm — not a quality benchmark. These are very different things.

Phase 1: audit your current programme before changing anything

The biggest mistake in QA transformation is starting with the solution rather than the problem. Before deploying any automation, spend two weeks doing a ground-truth audit of your existing programme.

Pull the last three months of QA scores across all agents. Map them against available outcome data: CSAT scores, complaint rate, escalation rate, repeat call rate. In most operations, you will find a weak or non-existent correlation between QA scores and these outcome metrics. This tells you one of two things: either your QA criteria are not measuring what drives outcomes, or your sample is too small to surface meaningful signal.

This audit matters because it gives you a baseline. When you later move to full coverage, you need to be able to show your leadership — and your client — that the new programme is delivering different insights, not just more of the same.

Phase 2: automate the compliance layer first

Full coverage does not mean full human review. It means full detection. The practical approach is to automate rule-based compliance checks across 100% of calls — disclosures, consent notices, script adherence, forbidden language — and reserve human review for calls that fail automated checks or fall into high-risk categories.

In regulated verticals like BFSI and healthcare, compliance checks account for roughly 30% to 40% of the typical QA scorecard by weight. Automating this layer immediately gives you evidence-based compliance coverage across all calls, not just the sample.

In a 250-seat BFSI BPO in the Philippines, automating the compliance layer of QA for all inbound calls reduced the average time per QA review by 40% — because analysts no longer had to check disclosures manually. They focused entirely on tone, resolution quality, and customer experience dimensions.

The most important design decision at this stage is rule specificity. Vague rules like "agent discussed product terms" produce high false-positive rates and erode analyst trust in the system. Specific rules — "agent mentioned APR between 00:30 and 03:00 of call duration" — are actionable. Work with your compliance and legal team to define each disclosure rule precisely before go-live.

Phase 3: restructure human review around exceptions

Once automated compliance checks are running on 100% of calls, your human QA team's role changes. Instead of sampling randomly, they review a curated queue: calls that failed one or more compliance checks, calls flagged for sentiment issues, high-value customer segments, and new agents in their first 60 days.

This exception-based model is not less rigorous than sampling. It is more rigorous where it matters. A random sample of 6 calls from a 400-call month is equally likely to pull good calls and bad ones. An exception queue pulls only the calls that need attention.

In practice, the exception queue typically represents 8% to 15% of total call volume. For a 100-agent operation with 40,000 calls per month, that is 3,200 to 6,000 calls requiring human review — more than a sampling approach would ever reach, but also more targeted.

Phase 4: build the coaching pipeline that actually uses the data

Full coverage only creates value if it changes how supervisors coach. The failure mode is generating excellent data that never reaches agents — call scores piling up in a dashboard that managers check quarterly.

The coaching pipeline that works in practice has three components. A weekly exception digest: each supervisor receives a summary of their team's flagged calls from the prior week, with the top three recurring patterns highlighted. A call library: for each recurring gap — missed rate disclosure, rushed close, handling a price objection badly — there is a clip library of examples, both failures and best practice. A 1:1 integration: supervisors are trained to open 1:1 sessions with a specific flagged call, not a general score summary.

The difference between coaching with a score and coaching with a call clip is the difference between "your compliance score was 72% this month" and "here is the exact moment at 2:14 of this call where you skipped the APR. Let's talk about what was happening." One is a report card. The other is a conversation.

How long does transformation take

Based on implementations across BFSI and telecom BPOs in Manila, Cebu, and Bengaluru, the typical timeline breaks into four phases: 30 days for baseline audit and rule design, 30 days for automated compliance layer go-live, 60 days for exception-based QA team restructuring, and 90+ days for coaching pipeline to show measurable outcome changes.

The outcome data is consistent: by month six, compliance miss rates are typically 60% to 75% lower than baseline. More importantly, the data is reliable — because it is based on all calls, not a sample.

Related: How to achieve 100% disclosure coverage and Turning missed closes into coaching moments.

READY TO TRY IT?

Start coaching every agent. Not just the sampled ones.

Live session. Your real call data. No mock demos.

CONTINUE READING