npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@vicistack/ai-call-center-quality-control

v1.0.0

Published

How AI Is Changing Call Center Quality Control (And Why Most Centers Are Still Stuck in 2015) — ViciStack call center engineering guide

Downloads

33

Readme

How AI Is Changing Call Center Quality Control (And Why Most Centers Are Still Stuck in 2015)

Here is the honest truth about quality assurance in most call centers: it is theater. A QA team listens to a handful of calls, fills out scorecards, delivers feedback two weeks late, and everyone pretends the operation has "quality control." Meanwhile, thousands of calls per day go completely unreviewed. Compliance violations slip through. Top performers get the same generic coaching as struggling agents. And the metrics your executives see are based on a sample so small it would make a statistician cringe. This has been the reality for over a decade. The tools existed to do better, but most call centers -- especially those running open-source platforms like VICIdial -- got stuck in a loop of manual processes, spreadsheet scorecards, and the quiet acceptance that real quality monitoring was too expensive to scale. That is finally changing. AI-powered quality control has matured past the buzzword phase and into something that actually works -- with real limitations you need to understand before you buy anything. This article is the no-nonsense breakdown: what AI QC does, what it does not do, what it costs, and how to implement it in a VICIdial environment without setting your operation on fire. ## The QA Crisis: Why 98% of Calls Go Unreviewed Let's start with the math, because this is where the entire argument for AI quality control begins. A typical outbound call center running VICIdial has 50 to 200 agents. Each agent handles somewhere between 40 and 80 calls per day depending on your dialer settings and campaign type. For a 100-agent operation at 60 calls per agent, that is 6,000 calls per day, or roughly 132,000 calls per month. Now look at your QA team. Most call centers have one QA analyst for every 50 to 75 agents. That gives our 100-agent center two QA analysts, maybe three if the operation takes quality seriously. A manual call evaluation takes 15 to 30 minutes. That includes listening to the recording, filling out the scorecard, documenting notes, and sometimes re-listening to sections. At the aggressive end, a QA analyst can complete about 15 to 25 evaluations per day. Let's be generous and say 20. Two QA analysts doing 20 evaluations per day, five days a week, produces 200 scored calls per week. Against a volume of 30,000 calls that same week, that is a review rate of 0.67%. Round up and call it 1 to 2 percent if your team is particularly efficient. This is not a controversial number. Industry data consistently shows that most contact centers manually review between 1 and 3 percent of customer interactions. Some organizations with larger QA teams reach 5 percent. Almost nobody gets above that through manual processes alone. The problem is not just coverage -- it is what you miss. When you are sampling 2% of calls, your quality data is essentially random noise dressed up as insight. You are making coaching decisions, identifying compliance risks, and evaluating agent performance based on a sample size that would not pass a first-year statistics course. A single bad day from an otherwise strong agent can tank their score. A consistently non-compliant agent can skate through months without being caught, simply because none of their calls happened to land in the review queue. Consider what this means for compliance. If you are in a regulated industry -- insurance, financial services, healthcare, debt collection -- every unreviewed call is a potential liability. The TCPA, state-level telemarketing regulations, mini-Miranda requirements for debt collection, CMS guidelines for Medicare sales -- these are not suggestions. They carry real penalties. And your 2% review rate means 98% of your compliance exposure goes completely unmonitored. The cost compounds from there. In our 100-agent example, annual QA labor costs run approximately $140,000 to $160,000 per year, assuming QA analysts at $30 to $35 per hour fully loaded. That is a significant investment to review less than 2% of your output. As your call center costs grow, QA spending scales linearly -- you need more humans to review more calls -- but your coverage percentage never improves. This is the fundamental structural problem that AI quality control solves. Not by making QA analysts faster at their jobs, but by changing the unit economics entirely so that 100% review coverage becomes the baseline rather than the aspiration. > Still relying on manual QA to catch compliance issues? A ViciStack audit can show you exactly what your current process is missing. Get your free audit and see the gaps for yourself. ## What AI Quality Control Actually Is (Not the Buzzword Version) Strip away the marketing language and AI quality control is a pipeline with three stages: transcription, analysis, and scoring. Every vendor in this space -- Observe.AI, CallMiner, Balto, Level AI, Cresta, and smaller players -- is running some variation of this same pipeline. The differences are in execution quality, not fundamental approach. Stage 1: Speech-to-Text Transcription. The call recording gets converted from audio to text. This uses automatic speech recognition (ASR) models -- the same underlying technology as Siri, Alexa, or Google's voice search, but tuned for telephony audio. The output is a time-stamped transcript showing who said what and when. Some systems also perform speaker diarization, which separates the agent's speech from the customer's speech, and channel separation, which uses the stereo recording to assign each side of the conversation to its own track. Stage 2: Natural Language Processing and Analysis. Once you have a transcript, NLP models analyze the text for specific elements. This is where the actual intelligence happens. The system identifies things like: Did the agent use the required compliance disclosures? Was the proper greeting and closing delivered? Did the customer express frustration, confusion, or satisfaction? Were objections handled? Was there dead air or excessive hold time? Did the agent attempt to upsell or cross-sell? Were prohibited phrases used? Modern systems use large language models (LLMs) for this analysis rather than the older keyword-spotting approach. The difference matters. A keyword system flags "cancel" as a negative event. An LLM understands that "I am not looking to cancel" and "I want to cancel immediately" have opposite meanings despite sharing the same keyword. Stage 3: Automated Scoring and Alerting. The analysis results get mapped against your scorecard criteria. Each call receives an automated quality score, broken down by category -- compliance, soft skills, process adherence, resolution effectiveness. Calls that score below thresholds trigger alerts. Patterns across multiple calls surface trends. The data feeds into dashboards and coaching workflows. What makes this "AI" rather than just "automation" is the analysis layer. Rule-based systems from the 2010s could do crude keyword spotting, but they generated so many false positives that QA teams spent more time reviewing false alerts than they saved. Modern AI-powered systems use transformer-based language models that understand context, intent, and conversational flow. They are not perfect -- we will get to limitations shortly -- but they are a categorical improvement over what existed five years ago. The practical result is that every single call gets evaluated against every single criterion on your scorecard, automatically, within minutes of the call ending. Your QA team stops spending their time listening to calls and filling out forms. Instead, they review AI-generated scores, validate flagged interactions, handle edge cases, and focus on coaching. The human role shifts from data entry to quality oversight and agent development. ## Speech-to-Text Accuracy: The Foundation That Has to Work First If the transcription is wrong, everything downstream is wrong. This is the part most AI QC vendors gloss over in their sales decks, but it is the single biggest technical risk in any implementation. Call center audio is not podcast audio. You are dealing with 8 kHz narrowband telephony -- the compressed audio format that phone networks use. It has roughly one-quarter the fidelity of a typical voice recording. Add background noise from a call floor, agents on headsets with varying quality, customers on speakerphone or driving, accents, dialects, cross-talk, and the occasional customer who mumbles through the entire conversation, and you have an extremely challenging environment for speech recognition. The 2025 Voicegain benchmark for 8 kHz call center audio tested five major ASR engines on 40 curated call center recordings featuring real-world conditions -- background noise, overlapping speech, and diverse accents. The results were sobering: - Amazon AWS Transcribe: 87.67% accuracy (highest) - Whisper Large V3: 86.17% accuracy - Voicegain Omega: 85.09% accuracy - Google Video: 68.38% accuracy (lowest) That means even the best-performing engine gets roughly one in eight words wrong on real call center audio. And this is on curated test files -- production environments with noisier conditions will see lower numbers. Why does this matter for quality control? Consider a compliance disclosure that an agent is required to read verbatim. If the ASR engine drops or misinterprets two words out of a 20-word disclosure, the AI scoring system might flag the call as non-compliant even though the agent said it correctly. Or worse, it might mark the disclosure as delivered when the agent actually skipped it, because the ASR hallucinated words that were not spoken. The accuracy gap varies dramatically based on conditions. One analysis showed the same ASR API performing at 92% accuracy on clean headset audio, 78% in conference room environments, and 65% on mobile calls with background noise. In a call center context, this means your accuracy will differ between agents with good headset discipline and those who do not, between quiet rooms and noisy floors, and between domestic and international customer populations. What you should demand from any AI QC vendor: - Accuracy benchmarks on your actual audio. Not clean demo recordings. Pull 100 representative calls from your VICIdial recordings, including your noisiest ones, and have the vendor run them through their pipeline. Compare their transcripts against human-verified ground truth. - Stereo recording support. Dual-channel recordings where the agent and customer are on separate audio channels dramatically improve both transcription accuracy and speaker identification. If your VICIdial instance is recording in mono, switch to stereo before implementing AI QC. - Domain-specific vocabulary tuning. If you are in insurance, your agents are saying words like "deductible," "copayment," and "underwriting" dozens of times per day. A general-purpose ASR model may struggle with industry jargon. Good vendors offer custom vocabulary lists or fine-tuned models for your vertical. - Word Error Rate (WER) reporting. Vendors should be transparent about their WER on telephony audio. If they cannot or will not share this number, that tells you something. For quality scoring purposes, you generally need 90% or higher accuracy for the system to be reliable enough to use without heavy human validation. The bottom line: speech-to-text accuracy in 2026 is good enough for AI quality control to work, but it is not good enough to work without human oversight. Any vendor telling you otherwise is overselling. ## What AI Can Score That Humans Score Inconsistently Here is where AI quality control genuinely outperforms manual QA, and it is not even close. Consistency across evaluators. Inter-rater reliability in manual QA programs typically falls between 65 and 75 percent. That means if two QA analysts score the same call independently, they will disagree on roughly a quarter to a third of the scorecard items. The disagreement is especially pronounced on subjective criteria -- did the agent show empathy? Was the tone professional? Was the objection handling effective? Different evaluators interpret these questions differently, and the same evaluator might score the same call differently on Monday versus Friday. AI does not have this problem. Given the same transcript and the same scoring criteria, the AI will produce the same score every time. This consistency alone is a massive win for a QA program because it eliminates one of the biggest sources of agent frustration: the perception (often accurate) that quality scores depend on which evaluator reviewed the call rather than how the call was actually handled. Binary compliance checks. Did the agent read the required disclosure? Did they verify the customer's identity? Did


Read the full article

About

Built by ViciStack — enterprise VoIP and call center infrastructure.

License

MIT