Verial
insightsvoice-aiivrtesting

Why Voice AI Agents Break on Real Healthcare IVR Calls

Why healthcare voice agents fail in production and what a deterministic test environment must include. Strategic overview of voice AI testing in healthcare.

S
Stan Liu · Co-founder, Verial
·7 min read
Share

TL;DR

  • Voice AI agents break in production because real payer calls are non-deterministic: IVR trees shift, hold times vary, and representatives behave unpredictably. You cannot regression-test against live phone systems.
  • Healthcare admin work still runs on phones. The CAQH Index 2024 estimates fully electronic transactions could save the US healthcare system $25 billion annually, with voice replacing most of what remains.
  • Physicians and staff complete an average of 39 prior authorizations per week per physician, spending 12 hours on them (AMA 2024 Prior Authorization Survey).
  • The answer is a deterministic simulation environment: configurable IVR trees, scripted representatives, full call recording, and regression-ready scenarios. For the tactical how-to, see voice agent IVR testing: a practical how-to guide.

The rise of voice AI in healthcare operations

Healthcare operations still run on phone calls. Prior authorizations, benefit verifications, claim status inquiries, prescription refills, scheduling. Despite decades of digitization, the volume of work that still requires dialing, navigating an IVR, waiting on hold, and speaking with a rep is staggering.

Companies like Infinitus, Thoughtful AI, and Cohere Health have built voice agents for exactly this. The pitch is simple: replace a 45-minute human call made 20 times a day with an 8-minute AI call made 24/7.

The tech works. Modern voice AI navigates DTMF menus, handles hold music and silence, responds to representative questions, and extracts structured data from speech. The hard part is not building the agent. The hard part is testing it.

"Voice is the last mile of healthcare interoperability. Until the phones go quiet, every AI agent has to pass a phone call."

Aaron Miri, Chief Digital and Information Officer, Baptist Health and former HHS HITAC member

Why voice agent testing is uniquely hard

IVR trees change without notice

Payers update menus regularly. "Press 3 for prior authorization" becomes "press 4." New options appear. Trees restructure after system upgrades. You discover the change when the agent fails in production.

Hold times are unpredictable

Some lines have 2-minute holds. Others have 45. Duration depends on time of day, day of week, call volume. The agent must detect hold music, wait patiently, recognize the rep pickup, and resume. Testing this against real numbers is expensive (per-minute telephony) and slow (you are literally on hold). You cannot configure the duration, so you cannot test a 60-minute hold edge case.

Representatives behave unpredictably

Reps do not follow scripts. They ask unexpected questions. They put you on hold mid-conversation. They transfer. They ask you to repeat. They mishear. A handful of real calls gives you a handful of data points, not coverage.

Audio quality varies

Some payers use VoIP with clear audio. Others run legacy systems with compression, noise, and echo. Speech recognition needs to work across the range.

You cannot regression test

Every real call is different. You cannot call the same payer twice and get the same tree, hold, and rep. You cannot verify that a code change did not break something. You are flying blind. This is the sandbox problem amplified by non-determinism.

What a deterministic voice test environment includes

Simulated IVR trees

A test system that responds to DTMF and voice with configurable trees. Define structure: "Press 1 for English, 2 for Spanish. Press 3 for prior auth." Pre-recorded prompts. Responds exactly as a real IVR would.

Configurability matters:

  • Different trees per payer (Aetna vs UnitedHealthcare)
  • Modified trees for edge cases (9 options instead of 4)
  • Injected errors (DTMF recognition fails)
  • Variable hold times (10 seconds vs 5 minutes)

Simulated representatives

After the IVR, the agent talks to a simulated rep with a defined persona and controlled variation:

  • Baseline. Asks for member ID, DOB, service. Provides a reference number.
  • Variations. Asks for extra info. Puts the agent on hold to check the system. Transfers to a supervisor. Mishears and asks to repeat.
  • Edge cases. System is down, call back later. Conflicting information. Fast speech or accents that challenge recognition.

Each variation is deterministic given the same config. That is what makes regression testing possible.

Full call recording

Every test call is recorded with transcripts, timing, and event logs. You can see exactly when the agent pressed each tone, when it spoke, when it detected hold music, when it resumed. This data feeds evaluation on whether the agent extracted the right reference number, navigated in optimal steps, and handled hold correctly.

A testing framework for IVR navigation

Test 1: Happy path navigation

For each payer tree, verify the agent reaches the correct department. Pass: reaches target within expected menu depth, no wrong turns.

Test 2: Error recovery

Inject errors at each step: bad DTMF recognition, unexpected options, "office is closed" messages. Pass: agent recovers from at least 90% without human help.

Test 3: Hold handling

Configure holds of 30 seconds, 5 minutes, 15 minutes at different points. Pass: zero false positives (speaking during hold) and zero false negatives (missing the rep pickup).

Test 4: Representative conversation

Run a full conversation with the simulated rep. Standard flow plus at least 5 variations: unexpected questions, repeat requests, transfers, mid-conversation holds, conflicting info. Pass: extract correct structured data from at least 95% of conversations.

Test 5: Population-scale

Run the same scenario against 100 simulated payer configurations: different trees, hold times, rep behaviors. Measure aggregate pass rate and identify systematic failures. Pass: overall above 90 to 95%, no single payer below 80%.

From testing to deployment

Voice testing is not one-time. Trees change. Payer behavior shifts. Capabilities evolve. As in the broader four-layer testing framework, tests run continuously:

  • On every code change. Happy path and error recovery for all supported payers.
  • Weekly. Full conversation tests with rep variations.
  • Monthly. Population-scale across all payer configurations.

If your agent makes hundreds or thousands of payer calls per day, the investment in deterministic testing pays for itself the first regression it catches before production.

Key Takeaways

  • Voice AI agents fail in production because production calls are non-deterministic. You cannot regression-test against real phone numbers.
  • A useful test environment has four ingredients: configurable IVR trees, scripted reps with variation, full recording, and a scenario runner.
  • Population-scale testing (100+ payer configurations) exposes systematic failures that small samples miss.
  • Run IVR navigation tests on every code change, conversation tests weekly, and population tests monthly.
  • For the step-by-step scenarios to build, see voice agent IVR testing: a practical how-to guide.

FAQ

Can I just test with real phone calls?

Not at scale. Real calls cost per-minute telephony, are slow, are non-deterministic, and cannot cover edge cases like 60-minute holds or rare rep behaviors. Real-call testing complements simulation but cannot replace it.

What pass rate is good enough?

A common bar is above 95% navigation accuracy, above 98% data extraction accuracy, and no single payer configuration below 80% pass rate.

Which scenarios should I build first?

Start with claim status, benefits verification, and prior auth for your top three payers by call volume. That typically covers 70 to 80% of real traffic. The how-to guide walks through each.

Does CMS-0057-F reduce the need for voice testing?

CMS-0057-F mandates FHIR-based prior auth APIs by 2027, but many eligibility, status, and provider-rep workflows will stay on the phone for years. Voice testing remains essential.

S
Stan Liu · Co-founder, Verial
·7 min read
Share