A/B testing is available for projects using the standard deployment pipeline. You must have at least one live deployment to start a test.
How it works
When you start an A/B test, live traffic is split between two agent deployments:- Control (A) — your current live version
- Variant (B) — a previous or alternative version you want to compare
Start an A/B test
Select a variant
Choose the version you want to test against the current live deployment. This becomes the variant (B group).
Configure the traffic split
Set the percentage of traffic routed to each version. Both versions run simultaneously on the live environment.
Review results
During and after a test, you can compare performance between the two groups:- Go to Analytics > Conversations to review calls from the test.
- Filter conversations by A/B test group to see which version each caller experienced.
- Compare key metrics across groups — containment rate, handoff rate, CSAT, and conversation length.
Stop a test
- Go to Deployments > Environments.
- Select Stop A/B test from the active test.
- All traffic returns to the current live version.
Best practices
- Run tests long enough to get meaningful data — short tests with few calls can produce misleading results. Aim for a statistically significant sample before making decisions.
- Test one change at a time — if you change multiple things between versions, you won’t know which change drove the difference.
- Monitor during the test — check conversation review regularly to catch issues early. If the variant is performing significantly worse, stop the test.
- Use test sets before A/B testing — validate your changes with the test suite before exposing them to live traffic. A/B testing is for measuring impact, not catching bugs.
Related pages
Environments
Understand the deployment pipeline and how versions move through environments.
Compare versions
Review differences between versions before starting a test.
Conversations
Review individual conversations and filter by A/B test group.

