Skip to main content

Viewing and Understanding A/B Test Results

Learn how to analyze A/B test results to see which variation drives better performance.

Rojen M Reji avatar
Written by Rojen M Reji
Updated over 2 weeks ago

Alia’s A/B testing analytics dashboard helps you confidently identify winning popup variations by measuring signup performance, conversion impact, and downstream revenue.

With advanced filtering capabilities, you can analyze results across specific shopper segments — including UTMs, device type, country, survey responses, browsing behavior, and more — allowing you to uncover insights that go far beyond surface-level metrics.

Step 1: Access A/B Test Results

  • Navigate to your campaign dashboard in Alia.

  • Click on "View A/B Test Results" to open the test results page.

You’ll land on the analytics dashboard where all experiment data is automatically aggregated.

Step 2: Review Test Timeline

The dashboard displays the full test duration, including:

  • Start date

  • End date (or current progress if still running)

If your test is active, metrics update in near real-time so you can monitor performance trends as data accumulates.

💡 Best practice: Avoid ending tests too early — stronger data leads to more reliable winners.

Step 3: Compare Control vs. Test Group

Each experiment includes:

  • Control: Your original popup

  • Variants: The versions being tested

    • You can have an unlimited number of variants

Review performance side-by-side to understand which experience drives stronger results.

👉 The percentage lift displayed between variants helps quickly identify directional winners.

Step 4: Segment Your Results with Advanced Filters

One of the most powerful features of Alia’s new analytics experience is the ability to filter performance across virtually any shopper dimension.

You can analyze results by:

Traffic & Marketing

  • UTM source, medium, and campaign

  • Initial path

  • Domain

Shopper Attributes

  • Country, region, and city

  • Shopify market and locale

  • Device type, browser, and OS

Behavior & Actions

  • Visitors matching targeting

  • Email submissions

  • Reward variants

  • Purchased products or collections

Zero-Party Data

  • Survey responses

  • Mini-quiz answers

  • Custom user properties

💡 Why this matters:

A test may appear neutral at a high level but reveal a clear winner within a high-value segment — such as paid traffic or mobile shoppers.

This is where the biggest optimization opportunities often live.

💡 The percentage difference between the control and test groups will help determine which performs better.

Step 5: Evaluate Key Performance Metrics

Your dashboard includes all primary analytics, plus sitewide impact metrics.

Core Metrics to Watch

  • Email & SMS submit rates

  • Opt-in rates

  • Attributed conversion rate

  • Attributed orders and sales

  • Average attributed order value

  • Time to purchase

Sitewide Impact Metrics

  • Sitewide conversion rate

  • Sitewide sales

  • Bounce rate

These help ensure your popup improves signup performance without harming the overall shopping experience.

💡 If statistical significance isn’t reached, allow the test to run longer for more reliable results.

Step 6: Check Statistical Significance

Each metric includes a significance indicator:

Check mark — Results are statistically significant
Clock icon — More data is needed

💡 Pro Tip: If significance hasn’t been reached, allow the test to run longer before making a decision. Ending tests prematurely can lead to false winners.

Step 7: Identify and Publish the Winning Variation

Once a clear winner emerges:

  1. Select the top-performing variant

  2. Publish it directly from the results page

Alia will automatically:

✅ Keep the winning popup live
✅ Move the losing variation to draft
✅ Preserve test data for future analysis


What’s Next?

Winning a test is just the beginning.

Use your insights to continuously improve performance:

🚀 Test new messaging or offers
🎨 Experiment with design changes
🎯 Refine targeting
📊 Optimize by traffic source
🧠 Personalize experiences using zero-party data

The highest-performing brands treat A/B testing as an ongoing growth engine — not a one-time exercise.

Did this answer your question?