View and Explore A/B Test Results

A/B testing is a powerful way to optimize your website by comparing two-page versions to determine which performs better. This article explains how to read and understand the results view of an A/B test in Crazy Egg.

Once you’ve published your A/B test, go to the A/B Testing Dashboard and click View the Results to check the goals met and the performance of your ideas.
SCR-20241210-klru.png

Overview of the Results Screen

When you open the results of an A/B test, you’ll see an overview that includes:

  • Goal: The specific action being measured (e.g., conversions, clicks).
  • Unique Visitors: The total number of visitors included in the test.
  • Conversions: The number of times the goal was completed.
  • Conversion Rate: The percentage of visitors who completed the goal.
  • Total Goal Value: The total monetary value of the goal (if assigned).

    Results Graph

    The line graph shows daily conversion trends for each variant.

    • X-axis: Test duration (e.g., Nov 10 – Dec 9).
    • Y-axis: Number of conversions per day.
    • Each line represents a test variant:
      • Control (baseline) is often displayed in one color.
      • Variant(s) appear in a different color.
    • Look for significant differences in conversion spikes and overall patterns.
    • Consistency over time indicates reliable results.

    Variants Table

    Below the graph, the table breaks down key metrics for each variant:

    ColumnExplanation
    Unique Visitors  Total number of visitors who saw each version of the page.
    Conversions  Number of users who completed the goal.
    Conversion Rate  Percentage of visitors who converted.
    Statistical Significance  The likelihood that the results are not due to random chance.
    Total Goal Value  If a monetary value is assigned, this shows the revenue or value  generated.

     

    Key Metrics to Compare

    • Conversion Rate: A higher rate indicates a better-performing variant.
    • Statistical Significance: For reliable results, aim for significance above 95%.
    • Total Goal Value: Helps prioritize the winning variant based on value generated.

    Example Breakdown

    In the image above:

    • Control (quickspout.com):
      • Conversion Rate: 3.58%
      • Conversions: 63
      • Total Goal Value: $61
    • Variant #1:
      • Conversion Rate: 4.08% (↑ 13.9% improvement)
      • Conversions: 65
      • Total Goal Value: $65

    Even with a small increase in conversions, Variant #1 shows a 13.9% improvement in conversion rate and slightly higher goal value.

    What to Do Next

    1. Check Statistical Significance: If results are significant, consider implementing the winning variant.
    2. Monitor Long-Term Results: Sometimes, short-term trends can shift.
    3. Document Learnings: Understand what changes impacted performance (e.g., wording, design tweaks).

    FAQs

    Q: What if the results are not statistically significant?
    A: Extend the test duration or analyze for additional data trends.

    Q: How do I know which variant is “winning”?
    A: Compare the conversion rates and goal values—look for the one with higher metrics and significance.

    Conclusion

    Interpreting A/B test results is essential for data-driven decision-making. By understanding metrics like conversion rate, statistical significance, and goal value, you can confidently identify which page variation delivers better results.

    Was this article helpful?

    Related Articles

    Need Support?

    Can’t find the answer you’re looking for? Don’t worry we’re here to help!

    Submit a request