ExperimentationAnalyzing experiments

Bayesian experiment results

Overview

This topic explains how to read and use the results table of a Bayesian experiment.

A Bayesian experiment's results table.

A Bayesian experiment's results table.

The winning variation, based on the probability to be best and the probability to beat control, is highlighted at the top of the Results tab when a Bayesian experiment has collected enough data. Click Ship it to stop the experiment iteration and ship the winning variation to everyone.

Filter options

You can filter an experiment’s results table by metric, variation, or attribute value.

The filter options for an experiment.

The filter options for an experiment.

Metrics filter

When you create an experiment, you can add one or more metrics for the experiment to measure. If an experiment is measuring more than one metric, you can filter the results to view only select metrics at a time. You cannot add additional metrics to an experiment after you create it.

To filter your results by metric, click All metrics and select the metric you want to view. The results table updates to show you results from only the metrics you selected.

Variations filter

To filter your results by variation, click All variations and select the variations you want to view. The results table updates to show you results from only the variations you selected.

Attributes filter

When you create an experiment, you can designate one or more context attributes as filterable for that experiment. If a context attribute is filterable, you can filter the experiment results by those attribute’s values. For example, if you designate the “Country” attribute as filterable, then you can narrow your results by users with a “Country” attribute value of Canada.

You cannot designate additional context attributes as filterable after you create an experiment.

To filter your results by context attribute value:

  • Click All attributes. A list of context attributes appears.
  • Hover on the context attribute you want to filter by. A list of available values for that attribute appears.
  • Search for and select the attribute values you want to view.

The results table updates to show you results from only contexts with the attribute value that you selected.

Visualization options

The results table has the following visualization options:

  • Probability density
  • Relative difference
  • Arm averages

Probability density

The probability density graph displays the distribution of the results for each metric.

The horizontal x-axis displays the unit of the primary metric included in the experiment. For example, if the metric is measuring revenue, the unit might be dollars, or if the metric is measuring website latency, the unit might be milliseconds.

If the unit you’re measuring on the x-axis is something you want to increase, such as revenue, account sign ups, and so on, then the farther to the right the curve is, the better. The variation with the curve farthest to the right means the unit the metric is measuring is highest for that variation.

If the unit you’re measuring on the x-axis is something you want to decrease, such as website latency, then the farther to the left the curve is, the better. The variation with the curve farthest to the left means the unit the metric is measuring is lowest for that variation.

How wide a curve is on the x-axis determines the credible interval. Narrower curves mean the results of the variation fall within a smaller range of values, so you can be more confident in the likely results of that variation’s performance.

The vertical y-axis measures probability. You can determine how probable it is that the metric will equal the number on the x-axis by how high the curve is.

Relative difference

The relative difference graph displays a time series of the relative difference between the treatment variation and the control. This graph is helpful for investigating trends in relative differences over time.

Arm averages

The arm averages graph displays the average value over time for each variation. This graph is useful for investigating trends that impact all experiment variations equally over time.

Results chart data

This section explains the data that displays in the experiment results chart.

Graph

The graph column displays the probability density, relative difference, or arm averages graph, depending on the visualization option you have selected. Click on the graph to view an enlarged version.

Probability to beat control

The probability to beat control represents the likelihood that this variation performs better than the control variation for a given metric. Probability to beat control is only relevant for treatment variations, as it measures the probability of outperforming the control, making it unnecessary for the control itself. When multiple treatment variations outperform the control, it’s helpful to also consider the probability to be best to determine the winning variation.

Probability to beat control in funnel optimization experiments

In funnel optimization experiments, the results table provides each variation’s probability to beat control for each step in the funnel, but the final metric in the funnel is the metric you should use to decide the winning variation for the experiment as a whole.

LaunchDarkly includes all end users that reach the last step in a funnel in the experiment’s winning variation calculations, even if an end user skipped some steps in the funnel. For example, if your funnel metric group has four steps, and an end user takes step 1, skips step 2, then takes steps 3 and 4, the experiment still considers the end user to have completed the funnel and includes them in the calculations for the winning variation.

Probability to be best

The probability to be best for a variation is the likelihood that it outperforms all other variations for a specific metric. However, the probability to be best alone doesn’t provide a complete view. For example, when multiple treatment variations outperform the control, the probability to be best may decrease, even if the treatments are generally effective.

To get a comprehensive understanding, pair the probability to be best with the probability to beat control and expected loss. These metrics together provide a clearer picture of performance and trade-offs. To learn more, read Decision making with Bayesian statistics.

Posterior mean

The posterior mean is the variation’s average numeric value that you should expect in this experiment, based on the data collected so far.

All of the data in the results table are based on a posterior distribution, which is the combination of the collected data and our prior beliefs about that data. To learn more about posterior distributions, read Frequentist and Bayesian modeling.

LaunchDarkly automatically performs checks on the results data, to make sure that actual context traffic matches the allocation you set. To learn more, read Understanding sample ratios.

The posterior mean displays only for numeric metrics. To learn more, read Custom numeric metrics.

Relative difference

The relative difference from the control variation measures how much a metric in the treatment variation differs from the control variation, expressed as a proportion of the control’s estimated value. LaunchDarkly calculates this by taking the difference between the treatment variation’s estimated value and the control variation’s estimated value, then dividing that difference by the control variation’s estimated value.

Expected loss

Ideally, shipping a winning variation would carry no risk. In reality, the probability for a treatment variation to beat the control variation is rarely 100%. This means there’s always some chance that a “winning” variation might not be an improvement over the control variation. To manage this, we need to measure the risk involved, which is called “expected loss.”

Expected loss represents the average potential downside of shipping a variation, quantifying how much one could expect to lose if it underperforms relative to the control variation. LaunchDarkly calculates this by integrating probability-weighted losses across all scenarios in which a given variation performs worse than control, with loss defined as the absolute difference between them.

A lower expected loss indicates lower risk, making it an important factor in choosing which variation to launch. The treatment variation with the highest probability to be best among those with a significant probability to beat control is generally considered the winner, but evaluating its expected loss clarifies the associated risk of implementing it.

For example, if you’re measuring conversion rate and have a winning variation with a 96% probability to beat control and an expected loss of 0.5%, this means there’s a strong likelihood of 96% that the winning variation will outperform the control variation. However, the 0.5% expected loss indicates that, on average, you’d expect a small 0.5% decrease in conversion rate if the winning variation were to underperform.

Expected loss displays only for metrics that use an “Average” analysis method.

Conversion rate (conversion metrics only)

The conversion rate column displays for all conversion metrics. Examples of conversions include clicking on a button, or entering information into a form.

Conversion metrics can be one of two types: count or binary.

Conversion rates in count conversion metrics

The value for each unit in a count conversion metric can be any positive value. The value equals the number of times the conversion occurred. For example, a value of 3 means the user clicked on a button three times.

The aggregated statistic for count conversion metrics is the average number of conversions across all units in the metric. For example, the average number of times users clicked on a button.

Count conversion metrics include:

  • Clicked or tapped metrics using the Count option
  • Custom conversion count metrics
  • Page viewed metrics using the Count option

Conversion rates in binary conversion metrics

The value for each unit in a binary conversion metric can be either 1 or 0. A value of 1 means the conversion occurred, such as a user viewing a web page, or submitting a form. A value of 0 means no conversion occurred.

The aggregated statistic for binary conversion metrics is the percentage of units with at least one conversion. For example, the percentage of users who clicked at least once.

Binary conversion metrics include:

  • Clicked or tapped metrics using the Occurrence option
  • Custom conversion binary metrics
  • Page viewed metrics using the Occurrence option

For funnel optimization experiments, the conversion rate includes all end users who completed the step, even if they didn’t complete a previous step in the funnel. LaunchDarkly calculates the conversion rate for each step in the funnel by dividing the number of end users who completed that step by the total number of end users who started the funnel. LaunchDarkly considers all end users in the experiment for whom the SDK has sent a flag evaluation event as having started the funnel.

Conversions (conversion metrics only)

Conversions is the total number of conversions for a conversion metric:

Mean (numeric and custom conversion count metrics only)

The mean is the variation’s average numeric value that you should expect in this experiment, based on the data collected so far. The mean displays only for numeric and custom conversion count metrics.

Total value (numeric metrics only)

The total value is the sum total of all the numbers returned by a numeric metric.

Exposures

The exposures column displays the total number of contexts that encountered the metric as part of the experiment.

To learn more about troubleshooting if your experiment hasn’t received any metric events, read Experimentation Results page status: “This metric has never received an event for this iteration”.