Holdouts

Last updated:

|Edit this page

Holdouts are randomly assigned lists of users who will be excluded from experiments. You can exclude these users from specific experiments or even all of your experiments.

Holdout testing is useful for understanding long term effects of product changes, typically over a few months. For example, a new onboarding flow might show great initial conversion rates, but a holdout group could reveal whether those users remain active in 3 months from now.

How to create a holdout

To create a new holdout:

  1. Navigate to Experiments and click on the Holdout groups tab.
  2. Click New holdout.
  3. Enter a descriptive name for your holdout, fill out the optional description to help you remember its purpose, and set the percentage of users you want to include in the holdout.
Screenshot of the form to create a new holdout

The holdout should be large enough to be statistically significant, but small enough as to not significantly reduce the pool of users available for experiments. We recommend between 1% and 10%.

Once created, you can assign the holdout to new or existing experiments. The holdout becomes locked once any experiment using it is launched and you will be unable to modify it. This is to prevent data inconsistencies that could affect results.

Screenshot of the form to create an experiment with the holdout field expanded

How to view holdout results

When assigned to an experiment, your holdout will be treated as another variant in experiment analysis. You'll be able to see the conversion rate, delta %, credible interval, and win probability alongside other experiment variants.

Screenshot of the experiment with the results displayed

Questions?

Was this page useful?

Next article

Using data warehouse tables in experiments

🚧 Note: Data warehouse integration is currently considered in beta . We are keen to gather as much feedback as possible so if you try this out please let us know. You can email daniel.b@posthog.com , send feedback via the in-app support panel , or use one of our other support options . Evaluating experiment metrics always depends on events. They rely on something happening at a certain point in time. If one of your data warehouse tables includes event-like data, you can use it as a…

Read next article