Test and launch your experiment

This article helps you:

  • QA your experiment before and after rollout

  • Launch your experiment to your users

Once you’ve designed your experiment and configured its delivery, you’re ready to test your experiment. Then, if all goes well, it’s time to launch it.

On your Experiment Overview page, review the Design and Delivery cards. Make sure everything is set the way you planned it. Then click Test Instrumentation to send the experiment’s variants to the testers you designated when you configured the experiment's delivery.

QA before rollout

Before any users see your experiment, you’ll want to make sure that the variants you’ve developed look and function exactly the way you intended.

Because Experiment allows you to allocate specific variants by user ID, device ID, or cohort, you can quickly and easily ensure that your test devices are served the relevant variants when they enter your experiment. All you have to do is open your product and trigger the exposure event for the experiment. You should see the variant you specified.

Launch your experiment

When you’re satisfied the implementation is as you intended, click Start Experiment. In the modal that opens, you can set an end date for the experiment, if you prefer.

Note

Start Experiment only activates the experiment once. Changing the start date will not automatically trigger the experiment to activate on the new start date.

Once the experiment is running, the button is relabeled Complete Experiment. You will be able to click this button again when you reach the experiment's end date, or when the experiment hits statistical significance. At that point, you can do one of three things:

  • roll out the winning variant
  • roll back everything and return to a pre-experiment state, or
  • continue the experiment

You can always revisit this decision after you've made it.

Note

You may receive the warning "Unable to analyze this metric, please check your metric definition or refresh this page" before launching an experiment or using a particular metric for the first time. This will not prevent you from running your experiment and testing your chosen parameters. 

QA after rollout

After rollout, you’ll be able to track how many of your users have been exposed to each variant on a daily basis. The Monitor card on the Experiment Overview page breaks this out for you in both chart and tabular form.

This is a useful way to QA the assignment process. If you notice that one variant is enrolling significantly more or significantly fewer users than you expected, it could indicate an issue you should investigate.

If you do spot some outliers or anomalies that concern you, click Root Cause Analysis to conduct a deep dive into the potential causes. To learn more about how the Root Cause Analysis feature works, see this article in the Help Center.

What happens when your experiment ends?

If you roll out your experiment to all users:

  • Percentage rollouts are set to 100%
  • Sticky bucketing is set to false
  • The rollout weights change, to 1 for the variant you're rolling out and to 0 for all other variants

If you roll out your experiment to “custom,” the automatic changes listed above will not occur. You will have to apply changes manually after confirming your rollout decision.

If you roll back your experiment:

  • The flag is turned off
  • Percentage rollouts are set to 0%

If you opt to continue running, your experiment, you can enter a new end date.

Now that you’ve rolled out your experiment, the next step is to learn from it.

Was this page helpful?

Thanks for your feedback!

April 30th, 2024

Need help? Contact Support

Visit Amplitude.com

Have a look at the Amplitude Blog

Learn more at Amplitude Academy

© 2024 Amplitude, Inc. All rights reserved. Amplitude is a registered trademark of Amplitude, Inc.