Nine Experimentation Best Practices

Serving over 200 billion feature flags daily to help software teams build better software, faster. LaunchDarkly helps eliminate risk for developers and operations teams from the software development cycle.

This post is by Dawn Parzych of LaunchDarkly

Running experiments helps you bridge gaps between technical and business teams as you learn about user engagement patterns and application behavior. Experiments can validate that your development teams’ efforts align with your business objectives. You can configure an experiment from any feature flag (front-end, back-end, or operational), but what makes a good experiment?

Experiments use metrics to validate or disprove gut feelings about whether a new feature is meeting customers’ expectations. With an experiment, you can create and test a hypothesis regarding any aspect of your feature development process.

  • Will adding white space to the page result in people spending more time on the site?
  • Do alternate images lead to increased sales?
  • Will adding sorting to the page significantly increase load times?

Experiments provide concrete reporting and measurements to ensure that you are launching the best version of a feature that positively impacts company metrics.

General experimentation best practices

Let’s first talk about best practices around creating an experiment. Even with a solid foundation of feature flagging in your organization, missteps in these areas can yield flawed results. Consider these best practices for experimentation.

1. Create a culture of experimentation

Experiments can help you prove or disprove a hypothesis, but only if you are willing to trust the outcome and not try to game the experiment. Creating a culture of experimentation means:

  • People feel safe asking questions and questioning the answer. Sometimes the results of an experiment may not be what you were expecting. It’s ok to question the results and explore anomalies.
  • Ideas are solicited from all team members—business stakeholders, data analysts, product managers, and developers.
  • A data-driven approach is used to put metrics first, not as something that is only useful in dashboards after a feature ships.

Part of creating a culture of experimentation is providing the tools and training to allow teams to test and validate their features.

Tools needed for experimentation:

  • Tools to help you collect relevant metrics can include monitoring, observability, and business analytics tools.
  • Tools to help you segment users.
  • Tools to clean and analyze the results.

2. Define what success looks like as a team

Experiments help you determine when a feature is good enough to release. How do you define what “good enough” is, and who is involved in creating the definition? Experiments can involve a cross-functional team of people or only a handful, depending on the focus of the experiment. If an experiment is on whether to add a “free trial” button to the home page, you may need to involve people from demand generation, design, and business development.

A good experiment requires well-defined and agreed-upon goals and metrics by all stakeholders. Ask yourself, “What does success look like?” Success means improving a specific metric by a specific amount. “Waiting to see what happens” is not a goal befitting a true experiment. Goals need to be concrete and avoid ambiguous words. Tie goals and metrics to business objectives like increasing revenue, time spent on a page, or growing the subscription base.

Examples of poor goals include vague and ambiguous statements:

  • Users will be happier with the home page.
  • The response time of search results will improve.

Examples of better goals include concrete statistics:

  • Paid conversions from a trial will improve by 7%.
  • The response time of search results will decrease by 450ms.

Looking at a single metric is good; looking at related metrics is even better. Identify complementary metrics and examine how they behave. If you get more new subscribers—that’s great, but you also want to look at the total number of subscribers. Say you get additional subscribers, but it inadvertently causes existing subscribers to cancel. Your total number of subscribers may be down as a result, does that make the experiment a success? I would say no.

3. Statistical significance

The number of samples will depend on your weekly or monthly active users and your application. No matter the size of the samples, it is important to maintain a relatively equal sample size. Significantly more samples in one cohort can skew the results.

Think about how and when users access your application when starting an experiment. If users primarily use your application on the weekends, experiments should include those days. If users visit a site multiple times over the course of a couple of weeks before converting and before the experiment starts, early results may be skewed, showing a positive effect when it was neutral or negative.

4. Proper segmentation

There are two aspects to segmentation. First, how will you segment your users, and second, how will you segment the data? A successful experiment needs two or more cohorts. How you segment your users will vary based on your business and users. Some ways to segment users:

  • Logged in vs. anonymous
  • By company
  • By geography
  • Randomly

However you segment users, make sure the sample sizes are balanced. Having skewed cohorts will result in skewed results.

When analyzing the data after an experiment, you may want to do additional segmentation to see if the results vary based on other parameters.

In AirBnB’s experimentation blog, they described an experiment where they received neutral results when they were expecting to see an uplift. Analyzing the results per browser, they uncovered a bug in IE that skewed the results. When the bug was fixed, they saw the improvement they expected. If there was not a culture of experimentation where it was ok to question the result and look for answers, this feature might not have rolled out.

Caution: don’t go digging to find data to support your hypothesis. The segmentation should be done to understand anomalies better, not make a case to call the experiment a success.

5. Recognize your biases

And this brings us to the subject of biases. Everybody is biased. This isn’t a bad thing; it is a matter of how our brains work. Understanding your biases and the biases of others around you will help when running experiments. Our biases influence how we process information, make decisions, and form judgments.

Some biases that may creep in:

  • Anchoring bias – our tendency to rely on the first piece of information we receive.
  • Confirmation bias – searching for information that confirms our beliefs, preconceptions, or hypotheses.
  • Default effect – the tendency to favor the default option when presented with choices.
  • And my personal favorite, the bias blind spot – our belief that we are less biased than others.

It is common in an experiment to scrub the data to remove outliers. This is acceptable, but make sure you aren’t eliminating outliers due to bias. Don’t invalidate the results of your experiment by removing data points that don’t fit the story.

6. Conduct a retro

Experiments are about learning. So after every experiment, hold a retrospective. Ask:

  • Was the experiment a success? Did we get the results we expected? Why or why not?
  • What questions were raised by the experiment? What questions were answered?
  • What did we learn?

And most importantly, should we launch the feature? You can still decide to launch a feature if the results of an experiment were neutral.

Experimentation feature flags best practices

Now that we’ve covered the basics of experiment best practices, let’s dive into best practices around using feature flags in your experiment. Well done feature flagging is a foundation for a well-run experiment.

If you’ve embraced feature management as part of your teams’ DNA, any group that wants to run an experiment can quickly do so without involving engineering. Follow these best practices to get the most from your existing feature flags.

7. Consider experiments during the planning phase

The decision of whether to wrap a feature in a flag begins during the planning phase. This is also the right time to think about experiments. When creating a feature, evaluate the metrics that will indicate whether the feature is a success—clicks, page load time, registrations, sales, etc.

Talk with the various stakeholders to determine whether an experiment may be necessary and provide them with the essential information on the flag, such as the name of the flag, to configure an experiment.

8. Empower others

When giving others the ability to run experiments, make sure you have the proper controls in place to avoid flags accidentally being changed or deleted. Empower other teams to create targeting rules, define segments, and configure roll-out percentages while preventing them from toggling or deleting flags.

9. Avoid technical debt

Experimentation flags should be short-lived. This means removing a flag after an experiment completes and is either rolled out to 100% of users or not released. Put processes and reminders in place to remove the flag. We highlighted some creative ways to do this in our short-term and permanent best practices blog.

To get started with your own experiments, check out the LaunchDarkly documentation.

Are you looking for other best practices? Check out the other blog posts in the series.

Long-term and short-term flags best practices

Operational flags best practices

Release management flags best practices

Serving over 200 billion feature flags daily to help software teams build better software, faster. LaunchDarkly helps eliminate risk for developers and operations teams from the software development cycle.
Tools mentioned in article
Open jobs at LaunchDarkly
IT Support Engineer
Oakland, CA
LaunchDarkly is a rapidly growing software company with a strong mission and vision carried out by a talented and diverse team of employees. Our goal is to help teams build better software, faster. You'll join a small team from companies like Atlassian, Intercom, and GitHub, and you'll have an immediate impact on our product and customers. We are looking for an IT Service Desk / Support Engineer to be a first line of defense for our growing IT support team. We currently have over 150 employees globally and will be adding additional headcount over the next year. This role will help to enable the scale of the company by being the face of the technical service desk and IT team. This position will be responsible for maintaining the service level of our service desk and will also be the first point of contact for all technical matters. This role will report directly to the IT Lead.
  • Be a primary responder for LaunchDarkly's service desk
  • Support a primarily Mac-based environment with some Windows / Linux devices
  • Install, configure and troubleshoot computer hardware
  • Assist with on-boarding and off-boarding LaunchDarkly employees
  • Support company AV suites and All Hands meetings
  • 1-3 years of experience as a technical support expert supporting users in a corporate environment
  • Excellent interpersonal and customer service skills. Interact well at all levels within the organization.
  • Well-organized and highly efficient
  • You bring a curiosity for technology trends and implementing cutting edge technical solutions
  • Strong documentation and communication skills
  • Okta
  • JAMF
  • Asset management
  • On-boarding/Off-boarding Automation
  • AV room and video conference platforms like Zoom, WebEx, BlueJeans
  • Supported a user base using SAAS based ticketing system
  • Supported Wireless 802.11X installations in corporate settings
  • Ability to troubleshoot complex systemic problems
  • Ability to troubleshoot Apple hardware
  • Systems administration for mac and Linux systems.
  • General understanding of ITIL, OSI Model, ISO 27001, SOC II
  • Scripting or Automating SaaS applications
  • GSuite administration
  • Atlassian administration
  • A+ Certification
  • Okta helpdesk certification
  • Solutions Engineer - Sydney, Australi...
    As a Solutions Engineer, you will educate and guide prospects on the proper implementation of LaunchDarkly's SaaS product and Private Instances. You will be the technical voice during our sale and ensure our customers are comfortable with the way our systems work. You are passionate about the developer tools space and helping development teams eliminate risk and deliver value. This role will support the Eastern timezone, as such candidates who are in New York or the East Coast are preferred. LaunchDarkly is a rapidly growing software company with a strong mission and vision carried out by a talented and diverse team of employees. Our goal is to help teams build better software, faster. You'll join a small team from companies like Atlassian, Intercom, and GitHub, and you'll have an immediate impact with our product and customers.
  • Evangelize and advise customers on the importance and different uses of feature flags and how to administer them
  • Create solutions to customer's challenges implementing feature flags across large monolith and microservice applications, large organizations, and different technology stacks
  • Become a domain expert on LaunchDarkly architecture
  • Demo LaunchDarkly product to technical and business audiences
  • Become a subject matter expert on LaunchDarkly and communicate our value and features to potential customers
  • Be the voice of the customer by translating, aggregating, and representing customer feedback to the Product and Engineering teams
  • You learn and synthesize large amounts of information with little context
  • You are an effective communicator and you can simplify complex technical concepts
  • You are a self‐starter and problem solver, willing to take on hard problems and work independently when necessary.
  • You have 4+ years experience consulting with enterprise customers and large development teams
  • You led successful technical proof of concepts and built strong customer relationships
  • You are passionate about trends and technologies involved in modern application development
  • You have worked with teams that underwent development process transformation
  • You are familiar with at least one of our supported languages: Java, .NET, GO, JS, Python, PHP, node, Ruby, Rails, iOS, or Android
  • You have worked with data persistence technologies like Varnish or Redis
  • Director, Infrastructure Delivery
    Oakland, CA
    Software powers the world, and LaunchDarkly empowers all teams to deliver and control their software. We serve hundreds of billions of feature flags daily, to help teams ship better software, faster and eliminate risk for companies big and small. We're based in downtown Oakland and growing quickly.  Infrastructure Delivery at LaunchDarkly is an engineering discipline responsible for building and delivering technology solutions for LaunchDarkly employees. The team's mission is to use technology to help make LaunchDarkly employees efficient and happy. It does this by delivering IT services and infrastructure (e.g. help desk, networking, physical security), building internal tools and services and delivering BI and data tools.  The ideal fit is someone who can dig into business challenges and ask core questions, ensure the effective use of data, and build convincing plans for action. The Director of Infrastructure Delivery will build and manage the department and report directly to the CTO. 
  • Hire and manage IT, Tools, and Data teams
  • Run the team with the same level of discipline, process, and rigor that we use on our product-focused engineering teams 
  • Help your team understand their mission and how they fit in with the overall business goals
  • Run a software team charged with crafting and maintaining internal LaunchDarkly tools using modern software development processes and stacks (React, Go, TypeScript, AWS, Terraform etc.)
  • Define and measure success criteria for your functional teams
  • Lead IT operations for all LaunchDarkly’s  locations, ensuring employee experience as measured by user satisfaction surveys and internal SLAs.
  • Manage the inventory of LaunchDarkly's IT assets, ensuring budget compliance, and best practices.
  • Determine business requirements for IT systems and coordinate activities to ensure Availability.
  • Effectively communicate with stakeholders regarding the effectiveness and progress of business systems and enterprise applications, as well as IT projects and performance.
  • Efficiently manage LaunchDarkly's global IT budget, anticipating for company growth as required.
  • Help the data team build an operate a scalable, reliable data pipeline and data warehouse.
  • Help identify and answer the business' most pressing questions.
  • Build tools and teach people how to find answers to their own questions using trustworthy data.
  • Establish and maintain strong relationships with vendors, MSPs, suppliers, and key consultants.
  • Three or more years’ experience leading IT or Engineering teams: hiring, motivating, growing, empowering, and performance managing.
  • A history of creating or maintaining great team cultures.
  • Skills and knowledge that will establish credibility with your team.
  • You understand build vs. buy dynamics and know when each is appropriate
  • Demonstrated analytical, problem-solving, organizational, interpersonal and communication skills.
  • Organized, thorough and diligent with an acute attention to detail.
  • Knowledge of middleware integration tools like (Bettercloud, Zapier, Celigo) 
  • Verified by
    Director Marketing
    VP of Product and Engineering
    You may also like