<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Inferentia: Experimentation]]></title><description><![CDATA[Experimentation is the engine of discovery and growth. On this page, you’ll learn how to design tests, analyse outcomes, and refine user experiences using data-driven insights—all to help you innovate faster and optimise results.]]></description><link>https://www.inferentia.in/s/experimentation</link><image><url>https://substackcdn.com/image/fetch/$s_!Ydl9!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20523561-1faf-45ef-81af-96c7aeb07db6_1120x1120.png</url><title>Inferentia: Experimentation</title><link>https://www.inferentia.in/s/experimentation</link></image><generator>Substack</generator><lastBuildDate>Sat, 11 Apr 2026 05:54:18 GMT</lastBuildDate><atom:link href="https://www.inferentia.in/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Piyush Ranjan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[inferentia@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[inferentia@substack.com]]></itunes:email><itunes:name><![CDATA[Piyush Ranjan]]></itunes:name></itunes:owner><itunes:author><![CDATA[Piyush Ranjan]]></itunes:author><googleplay:owner><![CDATA[inferentia@substack.com]]></googleplay:owner><googleplay:email><![CDATA[inferentia@substack.com]]></googleplay:email><googleplay:author><![CDATA[Piyush Ranjan]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Running and Analysing Experiments: An End-to-End Example]]></title><description><![CDATA[From Hypothesis to Insight]]></description><link>https://www.inferentia.in/p/running-and-analysing-experiments</link><guid isPermaLink="false">https://www.inferentia.in/p/running-and-analysing-experiments</guid><dc:creator><![CDATA[Piyush Ranjan]]></dc:creator><pubDate>Sun, 26 Jan 2025 14:22:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!XYod!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Introduction</h3><p>In today's data-driven world, making informed decisions about product changes is crucial. A/B testing offers a powerful framework for gathering actionable insights and making evidence-based decisions, particularly in web and app development. In this blog post, we delve into the essential principles of designing, executing, and analyzing an A/B test using a practical, real-world example.</p><h3>The Experimental Context: A Widget Store's Coupon Code Dilemma</h3><p>Our example centers on a fictional online commerce platform specializing in selling widgets. In an effort to boost sales, the marketing team proposes a strategy involving promotional emails that include a coupon code for discounts on widgets. This proposal marks a potential shift in the company's business model, as they have never offered coupon codes before. The team is cautious, however, due to findings from prior studies&#8212;such as Dr. Footcare's revenue loss following the introduction of coupon codes (Kohavi, Longbottom et al., 2009) and evidence from GoodUI.org suggesting that removing coupon codes can be beneficial (Linowski, 2018). These insights raise concerns that simply adding a coupon code field to the checkout page might negatively impact revenue, even if no valid codes are available. Users might become distracted searching for codes or even abandon their purchases altogether.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!opMW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!opMW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 424w, https://substackcdn.com/image/fetch/$s_!opMW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 848w, https://substackcdn.com/image/fetch/$s_!opMW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 1272w, https://substackcdn.com/image/fetch/$s_!opMW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!opMW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png" width="776" height="296" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:296,&quot;width&quot;:776,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:243182,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!opMW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 424w, https://substackcdn.com/image/fetch/$s_!opMW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 848w, https://substackcdn.com/image/fetch/$s_!opMW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 1272w, https://substackcdn.com/image/fetch/$s_!opMW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c268feb-2507-4678-bc79-f089dc43f97f_776x296.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Control and Treatment 1</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XYod!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XYod!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 424w, https://substackcdn.com/image/fetch/$s_!XYod!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 848w, https://substackcdn.com/image/fetch/$s_!XYod!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 1272w, https://substackcdn.com/image/fetch/$s_!XYod!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XYod!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png" width="870" height="334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:334,&quot;width&quot;:870,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:258783,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XYod!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 424w, https://substackcdn.com/image/fetch/$s_!XYod!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 848w, https://substackcdn.com/image/fetch/$s_!XYod!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 1272w, https://substackcdn.com/image/fetch/$s_!XYod!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af22e0-9276-4b71-8c73-2a39a69d74b4_870x334.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Treatment 2</figcaption></figure></div><p>To assess these potential effects, we adopt a &#8220;painted door&#8221; approach. This involves implementing a superficial change&#8212;adding a non-functional coupon code field to the checkout page. When users input any code, the system responds with &#8220;Invalid Coupon Code.&#8221; By focusing on this minimal implementation, we aim to evaluate the psychological and behavioral impact of merely displaying a coupon code field.</p><p>Given the simplicity of this change, we&#8217;ll test two distinct UI designs to compare their effectiveness. Testing multiple treatments alongside a control allows us to discern not only whether the idea of adding a coupon code is viable, but also which specific implementation is most effective. This targeted A/B test is a crucial step toward determining the feasibility of adopting coupon codes as part of the company&#8217;s broader business strategy.</p><h4></h4><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.inferentia.in/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.inferentia.in/subscribe?"><span>Subscribe now</span></a></p><h4>Hypothesis</h4><p><em>Adding a coupon code field to the checkout page will degrade revenue-per-user for users who start the purchase process.</em></p><p>To test this hypothesis, we consider two UI implementations. This A/B test is a critical step in assessing the feasibility of this business model.</p><h4>Goal Metrics</h4><p>The primary metric, or Overall Evaluation Criterion (OEC), is revenue-per-user, normalized to account for sample size variability. To measure the impact of the change, we define success metrics carefully. While revenue is an obvious choice, using the total revenue sum is not recommended due to sample size variations across variants. Instead, revenue-per-user provides a more accurate and normalized metric. For this metric, it is critical to determine the denominator:</p><ol><li><p>Including all site visitors introduces significant noise, as a large portion of users never initiate checkout. These users do not interact with the modified checkout process and, therefore, cannot provide meaningful data on the impact of the change. Including them dilutes the results, making it harder to detect any real differences caused by the experiment.</p></li><li><p>Restricting to only users who complete the purchase process presents a skewed perspective, as it inherently assumes that the modification influences purchase amounts without considering its effect on conversion rates. This approach excludes users who might abandon the process due to the change, potentially missing a critical aspect of the experiment's impact.</p></li><li><p>The best choice is users who start the purchase process, as they are directly exposed to the change at the checkout stage. This refined approach improves test sensitivity by ensuring that only the users who interact with the modified UI are analyzed. By excluding unaffected users, such as those who browse without adding items to the cart or initiating a purchase, the metric becomes more precise. This specificity helps isolate the true impact of the coupon code field, avoiding noise introduced by broader user behaviors that are irrelevant to the experiment.</p></li></ol><p>With this setup, our refined hypothesis becomes: &#8220;<strong>Adding a coupon code field to the checkout page will degrade revenue-per-user for users who start the purchase process.</strong>&#8221;</p><h4>Hypothesis Testing</h4><p>Before we can design, run, or analyze our experiment, let us go over a few foundational concepts relating to statistical hypothesis testing. First, we characterize the metric by understanding the baseline mean value&#8212;the average of the metric under normal, unaltered conditions&#8212;and the standard error of the mean. The standard error provides insight into the variability of our metric estimates and is crucial for determining the required sample size to detect meaningful differences. By accurately estimating this variability, we can size our experiment properly and assess statistical significance during analysis.</p><p>For most metrics, we measure the mean, but alternative summary statistics, such as medians or percentiles, may be more appropriate in specific contexts, such as highly skewed data distributions. Sensitivity, or the ability to detect statistically significant differences, improves when the standard error of the mean is reduced. This can be achieved by either increasing the traffic allocated to the experimental variants or running the experiment for an extended period. However, running longer experiments may yield diminishing returns after a few weeks due to sub-linear growth in unique users (caused by repeat visitors) and potential increases in variance for certain metrics over time.</p><p>To evaluate the impact of the experiment, we analyze revenue-per-user estimates from the Control and Treatment samples by computing the p-value for their difference. The p-value represents the probability of observing the measured difference, or a more extreme one, under the assumption that the Null hypothesis&#8212;that there is no true difference&#8212;is correct. A sufficiently small p-value allows us to reject the Null hypothesis and infer that the observed effect is statistically significant. But what constitutes a small enough p-value?</p><p>Typically, the scientific benchmark is a p-value less than 0.05. This threshold means there is less than a 5% probability of incorrectly concluding there is an effect when none actually exists, providing confidence in 95 out of 100 cases. Furthermore, another approach to determine significance is through confidence intervals. A 95% confidence interval defines a range where the true difference between Treatment and Control lies 95% of the time. If this interval does not include zero, it reinforces the conclusion that the effect is statistically significant. These tools collectively help establish the robustness of experimental findings, ensuring decisions are data-driven and reliable.</p><h4>Statistical Power</h4><p>Statistical power measures the ability of an experiment to detect a meaningful difference between variants when such a difference truly exists. In simple terms, it is the probability of correctly rejecting the null hypothesis when there is an actual effect. For instance, if a retailer is testing a new homepage layout, statistical power ensures that subtle but real increases in sales do not go unnoticed.</p><p>To achieve reliable results, experiments are often designed with 80-90% power, meaning there is a high likelihood of detecting true changes. Power is influenced by factors such as sample size and effect size; larger sample sizes tend to improve power, but overly small differences might still evade detection. For example, while a large e-commerce platform like Amazon might be interested in detecting a 0.2% increase in revenue-per-user due to its massive scale, a smaller startup might focus only on changes exceeding 5-10% because such increases are critical for their growth.</p><p>While statistical significance helps us understand whether an observed difference is likely due to chance, it does not always translate into practical significance. Practical significance asks a more business-oriented question: Is the observed change large enough to matter? For example, a 0.2% increase in revenue-per-user might be meaningful for billion-dollar platforms like Google or Bing, but for a small startup seeking rapid growth, a 2% increase might still fall short of expectations. Setting clear business thresholds for what constitutes a meaningful change is essential. For our hypothetical widget store, we define practical significance as a 1% or larger increase in revenue-per-user, recognising this as the minimum impact needed to justify potential costs or risks of implementation.</p><h3>Designing the Experiment</h3><p>We are now ready to design our experiment. We have a hypothesis, a practical significance boundary, and we have characterised our metric. We will use this set of decisions to finalize the design:</p><ol><li><p><strong>What is the randomisation unit?</strong> The randomisation unit for this experiment is the user.</p></li><li><p><strong>What population of randomisation units do we want to target?</strong> We will target all users and analyse those who visit the checkout page. Targeting a specific population allows for more focused results. For instance, if the new text in a feature is only available in certain languages, you would target users with those specific interface locales. Similarly, attributes such as geographic region, platform, and device type can guide targeting.</p></li><li><p><strong>How large does our experiment need to be?</strong> The experiment size directly impacts the precision of results. To detect a 1% change in revenue-per-user with 80% power, we will conduct a power analysis to determine the sample size. The following considerations influence size:</p><ul><li><p>Using a binary metric like purchase indicator (yes/no) instead of revenue-per-user can reduce variability, allowing for a smaller sample size.</p></li><li><p>Increasing the practical significance threshold&#8212;for example, detecting only changes larger than 1%&#8212;can also reduce sample size requirements.</p></li><li><p>Lowering the p-value threshold, such as from 0.05 to 0.01, increases sample size needs.</p></li></ul></li><li><p><strong>How long should the experiment run?</strong> To ensure robust results, we will run the experiment for at least one week to capture weekly cycles and account for day-of-week effects. External factors like seasonality and primacy or novelty effects are also important:</p><ul><li><p>User behavior can vary during holidays or promotional periods, affecting external validity.</p></li><li><p>Novelty effects (e.g., initial enthusiasm for a new feature) and adoption effects (e.g., gradual user adoption) may impact results over time.</p></li></ul></li></ol><h3>Final Experiment Design</h3><ol><li><p><strong>Randomization Unit</strong>: User</p></li><li><p><strong>Target Population</strong>: All users visiting the checkout page</p></li><li><p><strong>Experiment Size</strong>: Determined via power analysis to achieve 80% power for detecting a 1% change</p></li><li><p><strong>Experiment Duration</strong>: Minimum of one week to capture weekly cycles and extended if novelty or primacy effects are detected</p></li><li><p><strong>Traffic Split</strong>: 34/33/33% for Control, Treatment 1, and Treatment 2</p></li></ol><p>By carefully designing the experiment with these considerations, we can ensure that the results are both statistically and practically significant. Overpowering an experiment is often beneficial, as it allows for detailed segment analysis (e.g., by geographic region or platform). This approach not only improves the robustness of results but also helps businesses uncover nuanced insights. For example, identifying specific trends in user behavior across different demographics can inform future product iterations and marketing strategies.</p><p>Furthermore, running a well-structured A/B test fosters a culture of data-driven decision-making within the organization. By investing in comprehensive experiment design and analysis, businesses can mitigate risks, allocate resources effectively, and achieve sustainable growth. Ultimately, the insights derived from this experiment will not only validate the feasibility of introducing coupon codes but also set a benchmark for future experimentation, reinforcing the importance of innovation and customer-centric strategies in a competitive market.</p><h3>Running the Experiment and Getting Data</h3><p>Now let us run the experiment and gather the necessary data. To run an experiment, we need both:</p><ul><li><p><strong>Instrumentation</strong> to get logs data on how users are interacting with your site and which experiments those interactions belong to.</p></li><li><p><strong>Infrastructure</strong> to be able to run an experiment, ranging from experiment configuration to variant assignment.</p></li></ul><h3>Interpreting the Results</h3><p>The data collected from the experiment is the foundation for actionable insights, but ensuring its reliability is critical. Before diving into the revenue-per-user analysis, it is essential to validate the experiment's execution by examining invariant metrics, also known as guardrail metrics. These metrics serve two primary purposes:</p><ol><li><p><strong>Trust-related Guardrails</strong>: Metrics such as sample size consistency and cache-hit rates ensure that the Control and Treatment groups align with the experiment configuration. Any deviation here might indicate issues in randomization or assignment.</p></li><li><p><strong>Organizational Guardrails</strong>: Metrics like latency or system performance, which are crucial to business operations, should remain stable across variants. For example, significant changes in checkout latency would signal underlying problems unrelated to the coupon code test.</p></li></ol><p>If these metrics show unexpected changes, it suggests flaws in the experiment design, infrastructure, or data processing pipeline. Addressing these issues before analyzing the core results is vital to maintaining trust in the findings.</p><p>Once the guardrails are validated, the next step is to analyze and interpret the results with precision. For example, if the p-value for revenue-per-user in both Treatment groups is below 0.05, we reject the null hypothesis and conclude that the observed differences are statistically significant. However, statistical significance alone is insufficient. Practical significance&#8212;the magnitude of the observed effect&#8212;determines whether the change is worth implementing.</p><h4>Results Table</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!seBd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!seBd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 424w, https://substackcdn.com/image/fetch/$s_!seBd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 848w, https://substackcdn.com/image/fetch/$s_!seBd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 1272w, https://substackcdn.com/image/fetch/$s_!seBd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!seBd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png" width="1190" height="420" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:420,&quot;width&quot;:1190,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:67218,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!seBd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 424w, https://substackcdn.com/image/fetch/$s_!seBd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 848w, https://substackcdn.com/image/fetch/$s_!seBd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 1272w, https://substackcdn.com/image/fetch/$s_!seBd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f7ac163-f4b4-4bdf-83b0-ab7945bec07f_1190x420.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>From the table, we observe that both treatments significantly reduce revenue-per-user compared to the control group. While the p-values confirm statistical significance, the negative impact on revenue highlights the need to reassess introducing coupon codes.</p><h4>Decision-Making Framework</h4><p>Consider the context of your experiment:</p><ul><li><p><strong>Short-Term vs. Long-Term Impact</strong>: Changes with minimal downside risks, such as testing promotional headlines, may allow for lower thresholds of significance. Conversely, introducing high-cost features like a coupon code system requires higher thresholds due to long-term resource commitments.</p></li><li><p><strong>Balancing Metrics</strong>: A decrease in revenue may be acceptable if offset by an increase in user engagement, but only if the net impact aligns with organizational goals.</p></li></ul><p>Ultimately, the results must translate into a clear decision framework. For example:</p><ol><li><p>If the results are both statistically and practically significant, the decision to launch is straightforward.</p></li><li><p>If statistically significant but not practically meaningful, the change may not justify further investment.</p></li><li><p>If results are inconclusive, consider increasing sample size or re-evaluating the design.</p></li><li><p>The result is statistically significant, and likely practically significant. Like prior examples, it is possible that the change is not practically significant. In this situation, repeating the test with greater power is advisable. However, from a launch/no-launch perspective, choosing to launch is a reasonable decision. It is crucial to explicitly document the factors influencing this decision, particularly how they align with the practical and statistical significance boundaries. This clarity not only supports current decision-making but also establishes a solid foundation for future analyses.</p></li></ol><p>By grounding decision-making in robust analysis and broader business considerations, organizations can confidently use A/B testing as a tool for sustainable growth and innovation.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.inferentia.in/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Inferentia! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[A/B Testing 101: The Power of Experimentation]]></title><description><![CDATA[The $100M Experiment: How A/B Testing Transformed Bing&#8217;s Revenue]]></description><link>https://www.inferentia.in/p/ab-testing-101-a-beginners-guide</link><guid isPermaLink="false">https://www.inferentia.in/p/ab-testing-101-a-beginners-guide</guid><dc:creator><![CDATA[Piyush Ranjan]]></dc:creator><pubDate>Wed, 15 Jan 2025 03:30:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>"One accurate measurement is worth more than a thousand expert opinions."</em></p><p><em>                                                                                              &#8211; Admiral Grace Hopper</em></p><p>In the digital age, where data reigns supreme, the ability to measure the impact of changes accurately is crucial. The famous quote by Admiral Grace Hopper encapsulates the essence of why businesses like Microsoft's Bing have turned to A/B testing to drive innovation and profitability.</p><h4>The Bing Case Study</h4><p>In 2012, a small yet transformative idea within Microsoft&#8217;s Bing team provided a striking example of the power of data-driven experimentation. A simple suggestion to change ad headlines was overlooked for months, buried under bigger projects. When finally implemented, this idea became Bing&#8217;s most significant revenue-generating change, highlighting the profound impact of rigorous online controlled experiments (A/B tests).</p><p>The proposed change involved extending ad title lines by combining them with text from the line below, creating a more engaging headline. Initially deemed low priority, it wasn&#8217;t until a developer implemented it&#8212;due to its coding simplicity&#8212;that the idea was tested on real users. A/B testing split users between the existing and modified layouts, tracking interactions, clicks, and revenue metrics.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MK3T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MK3T!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 424w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 848w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 1272w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MK3T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png" width="970" height="1256" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1256,&quot;width&quot;:970,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:808949,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MK3T!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 424w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 848w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 1272w, https://substackcdn.com/image/fetch/$s_!MK3T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f8b6056-482f-414f-ac3b-2c463c9e6ada_970x1256.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Within hours of launch, an anomaly emerged: revenue surged unexpectedly. Teams braced for a potential bug, but upon verification, the results proved legitimate. The new ad layout increased Bing&#8217;s revenue by 12%, equating to over $100 million annually in the U.S. alone, highlighting several key lessons:</p><ul><li><p><strong>Value Assessment:</strong> Even seemingly minor ideas can have significant impacts, yet their value is often underestimated or overlooked.</p></li><li><p><strong>Impact of Small Changes:</strong> A small alteration can lead to massive financial returns if it aligns well with user behavior and expectations.</p></li><li><p><strong>Rarity of Big Wins:</strong> Not every change will yield such dramatic results; this was one of the few blockbuster successes from thousands of experiments conducted annually.</p></li><li><p><strong>Efficiency in Experimentation:</strong> The overhead of running an experiment must be small. Bing&#8217;s engineers had access to ExP, Microsoft&#8217;s experimentation system, which made it easy to scientifically evaluate the idea.</p></li></ul><h4>What is A/B Testing?</h4><p>A/B testing, also known as split testing, is a method of comparing two versions of a webpage, email, or other marketing asset to determine which one performs better. By splitting your audience into two groups and showing each group a different version (Version A vs. Version B), you can measure which variation drives more conversions, clicks, or other desired outcomes.</p><p>Companies like Airbnb, Amazon, Booking.com, eBay, Facebook, Google, LinkedIn, Lyft, Microsoft, and Netflix rely heavily on online controlled experiments (Gupta et al., 2019). These organisations conduct thousands to tens of thousands of experiments annually, often involving millions of users. They test a wide range of elements, such as user interface (UI) changes, relevance algorithms (including search, ads, personalization, and recommendations), latency and performance, content management systems, customer support systems, and more. Experiments span various channels, including websites, desktop applications, mobile apps, and email.</p><p>For example, if you&#8217;re testing a landing page, you might experiment with two different headlines. Group A sees &#8220;Save Big on Your Next Adventure,&#8221; while Group B sees &#8220;Discover Affordable Travel Deals.&#8221; By analysing which headline leads to more sign-ups or sales, you gain valuable insights into what resonates with your audience.</p><p>In typical online controlled experiments, users are randomly assigned to different variants in a consistent manner, ensuring they experience the same variant across multiple visits. In the Bing example, the Control group saw the original ad display, while the Treatment group saw ads with longer titles. User interactions with the Bing website were tracked and logged, enabling the calculation of metrics from the logged data. These metrics were then used to evaluate the differences between the variants.</p><p><strong>Understanding A/B Testing Terminology</strong></p><p>A/B testing, or controlled experiments, involves several terms:</p><ul><li><p><strong>Overall Evaluation Criterion (OEC):</strong> A metric that encapsulates the experiment's goal, like increase in revenue balanced with user experience metrics. The OEC should be measurable within the short timeframe of an experiment while being expected to causally influence long-term strategic goals. </p><p>Experiments may have multiple objectives, and analysis can employ a balanced scorecard approach. However, it is highly recommended to select a single metric, potentially as a weighted combination of these objectives, to streamline evaluation (Roy 2001, 50, 405&#8722;429).</p></li><li><p><strong>Parameters:</strong> A controllable experimental variable, often called a parameter, is a factor that can be adjusted during an experiment and is believed to influence the Overall Evaluation Criterion (OEC) or other key metrics of interest. These parameters are assigned specific values or levels, representing the variations being tested. Understanding how these parameters affect outcomes allows experimenters to optimize for the best results.</p><ul><li><p>Simple A/B Tests:<br>In an A/B test, there is typically one parameter with two levels. A single parameter like button color might have two levels: blue (A) and red (B), with the goal of determining which generates higher click-through rates.</p></li><li><p>Univariable Tests with Multiple Levels:<br>A test with a single parameter that has more than two levels.A parameter like button placement could have levels such as top of the page (A), middle (B), bottom (C), or sidebar (D), to identify the most effective location.</p></li><li><p>Multivariable Tests (Multivariate Tests - MVTs):<br>MVTs test multiple parameters simultaneously, making it possible to analyze how their interactions impact results.Multiple parameters are tested together, such as button shape (round, square) and button text (&#8220;Buy Now,&#8221; &#8220;Add to Cart&#8221;), to evaluate combinations and uncover interactions that maximize conversions.</p></li></ul><p>While simple A/B tests or univariable designs are effective for evaluating straightforward changes, MVTs are particularly useful when multiple factors interact in non-obvious ways. For instance, a font size that performs well with one color might perform poorly with another. By testing these combinations, experimenters can discover a <strong>global optimum</strong>&#8212;the best overall combination of changes that maximizes the desired outcome.</p></li><li><p><strong>Variants:</strong> A user experience being tested is defined by assigning specific values to parameters, creating distinct variants for comparison. In a typical A/B test, these variants are labeled as Control and Treatment, with the Control representing the existing version (baseline) and the Treatment reflecting the modified version being tested. While some literature uses "variant" solely for the Treatment, the Control is also considered a critical variant, serving as the benchmark for measuring changes in key metrics.</p><p>For example, in an experiment testing button colors, the Control might feature the current blue button, while the Treatment introduces a red button. If a bug or unexpected issue arises during the experiment, it is standard practice to abort the test and ensure all users are reverted to the Control variant, thereby minimizing the potential for adverse impacts. This process safeguards the user experience while maintaining the integrity of the baseline performance data for future analysis.</p></li><li><p><strong>Randomisation Unit:</strong> Randomization is a critical component of controlled experiments. A pseudo-randomization process, such as hashing, is applied to units (e.g., users or pages) to assign them to different variants. Proper randomization ensures that the populations assigned to each variant are statistically similar, enabling causal effects to be determined with high confidence. The mapping of units to variants must be both persistent and independent. For example, if a user is the randomization unit, that user should consistently experience the same variant throughout the experiment, and their assignment should reveal nothing about how other users are assigned.</p><p>Using users as the randomization unit is highly recommended for online experiments targeting audiences across websites, apps, or other digital platforms. However, alternative randomization units are sometimes employed based on the experiment's goals. These can include:</p><ul><li><p><strong>Pages</strong>: Randomising content displayed on specific pages.</p></li><li><p><strong>Sessions</strong>: Assigning a variant to a single user session but allowing different experiences across sessions.</p></li><li><p><strong>User-Days</strong>: Ensuring a consistent experience for a user within a specific 24-hour period defined by the server.</p></li></ul><p>Proper randomisation is essential to maintain the integrity of the experiment. In cases where each variant is assigned an equal proportion of users, every user must have an equal probability of being assigned to any variant. This deliberate process eliminates potential biases that could distort results.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LflP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LflP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 424w, https://substackcdn.com/image/fetch/$s_!LflP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 848w, https://substackcdn.com/image/fetch/$s_!LflP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 1272w, https://substackcdn.com/image/fetch/$s_!LflP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LflP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png" width="1456" height="546" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:546,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1117466,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LflP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 424w, https://substackcdn.com/image/fetch/$s_!LflP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 848w, https://substackcdn.com/image/fetch/$s_!LflP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 1272w, https://substackcdn.com/image/fetch/$s_!LflP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0230b97a-bb06-485c-bcb2-fa5e0b7a5304_2770x1038.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h4>Why is A/B Testing Important?</h4><p>The challenge with data interpretation lies in distinguishing correlation from causation. Observational data can mislead; for instance, users experiencing more errors might show lower churn rates, not because errors are beneficial, but because they are heavy users. Controlled experiments help establish causality by systematically varying one thing at a time while keeping everything else constant.</p><p><strong>Example: Microsoft Office 365</strong></p><p>Consider the case of Microsoft Office 365, a subscription-based software service. Observational data might reveal that users who encounter more error messages and software crashes have lower churn rates compared to others. At a glance, one might conclude that introducing more errors or lowering the software's quality could reduce customer churn, which is clearly not the logical step to take.</p><ul><li><p><strong>Correlation:</strong> The data shows a correlation between seeing error messages, experiencing crashes, and lower churn rates.</p></li><li><p><strong>Misleading Interpretation:</strong> One might mistakenly infer that errors or crashes somehow keep users engaged or less likely to unsubscribe.</p></li><li><p><strong>Actual Causation:</strong> However, the real underlying factor here is "usage". Heavy users of the product are more likely to encounter errors due to the frequency and intensity of their use. These heavy users also tend to have lower churn rates because they find more value in the product or are more invested in it, not because the errors are beneficial.</p></li></ul><p>This example with Microsoft Office 365 illustrates why jumping to conclusions from mere correlations can lead to flawed strategies. <strong>Controlled experiments provide a structured approach to validate hypotheses, ensuring that decisions are based on causal relationships rather than coincidental correlations.</strong> By isolating variables and observing the direct impact of changes, businesses can avoid costly mistakes and invest in strategies that genuinely contribute to better user experiences and business outcomes.</p><h4>Key Elements of an A/B Test</h4><p>For A/B testing to be effective:</p><ol><li><p><strong>Experimental Units (</strong>or Users<strong>):</strong> Must be assignable to different variants without interference.</p></li><li><p><strong>Sufficient Scale:</strong> Enough units to detect even small effects statistically.</p></li><li><p><strong>Metrics:</strong> Key metrics like an OEC must be clearly defined and measurable.</p></li><li><p><strong>Ease of Change:</strong> Software changes should be implementable without significant overhead.</p></li></ol><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.inferentia.in/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.inferentia.in/subscribe?"><span>Subscribe now</span></a></p><p></p><h4>Lessons for Organizations</h4><ol><li><p><strong>Data-Driven Decision Making:</strong> Establish a clear OEC that aligns with strategic goals, ensuring measurable and actionable insights.</p></li><li><p><strong>Invest in Experimentation Infrastructure:</strong> Reliable systems are crucial for managing experiments, logging data, and deriving trustworthy conclusions. Also, implementing software changes with minimal overhead is a critical enabler for effective experimentation. Agile systems and processes should allow teams to make swift updates, test hypotheses, and adapt based on findings. This approach reduces delays, fosters innovation, and ensures experiments can iterate rapidly to achieve meaningful insights without bottlenecks.</p></li><li><p><strong>Embrace Humility in Idea Assessment:</strong> Most ideas fail to deliver the anticipated impact. Controlled experiments help validate assumptions, reducing wasted effort and guiding resources toward high-value initiatives.Common A/B Testing Mistakes to Avoid</p><p></p></li></ol><h4><strong>Strategic and Tactical Benefits</strong></h4><ul><li><p><strong>Strategy Validation:</strong> Experiments can affirm or challenge strategic directions through tangible outcomes.</p></li><li><p><strong>Tactical Optimization:</strong> Small, iterative changes can lead to substantial cumulative gains.</p></li><li><p><strong>Pivoting:</strong> When experiments hint at strategic misalignments, they can lead to strategic pivots or reassessments.</p></li></ul><h4><strong>Conclusion</strong></h4><p>The Bing example serves as a powerful illustration of how A/B testing can revolutionize digital products and experiences. It highlights the importance of fostering a culture that prioritizes data-driven decision-making, backed by robust infrastructure to conduct rigorous experiments at scale. Moreover, it emphasizes the value of intellectual humility&#8212;recognizing that even the most promising ideas must be validated through empirical evidence, as our initial intuitions are often flawed.</p><p>As digital platforms continue to evolve in complexity and scale, the reliance on controlled experiments will become even more indispensable. These experiments not only validate innovations but ensure they deliver measurable value to both businesses and users. By systematically testing and learning, organizations can refine their offerings, reduce risks, and make decisions grounded in objective insights rather than assumptions. In this way, A/B testing acts as a critical enabler of innovation, ensuring that progress is not only creative but also meaningful and impactful.</p><h3>Reference</h3><ol><li><p>Kohavi, Ron, Diane Tang, and Ya Xu. 2020. &#8203;Trustworthy Online Controlled Experiments: A Practical Guide to A/B Testing. Cambridge University Press.</p></li><li><p>Gupta, Somit, Ronny Kohavi, Diane Tang, Ya Xu, and etal. 2019. &#8220;Top Challenges from the first Practical Online Controlled Experiments Summit.&#8221; Edited by Xin Luna Dong, Ankur Teredesai and Reza Zafarani. SIGKDD Explorations (ACM) 21 (1). https://bit.ly/OCESummit1.</p></li><li><p>Roy, Ranjit K. 2001. Design of Experiments using the Taguchi Approach : 16 Steps to Product and Process Improvement. John Wiley &amp; Sons, Inc.</p></li></ol>]]></content:encoded></item></channel></rss>