Skip to content
CMO & CTO
CMO & CTO

Closing the Bridge Between Marketing and Technology, By Luis Fernandez

  • Digital Experience
    • Experience Strategy
    • Experience-Driven Commerce
    • Multi-Channel Experience
    • Personalization & Targeting
    • SEO & Performance
    • User Journey & Behavior
  • Marketing Technologies
    • Analytics & Measurement
    • Content Management Systems
    • Customer Data Platforms
    • Digital Asset Management
    • Marketing Automation
    • MarTech Stack & Strategy
    • Technology Buying & ROI
  • Software Engineering
    • Software Engineering
    • Software Architecture
    • General Software
    • Development Practices
    • Productivity & Workflow
    • Code
    • Engineering Management
    • Business of Software
    • Code
    • Digital Transformation
    • Systems Thinking
    • Technical Implementation
  • About
CMO & CTO

Closing the Bridge Between Marketing and Technology, By Luis Fernandez

Ab testing without drama

Posted on September 2, 2016 By Luis Fernandez

AB testing without drama sounds like a fantasy on some teams. Product wants to ship fast. Marketing wants that win on the dashboard. Data folks keep saying the words sample size and significance while everyone checks the experiment every hour like it is a soccer match. Let’s dial it down. This is about perspective, decisions, and practical tradeoffs so we get answers we trust and keep shipping work that matters.

Problem framing

Right now we have great tools. Optimizely and VWO are steady. Google Analytics Content Experiments is good enough for simple splits. Mobile teams are rolling feature flags to stage changes. With all that, drama sneaks in because speed fights with certainty. The boss wants the lift today. The data needs a week or two. People peek at the dashboard and call it. Then the next month the lift is gone and trust drops.

AB testing is not about hunting for a magic color on a button. It is about reducing risk on real decisions. You still need a clear question, a primary metric, a plan for timing, and guardrails so you do not break what already works. You also need to accept that sometimes the answer is we did not learn enough. That is not failure. That is a nudge to try a bolder variant or gather more traffic.

Two common sources of pain: peeking and scope creep. Peeking is the habit of watching p values swing and declaring victory on a good day. Scope creep is changing copy or traffic mid run. Both wreck the math and the trust. If we want less drama, we decide up front what we will measure, how long we will run, and what it takes to ship.

Patterns and antipatterns

Patterns that keep teams calm:

  • Pick one primary metric. Example: conversion rate to purchase or qualified lead. Keep it simple. Secondary metrics are for context, not for changing the verdict.
  • Set guardrails. Track revenue per session, average order value, bounce, and error rate. If the primary metric is up but a guardrail craters, do not ship.
  • Decide your minimum detectable lift before you start. If you only care about five percent or more, say it. Small lifts need big samples. Be honest about your traffic.
  • Plan duration. Run through at least one full business cycle so you cover weekday and weekend. Write the stop date in the test card and stick to it.
  • Freeze the variant. No edits during the run. If you must fix a bug, log it and reset the clock.
  • Segment after you decide. Make the call on the full target first. Then inspect mobile, desktop, new, returning. This avoids cherry picking.
  • Keep a changelog. One page with test name, start, stop, audience, primary metric, winner, link to data. It beats searching email and keeps the team aligned.
  • Use holdouts in production. Ship the winner to ninety five percent and keep five percent as a control for a few days. This catches seasonality and novelty spikes.
  • Lean on known stats choices. Optimizely Stats Engine is built to control false positives while you look. If you are on a simpler stack, resist the urge to refresh all day.

Antipatterns that crank up the drama:

  • Peeking and flipping the call. Today A wins, tomorrow B wins, the next day someone moves the target. That is not learning. That is noise.
  • Declaring on 90 percent and calling it a day. If you pick a threshold, stick with it. Do not drop the bar when the clock runs out.
  • Running five tests on the same audience at the same time. Cross talk can mask or fake effects. Stagger them or split audiences cleanly.
  • Mixing traffic allocation during the run. Going from 50 to 90 to push a winner early changes the math. If you must ramp, plan it and record it.
  • Ignoring device splits. Many sites are majority mobile now. A clean win on desktop can be a loss on phones. Always check the big device groups.
  • No QA. Shipping a broken variant and calling the result is worse than no test. Run through key flows on real devices before launch.

Case vignette

A mid size retail site wanted to raise checkout starts. The team had about eighty thousand sessions per week to the product page. They proposed a new layout with a larger primary button, fewer distractions, and clearer shipping copy. The primary metric was checkout initiation rate. Guardrails were revenue per session and support tickets tagged with checkout.

They set a target of a three percent lift and wrote a two week plan to cover two weekends. On day three the dashboard showed a big jump. The CMO pinged the team to ship. The data lead said no peeking call and pointed to the plan. By day ten, mobile was up seven percent, desktop was down two percent. On day fourteen, the split held. The team shipped the new layout to mobile only and kept a five percent holdout for four more days. Revenue per session stayed flat. Support tickets stayed flat. The change went live for all mobile traffic. Desktop got a lighter tweak and a new test card.

Result: clean lift on mobile, no damage on guardrails, less drama. The CMO got a clear story for the next board deck. The product team kept trust to run the next test without Slack fires.

Lessons learned

  • Pick bold variants. Small tweaks need huge samples. If your traffic is limited, test clearer value, shorter flows, or different framing.
  • Write the rules before you start. Primary metric, threshold, duration. Lock them in the test card so you do not move the goalposts.
  • Use guardrails to avoid winning the wrong game. Lifts that hurt revenue or quality are not wins.
  • Plan device outcomes. Be ready to ship per device. That can turn a mixed result into a real gain.
  • Keep a small holdout after you ship. It de risks seasonality, ad bursts, and novelty spikes.
  • Share results fast. One page with the story and the numbers kills drama and cuts the urge to relitigate the call.

AB testing is a habit. Keep the plan simple, keep the team honest, and keep shipping. Do that and the test tool becomes a quiet partner, not a source of late night emergencies.

Analytics & Measurement Digital Experience Experience Strategy Marketing Technologies Ab TestingConversion OptimizationmetricsMultivariate Testing

Post navigation

Previous post
Next post
  • Digital Experience (94)
    • Experience Strategy (19)
    • Experience-Driven Commerce (5)
    • Multi-Channel Experience (9)
    • Personalization & Targeting (21)
    • SEO & Performance (10)
  • Marketing Technologies (92)
    • Analytics & Measurement (14)
    • Content Management Systems (45)
    • Customer Data Platforms (4)
    • Digital Asset Management (8)
    • Marketing Automation (6)
    • MarTech Stack & Strategy (10)
    • Technology Buying & ROI (3)
  • Software Engineering (310)
    • Business of Software (20)
    • Code (30)
    • Development Practices (52)
    • Digital Transformation (21)
    • Engineering Management (25)
    • General Software (82)
    • Productivity & Workflow (30)
    • Software Architecture (85)
    • Technical Implementation (23)
  • 2025 (12)
  • 2024 (8)
  • 2023 (18)
  • 2022 (13)
  • 2021 (3)
  • 2020 (8)
  • 2019 (8)
  • 2018 (23)
  • 2017 (17)
  • 2016 (40)
  • 2015 (37)
  • 2014 (25)
  • 2013 (28)
  • 2012 (24)
  • 2011 (30)
  • 2010 (42)
  • 2009 (25)
  • 2008 (13)
  • 2007 (33)
  • 2006 (26)

Ab Testing Adobe Adobe Analytics Adobe Target AEM agile-methodologies Analytics architecture-patterns CDP CMS coding-practices content-marketing Content Supply Chain Conversion Optimization Core Web Vitals customer-education Customer Data Platform Customer Experience Customer Journey DAM Data Layer Data Unification documentation DXP Individualization java Martech metrics mobile-development Mobile First Multichannel Omnichannel Personalization product-strategy project-management Responsive Design Search Engine Optimization Segmentation seo spring Targeting Tracking user-experience User Journey web-development

©2025 CMO & CTO | WordPress Theme by SuperbThemes