Experiment Bet Sizing Using Revenue Per Session (RPS)
If you're running experiments under pressure, the hardest part isn't ideas. It's bet sizing: deciding how big a bet to place based on expected value, and how much traffic to risk.
Practical A/B testing frameworks, behavioral science, and conversion optimization — for growth leaders responsible for revenue.
If you're running experiments under pressure, the hardest part isn't ideas. It's bet sizing: deciding how big a bet to place based on expected value, and how much traffic to risk.
If you've ever run a quick test without an experiment brief template that somehow turned into six weeks of meetings, rework, and second-guessing, you're not alone.
If stakeholders keep rewriting your experiment doc, it's not because they're picky. It's because your brief doesn't answer the questions they get judged on.
Most teams don't fail because they ship nothing. They fail because they ship a lot of work that never moves the numbers, incurring shipping costs from unsuccessful features. When I'm under pressure, the trap is simple: I treat "a good idea" as "a shippable idea," blind to the complexities akin to in
If you run home pregnancy tests frequently after a suspected conception, you'll feel the temptation: the result looks promising on day three, excitement is building, and you want the confirmation. Or the opposite, the home pregnancy test result is negative, and you want to pull the plug before you "waste" more on excessive testing frequency.
If you run A/B testing on real revenue flows, variance isn't a stats problem. It's a cash problem. Every extra week you wait for confidence is a week you keep a worse checkout, a weaker onboarding, or a lower-priced plan. That slows Decision making, and it quietly taxes your growth strategy. The CUPED method is one of the few techniques that can shorten that wait.
Stakeholder pressure in business strategy doesn't break your metric tree because people are unreasonable. It breaks because the tree isn't tied to a decision anyone is willing to defend. I've been in the room when revenue misses, the board wants answers, and every exec grabs the nearest metric to justify their plan.
If your A/B test says "+6% conversion," the first question I ask to ensure trustworthy results isn't "Is it significant?" It's "Did you actually perform random allocation the way you think you did?" Sample ratio mismatch (SRM) is the quiet failure mode that turns clean-looking results into expensive mistakes.
When my experiment backlog gets long, my decision quality drops fast. Everything looks “important,” every stakeholder has a favorite, and the loudest idea starts to win. That’s when I fall back on the expected value framework. Not because it’s fancy, but because it forces one thing: dollars first, o
Most teams don’t get burned by a bad idea, they get burned by a good idea with hidden damage. That’s why experiment guardrails matter. In A/B testing, you’re not only asking about primary success metrics like “Did conversion go up?”, you’re also asking about unintended consequences: “Did we quietly
If you’re a product manager and your experiment roadmap isn’t tied to revenue growth, it turns into a list of “interesting” tests that never earn their keep. I’ve watched teams run months of A/B testing, learn a few things, and still miss the quarter because nothing connected back to dollars. The fi
If your team runs experimentation, you already know the ugly part: the results meeting turns into a debate about which metric “matters.” Someone points at conversion. Someone else points at retention. Finance wants revenue. Product wants engagement. When you don’t have a single North Star Metric, ev
If your experiment backlog is full but your learning feels thin, it’s usually not a testing problem. It’s a memory problem. Teams run dozens of tests, then six months later no one can find what happened, why it happened, or whether it’s safe to try again. A solid ab test repository fixes that, but o
If your team runs enough tests, you eventually hit the same frustrating problem: two “Checkout CTA” experiments, three different names, and nobody can tell which result was real. It’s like trying to run a library where books don’t have ISBNs. A scalable experiment ID system fixes that by giving ever
If your experimentation program feels busy but not productive, the problem often isn’t idea volume. It’s flow. Tests get created, half-built, re-prioritized, and then quietly die in a backlog, a spreadsheet tab, or someone’s memory. A well-run A/B test repository fixes that by treating experiments l
If your A/B test history lives in Notion, you’ve probably felt the pain. Tests get logged, but results are hard to compare. Metrics drift. People rename fields. Old pages turn into dead ends no one trusts. A real experiment library fixes that, but the move can get messy fast. Not because the export
If your team runs a lot of experiments, you’ve felt the pain: the results live in someone’s spreadsheet, the “why” is buried in a Jira ticket, and the final decision is in a Slack thread that no one can find later. Everyone moves fast, but learning moves slow. A solid A/B test repository fixes the [
If a new PM asks, “Have we tested trust badges in checkout?”, the answer shouldn’t be a 30-minute Slack archaeology session. It should be a quick search, a clear summary, and links to the original assets, data, and decision. That’s what an experiment library taxonomy is for. It turns messy, one-off
If your team has more than a handful of testers, duplicates don’t show up as one obvious mistake. They show up as slow bleed, the same “new idea” getting shipped again with a slightly different headline, a different Jira ticket, and no memory of why it failed last time. That’s why experiment naming
Spreadsheets are the duct tape of experimentation ops. When a program is young, a single Google Sheet can feel like a perfect source of truth. Everyone can edit it, it’s searchable enough, and it’s “good for now”. Then “now” becomes six months, the team triples, and someone asks a simple question: H
If your experimentation program is growing, your biggest risk isn’t running fewer tests. It’s repeating work you already paid for, forgetting why something worked, and losing the confidence to act on results. That’s why a real A/B test repository matters. Not a folder of screenshots. Not a “Tests” s
An ROI calculator can be your best “middle-of-funnel closer”… or a silent leak that turns high-intent visitors into bounce traffic. Most teams focus on the math, then wonder why demo requests don’t move. In practice, conversion is usually won or lost in three places: how many inputs you ask for, wha
Your top navigation is the set of street signs on your website. When the signs are clear, buyers keep moving. When they’re vague or crowded, they stop, hesitate, and bounce. In 2026 B2B SaaS buying, that hesitation costs more than it used to. Prospects arrive with opinions, they skim fast, and they
Most teams treat their app marketplace listing like a one-time launch task. Write a description, upload a few screenshots, hit publish, move on. That’s how you end up with “nice traffic” and no pipeline. Marketplace visitors are already in a buying mood. They’re comparing options, checking trust sig
Practical A/B testing frameworks, behavioral science, and CRO strategies for growth leaders responsible for revenue. Practical. Free. Weekly.
Free · No spam · Unsubscribe anytime