The Best Practice That Was Not

Somewhere right now, a product team is implementing a design change because a blog post said it was a best practice. Bigger buttons. Shorter forms. Social proof near the call to action. Green instead of red. Single-column layouts.

These recommendations come with impressive-sounding case studies. One company increased conversion by a dramatic amount just by changing their button color. Another saw enormous gains from reducing form fields.

So the team implements the best practice, runs an A/B test, and gets a flat or negative result.

This is not a bug in the experimentation process. It is a feature of how context shapes human behavior. And understanding why best practices fail is one of the most important lessons in applied experimentation.

Why Universal Rules Do Not Exist in Optimization

The fundamental problem with best practices is that they are extracted from one context and applied to another. The company that improved conversion by changing a button was operating in a specific environment: a particular product category, a specific audience demographic, a unique traffic source mix, an existing design that the change interacted with, and a moment in time with particular competitive dynamics.

Change any of those variables and the result might reverse.

This is the external validity problem from research methodology. A finding that is true in one setting does not automatically generalize to other settings. Behavioral science has repeatedly demonstrated that even robust psychological effects vary dramatically across cultures, contexts, and demographics.

The Context Variables That Override Best Practices

Audience Sophistication

What works for a mass-market consumer audience often fails for a sophisticated B2B audience, and vice versa.

A prominent "Buy Now" button with urgency language performs well on impulse-purchase consumer sites. The same approach on a B2B SaaS site serving enterprise procurement teams signals desperation and erodes trust. These buyers have long decision cycles, involve multiple stakeholders, and respond to credibility signals, not urgency cues.

Conversely, the detailed technical specifications and case studies that drive enterprise sales can overwhelm and confuse consumer audiences looking for a quick, emotional decision.

The elaboration likelihood model from psychology explains this: high-involvement decisions are influenced by argument quality (central route), while low-involvement decisions are influenced by surface cues (peripheral route). Best practices that target the wrong processing route for your audience will fail.

Product Category Norms

Every product category has established conventions that visitors expect. Deviating from these norms — even when the deviation is objectively better — can reduce trust and increase cognitive load.

Luxury brands that follow e-commerce best practices (prominent pricing, urgency timers, comparison charts) undermine the exclusivity perception that drives their value proposition. Budget brands that follow luxury conventions (minimalist layouts, hidden pricing, curated experiences) frustrate price-sensitive shoppers who want information fast.

Category norms are not arbitrary. They evolved because they match the decision-making process of the category's buyers. Cognitive fluency — the ease with which information is processed — is highest when experiences match expectations.

Existing Design Context

Best practices describe a destination, not a journey. The impact of any change depends on what you are changing from.

Adding social proof to a page that already has five forms of social proof will produce a different result than adding it to a page with none. Shortening a form that has twenty fields will produce a larger effect than shortening one that has five. The same change applied to different starting points produces different outcomes because the marginal value of each improvement depends on the current state.

This is diminishing marginal returns in action. The first trust signal has more impact than the fifth. The first step removed from a form has more impact than the sixth.

Traffic Source and Intent

Visitors arriving from different channels carry different levels of awareness, intent, and trust. A best practice calibrated for one traffic profile may fail completely for another.

  • Paid search visitors have high intent and low patience — they searched for something specific and expect to find it immediately
  • Social media visitors have low intent and high distractibility — they were not looking for you and can easily return to their feed
  • Email subscribers have existing trust and moderate intent — they already know your brand and responded to a specific message
  • Organic search visitors vary widely but typically have informational needs that must be satisfied before they consider converting

A landing page optimized for paid search traffic (fast, direct, minimal distraction) may perform terribly when organic search traffic lands on it (needs context, explanation, and credibility building).

Competitive Landscape

Best practices assume a vacuum. Real optimization happens in a competitive context where what your competitors do affects how visitors perceive and respond to your experience.

If every competitor in your space uses a free trial model, your best practice of requiring a credit card at signup creates friction that your competitors do not impose. The best practice might be correct in isolation but wrong in context.

This is the reference dependence principle from behavioral economics. People do not evaluate your experience in absolute terms. They evaluate it relative to their alternatives and expectations.

The Pattern of Failed Best Practices

Certain best practices fail more frequently than others because they are most context-dependent.

"Remove form fields to increase completion." Fails when the removed fields served a qualifying function, when users equate thoroughness with seriousness, or when the form is already minimal.

"Add social proof near the conversion point." Fails when the social proof is generic, when the audience is skeptical of testimonials, or when the proof contradicts the visitor's experience.

"Use urgency and scarcity cues." Fails with sophisticated audiences who recognize manufactured urgency, in categories where rushed decisions are punished, or when overused to the point where visitors have developed immunity.

"Simplify the page to reduce distraction." Fails when visitors need information to build confidence, when the product is complex and requires explanation, or when the simplification removes trust signals.

"Make the CTA button bigger and more prominent." Fails when the button was already visible, when the problem is motivation rather than findability, or when the prominent button creates pressure that increases abandonment.

How to Build Context-Specific Hypotheses

Instead of applying best practices directly, use them as starting points for hypotheses that account for your specific context.

Step 1: Identify the Principle Behind the Practice

Every best practice is based on a behavioral principle. "Remove form fields" is based on the principle that friction reduces completion. "Add social proof" is based on the principle that people follow others' behavior.

Extracting the principle lets you apply it in a way that fits your context rather than blindly copying the implementation.

Step 2: Map the Principle to Your Audience

Ask: does this behavioral principle apply to my specific audience in my specific context? People who exhibit strong herding behavior will respond to social proof. People who pride themselves on independent thinking may react against it.

Step 3: Diagnose Before You Prescribe

Before implementing any change, verify that the problem exists. If your form completion rate is already high, reducing fields will not produce a meaningful lift. If visitors are not reaching your CTA, making it bigger will not help — the problem is upstream.

Use analytics data, session recordings, and user research to identify the actual friction points. Then design treatments that address those specific points.

Step 4: Test, Do Not Assume

The only way to know whether a change works in your context is to test it. The entire purpose of experimentation is to replace assumption with evidence.

This might sound obvious, but teams routinely skip testing for changes labeled as best practices because they believe the outcome is predetermined. The data consistently shows otherwise.

Frequently Asked Questions

If best practices are unreliable, how do I decide what to test?

Use your own data as the primary source of test ideas. Analytics identify where visitors drop off. Session recordings show what confuses them. Surveys reveal what they need. Best practices become useful when they offer potential solutions to problems you have already identified in your data.

Should I ignore all industry benchmarks and case studies?

No. Case studies and benchmarks provide useful hypotheses and directional guidance. The mistake is treating them as conclusions rather than starting points. Read them for the underlying behavioral principle, not the specific implementation.

Why do some best practices work for competitors but not for me?

Because your context differs in ways that matter. Different audience demographics, different traffic sources, different product complexity, different brand perception, and different starting-point designs all influence how a change performs. Your competitor's win is a hypothesis for you, not a guarantee.

How do I build institutional knowledge about what works for my specific audience?

Maintain a detailed experiment catalog that documents not just outcomes but the context variables: audience segments, traffic sources, current design state, and competitive landscape at the time of the test. Over time, patterns emerge that are far more reliable than generic best practices because they are grounded in your specific reality.

Share this article
LinkedIn (opens in new tab) X / Twitter (opens in new tab)
Written by Atticus Li

Revenue & experimentation leader — behavioral economics, CRO, and AI. CXL & Mindworx certified. $30M+ in verified impact.