How Not To Fail At A Website Redesign
Your most important pages—product listings, pricing grids, checkout flows—are also your riskiest to change. They're dense, high-pressure, and layered with CX decisions.
Yet when performance dips, what do most teams do? Burn it down and redesign.
Teams collect qualitative feedback, skim disconnected analytics, launch a “Version B,” and hope an A/B test gives it the green light.
But here’s the problem: guesswork wrapped in good intentions isn’t optimization.

🚫 Stop Betting on Redesigns
Most redesign failures follow a familiar pattern:
1
Collect qualitative feedback
(“Users said the page feels cluttered.”)
2
Make sweeping visual changes
(“Let’s simplify everything!”)
3
Validate with a binary A/B test
(“Did it win or lose?”)
Here's an example
Industry: Media
In this scenario, a media brand came to Evolv AI with a redesigned pricing page inspired by qualitative feedback looking to run a standard A/B test to validate.
The assumption: Sleeker layout = higher conversions
Validation method: A/B Test
The result: ~15% drop in conversions


What went wrong?
Like most teams, they:
- Built ideas with qualitative data
- Validated them with quantitative tests
- Never linked the two
And, the A/B test gave them a binary outcome with few insights on which changes hurt performance. Was it the layout? The CTA? The copy? Which were actually the right design choices? They had no visibility - starting them back at square one.
Your most important pages—product listings, pricing grids, checkout flows—are also your riskiest to change. They're dense, high-pressure, and layered with CX decisions.
Yet when performance dips, what do most teams do? Burn it down and redesign.
Teams collect qualitative feedback, skim disconnected analytics, launch a “Version B,” and hope an A/B test gives it the green light.
But here’s the problem: guesswork wrapped in good intentions isn’t optimization.
The Problem
Most redesign failures follow a familiar pattern:
- Collect qualitative feedback (“Users said the page feels cluttered.”)
- Make sweeping visual changes (“Let’s simplify everything!”)
- Validate with a binary A/B test (“Did it win or lose?”)
In the following example, we explore how Evolv AI would approach a media company.
- Scenario: A brand came to Evolv AI with a redesigned pricing page inspired by qualitative feedback looking to run a standard A/B test to validate.
- Assumption: Sleeker layout = higher conversions.
- Validation Method: A/B Test
- Result: –15% drop in conversion drop.
What went wrong?
Like most teams, they:
- Built ideas with qualitative data
- Validated them with quantitative tests
- Never linked the two
And, the A/B test gave them a binary outcome with few insights on which changes hurt performance. Was it the layout? The CTA? The copy? Which were actually the right design choices? They had no visibility - starting them back at square one.
How Evolv AI Responded
Step 1: Deconstruct the Redesign
Break the pricing page experience into modular UX “blades”—like CTA, pricing section, and social proof placement.
Step 2: Unify All Relevant Data
- The original qualitative insights
- The results from the A/B test
- Business intelligence from Evolv AI, Adobe Experience Manager, GA4, and Heap
Step 3: Form Hypotheses per Blade
Using the synthesized insights to assess each blade, Evolv AI’s platform and UX strategists developed testable ideas related to:
- CTA wording and color
- Pricing section placement
- Plan comparison format (grid vs. accordion)
- Copy tone and length
- Visual type: icons vs. text
- Placement of social proof and trust badges
From Static A/B to Dynamic MVT
The New MVT Strategy
Shifting the customer away from A/B testing, we opted for a multivariate experiment:
- 12 key UX elements
- 23 variants per element
With Evolv AI’s Active Learning engine: (alt: Application of Active Learning)
- Variants were dynamically served in real-time
- Winning combinations of variants were surfaced automatically
- Losing variants were phased out
- Mid-test changes occurred without restarting
The outcome
- A sustained 5–6% relative lift vs -15% drop
- Clear attribution for every UX element
- Ability to reuse insights across other key touchpoints in the journey
- Deeper understanding of what resonates—enabling smarter, targeted testing
- Company-wide shift toward iterative, data-first optimization
The Redesign Optimization Playbook
Here’s how you can apply this approach to your next redesign:
1. Right-Size Your Experimentation Strategy
- Test volume should match your traffic capacity.
- Too many variants on low traffic = noise.
- Too few on high traffic = missed opportunity.
2. Expect Initialized Volatility
Early data will swing. That’s not failure. That’s signal. Evolv AI’s models dynamically reweight experiences to stabilize insights faster.
3. Pair qualitative and quantitative data together
Most teams use qualitative insights to ideate, and quantitative to validate. But without linking them, you never truly know why something worked.
With Evolv AI, your voice-of-customer data is embedded into hypothesis generation and prioritized for testing.
The Assumption
Sleeker layout = higher conversions
Validation Method
A/B test
The Result
~15% drop in conversions
How Evolv AI responded
1
Break Down the Experience
& Connect the Data
Break the experience into modular UX “blades”—like CTA, pricing, and social proof sections.
Evolv AI unifies qualitative research, A/B outcomes, and behavioral analytics—creating a single intelligence layer for smarter decisions.
2
Generate Blade-Level
Hypotheses That Matter
Evolv AI’s platform turns that intelligence into prioritized hypotheses for each blade.
By focusing on what matters most, the platform can rapidly surface the highest-impact ideas to test and optimize—automatically and at scale.
These hypotheses directly explore high-impact UX changes, such as:
1
Deconstruct the Redesign
Break the pricing page experience into modular UX “blades”—like CTA, pricing section, and social proof placement.
2
Unify All Relevant Data
Evolv AI synthesized the original qualitative insights, results from the A/B test, plus business intelligence from Evolv AI, Adobe Experience Manager, GA4, and Heap.
3
Form Hypotheses Per Blade
Using the synthesized insights to assess each blade, Evolv AI’s platform and UX strategists developed testable ideas related to:

CTA wording and color

Pricing section placement

Plan comparison layout

Content tone and length

Visual type: icons vs. text

Placement of social proof
AI-Led Experience Optimization
From Hypothesis to High-Impact
Evolv AI automates the heavy lifting of experimentation, learning in real time to continuously surface what works—so your team can focus on driving growth.
From Static A/B to Dynamic MVT
The New MVT Strategy
Shifting the customer away from A/B testing, we opted for a multivariate experiment: 12 key UX elements with 23 variants per element.

The outcome was overwhelmingly positive with Evolv AI’s Active Learning engine:
- Variants were dynamically served in real-time
- Winning combinations of variants were surfaced automatically
- Losing variants were phased out
- Mid-test changes occurred without restarting

Sustained 5–6% lift compared to a -15% decline

Every UX element clearly attributed to impact

Reusable insights applied across key touchpoints

Smarter, targeted testing from clearer resonance

Organization-wide shift to iterative, data-led optimization
The Redesign Optimization Playbook
Here’s how you can apply this approach to your next redesign:
1
Right-Size Your Experimentation Strategy
Match variant count to your traffic. Too many ideas with low traffic = noise. Too few with high traffic = missed potential.
Evolv AI dynamically calibrates variant complexity to traffic volume, ensuring statistical power without slowing velocity.
2
Expect Initialized Volatility
Early data shifts don’t mean failure—they’re insight in disguise. Lean into them to learn faster.
Evolv AI applies adaptive reweighting to stabilize early signals and accelerate insight discovery.
3
Pair Qualitative + Quantitative Insights
Don’t just ideate with one and validate with the other—connect both to understand why your ideas actually work.
Evolv AI combines VoC, analytics, and experimentation data to uncover why your highest-impact ideas succeed.
Resources
Company
2261 Market Street STE 22002 San Francisco, CA 94114 USA