← Back to News

The Framing Effect: How You Ask Changes What People Choose

coin

In the early 1980s, psychologists Amos Tversky and Daniel Kahneman presented people with a scenario:

Imagine the U.S. is preparing for an outbreak of an unusual disease expected to kill 600 people. Two programs are proposed:

Positive frame (gain):

  • Program A: 200 people will be saved.

  • Program B: 1/3 probability that 600 people will be saved, 2/3 probability that no one will be saved.

72% chose Program A. The sure thing.

Negative frame (loss):

  • Program C: 400 people will die.

  • Program D: 1/3 probability that nobody will die, 2/3 probability that 600 people will die.

78% chose Program D. The gamble.

A and C are identical. So are B and D. The only difference is whether outcomes are described as lives saved or lives lost.

When framed as gains (lives saved), people are risk-averse. They take the sure thing.

When framed as losses (lives lost), people are risk-seeking. They gamble to avoid the sure loss.

Same math. Different frames. Opposite choices.

How Framing Works

Framing effects operate through several mechanisms:

Reference Point Shifting

People don't evaluate options in absolute terms. They evaluate them relative to a reference point.

"Save $50" feels different than "avoid losing $50," even though both result in the same $50 in your pocket.

The frame sets the reference point. Gain frames anchor you at zero, making the gain feel like a bonus. Loss frames anchor you at having something, making the potential loss feel painful.

Loss Aversion

Losses loom larger than equivalent gains. Losing $100 feels roughly twice as bad as gaining $100 feels good.

This asymmetry means loss-framed messages can be more motivating—but also more anxiety-inducing. "Don't miss out" hits harder than "get this benefit," even when they describe the same outcome.

Attribute Highlighting

The way you frame information determines which attributes people attend to.

"90% fat-free" makes you think about the lean meat. "10% fat" makes you think about the fat. Your brain fixates on the number that's presented, not the complement.

Same percentage, different focus, different appeal.

Why This Wrecks Research

Framing effects aren't just interesting psychology. They're a practical nightmare for anyone trying to test messages, pricing, or product positioning.

The Pricing Research Trap

You're testing price sensitivity for a subscription service. You ask:

Frame 1: "How much would you be willing to pay per month?"

Frame 2: "Our service costs $X per month. How likely are you to subscribe?"

Frame 3: "For less than $Y per day, you get [benefits]. How appealing is this?"

Three framings. Three different numbers. Which one is "true"?

The monthly frame anchors people to monthly spending. The daily frame makes the cost seem trivial. The benefits-first frame primes value before price.

All three are measuring "willingness to pay," but they're measuring it through different psychological lenses. Your pricing strategy will be radically different depending on which frame you test.

Message Testing's Hidden Variable

A financial services company wants to test messaging for a retirement product. They develop two concepts:

Message A (gain frame): "Start saving today and enjoy a comfortable retirement."

Message B (loss frame): "Don't risk running out of money in retirement. Start saving today."

Message B tests better. They run with it.

But what if they'd tested:

Message C (gain frame, concrete): "Save now and retire with an extra $200,000."

Message D (loss frame, concrete): "Without savings, you could face a retirement shortfall of $200,000."

Would the loss frame still win? Or does specificity override frame?

You don't know unless you test both the frame and the specificity. That's four messages, not two. Four cells, four sample requirements, twice the cost.

And that's just two dimensions. What about:

  • Personal vs. social framing

  • Immediate vs. future framing

  • Emotional vs. rational framing

  • Absolute vs. relative framing

Each dimension multiplies your study design. Each combination requires sample size. Costs explode.

The Feature Positioning Problem

You're launching a product feature. You could frame it as:

  1. What you gain: "Get 2x faster processing speed."

  2. What you avoid: "Stop wasting time on slow processing."

  3. What you keep: "Preserve your productivity with faster processing."

  4. Comparison gain: "Process files 2x faster than the competition."

  5. Comparison loss: "Don't fall behind competitors with slow processing."

Five framings. All describing the same feature. All likely to produce different purchase intent scores.

Traditional research would pick one, maybe two framings to test. But then you're making a multi-million dollar bet that you happened to test the right frame.

The Multiplier Problem

Framing isn't a simple A/B choice. It's a multi-dimensional space:

  • Gain vs. loss

  • Concrete vs. abstract

  • Emotional vs. rational

  • Personal vs. social

  • Immediate vs. delayed

  • Absolute vs. relative

  • Simple vs. detailed

If you want to test just three framings across four product benefits, that's 12 versions. At 400 responses per cell for statistical power, you need 4,800 completes.

At $5 per complete (a reasonable blended rate), that's $24,000. For one round of testing. Before optimization.

Most research budgets can't absorb that. So companies test one or two framings and hope they picked the right ones.

Documented Disasters

The 93% Fat-Free Study

Research on food labeling consistently shows that "X% fat-free" dramatically outperforms "Y% fat" in purchase intent and taste perception, even when people can do the math.

In one study, ground beef labeled "75% lean" was rated as less greasy, better tasting, and higher quality than identical beef labeled "25% fat." Same beef. Different labels. Different sensory experience.

The food industry has known this for decades. That's why you rarely see "10% fat" labels. Everyone uses the gain frame.

The Surgery Framing Effect

Doctors presented patients with surgery options framed two ways:

Survival frame: "Of 100 people who have this surgery, 90 are alive after five years."

Mortality frame: "Of 100 people who have this surgery, 10 are dead after five years."

Same statistic. When framed as survival, 84% of patients chose surgery. When framed as mortality, only 56% chose surgery.

28 percentage point swing from frame alone.

This isn't hypothetical. This affects real medical decisions. Patients who would benefit from surgery decline it because of how the risk is framed.

Why Traditional Solutions Fall Short

Testing Multiple Frames

The obvious answer: test all the frames.

The problem: sample size and budget explode. If you're testing five framings with 400 respondents each, you need 2,000 completes. If you're testing those five framings across three audience segments, you need 6,000 completes.

At $5 per complete, that's $30,000. Before you've tested any actual message variants, pricing tiers, or creative executions.

Sequential Testing

Another approach: test a few framings, pick a winner, then optimize within that frame.

The problem: you're committing to a frame before you know if it's the right one. If the "gain frame" wins your initial test but only because you wrote a weak loss-framed version, you've locked yourself into a sub-optimal path.

You also can't go back. Once you've spent the budget on frame testing, there's no money left to revisit the choice.

Expert Judgment

Many researchers just pick a frame based on category norms or intuition.

"Everyone in financial services uses loss framing for retirement, so we will too."

The problem: category norms might be wrong. Just because everyone does it doesn't mean it's optimal. Maybe everyone is copying everyone else's unvalidated choice.

The Hidden Interaction Effects

Framing doesn't exist in isolation. It interacts with:

Audience characteristics: Loss framing might work for risk-averse audiences but backfire with risk-seeking ones.

Product category: Gain framing might dominate in aspirational categories (luxury, wellness) while loss framing wins in protection categories (insurance, security).

Purchase stage: Early awareness might respond to gain frames, while late-stage conversion responds to loss frames.

Cultural context: Some cultures are more responsive to collectivist frames ("help your community") while others respond to individualist frames ("achieve your goals").

You can't know these interactions without testing. But testing all the combinations is prohibitively expensive.

A Different Approach: Rapid Frame Testing

What if you could test frame sensitivity during message development, not after?

Run the same core message in five different frames on synthetic respondents. See which frames are stable and which ones produce wild swings.

If gain and loss frames produce nearly identical results, you know framing isn't your key variable—focus on other message elements.

If gain frames outperform loss frames by 30 points, you know frame is critical. Invest in optimizing the gain-framed version.

If results vary by audience segment (loss frames win with older audiences, gain frames win with younger), you know you need segmented messaging.

This isn't theoretical. With synthetic respondents, you can test 10 framings across 5 audience segments in an afternoon. 50 cells, 500 responses each. Total cost: a fraction of one traditional framing test.

You're not replacing your final validation study. You're narrowing the hypothesis space before you spend real money.

Design Principles That Help

While you can't eliminate framing sensitivity, you can design research that accounts for it:

1. Test the Frame, Not Just the Message

Don't ask "Which message is better?" Ask "Which frame is more effective for this message?"

If your two test messages use different frames, you don't know if you're measuring message quality or frame sensitivity.

Control for frame, then optimize within it.

2. Match Frame to Decision Context

Think about how people will encounter your message in the real world.

If they're comparison shopping, they're in an analytical mindset. Rational frames and concrete numbers will resonate.

If they're scrolling social media, they're in a fast-think mode. Emotional frames and simple contrasts will land better.

Test frames that match the decision context, not just frames that test well in a sterile survey.

3. Test Extreme Versions First

Don't start by testing subtle variations. Start with the most different frames:

  • Pure gain vs. pure loss

  • Highly concrete vs. highly abstract

  • Deeply emotional vs. purely rational

If extreme versions show no difference, subtle variations won't either. You've saved yourself a lot of testing.

If extreme versions show big differences, you know framing matters and you can zero in on the optimal zone.

4. Separate Frame From Execution

Framing is conceptual. Execution is creative.

"Don't miss out" (loss frame) could be executed as:

  • A calm, informative message

  • An urgent, scarcity-driven message

  • A supportive, helpful message

If you test poorly-executed loss frames against well-executed gain frames, you'll conclude loss frames don't work. But maybe you just wrote bad loss-frame copy.

Test frames in neutral, equivalent executions first. Optimize execution after you've picked a frame.

When Framing Effects Actually Matter

Not every message decision requires frame testing. The stakes matter.

High stakes:

  • Brand taglines and positioning statements that will run for years

  • High-investment ad campaigns

  • Pricing and packaging that shape willingness to pay

  • Public health messaging where uptake drives outcomes

Lower stakes:

  • Transactional emails where the frame is dictated by function

  • Internal communications where clarity matters more than persuasion

  • Content marketing where multiple framings can co-exist

If you're spending $5M on a campaign, spending $50K to validate the frame is cheap insurance.

If you're writing a one-off blog post, pick a frame and move on.

The Real Cost of Guessing

Most messaging research doesn't test framing. It picks a frame—usually based on category norms or intuition—and optimizes within that frame.

But if you picked the wrong frame, you're optimizing the wrong thing.

A brilliantly-executed loss-framed message might still underperform a mediocre gain-framed message if your audience is loss-averse.

You're leaving performance on the table. Maybe a lot of it.

The traditional solution—test all the frames with large samples—is correct but unaffordable.

But the alternative—picking a frame and hoping—isn't acceptable for high-stakes decisions.

Testing as a Filter, Not a Final Answer

The insight: framing sensitivity isn't something you discover in your final validation study.

It's something you map during development, when it's cheap to pivot.

Run a quick frame sensitivity test early. Learn which frames are promising. Kill the ones that aren't. Invest your validation budget in optimizing the frames that matter.

You're not eliminating the need for traditional research. You're making it more efficient by testing fewer, better-chosen options.

Traditional research tells you which message won.

Better research tells you which frame wins, why it wins, and how stable that win is across audiences and contexts.

Your message isn't neutral. How you frame it determines what people hear, what they feel, and what they choose.

Knowing that before you commit—rather than discovering it after you launch—is the difference between strategic messaging and expensive guessing.

Andreas Duess

About the author

Andreas Duess

Andreas Duess builds AI tools that turn consumer behavior into fast, usable signal. He started his career in London, UK, working with Cisco Systems, Sony, and Autonomy before co-founding and scaling Canada’s largest independent communications agency focused on food and drink.

After exiting the agency, Andreas co-founded ditto, born from a clear gap he saw firsthand: teams needed faster, more accurate ways to map consumer behavior and pressure-test decisions before committing time and capital.

Related Articles


Ready to Experience Synthetic Persona Intelligence?

See how population-true synthetic personas can transform your market research and strategic decision-making.

Book a Demo

Ditto Newsletter - Subscribe

Get insights that don't slow you down. Research updates, case studies, and market intelligence—delivered monthly.