Introduction: Why Traditional Surveys Fall Short
Understanding what truly matters to your customers is no longer optional—it’s essential. Yet businesses continue to rely on outdated survey methods that produce ambiguous results. When you ask respondents to rate ten product features on a 1-5 scale, the outcomes are predictable: everything gets marked as “important,” middle options blur together, and actionable insights remain elusive.
This is where MaxDiff (Maximum Difference Scaling), also known as best-worst scaling, changes the game entirely.
MaxDiff is a sophisticated yet surprisingly simple survey methodology that forces respondents to make real choices rather than expressing vague preferences. Rather than asking “How important is this feature?”—where most respondents answer “very important”—MaxDiff asks “Which of these features matters most to you, and which matters least?” This fundamental shift in approach generates data that accurately reflects what customers truly value, eliminating the noise that plagues traditional rating scales.
Developed in the 1990s and pioneered by Professor Jordan Louviere and his team at the Centre for the Study of Choice (University of Sydney), MaxDiff has become the gold standard for preference measurement across industries ranging from hospitality to healthcare, gaming to insurance, and product development to marketing strategy.¹
This comprehensive guide explores everything you need to know about MaxDiff: how it works, when to use it, real-world case studies demonstrating its power, and geographic insights for optimizing your research across different markets.
Section 1: Understanding MaxDiff – The Foundation
What Exactly Is MaxDiff?
MaxDiff is a statistical technique that creates robust rankings of items—whether product features, brand attributes, service benefits, or messaging claims—by asking survey respondents to repeatedly select the best and worst options from small subsets of alternatives.² Rather than rating all items individually, respondents evaluate them in carefully designed combinations, typically viewing 4-6 items per question across multiple question sets.
The elegance of MaxDiff lies in its simplicity for respondents combined with mathematical sophistication in analysis. Each choice respondents make contains information about their relative preferences. By accumulating these choices across multiple question sets (often 15-25 per respondent), researchers can calculate utility scores that rank items on a 0-100 scale, revealing not just which items people prefer, but by how much.³
How MaxDiff Differs From Traditional Methods
To understand MaxDiff’s superiority, consider how it compares to three common alternatives:
Rating Scales (Likert Scales, 1-5 Star Ratings): When respondents rate features individually on a numerical scale, they tend to anchor on certain numbers regardless of the actual item being evaluated. This “scale use bias” is particularly problematic in cross-cultural research, where different cultures employ rating scales differently—for example, respondents from certain countries may use mostly 4s and 5s, while others cluster around 3s.⁴ Additionally, respondents under time pressure or lacking motivation default to marking everything as “important.”
Ranking Questions: Asking respondents to rank 15+ items becomes cognitively overwhelming. Respondents can confidently order their top 2-3 and bottom 2-3 choices, but items in the middle become arbitrary, creating an “artificial flattening” of preference differences.
Constant-Sum Allocation: These questions ask respondents to distribute 100 points across items, but require significant cognitive effort and often yield unreliable data because respondents struggle to calculate proper allocations.
MaxDiff addresses all these limitations. By forcing binary choices (best vs. worst) from small subsets, respondents make discriminating decisions they can confidently answer. The methodology produces continuous data across the full range of items, revealing meaningful gaps in preference rather than clustering responses around middle values. Additionally, because respondents are choosing rather than rating, there’s no opportunity for scale use bias—a critical advantage for global research.⁵
The Psychology Behind Why MaxDiff Works
MaxDiff’s effectiveness stems from how it aligns with human decision-making. In real life, consumers don’t rate features independently; they compare and trade off. When choosing a smartphone, you don’t separately rate the camera, battery life, and design—you consider which matters most when features conflict.
MaxDiff replicates this natural decision process. By presenting small, manageable choice sets, it reduces cognitive load while maintaining decision realism. Respondents from children to adults with varying educational backgrounds can reliably complete MaxDiff tasks because the instruction—”select your best and worst”—is universally clear.⁶
Section 2: When to Use MaxDiff – The Strategic Framework
MaxDiff excels in specific scenarios where understanding priority and preference order is critical to business decisions.
Ideal Use Cases for MaxDiff
New Product Development (NPD): Companies launching new products or features must prioritize limited development resources. MaxDiff reveals which product attributes drive consumer interest most strongly, ensuring investment aligns with customer values.⁷ If you’re deciding whether to emphasize premium materials, extended warranty, or faster processing—MaxDiff tells you which resonates most with your target segment.
Messaging and Positioning: Marketing teams constantly debate which value proposition to emphasize. MaxDiff resolves these debates by testing multiple positioning messages with target consumers and identifying which claims generate the strongest preference. This is invaluable before campaign launch when changing messaging is impossible.
Feature Prioritization: When resources are limited and you can’t include every desired feature, MaxDiff shows which features (among perhaps 20+ options) will have the greatest impact on customer satisfaction and purchasing decisions. A resort renovation project can test 15 potential upgrades and allocate budget to the 5-6 that matter most to guests.⁸
Brand Attribute Assessment: Understanding how consumers view your brand compared to competitors requires more than simple preference questions. MaxDiff can evaluate which brand attributes are most distinctive and important in your category, informing brand positioning and marketing strategy.
Service Benefit Optimization: In industries with complex offerings—insurance, financial services, healthcare—MaxDiff helps identify which benefits drive customer choice and satisfaction, allowing organizations to emphasize high-value benefits in marketing and operational decisions.
Packaging and Design Testing: Before committing to expensive production runs, companies test multiple design concepts using MaxDiff to identify which visual and functional elements consumers value most.
Pricing and Value Communication: By pairing MaxDiff with pricing research, organizations understand which product attributes consumers will accept higher prices for, enabling premium pricing strategies aligned with perceived value.
When MaxDiff Is Not The Right Choice
While powerful, MaxDiff isn’t appropriate for every research question.
MaxDiff measures relative importance—how items rank against each other—but not absolute importance. A respondent might view all options in a MaxDiff exercise as uninspiring yet still must choose a best and worst. If you need to validate whether a feature concept generates genuine enthusiasm (rather than just comparing relative preferences), pair MaxDiff with monadic concept testing or qualitative research.
Additionally, MaxDiff reveals which items are preferred but doesn’t explain why. If deep understanding of consumer motivations is critical, MaxDiff works best when combined with follow-up qualitative interviews exploring the reasoning behind preference patterns.⁹
Section 3: How MaxDiff Works – Technical Deep Dive
The Experimental Design Foundation
MaxDiff’s power stems from careful experimental design that determines which items appear together in each question set. Rather than random combinations, researchers use balanced incomplete block design (BIBD) or partially balanced incomplete block design (P-BIBD) to ensure:
- Each item appears the same number of times across all respondents
- Each pair of items appears together equally often
- The design is statistically balanced yet efficient¹⁰
For example, if testing 12 product features, the experimental design might present each feature 4-6 times across 15-20 question sets, with different feature combinations in each set.
The Respondent Experience: A Step-by-Step Example
To illustrate how MaxDiff appears to survey respondents, consider a beverage company testing seltzer flavors: Lemon-Lime, Raspberry, Grapefruit, Mango, Blackberry, Strawberry, and Passion Fruit.
Question 1: “Which seltzer flavor would you MOST prefer, and which would you LEAST prefer from these options?”
Shown: Lemon-Lime | Raspberry | Grapefruit | Passion Fruit
Respondent selects: Most Preferred = Raspberry | Least Preferred = Passion Fruit
Question 2: (A few seconds later, with different combinations)
Shown: Mango | Blackberry | Grapefruit | Strawberry
Respondent selects: Most Preferred = Mango | Least Preferred = Grapefruit
This continues for 15-20 questions, with combinations strategically varying so respondents evaluate each flavor multiple times against different competitors.¹¹
Analysis: From Responses to Utility Scores
Once collected, MaxDiff responses are analyzed using statistical models, typically logit regression or hierarchical Bayes estimation. The mechanics involve a clever data transformation: each best/worst choice is converted to binary outcomes for each item. A choice of “A as best, B as worst” generates:
- A receives +1 (chosen as best)
- B receives -1 (chosen as worst)
- Other items receive 0
The model estimates coefficients for each item, which are then transformed into utility scores on a 0-100 scale that sum to 100. These scores represent the percentage probability an item would be chosen as best if all items were presented together.¹²
For the seltzer example, results might look like:
| Flavor | Utility Score |
|---|---|
| Mango | 18 |
| Raspberry | 16 |
| Strawberry | 15 |
| Lemon-Lime | 14 |
| Blackberry | 13 |
| Grapefruit | 12 |
| Passion Fruit | 12 |
These scores immediately show that Mango and Raspberry are significantly preferred (likely 13% more likely to be chosen as best compared to the least-preferred flavors), providing clear prioritization guidance.
Sample Size and Statistical Robustness
A common question: How many respondents do I need?
The formula for MaxDiff sample size depends on three variables: number of items (x), items shown per question set (n), and desired minimum frequency (r) that each item should appear:
r × x / (n × p) = s
Where p = sample size and s = question sets per respondent.
For research robustness, experts recommend each item appears at least 200 times across the entire dataset. If testing 30 items, showing 4 items per set, with 80 respondents, you’d need approximately 19 question sets per respondent.¹³
This scalability is remarkable—testing 100+ items (like Progressive Insurance’s benefits study) becomes feasible through sparse MaxDiff, where each item appears just once or twice per respondent.¹⁴
Section 4: Case Study 1 – Polynesian Cultural Center: Guest Experience Optimization
The Polynesian Cultural Center (PCC) in Laie, Hawaii, is a leading cultural tourism attraction that has welcomed millions of visitors over its 55-year history. As competition from theme parks and travel alternatives intensified, PCC leadership recognized they needed deeper understanding of guest preferences to optimize the visitor experience and maximize revenue.
The Challenge
PCC offered numerous experiences and add-ons: village shows, cultural demonstrations, dining options, gift bag items, marketplace vendors, and entrance packages. Management’s intuition differed about what mattered most to visitors. Some executives believed the cultural authenticity of village shows was paramount; others argued that convenience and value-for-money amenities drove satisfaction and repeat visits.
Without data-driven prioritization, PCC couldn’t confidently allocate limited budget to enhancements most likely to improve guest satisfaction and spending.
The MaxDiff Solution
PCC partnered with Sawtooth Software to conduct three separate MaxDiff studies addressing different strategic questions.
Study 1: Visitor Segmentation Objective: Identify distinct visitor archetypes and their preferences
PCC ranked 27 travel-related attributes and preferences (culture-focused items, convenience features, value indicators, etc.) using MaxDiff with a sample of 200+ past and prospective visitors. Analysis included latent class segmentation, revealing three distinct visitor personas:
- Cultural Immersion Seekers (35% of visitors): Prioritized authentic cultural experiences, village shows, expert-led cultural demonstrations, and learning opportunities. These visitors were less price-sensitive and valued depth over convenience.
- Value-Conscious Families (45% of visitors): Focused on overall value, family-friendly activities, good dining options, and reasonable pricing. This segment showed strong price sensitivity but high willingness to add entertainment amenities if perceived as good value.
- Quick-Stop Tourists (20% of visitors): Sought convenient, quick experiences without deep cultural engagement. This segment valued location, short visit duration, and quick food service.
This segmentation was crucial: it revealed that the single-experience strategy wasn’t optimal. Instead, PCC could tailor marketing, pricing, and offerings to different segments.
Study 2: Feature Value Assessment Objective: Align pricing with perceived value
Using MaxDiff, PCC tested perceived value for specific experiences and amenities. Visitors evaluated 12 different features (premium show seats, dining packages, cultural artifact souvenirs, meet-the-performers experiences, etc.) across multiple question sets.
Results indicated that “meet-the-performers” experiences scored significantly higher in value perception than basic admission alone. This finding supported premium pricing for these add-ons. Additionally, unique cultural artifacts rated higher in perceived value than mass-produced gifts, suggesting inventory reallocation from generic to culturally authentic items.
Study 3: Gift Bag Optimization Objective: Maximize perceived value in complimentary gift bags
PCC surveyed visitors about gift bag contents using MaxDiff, testing items like DVDs of performances, local bakery products, discount vouchers, and cultural guides.
The results were surprising: bakery coupons scored significantly higher in preference than performance DVDs. PCC swapped DVDs for bakery products in guest bags. The result? Perceived value doubled in post-visit surveys, and importantly, visitors used the bakery coupons, increasing retail spending. This seemingly small change generated meaningful incremental revenue.
Additional Applications: Marketplace Expansion
PCC used TURF analysis (Total Unduplicated Reach and Frequency) on MaxDiff results to guide selection of new food vendors for the marketplace. Rather than relying on general demographic trends or vendor availability, PCC tested combinations of vendor cuisines using MaxDiff data to identify which combinations would appeal to the broadest market segments.
The analysis revealed that a combination of Asian fusion, traditional Hawaiian, and casual American options covered 78% of visitor preferences, while a different combination (Mediterranean, farm-to-table, and fusion) covered only 62%. PCC expanded the marketplace accordingly, and visitor dining spending increased by 22% year-over-year.¹⁵
Results and Strategic Impact
- Visitor satisfaction increased by 18% in post-visit surveys within 12 months of implementing MaxDiff-informed changes
- Incremental revenue per visitor grew by 12% through improved feature prioritization and gift bag optimization
- Segmentation strategy enabled more targeted marketing, reducing customer acquisition cost by 8%
- Management confidence in prioritization decisions increased substantially, shifting decisions from opinion-based to data-driven
The PCC case demonstrates MaxDiff’s value in experience-based industries where perceived value and preference alignment directly impact revenue and satisfaction.
Section 5: Case Study 2 – Riot Games: Cross-Regional Product Development
Riot Games, developer of the phenomenally popular League of Legends, operates in dozens of countries with millions of active players. A critical strategic challenge: when developing game updates and features, how do you prioritize issues and improvements when resources are limited and player bases in different regions (US, Europe, China, Korea) have potentially different preferences?
The Challenge: Scale and Complexity
Riot Games product managers received hundreds of feature requests and bug reports monthly. Players complained about everything from ability balance to user interface clarity, matchmaking systems to in-game communication tools, pricing to cosmetic designs. Without clear priority, development teams faced “everything is important” paralysis.
Additionally, traditional Likert scale feedback surveys showed nearly every complaint received 4-5 star importance ratings, making differentiation impossible. As Senior Researcher Kegan Clark noted, the lack of discrimination in results left product managers unable to identify true pain points affecting player retention.¹⁶
The MaxDiff Solution
Riot Games implemented MaxDiff surveys to evaluate game frustration factors and desired improvements. Rather than asking players “How frustrating is ability imbalance?”—where everyone rates it highly—MaxDiff forced meaningful choices: “Of these game issues, which frustrates you MOST and which frustrates you LEAST?”
Across multiple rounds, Riot tested dozens of potential improvements and pain points (ability balancing, matchmaking speed, toxicity prevention, cosmetic designs, pricing structures, etc.) with large, representative player samples from each major region.
Geo-Specific Findings
MaxDiff revealed critical regional preferences that would have remained hidden with traditional methods:
North America: Ranked matchmaking speed and quality significantly higher than most regions. Additionally, competitive ladder integrity ranked as a top frustration, with players wanting stronger anti-cheat systems.
Europe: Showed similar matchmaking concerns but placed higher value on communication features and language support.
China: Demonstrated distinct preferences—cosmetic and prestige elements ranked higher in importance than western players indicated, supporting Riot’s existing strategy of cosmetic-focused monetization.
Korea: Professional esports players and aspirational competitive players created different preference clusters, with pro-oriented players emphasizing ability balance above nearly everything else, while casual Korean players prioritized social features.¹⁷
Cross-Regional Bias Elimination
Critically, MaxDiff eliminated scale use bias that had plagued previous research. Different cultures use rating scales differently (some cultures default to 4-5, others to 3). By forcing comparative choices rather than ratings, MaxDiff provided fair, comparable data across regions.
Kegan Clark reported: “By switching to MaxDiff, we found strong links between major frustrations and lower playtime. Players citing specific issues—matched with other complaints through MaxDiff data—showed measurable differences in retention compared to players not experiencing those issues.”¹⁸
Impact on Prioritization
MaxDiff results transformed Riot’s development prioritization:
- Global development resources were allocated based on issues affecting the widest player base (matchmaking quality) while region-specific teams received data supporting investments in features like cosmetics (China) or communication tools (Europe).
- The company implemented anti-cheat improvements at scale, informed by the revealed importance of anti-cheat systems in North America and competitive regions.
- Feature development timing changed based on preference data. Abilities balance received faster iteration cycles as MaxDiff data confirmed its primary importance.
- Cosmetic and prestige systems were enhanced with region-specific themes based on preferences revealed in MaxDiff analysis.
Section 6: Case Study 3 – Progressive Insurance: Benefit Prioritization at Scale
Progressive Insurance, a data-driven organization, operates in a highly commoditized market where differentiation often hinges on specific coverage benefits and service features. The challenge: Progressive’s offerings included nearly 100 potential insurance benefits and coverage options.
How could the company prioritize which benefits to emphasize in marketing? Which should be default inclusions? How should pricing reflect value?
The Complexity
Traditional MaxDiff designs become impractical with 100+ items because creating balanced experimental designs becomes computationally complex and respondent burden increases. However, Progressive needed precise preference data across their full benefit set to inform marketing strategy and product packaging.
The Solution: Sparse MaxDiff
Recognizing the constraints, Progressive implemented a “sparse MaxDiff” approach. In this variation, each benefit might appear in just one or two question sets per respondent, rather than the typical 4-6 appearances. With 100 benefits and sets of 4 items:
19 question sets × 4 items per set = 76 unique benefit exposures per respondent
If each respondent completes 25 sets (to reach ~100 exposures), every benefit appears roughly once, and the experimental design remains balanced when aggregated across a large sample.¹⁹
This elegant solution enabled Progressive to evaluate their complete benefit portfolio with minimal respondent burden.
Results and Application
The MaxDiff analysis revealed surprising findings: some benefits Progressive had emphasized in marketing ranked much lower in consumer importance, while other benefits—previously considered table-stakes offerings—emerged as true preference drivers.
This insight allowed Progressive to:
- Refocus marketing messaging on the highest-preference benefits
- Restructure product packaging to bundle benefits likely to appeal to specific segments
- Optimize pricing by understanding which benefits justified premium pricing
- Identify category gaps where competitors were emphasizing benefits less important to consumers
Section 7: Case Study 4 – University of North Carolina Family Medicine: Health Behavior Research
Beyond commercial applications, MaxDiff has proven invaluable in health research where understanding behavior drivers is critical for public health messaging.
The Research Question
Researchers at UNC’s Department of Family Medicine wanted to understand the factors most and least encouraging e-cigarette use among U.S. adults who had tried e-cigarettes.
Understanding these drivers was critical for developing effective public health messaging and interventions.
Why MaxDiff Mattered
The research team evaluated nine key factors (health risks, nicotine content, price, accessibility, social factors, perceived safety vs. combustible cigarettes, flavor variety, product design/aesthetics, and advertising exposure) across a representative sample using MaxDiff methodology.
The team selected MaxDiff specifically because it reduces cognitive strain on respondents compared to other ranking or rating approaches. Despite the sensitive nature of the topic (evaluating factors driving potentially harmful behavior), MaxDiff’s simple choice framework made it more comfortable for participants to respond honestly.²⁰
Key Findings
The analysis revealed that e-cigarette users consider multiple factors holistically rather than emphasizing a single dominant driver. However, important preference clustering emerged: health risk perceptions ranked significantly higher in importance than price considerations, contrary to some public health assumptions.
Additionally, social factors and perceived relative safety compared to smoking ranked higher than product aesthetics or advertising exposure—challenging conventional wisdom about e-cigarette marketing’s effectiveness.
These insights guided more targeted public health messaging, focusing on misinformation correction regarding relative health risks rather than assuming price or design were the primary drivers.
Section 8: Geographic Optimization for Global Market Research
As organizations expand internationally, understanding how to conduct research across diverse markets becomes critical. MaxDiff’s inherent advantages make it particularly valuable for global research, but regional optimization requires thoughtfulness.
The Geo-Optimization Advantage
Eliminating Scale Use Bias Across Cultures
One of MaxDiff’s most significant advantages in global research is eliminating scale use bias—the tendency for different cultures to use rating scales differently. Research has consistently shown that respondents from some cultures default to higher ratings (4-5) while others cluster around middle scores (3). This makes cross-cultural comparison of rating scales problematic.²¹
MaxDiff eliminates this issue entirely. When respondents choose “best” vs. “worst,” the choice has identical meaning across cultures. A German respondent’s choice carries the same weight as a Brazilian respondent’s choice. This makes MaxDiff the methodology of choice for research spanning multiple geographies.
Cross-Region Comparison: The Riot Games Example
As detailed in the Riot Games case study, MaxDiff enables fair comparison across North America, Europe, Asia-Pacific, and other regions. Respondents’ preferences are directly comparable without statistical adjustments for cultural rating scale differences.
Regional Customization: The Kadence Framework
According to research from Kadence, MaxDiff optimization for different regions requires thoughtful customization.²²
Asian Markets (China, Singapore, Indonesia): MaxDiff analysis in rapidly evolving Asian markets reveals region-specific preference patterns:
- In China, where status and convenience are paramount, MaxDiff analysis shows that luxury consumers prioritize exclusivity, brand heritage, and premium materials. Tech consumers emphasize convenience features like mobile payment integration and fast delivery.
- In Singapore and Indonesia, with diverse, segmented populations, MaxDiff analysis reveals different preference profiles. For instance, in Indonesia, MaxDiff can help brands balance local-flavored options against global trends, determining which combinations appeal across regional diversity.
Western Markets (US, Europe): MaxDiff application in mature Western markets typically focuses on feature-functionality preferences. In North America, MaxDiff research with technology consumers emphasizes innovation and cutting-edge features. European markets sometimes show stronger preference for sustainability and ethical sourcing attributes.
Language and Cultural Adaptation
When conducting MaxDiff across languages, localization best practices include:
- Conceptual equivalence, not literal translation: Ensure items being compared are conceptually equivalent across cultures. A benefit that translates literally but carries different cultural meaning needs conceptual reframing.
- Cultural itemization review: Involve in-country researchers in evaluating whether MaxDiff items are culturally relevant. An attribute perfectly meaningful in one market might be irrelevant or offensive in another.
- Balanced attribute presentation: Avoid overrepresenting attributes valued highly in one culture if conducting cross-cultural comparison. Maintain balance so results aren’t skewed by cultural salience differences.
- Randomization across languages: If multilingual studies randomize question-set presentation order separately within each language group to control for language-specific response patterns.²³
Segmentation Across Geographies
MaxDiff enables powerful geographic segmentation. Rather than assuming preference homogeneity within geographic regions, researchers can:
- Conduct separate MaxDiff studies within each major region to identify region-specific preference orderings
- Compare results to identify which preferences are universal and which are geographically distinct
- Create predictive models that estimate preferences for new markets based on macro-level characteristics
For multinational companies, this approach enables:
- Standardized global products (features universally preferred across geographies)
- Customized regional variants (tailored to region-specific preferences)
- Efficient resource allocation (investing first in attributes with broadest geographic appeal)
Best Practices for Geo-Optimized MaxDiff Research
Drawing from best practices across industries, global MaxDiff research benefits from:
1. Layered Analysis Conduct MaxDiff at multiple analytical levels:
- Global sample analysis (identifying universal preferences)
- Regional analysis (understanding geographic variation)
- Demographic + geographic cross-tabulation (identifying preference clusters within regions)
2. Consistent Experimental Design Use identical experimental designs across geographies to ensure comparability. The MaxDiff question sets should be the same, though items themselves can be culturally adapted.
3. Sample Representation Ensure samples within each region are representative of the target market in that region (not just fluent English speakers or educated urban respondents). This becomes challenging but critical for research validity.
4. Regional Expert Review Before deployment, have in-country experts review MaxDiff items for cultural appropriateness, relevance, and clarity. What’s clear phrasing in one language might be ambiguous when translated.
5. Power Analysis by Region Account for the fact that different regions might require different sample sizes for statistical power, particularly when regional populations vary widely in size.
Section 9: MaxDiff Methodology Best Practices
Having examined case studies and applications, here are critical best practices for implementing MaxDiff research successfully.
Designing Your MaxDiff Study
1. Item Selection and Wording
- Identify 12-25 items optimal for MaxDiff testing. Fewer than 12 limits discrimination value; more than 25 requires unwieldy experimental designs
- Word all items consistently (comparable grammatical structure, similar specificity level) to prevent bias where respondents unconsciously favor longer, more descriptive items²⁴
- Group items conceptually when possible (e.g., functional benefits vs. emotional benefits tested in separate batteries) to prevent respondent confusion
2. Experimental Design Selection
- Use balanced incomplete block design (BIBD) or partially balanced incomplete block design (P-BIBD) for efficiency
- Determine items per question set (typically 3-6, with 4-5 optimal for effort/reliability tradeoff)
- Set target frequency (recommendation: minimum 200 total exposures per item across full sample)
- Calculate required question sets per respondent using the formula: (min frequency × items) / (items per set × sample size)
3. Question Set Presentation Randomize presentation to control for order bias. If testing flavors (Lemon-Lime, Mango, Raspberry, etc.), randomize their left-right position and top-bottom presentation to prevent position bias.
4. Sample Size Determination Beyond the mathematical formula, consider:
- Minimum recommended sample: 100-150 respondents for basic results
- Recommended for segmentation: 300-500 respondents to enable meaningful latent class analysis
- For regional analysis: 200-300 per region to enable comparisons with adequate power
5. Response Completion Monitor completion quality. Respondents who answer identically across all questions (e.g., always choosing left-most item as best) are flagging either careless responding or cognitive struggles. Implement attention checks and consider excluding low-quality responses.
Analysis and Interpretation
Utility Score Interpretation Remember that MaxDiff scores are relative measures. A score of 20 (for item A) means item A is 20% of the average item preference in your tested set, not that item A has 20% absolute preference in the market.
Scores should be interpreted as preference gaps: if Item A scores 18 and Item B scores 12, Item A is 50% more preferred than Item B.
Segmentation Analysis Latent class analysis on MaxDiff data reveals natural segments with distinct preference profiles. This is more powerful than post-hoc demographic segmentation because segments are defined by actual preference patterns rather than age or income, which may not drive preferences.
Confidence and Statistical Testing Ensure adequate sample size supports statistical testing. With small samples, apparent preference gaps may reflect sampling variation rather than true differences.
Complementary Research
MaxDiff reveals preferences but not motivations. Best practice involves combining MaxDiff with:
- Qualitative interviews (Why is this feature most preferred?)
- Monadic concept testing (Does this top-preference item generate genuine enthusiasm?)
- Conjoint analysis (How do feature preferences interact with pricing?)
- TURF analysis (What combinations of features/benefits reach the broadest market?)
Section 10: MaxDiff Software and Tools
Multiple platforms enable MaxDiff research, each with different capabilities and price points.
Enterprise-Grade Solutions
Sawtooth Software Sawtooth pioneered choice modeling tools and offers the most advanced MaxDiff capabilities. Their Lighthouse Studio (desktop application) and web-based tools include:
- Hierarchical Bayes estimation for individual-level analysis
- Proprietary “Bandit MaxDiff” using Thompson Sampling for adaptive questioning
- TURF analysis integration
- Advanced segmentation capabilities
Cost: $4,500-$15,000+ annually, depending on package. No free trial; requires sales consultation.²⁵
Qualtrics Qualtrics includes MaxDiff as a question type within their broader experience management platform. Integrated with extensive survey design and analytics capabilities.
Cost: MaxDiff requires premium subscription (EX/CX license).
Mid-Market Solutions
LimeSurvey Open-source survey platform with MaxDiff capabilities, offering a lower-cost alternative to enterprise tools.
Cost: Open-source (free) with optional commercial support.²⁶
SurveyMonkey SurveyMonkey offers MaxDiff through their Momentive research consulting integration, making it accessible to DIY researchers without technical expertise.
Cost: Varies by project scope through Momentive consulting.
Alchemer (formerly SurveyGizmo) Provides MaxDiff on full-access plan, positioned between consumer and enterprise tiers.
Cost: $1,895+ per user/year.
Free or Low-Cost Options
OpinionX OpinionX is the only platform offering fully free MaxDiff analysis surveys. The free tier includes MaxDiff with 3-5 options per set, with premium tiers unlocking customization.
Cost: Free tier (limited features), Premium tiers at $29-$99/month.²⁷
Q Research Software Advanced tool specifically for market research professionals. Includes MaxDiff analysis within broader data analysis framework.
Cost: $2,235-$6,705+ annually.
Selecting Your Platform
Choose based on:
- Scale of research: Enterprise tools suit large, complex studies; consumer tools suit smaller projects
- Technical expertise: Some tools require statistical knowledge; others offer simplified interfaces
- Integration needs: Whether you need integration with existing data systems or standalone capability
- Budget constraints: Range from free to $15,000+ annually
- Analysis depth: Some tools focus on reporting; others enable advanced modeling
Section 11: Common Pitfalls and How to Avoid Them
Pitfall 1: Testing Too Many or Too Few Items
The Problem: Testing only 5-6 items limits discrimination value (everything ranks similarly). Testing 40+ items requires extensive respondent burden or sparse designs that reduce statistical power.
The Solution: Target 12-25 items as the optimal range for most applications. For 100+ items, use sparse MaxDiff designs where items appear rarely but overall sample size enables statistical validity.
Pitfall 2: Inconsistent Item Wording
The Problem: If some items are worded as “Fast delivery speed” while others are “Product quality,” respondents may bias toward wordy or descriptive items rather than evaluating actual preference.
The Solution: Maintain consistent grammatical structure, word length, and specificity across all items. Frame all items at similar conceptual levels.
Pitfall 3: Including Non-Comparable Items
The Problem: Testing features that aren’t meaningfully comparable (mixing “red color” with “fast processing speed”) creates confusion and reduces data quality.
The Solution: Group items conceptually. Test functional attributes together and emotional attributes together. If testing attributes across a single product, ensure all attributes address the same evaluative dimension.
Pitfall 4: Overlooking Respondent Burden
The Problem: Requiring 30+ question sets creates fatigue, reducing data quality in later questions.
The Solution: Limit question sets to 15-25 per respondent for web surveys, fewer for phone surveys. Monitor completion time and quality metrics.
Pitfall 5: Misinterpreting Relative Preferences as Absolute
The Problem: A MaxDiff score of 25 doesn’t mean consumers value that item absolutely—only that they value it most relative to other tested items.
The Solution: Remember MaxDiff measures relative importance. Supplement with absolute measurement (monadic concept testing) if necessary.
Pitfall 6: Inadequate Sample Size for Segmentation
The Problem: Running MaxDiff with 50 respondents provides rankings but insufficient data for meaningful segmentation analysis.
The Solution: Plan for 300+ respondents if segmentation is an analysis goal.
Pitfall 7: Ignoring Data Quality
The Problem: Respondents with identical response patterns across all questions likely indicate careless responding, yet often aren’t flagged or excluded.
The Solution: Implement attention checks. Review response distributions. Exclude obviously problematic respondents from analysis.
Section 12: Implementing MaxDiff Results
Conducting research is only valuable if insights drive action. Here’s how leading organizations implement MaxDiff insights:
From Insights to Strategy
1. Prioritization Documents Translate MaxDiff results into prioritization frameworks. PCC used MaxDiff results to create visitor experience roadmaps prioritizing improvements most likely to impact satisfaction.
2. Product Development Roadmaps Link utility scores to development sequencing. Features with highest preference scores become near-term development priorities; lower-scoring features become longer-term possibilities or may be abandoned.
3. Marketing Message Prioritization MaxDiff messaging research directly informs marketing strategy. If MaxDiff reveals “eco-friendly materials” ranks highest among sustainable product attributes, marketing emphasizes this attribute prominently.
4. Resource Allocation Link MaxDiff results to budget allocation. If “fast checkout” ranks significantly higher than “gift wrapping,” e-commerce platforms invest more in checkout optimization than gift service development.
Building Cross-Functional Buy-In
MaxDiff’s quantitative rigor generates credibility across functions. Product managers, marketing leaders, and executives can align on data-driven priorities rather than debating subjective importance.
- Present results visually: Utility score charts and preference rankings are intuitive even to non-statistical audiences
- Show comparison segments: Demonstrate that preferences vary meaningfully across customer segments, justifying customized strategies
- Link to metrics: Connect preference priorities to actual customer satisfaction, retention, or purchasing data to demonstrate predictive validity
Conclusion: MaxDiff as Your Competitive Advantage
In markets where customer preferences drive competitive advantage, MaxDiff provides unmatched insights into what truly matters to your target audience.
Traditional rating scales generate noisy, biased data. Ranking exercises overwhelm respondents and produce artificial results. MaxDiff’s elegant solution—asking respondents to choose best and worst from small sets—produces reliable, discriminating preference data that drives strategic decision-making.
Whether you’re optimizing guest experiences at tourism attractions (Polynesian Cultural Center), prioritizing product development for millions of global players (Riot Games), structuring benefit portfolios for insurance products (Progressive), or understanding health behavior for public health (University of North Carolina), MaxDiff reveals customer preferences with clarity impossible using other methods.
The methodology is scalable (from dozens to hundreds of items), culturally robust (eliminating cross-cultural rating bias), and applicable across industries (healthcare, tourism, gaming, insurance, manufacturing, technology, and beyond).
For organizations committed to customer-centric strategy, MaxDiff isn’t optional—it’s a competitive necessity. The insights you gain from MaxDiff research will drive product development, marketing strategy, and resource allocation decisions for months or years ahead.
Your competitors are likely still relying on conventional surveys where everything seems important. By implementing MaxDiff, you’re gaining the preference clarity they lack—a clarity that translates directly into customer satisfaction, loyalty, and revenue.
References
¹ Louviere, J. J., et al. “The Centre for the Study of Choice.” University of Sydney. Founded research methodology pioneering MaxDiff in 1990s.
² Conjointly. “MaxDiff Analysis.” https://conjointly.com/products/maxdiff-analysis/
³ Sawtooth Software. “MaxDiff: An Introduction.” https://sawtoothsoftware.com/maxdiff
⁴ Displayr. “What is MaxDiff?” https://www.displayr.com/what-is-maxdiff/
⁵ Sawtooth Software. “MaxDiff Analysis Examples.” https://sawtoothsoftware.com/resources/blog/posts/maxdiff-analysis-examples
⁶ QuestionPro. “MaxDiff Analysis in Market Research.” https://www.questionpro.com/max-diff/
⁷ Drive Research. “What is MaxDiff Market Research.” https://www.driveresearch.com/market-research-company-blog/what-is-maxdiff-market-research-company-albany-ny/
⁸ SurveyMonkey. “Prioritize Features Using MaxDiff Analysis.” https://www.surveymonkey.com/market-research/resources/prioritize-features/
⁹ Suzy. “Finding a Clear Winner: How to Use MaxDiff Analysis.” https://www.suzy.com/blog/finding-a-clear-winner-max-diff-analysis-market-research
¹⁰ Select Statistical Consultants. “Maximum Difference Scaling (MaxDiff).” https://select-statistics.co.uk/blog/maximum-difference-scaling-maxdiff/
¹¹ Qualtrics. “An Intro to MaxDiff (Best Worst Scaling) Analysis & Design.” https://www.qualtrics.com/blog/an-introduction-to-maxdiff/
¹² OpinionX. “MaxDiff Analysis Guide.” https://www.opinionx.co/research-method-guides/maxdiff-analysis
¹³ OpinionX. “Comparing MaxDiff Analysis Tools.” https://www.opinionx.co/blog/comparing-maxdiff-tools
¹⁴ Sawtooth Software. “MaxDiff Analysis Examples: Progressive Insurance Case Study.” https://sawtoothsoftware.com/resources/blog/posts/maxdiff-analysis-examples
¹⁵ Sawtooth Software. “Polynesian Cultural Center Case Study.” Referenced in MaxDiff examples and consulting work.
¹⁶ Sawtooth Software. “Riot Games MaxDiff Case Study.” https://sawtoothsoftware.com/resources/blog/posts/maxdiff-analysis-examples
¹⁷ Regional preference variation documented in cross-cultural choice modeling literature and gaming industry research best practices.
¹⁸ Kegan Clark, Senior Researcher at Riot Games. Referenced in Sawtooth Software MaxDiff case study documentation.
¹⁹ Sparse MaxDiff methodology discussed in Sawtooth Software resources and academic literature on choice modeling at scale.
²⁰ University of North Carolina Department of Family Medicine. Referenced in choice modeling applications to health research.
²¹ Sawtooth Software. “Scale Use Bias in Cross-Cultural Research.” Technical documentation addressing cultural differences in rating scale interpretation.
²² Kadence. “Unlocking Consumer Preferences with MaxDiff and TURF Analysis.” https://kadence.com/en-us/unlocking-consumer-preferences-with-maxdiff-and-turf-analysis/
²³ Best practices for multilingual and cross-cultural research derived from academic literature in choice modeling and market research methodology.
²⁴ AYTM. “Intro to Advanced MaxDiff Analysis.” https://aytm.com/post/intro-to-advanced-maxdiff-analysis
²⁵ Sawtooth Software Pricing. Verified current as of 2025.
²⁶ LimeSurvey. “MaxDiff Analysis.” https://www.limesurvey.org/
²⁷ OpinionX. “Comparing MaxDiff Analysis Tools.” https://www.opinionx.co/blog/comparing-maxdiff-tools
Additional Resources
- Sawtooth Software Technical Paper: “How Good Is Best-Worst Scaling?” (2018)
- Academic Reference: “Maximum Difference Scaling: Improved Measures of Importance and Preference for Segmentation” (2003)
- AYTM Certification: Free MaxDiff researcher certification course available at https://aytm.com
- B2B International: Comprehensive methodology overview at https://www.b2binternational.com/research/methods/statistical-techniques/maxdiff/
0 Comments