Should you use Advantage+ or traditional campaign structures? The only reliable answer comes from testing. Here's a rigorous framework for comparing performance and making data-driven decisions.
Why Testing Matters
Meta recommends Advantage+ for most advertisers, but blanket recommendations don't account for your specific:
- Product type and price point
- Conversion volume and patterns
- Creative assets and diversity
- Historical campaign performance
- Business objectives and constraints
Testing reveals what works for your account, not what works on average.
Test Design Principles
Isolate the Variable
A valid test changes only one thing — campaign structure. Everything else must match:
- Same creative assets
- Same conversion event
- Same geographic targeting
- Same daily budget
- Same launch timing
Statistical Significance
Don't make decisions on small sample sizes. Plan for enough conversions:
- Minimum: 50 conversions per campaign
- Recommended: 100+ conversions per campaign
- Ideal: 200+ conversions for confident decisions
This usually means running tests for 2-4 weeks depending on your conversion volume. See our guide to algorithmic learning.
Account for Learning Phase
Both campaigns need time to optimize:
- Exclude first 7 days from analysis (learning period)
- Analyze performance from day 8 onward
- If learning phase extends, adjust analysis window
Test Setup: Traditional Campaign
Structure Options
Choose your best-performing traditional structure:
- CBO with multiple ad sets: Campaign budget optimization across audiences
- ABO with controlled budgets: Fixed budget per audience segment
- Single ad set with detailed targeting: Your best-performing audience
Configuration
- Use your proven audience targeting
- Manual placements or Advantage+ placements (match Advantage+)
- Same creative as Advantage+ test
- Optimize for your standard conversion event
Test Setup: Advantage+ Campaign
Advantage+ Shopping (Ecommerce)
- Connect product catalog
- Define existing customer audience carefully
- Set existing customer budget cap (suggest 25-30%)
- Upload same creative as traditional test
Advantage+ Audience (Non-Shopping)
- Choose whether to add suggestions (test both if possible)
- Use same geographic targeting as traditional
- Same creative assets
- Same conversion optimization
Measurement Framework
Primary Metrics
- New Customer CPA: Cost to acquire new buyers
- New Customer ROAS: Return from new customers only
- Blended ROAS: Overall campaign efficiency
New customer metrics matter most because they show true acquisition efficiency. Learn about measurement in our broad targeting guide.
Secondary Metrics
- CPM: Cost efficiency of delivery
- CTR: Creative engagement
- Frequency: Audience saturation
- Reach: Audience expansion
Operational Metrics
- Learning phase duration: Time to exit learning
- Scaling behavior: Performance at higher budgets
- Management time: Hours required for optimization
Analysis Methodology
Week-by-Week Tracking
Create a comparison dashboard tracking:
- Daily and weekly CPA trends
- ROAS progression over time
- Budget utilization
- Creative performance within each structure
Segment Analysis
Break down results by:
- New vs existing customers
- Placement performance
- Creative asset performance
- Day of week patterns
Decision Criteria
Determine winners based on:
- Clear winner: 15%+ difference in primary metrics, consistent over time
- Marginal difference: Under 10% difference may be noise — consider other factors
- Trade-offs: One wins on CPA, other wins on ROAS — choose based on business priority
Extended Testing Considerations
Scaling Tests
Performance at test budget may not match performance at scale. After initial winner is identified:
- Increase budget by 2x on winning structure
- Monitor for performance changes
- Test at your target scale before full commitment
Seasonal Factors
Test results during one season may not apply to another. Consider retesting during:
- Peak season vs off-season
- Different quarters
- After major algorithm updates
Creative Interaction
Different structures may favor different creative. After structure testing, test creative variations within winning structure. See creative diversification strategies.
How ROASPIG Helps
Valid A/B testing requires identical creative across test variants. ROASPIG ensures creative consistency:
- Asset Management: Organize and deploy identical creative across test campaigns
- Performance Tracking: Compare creative performance within each campaign structure
- Diversity Assurance: Ensure both tests have sufficient creative diversity
- Refresh Coordination: Update creative simultaneously across test variants
- Attribution Analysis: Understand which creative concepts drive results in each structure
The Bottom Line
Testing Advantage+ vs traditional structures is essential — don't rely on assumptions or general recommendations. Design clean tests with isolated variables, sufficient sample sizes, and clear decision criteria. Measure what matters for your business, not just what's easy to track.
Remember: the winner for your account may not match industry averages. Test rigorously, decide based on data, and be willing to retest as conditions change.
Frequently Asked Questions About A/B Testing Advantage+
Run tests until you have 100+ conversions per campaign, typically 2-4 weeks. Exclude the first 7 days (learning phase) from analysis. Shorter tests risk decisions based on statistical noise rather than true performance differences.
New customer CPA and new customer ROAS are most important because they show true acquisition efficiency. Blended ROAS can be misleading if one structure is retargeting more. Always segment by new vs existing customers.
Use identical creative, same conversion event, same geographic targeting, same budget, and launch simultaneously. The only difference should be campaign structure. Any other differences invalidate the comparison.
This indicates a trade-off between volume and efficiency. Choose based on business priorities: if growth is primary, favor lower CPA (Advantage+). If profitability is primary, favor higher ROAS (traditional). Consider blended performance too.
Retest when significant changes occur: new creative, different seasons, algorithm updates, or if performance declines. Quarterly retesting is reasonable for accounts with sufficient volume. Lower-volume accounts may test annually.