A/B Testing Methodologies

A/B testing methodologies in game monetization represent a systematic, data-driven approach to optimizing revenue generation by comparing two or more variants of game features, pricing models, or in-game purchase offerings to determine which performs better with players 12. This experimental framework enables game developers and publishers to make evidence-based decisions about monetization strategies by exposing different player segments to controlled variations and measuring their behavioral responses 3. In the highly competitive gaming industry, where player acquisition costs continue to rise and retention remains challenging, A/B testing has become indispensable for maximizing lifetime value (LTV) while maintaining player satisfaction 12. The methodology matters critically because even marginal improvements in conversion rates, average revenue per user (ARPU), or retention can translate into millions of dollars in additional revenue for successful titles 56.

Overview

The emergence of A/B testing in game monetization reflects the broader evolution of the gaming industry from premium, one-time purchase models to free-to-play and games-as-a-service approaches that require continuous optimization 67. As mobile gaming exploded in the 2010s and free-to-play models became dominant, developers faced the fundamental challenge of balancing revenue generation with player experience—aggressive monetization could drive short-term revenue but damage long-term retention and brand reputation 28. Traditional intuition-based pricing and feature decisions proved insufficient in this complex environment, where player behavior varied dramatically across segments and small changes could have outsized impacts on business outcomes 15.

The practice has evolved significantly from simple price testing to sophisticated experimentation programs encompassing every aspect of monetization design 67. Early implementations focused primarily on testing discrete elements like price points or promotional offers, but modern approaches employ multivariate testing, sequential experimentation, and machine learning-driven personalization to optimize entire monetization ecosystems 28. Leading game companies now run dozens or hundreds of concurrent experiments, treating their games as living laboratories where continuous testing drives incremental improvements that compound into substantial competitive advantages 610.

Key Concepts

Randomized Controlled Experiments

Randomized controlled experiments form the foundation of A/B testing, involving the random assignment of players to different groups that experience distinct variants of monetization features 13. The fundamental principle requires randomly dividing the player base into a control group (A) experiencing the current implementation and one or more treatment groups (B, C, etc.) experiencing modified versions, ensuring that any observed differences in behavior can be attributed to the variant rather than pre-existing player characteristics 28.

Example: A mobile puzzle game developer wants to test whether reducing the price of their "Starter Bundle" from $4.99 to $2.99 will increase overall revenue. They randomly assign 50% of new players installing the game on Monday to see the $4.99 price (control group) and 50% to see the $2.99 price (treatment group). After two weeks with 10,000 players in each group, they find that the $2.99 variant achieved a 4.2% conversion rate compared to 2.1% for the $4.99 variant, and despite the lower price point, generated 18% higher total revenue due to the doubled conversion rate.

Statistical Significance and Confidence Intervals

Statistical significance determines whether observed differences between variants represent genuine effects or merely random variation, typically using a threshold of p < 0.05 to conclude that results are unlikely to have occurred by chance 12. Confidence intervals quantify the uncertainty around measured effects, providing a range within which the true effect likely falls and enabling more nuanced decision-making than binary significance tests alone 38.

Example: A strategy game tests two different battle pass pricing structures—$9.99 versus $12.99—and observes that the $9.99 variant generates an average of $2.45 per player while the $12.99 variant generates $2.38 per player. While the $9.99 variant appears superior, the 95% confidence interval for the difference ranges from -$0.15 to +$0.29, crossing zero and indicating the result is not statistically significant (p = 0.18). The team decides not to implement the change, recognizing they cannot confidently conclude which price is truly better.

Primary, Secondary, and Guardrail Metrics

Effective A/B testing requires a hierarchical metric framework consisting of primary metrics (the main success criteria), secondary metrics (supporting indicators), and guardrail metrics (safeguards against negative impacts) 25. Primary metrics directly measure the business objective, such as 7-day ARPU or conversion rate, while secondary metrics provide additional context about player behavior, and guardrail metrics ensure optimizations don't damage player experience or long-term retention 18.

Example: A role-playing game tests a more aggressive daily deal system. Their primary metric is 14-day ARPU, their secondary metrics include daily active users (DAU), session length, and purchase frequency, and their guardrail metrics include day-7 retention and player satisfaction scores from in-game surveys. The test shows a 12% increase in 14-day ARPU (primary metric success) and 8% higher purchase frequency (positive secondary signal), but also reveals a 5% decrease in day-7 retention (guardrail violation). Despite the revenue gains, the team rejects the variant due to the retention impact, recognizing that long-term player value depends on sustained engagement.

Sample Size and Statistical Power

Sample size calculations determine how many players must be exposed to each variant to reliably detect meaningful differences, while statistical power (typically set at 80%) represents the probability of detecting a true effect if one exists 12. Insufficient sample sizes lead to underpowered tests that produce inconclusive results, while excessive sample sizes waste time and resources testing beyond the point of useful learning 38.

Example: A card game wants to test whether adding a "limited time" countdown timer to special offers increases conversion by at least 15% (their minimum detectable effect). Using a power analysis calculator with baseline conversion rate of 3.5%, desired power of 80%, and significance level of 0.05, they determine they need 4,200 players per variant. Given their daily new player rate of 600, they plan to run the test for 14 days to accumulate sufficient sample size, rather than making premature decisions based on incomplete data.

Player Segmentation and Heterogeneous Treatment Effects

Player segmentation recognizes that monetization changes may affect different player groups differently, requiring analysis across cohorts such as new versus returning players, geographic regions, spending tiers, or device types 26. Heterogeneous treatment effects occur when a variant performs better for some segments but worse for others, insights that aggregate analysis would miss 58.

Example: A shooter game tests a new weapon skin pricing strategy and finds that overall revenue increases by 6%. However, segmented analysis reveals that the new pricing increases revenue by 22% among players in their first 30 days but decreases revenue by 8% among veteran players (90+ days). This heterogeneous treatment effect leads the team to implement a hybrid approach: new players see the new pricing structure, while veteran players retain the original pricing, maximizing revenue across both segments.

Novelty Effects and Long-Term Impact

Novelty effects occur when players respond positively to changes simply because they're new and different, not because they're genuinely better, creating misleading short-term results that don't reflect long-term performance 28. Distinguishing between genuine improvements and temporary novelty requires extended testing periods that capture player behavior after the initial excitement fades 16.

Example: A farming simulation game introduces a redesigned shop interface with animated characters and dynamic promotions. Initial two-week results show a 25% increase in shop visits and 15% higher conversion rates. However, the team extends the test to six weeks and observes that by weeks 5-6, shop visits have declined to only 8% above baseline and conversion rates show no significant difference. Recognizing the novelty effect, they conclude the redesign's true impact is much smaller than initial results suggested and decide the implementation cost isn't justified by the modest long-term gains.

Multi-Armed Bandit Algorithms

Multi-armed bandit algorithms represent an adaptive testing approach that continuously allocates more traffic toward better-performing variants during the experiment, balancing exploration (learning which variants work best) with exploitation (maximizing revenue by favoring winners) 28. Unlike traditional A/B tests with fixed traffic allocation, bandits dynamically adjust based on accumulating evidence, reducing the opportunity cost of exposing players to inferior variants 67.

Example: A match-3 game runs a bandit test on four different promotional offer bundles during a weekend event. The algorithm starts with equal 25% traffic allocation to each variant but, as data accumulates, progressively shifts traffic toward the best performers. By Sunday evening, the top-performing bundle receives 55% of traffic, the second-best receives 30%, and the two underperformers receive only 7.5% each. This adaptive allocation generates 11% more revenue during the event compared to a traditional A/B test with fixed allocation, while still gathering sufficient data to identify the winning variant for future events.

Applications in Game Monetization

In-App Purchase Pricing Optimization

A/B testing enables systematic optimization of in-app purchase pricing across the full spectrum of offerings, from small consumable purchases to premium bundles 15. Developers test individual price points, bundle compositions, discount percentages, and pricing tiers to identify optimal structures that maximize revenue while maintaining perceived value 26. This application proves particularly critical for mobile games where pricing psychology and competitive positioning significantly influence purchase decisions 810.

Mobile game publishers frequently test price localization strategies, recognizing that optimal pricing varies dramatically across geographic markets based on purchasing power, competitive landscapes, and cultural factors 56. A successful implementation might involve testing three price points for a premium currency bundle in each major market—for example, $4.99, $5.99, and $6.99 in the United States, while simultaneously testing ₹349, ₹399, and ₹449 in India—then analyzing conversion rates and revenue to identify market-specific optimal prices that can increase global revenue by 15-25% compared to uniform pricing 210.

Promotional Offer Design and Timing

Game developers extensively test promotional offer structures, including discount depths, bundle compositions, visual presentation, and timing triggers 16. These tests examine how different promotional mechanics—such as first-time user discounts, time-limited flash sales, or milestone-based offers—impact both immediate conversion and long-term player value 28. The application extends to testing offer frequency and cadence, balancing the revenue boost from promotions against potential conditioning effects where players learn to wait for discounts rather than purchasing at full price 57.

A role-playing game might test whether offering a "welcome bundle" immediately upon tutorial completion versus after the first gameplay session produces better results 26. The immediate offer variant might achieve higher visibility (95% of players see it versus 78% who complete the first session), but the delayed offer variant could demonstrate stronger conversion (8.2% versus 6.1%) because players have developed more investment in the game and better understand the value proposition, ultimately generating 24% higher revenue per new player despite lower reach 18.

Progression System and Economy Balancing

A/B testing informs fundamental game economy decisions by testing different progression curves, resource generation rates, and monetization pressure points 26. Developers test variations in how quickly players earn free currency, how much premium currency purchases accelerate progress, and where friction points encourage spending 58. This application requires careful attention to both short-term monetization metrics and long-term retention, as overly aggressive economies can drive immediate revenue but damage player satisfaction and lifetime value 17.

A city-building game might test two different building time curves: Variant A where building times increase gradually (1 minute at level 1, 30 minutes at level 10, 4 hours at level 20) versus Variant B with steeper progression (1 minute at level 1, 2 hours at level 10, 24 hours at level 20) 26. While Variant B creates stronger monetization pressure and increases premium currency purchases by 18%, it also reduces day-14 retention by 7% and generates 12% more negative reviews mentioning "pay to win," leading the team to implement a middle-ground approach that balances monetization and player experience 810.

Battle Pass and Seasonal Content Monetization

Battle pass systems undergo rigorous A/B testing of tier structures, reward distributions, pricing, and progression requirements to optimize both purchase conversion and completion rates 26. Developers test different price points, free versus premium tier reward ratios, and the effort required to complete the pass, recognizing that completion rates drive both player satisfaction and likelihood of purchasing future passes 58. This application has become increasingly important as battle passes have emerged as a dominant monetization model across multiple game genres 17.

A competitive multiplayer game tests two battle pass structures: a $9.99 pass with 50 tiers requiring approximately 40 hours of gameplay to complete versus a $12.99 pass with 75 tiers requiring 60 hours 26. Analysis reveals that while the premium pass achieves 8% lower initial purchase conversion, players who buy it spend 22% more time in-game and are 31% more likely to purchase the subsequent season's pass, resulting in 19% higher lifetime value over a six-month period despite the lower initial conversion 810.

Best Practices

Pre-Register Analysis Plans to Prevent P-Hacking

Establishing analysis plans before launching tests prevents p-hacking and post-hoc rationalization by specifying in advance which metrics will be examined, how they'll be calculated, what segmentation will be performed, and what decision criteria will be applied 12. This practice ensures that teams don't selectively report favorable results while ignoring unfavorable ones, or continuously slice data until finding a significant result by chance 38. Pre-registration creates accountability and scientific rigor, distinguishing genuine discoveries from statistical artifacts.

Implementation Example: Before launching a test of three different starter pack configurations, a team documents their analysis plan specifying that the primary decision metric will be 7-day ARPU, measured from first session to day 7, including all in-app purchases but excluding ad revenue. They commit to analyzing results after reaching 5,000 players per variant, examining three pre-specified segments (geographic region, device type, and acquisition source), and implementing the winning variant only if it shows at least 10% improvement with p < 0.05. This pre-commitment prevents them from later cherry-picking favorable metrics or segments if their primary analysis shows no significant difference 128.

Account for Multiple Comparisons When Running Concurrent Tests

Running multiple simultaneous tests increases the probability of false positives, requiring statistical adjustments such as Bonferroni correction or false discovery rate control to maintain valid inference 23. Without correction, running 20 independent tests at p < 0.05 would be expected to produce one false positive by chance alone, leading teams to implement changes that don't actually improve outcomes 18. Best practice involves either limiting the number of concurrent tests, applying appropriate statistical corrections, or using hierarchical testing frameworks that control family-wise error rates.

Implementation Example: A game studio runs a weekly experimentation program with an average of 12 active tests. Rather than evaluating each test independently at p < 0.05 (which would generate frequent false positives), they apply a Bonferroni correction by dividing their significance threshold by the number of tests: 0.05 / 12 = 0.0042. This means they only consider results significant if p < 0.0042, substantially reducing false positives while requiring larger sample sizes or effect sizes to reach significance. Alternatively, they might prioritize tests and apply sequential testing, where only the most promising experiments receive full statistical analysis 238.

Monitor Sample Ratio Mismatch to Detect Technical Issues

Sample ratio mismatch (SRM) occurs when the observed allocation of players to variants differs significantly from the intended allocation, signaling technical implementation problems that can bias results 12. For example, if a test designed to split traffic 50/50 actually delivers 52/48 or 55/45, this indicates potential issues with randomization, variant assignment, or data collection that may invalidate the experiment 8. Best practice involves automatically monitoring SRM for all tests and investigating any significant deviations before analyzing results.

Implementation Example: A mobile game's experimentation platform automatically calculates a chi-square test for each experiment, comparing observed traffic allocation to intended allocation. When a test designed for 50/50 split shows 10,847 players in the control group and 9,203 in the treatment group (54/46 split), the platform flags a significant SRM (p < 0.001). Investigation reveals that a bug in the variant assignment code caused players on certain device models to be preferentially assigned to the control group. The team halts the test, fixes the bug, and restarts with proper randomization, avoiding the biased conclusions they would have drawn from the corrupted data 128.

Test for Sufficient Duration to Capture Player Lifecycle Effects

Monetization tests must run long enough to capture relevant player lifecycle patterns, including weekly cycles, progression milestones, and delayed monetization events 26. Tests that conclude too quickly may miss important effects, such as how a change impacts players who reach mid-game content or how weekend versus weekday behavior differs 58. Best practice involves setting test duration based on the player lifecycle stage being affected, typically requiring at least one full week and often 2-4 weeks for meaningful monetization tests.

Implementation Example: A strategy game tests a change to their mid-game economy that affects players who reach city level 15, which typically occurs 8-12 days after install. Rather than running a 7-day test (which would capture only early reactions), they commit to a 21-day test duration that allows most players to reach the affected content and exhibit their full behavioral response. This extended duration reveals that while the change shows promising 7-day results (+8% revenue), the effect diminishes by day 14 (+3% revenue) and becomes negative by day 21 (-2% revenue) as players encounter downstream consequences of the economy change, leading to rejection of a variant that shorter testing would have falsely validated 268.

Implementation Considerations

Experimentation Platform Selection and Technical Infrastructure

Choosing appropriate experimentation platforms involves balancing functionality, integration complexity, and cost across options ranging from third-party services to custom-built solutions 28. Third-party platforms like Firebase Remote Config, Optimizely, or specialized game analytics services (GameAnalytics, deltaDNA) offer rapid implementation and proven reliability but may have limitations in customization or data ownership 16. Custom-built solutions provide maximum flexibility and control but require substantial engineering investment and statistical expertise to implement correctly 57.

For a small indie studio launching their first mobile game, Firebase Remote Config might provide the optimal balance—offering free tier access, straightforward integration with Unity or Unreal Engine, and sufficient functionality for basic price testing and feature flags without requiring dedicated data engineering resources 28. In contrast, a large publisher managing dozens of live games might invest in a custom experimentation platform that integrates deeply with their proprietary analytics infrastructure, supports complex multivariate tests and bandit algorithms, and provides unified reporting across their entire portfolio 16.

Metric Definition and Measurement Consistency

Establishing clear, consistent metric definitions prevents confusion and ensures reproducibility across experiments and teams 12. Critical decisions include whether to measure revenue at transaction time or recognition time, how to handle refunds and chargebacks, whether to include or exclude outliers, and how to attribute revenue to specific player cohorts 58. Documentation of these definitions and automated calculation pipelines ensure that different analysts examining the same test reach identical conclusions.

A game company might establish a standard metrics dictionary defining "7-day ARPU" as total revenue from all in-app purchases (excluding advertising) generated by a player cohort from their first session through 168 hours later, measured at transaction time in USD, including refunds as negative revenue, and calculated as total cohort revenue divided by total cohort size including non-payers 12. This precise definition eliminates ambiguity—for example, clarifying that a player who installs at 2:00 PM on Monday contributes revenue through 2:00 PM the following Monday, and that refunds reduce ARPU rather than being excluded 8.

Segmentation Strategy and Personalization Capabilities

Determining appropriate segmentation strategies involves balancing the insight value of granular analysis against the statistical power costs of splitting samples into smaller groups 26. Standard segments include new versus returning players, geographic regions, device types (iOS versus Android, phone versus tablet), acquisition sources, and spending tiers 58. Advanced implementations may incorporate behavioral segments based on play patterns, progression speed, or engagement levels, enabling personalized monetization strategies that optimize for each player type.

A puzzle game might implement a three-tier segmentation strategy: Level 1 (always analyzed) includes platform (iOS/Android) and player tenure (0-7 days, 8-30 days, 31+ days); Level 2 (analyzed for major tests) adds geographic region (North America, Europe, Asia, Other) and spending tier (non-payer, low spender <$5, medium spender $5-$50, high spender >$50); Level 3 (analyzed only for strategic tests with large samples) incorporates behavioral segments based on progression speed and session patterns 26. This hierarchical approach ensures they always examine critical segments while avoiding excessive fragmentation that would underpower their analyses 8.

Organizational Processes and Cross-Functional Collaboration

Successful A/B testing programs require robust organizational processes coordinating data science, engineering, product management, and design teams 12. Establishing clear ownership, review processes, and decision-making authority prevents tests from languishing in analysis paralysis or being implemented without proper validation 68. Regular experimentation reviews, shared dashboards, and documented learnings create institutional knowledge that compounds over time.

A mid-size game studio might implement a weekly experimentation review meeting where the data science team presents results from completed tests, the product team proposes new experiments based on roadmap priorities, and engineering provides implementation feasibility assessments 28. They maintain a shared experiment registry tracking all active and completed tests, their hypotheses, results, and decisions, creating a searchable knowledge base that prevents redundant testing and enables new team members to understand the evidence base behind current monetization strategies 16. Clear decision criteria—such as requiring both statistical significance (p < 0.05) and practical significance (>10% improvement) plus positive or neutral guardrail metrics—streamline decision-making and reduce subjective debates 5.

Common Challenges and Solutions

Challenge: Insufficient Statistical Power and Inconclusive Results

Many game monetization tests fail to reach statistical significance because they lack sufficient sample size to detect realistic effect sizes, particularly when testing changes with modest impacts or when measuring high-variance metrics like revenue 12. Games with smaller player bases face particular challenges, as accumulating adequate samples may require impractically long test durations 68. Underpowered tests waste resources and create frustration when teams invest in experiments that yield inconclusive results, neither validating nor rejecting hypotheses.

Solution:

Conduct prospective power analysis before launching tests to ensure adequate sample sizes, using realistic effect size estimates based on historical data or industry benchmarks rather than wishful thinking 12. For games with limited traffic, focus testing on higher-impact changes more likely to produce detectable effects, use proxy metrics with lower variance that reach significance faster (such as conversion rate instead of revenue), or extend test durations to accumulate sufficient samples 68. Consider sequential testing methods that allow valid early stopping when strong effects emerge quickly while continuing tests that show promising but not yet significant trends. For very small games, accept that rigorous A/B testing may not be feasible and rely instead on careful rollouts with before/after analysis or qualitative player feedback 5.

Challenge: Novelty Effects and Short-Term Versus Long-Term Impact

Players often respond positively to changes simply because they're new and different, creating misleading short-term results that don't reflect long-term performance 28. A redesigned shop interface might generate excitement and increased engagement initially, but these effects may fade as players habituate to the new design 16. Conversely, some changes may show negative short-term impacts as players adjust to new systems but prove beneficial long-term once learning curves are overcome. Distinguishing genuine improvements from temporary novelty requires extended observation periods that many teams lack patience for.

Solution:

Extend test durations beyond initial reaction periods to capture stabilized behavior, typically running monetization tests for at least 2-4 weeks rather than concluding after a few days 26. Analyze results in time-segmented windows (days 1-7, days 8-14, days 15-21) to detect whether effects strengthen, weaken, or reverse over time, providing evidence of novelty versus sustained impact 18. For major changes, implement holdout groups that remain on the control variant indefinitely, enabling long-term comparison of cumulative effects over months 5. When novelty effects are suspected, consider running follow-up tests that re-randomize players who experienced the initial test, checking whether the effect persists when the variant is no longer novel 2.

Challenge: Multiple Testing and False Discovery Rates

Running many concurrent experiments increases the probability of false positives—purely by chance, approximately 1 in 20 tests will show significant results at p < 0.05 even when no true effect exists 13. Organizations running aggressive experimentation programs with dozens of simultaneous tests face substantial false discovery rates, potentially implementing changes that don't actually improve outcomes 28. The problem compounds when teams examine multiple metrics, segments, or time windows for each test, further inflating false positive risks through multiple comparisons.

Solution:

Apply appropriate statistical corrections such as Bonferroni adjustment (dividing significance threshold by number of tests) or false discovery rate control methods that maintain acceptable error rates across multiple comparisons 23. Implement hierarchical testing frameworks where primary metrics must show significance before examining secondary metrics, reducing unnecessary comparisons 18. Require replication of surprising or high-impact results before implementation, running confirmatory tests that validate initial findings and substantially reduce false positive risks 6. Establish higher evidence standards for major strategic decisions, such as requiring p < 0.01 instead of p < 0.05, or demanding consistent effects across multiple segments and metrics rather than relying on single significant results 5.

Challenge: Sample Ratio Mismatch and Technical Implementation Errors

Technical bugs in randomization, variant assignment, or data collection can invalidate experiments by creating biased samples or incorrect measurements 12. Sample ratio mismatches—where observed traffic allocation differs from intended allocation—signal potential problems but often go undetected without systematic monitoring 8. Implementation errors might cause certain player types to be preferentially assigned to specific variants, variant experiences to differ in unintended ways beyond the tested change, or metrics to be calculated incorrectly, all of which can lead to false conclusions.

Solution:

Implement automated sample ratio mismatch detection for all experiments, using chi-square tests to compare observed versus expected traffic allocation and flagging significant deviations for investigation 12. Establish rigorous quality assurance processes including code review of experiment implementations, manual testing of variant experiences across different devices and player states, and validation that data collection captures all relevant events correctly 8. Create staging environments where experiments can be tested with synthetic traffic before production deployment, catching technical issues before they affect real players 6. When SRM or other technical issues are detected, immediately halt the experiment, investigate root causes, fix implementation problems, and restart with clean data rather than attempting to salvage corrupted results 25.

Challenge: Balancing Short-Term Revenue Optimization with Long-Term Player Health

Aggressive monetization variants often boost immediate revenue metrics while potentially damaging long-term retention, player satisfaction, and brand reputation 26. Tests focusing exclusively on short-term revenue KPIs may lead to implementing exploitative practices that maximize immediate extraction at the expense of sustainable player relationships 18. The challenge intensifies because long-term effects take months to fully manifest, requiring patience and sophisticated measurement approaches that many organizations struggle to maintain under pressure for quarterly results.

Solution:

Implement comprehensive metric frameworks that include guardrail metrics protecting player experience alongside revenue metrics, requiring that winning variants show positive or neutral effects on retention, engagement, and satisfaction 26. Extend test durations and analysis windows to capture medium-term effects (30-90 days) rather than only immediate impacts, revealing whether revenue gains prove sustainable or merely pull forward spending 18. Maintain long-term holdout groups that enable measuring cumulative effects of monetization strategies over 6-12 month periods, quantifying whether optimization programs genuinely increase lifetime value or simply optimize short-term extraction 5. Incorporate qualitative feedback through player surveys, community sentiment analysis, and customer support ticket monitoring to detect negative reactions that quantitative metrics might miss 610.

References

  1. Game Developer. (2019). How to A/B Test Your Game Monetization. https://www.gamedeveloper.com/business/how-to-a-b-test-your-game-monetization
  2. Unity Technologies. (2021). A/B Testing Best Practices for Game Developers. https://blog.unity.com/games/ab-testing-best-practices-for-game-developers
  3. ScienceDirect. (2019). A/B Testing in Mobile Games. https://www.sciencedirect.com/science/article/pii/S1875952119300142
  4. ACM Digital Library. (2019). Experimental Design for Game Monetization. https://dl.acm.org/doi/10.1145/3290605.3300537
  5. GamesIndustry.biz. (2020). How to Use A/B Testing to Improve Your Game's Monetisation. https://www.gamesindustry.biz/how-to-use-ab-testing-to-improve-your-games-monetisation
  6. Deconstructor of Fun. (2019). A/B Testing in Mobile Games. https://www.deconstructoroffun.com/blog/2019/1/15/ab-testing-in-mobile-games
  7. VentureBeat. (2021). How Game Developers Use A/B Testing to Boost Monetization. https://venturebeat.com/games/how-game-developers-use-ab-testing-to-boost-monetization/
  8. GDC Vault. (2019). A/B Testing for Game Monetization. https://www.gdcvault.com/play/1026282/A-B-Testing-for-Game
  9. IEEE Xplore. (2019). Statistical Methods for Game Analytics. https://ieeexplore.ieee.org/document/8847987
  10. PocketGamer.biz. (2020). A/B Testing Mobile Games Monetization Strategy. https://www.pocketgamer.biz/comment-and-opinion/74523/ab-testing-mobile-games-monetization-strategy/