Interference, Bias, and Variance in Two-Sided Marketplace Experimentation: Guidance for Platforms

Two-sided marketplace platforms often run experiments to test the effect of an intervention before launching it platform-wide. A typical approach is to randomize individuals into the treatment group, which receives the intervention, and the control group, which does not. The platform then compares the performance in the two groups to estimate the effect if the intervention were launched to everyone. We focus on two common experiment types, where the platform randomizes individuals either on the supply side or on the demand side. The resulting estimates of the treatment effect in these experiments are typically biased: because individuals in the market compete with each other, individuals in the treatment group affect those in the control group and vice versa, creating interference. We develop a simple tractable market model to study bias and variance in these experiments with interference. We focus on two choices available to the platform: (1) Which side of the platform should it randomize on (supply or demand)? (2) What proportion of individuals should be allocated to treatment? We find that both choices affect the bias and variance of the resulting estimators but in different ways. The bias-optimal choice of experiment type depends on the relative amounts of supply and demand in the market, and we discuss how a platform can use market data to select the experiment type. Importantly, we find in many circumstances, choosing the bias-optimal experiment type has little effect on variance. On the other hand, the choice of treatment proportion can induce a bias-variance tradeoff, where the bias-minimizing proportion increases variance. We discuss how a platform can navigate this tradeoff and best choose the treatment proportion, using a combination of modeling as well as contextual knowledge about the market, the risk of the intervention, and reasonable effect sizes of the intervention.

[1]  G. Imbens,et al.  Exact p-Values for Network Interference , 2015, 1506.02084.

[2]  Ron Kohavi,et al.  Trustworthy Online Controlled Experiments , 2020 .

[3]  Ya Xu,et al.  SQR: Balancing Speed, Quality and Risk in Online Experiments , 2018, KDD.

[4]  David Holtz,et al.  Reducing Interference Bias in Online Marketplace Pricing Experiments , 2020, 2004.12489.

[5]  Ramesh Johari,et al.  Adaptive Experimental Design with Temporal Interference: A Maximum Likelihood Approach , 2020, NeurIPS.

[6]  Kuang Xu,et al.  Experimenting in Equilibrium , 2019, Manag. Sci..

[7]  Guillaume W. Basse,et al.  Randomization tests of causal effects under interference , 2019, Biometrika.

[8]  Jon M. Kleinberg,et al.  Graph cluster randomization: network exposure to multiple universes , 2013, KDD.

[9]  Viet Ha-Thuc,et al.  A Counterfactual Framework for Seller-Side A/B Testing on Marketplaces , 2020, SIGIR.

[10]  Randall Wright,et al.  Pricing and Matching with Frictions , 2001, Journal of Political Economy.

[11]  Edoardo M. Airoldi,et al.  Detecting Network Effects: Randomizing Over Randomized Experiments , 2017, KDD.

[12]  Corwin M Zigler,et al.  Bipartite Causal Inference with Interference. , 2018, Statistical science : a review journal of the Institute of Mathematical Statistics.

[13]  Diane Lambert,et al.  Randomization and The Pernicious Effects of Limited Budgets on Auction Experiments , 2016, AISTATS.

[14]  Thomas Blake,et al.  Why marketplace experimentation is harder than it seems: the role of test-control interference , 2014, EC.

[15]  D. Rubin,et al.  Causal Inference for Statistics, Social, and Biomedical Sciences: An Introduction , 2016 .

[16]  Charles F. Manski,et al.  Identification of Treatment Response with Social Interactions , 2013 .

[17]  R. Johari,et al.  Experimental Design in Two-Sided Platforms: An Analysis of Bias , 2020, EC.

[18]  Vahab Mirrokni,et al.  Variance Reduction in Bipartite Experiments through Correlation Clustering , 2019, NeurIPS.

[19]  David Holtz,et al.  Limiting Bias from Test-Control Interference in Online Marketplace Experiments , 2020, 2004.12162.

[20]  Andrey Fradkin,et al.  Search Frictions and the Design of Online Marketplaces , 2015, AMMA 2015.