How Mega Is the Mega? Measuring the Spillover Effects of WeChat by Machine Learning and Econometrics

WeChat, an instant messaging app, is considered a mega app due to its dominance in terms of usage among Chinese smartphone users. Nevertheless, little is known about its externality in regard to the broader app market. Our work estimates the spillover effects of WeChat on the other top-50 most frequently used apps in China through data on users’ weekly app usage. Given the challenge of determining causal inference from observational data, we apply a graphical model and econometrics to estimate the spillover effects through two steps: (1) we determine the causal structure by estimating a partially ancestral diagram, using a Fast Causal Inference (FCI) algorithm; (2) given the causal structure, we find a valid adjustment set and estimate the causal effects by an econometric model with the adjustment set as controlling non-causal effects. Our findings show that the spillover effects of WeChat are limited; in fact, only two other apps, Tencent News and Taobao, receive positive spillover effects from WeChat. In addition, we show that, if researchers fail to account for the causal structure that we determined from the graphical model, it is easy to fall into the trap of confounding bias and selection bias when estimating causal effects.

[1]  Jean-Baptiste Denis,et al.  Bayesian Networks , 2014 .

[2]  Judea Pearl,et al.  Complete Identification Methods for the Causal Hierarchy , 2008, J. Mach. Learn. Res..

[3]  Gregory F. Cooper,et al.  A Bayesian method for the induction of probabilistic networks from data , 1992, Machine Learning.

[4]  Sang Pil Han,et al.  Estimating Demand for Mobile Applications in the New Economy , 2014, Manag. Sci..

[5]  Jiji Zhang,et al.  On the completeness of orientation rules for causal discovery in the presence of latent confounders and selection bias , 2008, Artif. Intell..

[6]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[7]  John A. W. McCall,et al.  A chain-model genetic algorithm for Bayesian network structure learning , 2007, GECCO '07.

[8]  Luis M. de Campos,et al.  A new approach for learning belief networks using independence criteria , 2000, Int. J. Approx. Reason..

[9]  P. Spirtes,et al.  Causation, prediction, and search , 1993 .

[10]  J. Pearl [Bayesian Analysis in Expert Systems]: Comment: Graphical Models, Causality and Intervention , 1993 .

[11]  P. Spirtes,et al.  Ancestral graph Markov models , 2002 .

[12]  Jiji Zhang,et al.  Causal Reasoning with Ancestral Graphs , 2008, J. Mach. Learn. Res..

[13]  Judea Pearl,et al.  Equivalence and Synthesis of Causal Models , 1990, UAI.

[14]  Peter Bühlmann,et al.  Causal Inference Using Graphical Models with the R Package pcalg , 2012 .

[15]  Judea Pearl,et al.  The International Journal of Biostatistics C AUSAL I NFERENCE An Introduction to Causal Inference , 2011 .

[16]  Peter Bühlmann,et al.  Estimating High-Dimensional Directed Acyclic Graphs with the PC-Algorithm , 2007, J. Mach. Learn. Res..

[17]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[18]  Judea Pearl,et al.  Reverend Bayes on Inference Engines: A Distributed Hierarchical Approach , 1982, AAAI.

[19]  Marco Valtorta,et al.  Identifiability in Causal Bayesian Networks: A Sound and Complete Algorithm , 2006, AAAI.

[20]  David Maxwell Chickering,et al.  Learning Bayesian Networks: The Combination of Knowledge and Statistical Data , 1994, Machine Learning.

[21]  Pedro Larrañaga,et al.  Structure Learning of Bayesian Networks by Hybrid Genetic Algorithms , 1995, AISTATS.

[22]  Diego Colombo,et al.  A generalized backdoor criterion , 2013, ArXiv.

[23]  Octavian Carare,et al.  The Impact of Bestseller Rank on Demand: Evidence from the App Market , 2011 .

[24]  J. Pearl Causal diagrams for empirical research , 1995 .

[25]  Richard E. Neapolitan,et al.  Learning Bayesian networks , 2007, KDD '07.

[26]  Constantin F. Aliferis,et al.  The max-min hill-climbing Bayesian network structure learning algorithm , 2006, Machine Learning.