Mixed Sum-Product Networks: A Deep Architecture for Hybrid Domains

While all kinds of mixed data—from personal data, over panel and scientific data, to public and commercial data—are collected and stored, building probabilistic graphical models for these hybrid domains becomes more difficult. Users spend significant amounts of time in identifying the parametric form of the random variables (Gaussian, Poisson, Logit, etc.) involved and learning the mixed models. To make this difficult task easier, we propose the first trainable probabilistic deep architecture for hybrid domains that features tractable queries. It is based on Sum-Product Networks (SPNs) with piecewise polynomial leaf distributions together with novel nonparametric decomposition and conditioning steps using the Hirschfeld-Gebelein-Rényi Maximum Correlation Coefficient. This relieves the user from deciding a-priori the parametric form of the random variables but is still expressive enough to effectively approximate any distribution and permits efficient learning and inference. Our experiments show that the architecture, called Mixed SPNs, can indeed capture complex distributions across a wide range of hybrid domains.

[1]  Kristian Kersting,et al.  Poisson Sum-Product Networks: A Deep Architecture for Tractable Multivariate Poisson Distributions , 2017, AAAI.

[2]  Zoubin Ghahramani,et al.  Automatic Discovery of the Statistical Types of Variables in a Dataset , 2017, ICML.

[3]  Gal Elidan,et al.  Copula Bayesian Networks , 2010, NIPS.

[4]  Pradeep Ravikumar,et al.  Mixed Graphical Models via Exponential Families , 2014, AISTATS.

[5]  Barnabás Póczos,et al.  Copula-based Kernel Dependency Measures , 2012, ICML.

[6]  Constantin F. Aliferis,et al.  The max-min hill-climbing Bayesian network structure learning algorithm , 2006, Machine Learning.

[7]  Bernhard Schölkopf,et al.  The Randomized Dependence Coefficient , 2013, NIPS.

[8]  Prakash P. Shenoy,et al.  Inference in hybrid Bayesian networks using mixtures of polynomials , 2011, Int. J. Approx. Reason..

[9]  Floriana Esposito,et al.  Simplifying, Regularizing and Strengthening Sum-Product Network Structure Learning , 2015, ECML/PKDD.

[10]  Floriana Esposito,et al.  Visualizing and understanding Sum-Product Networks , 2018, Machine Learning.

[11]  Daniel Lowd,et al.  Learning Sum-Product Networks with Direct and Indirect Variable Interactions , 2014, ICML.

[12]  H. Gebelein Das statistische Problem der Korrelation als Variations‐ und Eigenwertproblem und sein Zusammenhang mit der Ausgleichsrechnung , 1941 .

[13]  Serafín Moral,et al.  Mixtures of Truncated Exponentials in Hybrid Bayesian Networks , 2001, ECSQARU.

[14]  Guy Van den Broeck,et al.  Probabilistic Inference in Hybrid Domains by Weighted Model Integration , 2015, IJCAI.

[15]  N. Wermuth,et al.  Graphical Models for Associations between Variables, some of which are Qualitative and some Quantitative , 1989 .

[16]  Denny Borsboom,et al.  Multicausal systems ask for multicausal approaches: A network perspective on subjective well-being in individuals with autism spectrum disorder , 2017, Autism : the international journal of research and practice.

[17]  Pedro M. Domingos,et al.  Sum-product networks: A new deep architecture , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[18]  Lourens J. Waldorp,et al.  mgm: Estimating Time-Varying Mixed Graphical Models in High-Dimensional Data , 2015, Journal of Statistical Software.

[19]  Guy Van den Broeck,et al.  Tractable Learning for Complex Probability Queries , 2015, NIPS.

[20]  Sebastian Tschiatschek,et al.  On Theoretical Properties of Sum-Product Networks , 2015, AISTATS.

[21]  Guy Van den Broeck,et al.  Hashing-Based Approximate Probabilistic Inference in Hybrid Domains , 2015, UAI.

[22]  Pedro M. Domingos,et al.  Learning the Structure of Sum-Product Networks , 2013, ICML.

[23]  V. Pawlowsky-Glahn,et al.  Latent Compositional Factors in The Llobregat River Basin (Spain) Hydrogeochemistry , 2005 .

[24]  Adnan Darwiche,et al.  On Relaxing Determinism in Arithmetic Circuits , 2017, ICML.

[25]  AI Koan,et al.  Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning , 2008, NIPS.

[26]  Christian Bauckhage,et al.  Descriptive matrix factorization for sustainability Adopting the principle of opposites , 2011, Data Mining and Knowledge Discovery.

[27]  Ursula Gather,et al.  Combining regular and irregular histograms by penalized likelihood , 2010, Comput. Stat. Data Anal..

[28]  J. Gower A General Coefficient of Similarity and Some of Its Properties , 1971 .

[29]  Scott Sanner,et al.  Symbolic Variable Elimination for Discrete and Continuous Graphical Models , 2012, AAAI.

[30]  Wei-Chen Cheng,et al.  Language modeling with sum-product networks , 2014, INTERSPEECH.

[31]  Reza Modarres,et al.  Measures of Dependence , 2011, International Encyclopedia of Statistical Science.

[32]  David M. Blei,et al.  Probabilistic topic models , 2012, Commun. ACM.

[33]  David Heckerman,et al.  Learning Bayesian Networks: A Unification for Discrete and Gaussian Domains , 1995, UAI.

[34]  C. Ji An Archetypal Analysis on , 2005 .

[35]  Andrea Passerini,et al.  Efficient Weighted Model Integration via SMT-Based Predicate Abstraction , 2017, IJCAI.

[36]  Han Zhao,et al.  On the Relationship between Sum-Product Networks and Bayesian Networks , 2015, ICML.

[37]  Rafael Rumí,et al.  Mixtures of truncated basis functions , 2012, Int. J. Approx. Reason..