Learning Credal Sum-Product Networks

Probabilistic representations, such as Bayesian and Markov networks, are fundamental to much of statistical machine learning. Thus, learning probabilistic representations directly from data is a deep challenge, the main computational bottleneck being inference that is intractable. Tractable learning is a powerful new paradigm that attempts to learn distributions that support efficient probabilistic querying. By leveraging local structure, representations such as sum-product networks (SPNs) can capture high tree-width models with many hidden layers, essentially a deep architecture, while still admitting a range of probabilistic queries to be computable in time polynomial in the network size. While the progress is impressive, numerous data sources are incomplete, and in the presence of missing data, structure learning methods nonetheless revert to single distributions without characterizing the loss in confidence. In recent work, credal sum-product networks, an imprecise extension of sum-product networks, were proposed to capture this robustness angle. In this work, we are interested in how such representations can be learnt and thus study how the computational machinery underlying tractable learning and inference can be generalized for imprecise probabilities.

[1]  Toniann Pitassi,et al.  Solving #SAT and Bayesian Inference with Backtracking Search , 2014, J. Artif. Intell. Res..

[2]  Pedro M. Domingos,et al.  Learning Relational Sum-Product Networks , 2015, AAAI.

[3]  Michael I. Jordan,et al.  Thin Junction Trees , 2001, NIPS.

[4]  G. Shafer The Enterprise of Knowledge: An Essay on Knowledge, Credal Probability, and Chance , 1982 .

[5]  Fabio Gagliardi Cozman,et al.  Credal networks , 2000, Artif. Intell..

[6]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[7]  L. Zadeh Fuzzy sets as a basis for a theory of possibility , 1999 .

[8]  K. Wagstaff Clustering with Missing Values: No Imputation Required , 2004 .

[9]  Kristian Kersting,et al.  Mixed Sum-Product Networks: A Deep Architecture for Hybrid Domains , 2018, AAAI.

[10]  Salem Benferhat,et al.  Approximating MAP Inference in Credal Networks Using Probability-Possibility Transformations , 2017, 2017 IEEE 29th International Conference on Tools with Artificial Intelligence (ICTAI).

[11]  Pascal Poupart,et al.  Online Structure Learning for Sum-Product Networks with Gaussian Leaves , 2017, ICLR.

[12]  Denis Deratani Mauá,et al.  Credal Sum-Product Networks , 2017, ISIPTA.

[13]  Pedro M. Domingos,et al.  Discriminative Learning of Sum-Product Networks , 2012, NIPS.

[14]  Stef van Buuren,et al.  Flexible Imputation of Missing Data , 2012 .

[15]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[16]  Pedro M. Domingos,et al.  Learning the Structure of Sum-Product Networks , 2013, ICML.

[17]  Karim Tabia,et al.  Learning the Parameters of Possibilistic Networks from Data: Empirical Comparison , 2017, FLAIRS Conference.

[18]  Vaishak Belle,et al.  Tractable Querying and Learning in Hybrid Domains via Sum-Product Networks , 2018, ArXiv.

[19]  Anton Wallner Extreme points of coherent probabilities in finite spaces , 2007, Int. J. Approx. Reason..

[20]  Luc De Raedt,et al.  Inference in Probabilistic Logic Programs using Weighted CNF's , 2011, UAI.

[21]  Vaishak Belle,et al.  Open-Universe Weighted Model Counting , 2017, AAAI.

[22]  Adnan Darwiche,et al.  On probabilistic inference by weighted model counting , 2008, Artif. Intell..

[23]  Glenn Shafer,et al.  A Mathematical Theory of Evidence , 2020, A Mathematical Theory of Evidence.

[24]  Leslie G. Valiant,et al.  The Complexity of Enumeration and Reliability Problems , 1979, SIAM J. Comput..

[25]  Guy Van den Broeck,et al.  Learning the Structure of Probabilistic Sentential Decision Diagrams , 2017, UAI.

[26]  Adnan Darwiche,et al.  A differential approach to inference in Bayesian networks , 2000, JACM.

[27]  Manfred Jaeger,et al.  Compiling relational Bayesian networks for exact inference , 2006, Int. J. Approx. Reason..

[28]  Joseph Y. Halpern Reasoning about uncertainty , 2003 .

[29]  Thierry Denoeux,et al.  Maximum Likelihood Estimation from Uncertain Data in the Belief Function Framework , 2013, IEEE Transactions on Knowledge and Data Engineering.

[30]  Pedro M. Domingos,et al.  Sum-product networks: A new deep architecture , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[31]  Luc De Raedt,et al.  Lifted Probabilistic Inference by First-Order Knowledge Compilation , 2011, IJCAI.

[32]  Martin Grohe,et al.  Probabilistic Databases with an Infinite Open-World Assumption , 2018, PODS.

[33]  Guy Van den Broeck,et al.  Open World Probabilistic Databases (Extended Abstract) , 2016, Description Logics.

[34]  Leonid Libkin,et al.  On Querying Incomplete Information in Databases under Bag Semantics , 2017, IJCAI.

[35]  Leonid Libkin,et al.  SQL’s Three-Valued Logic and Certain Answers , 2016, TODS.

[36]  Prasoon Goyal,et al.  Probabilistic Databases , 2009, Encyclopedia of Database Systems.

[37]  Stuart J. Russell,et al.  Unifying logic and probability , 2015, Commun. ACM.

[38]  Matthias C. M. Troffaes,et al.  Introduction to imprecise probabilities , 2014 .

[39]  Raymond Reiter On Closed World Data Bases , 1977, Logic and Data Bases.