Learning and T-Norms Theory

Deep learning has been shown to achieve impressive results in several domains like computer vision and natural language processing. Deep architectures are typically trained following a supervised scheme and, therefore, they rely on the availability of a large amount of labeled training data to effectively learn their parameters. Neuro-symbolic approaches have recently gained popularity to inject prior knowledge into a deep learner without requiring it to induce this knowledge from data. These approaches can potentially learn competitive solutions with a significant reduction of the amount of supervised data. A large class of neuro-symbolic approaches is based on First-Order Logic to represent prior knowledge, that is relaxed to a differentiable form using fuzzy logic. This paper shows that the loss function expressing these neuro-symbolic learning tasks can be unambiguously determined given the selection of a t-norm generator. When restricted to simple supervised learning, the presented theoretical apparatus provides a clean justification to the popular cross-entropy loss, that has been shown to provide faster convergence and to reduce the vanishing gradient problem in very deep structures. One advantage of the proposed learning formulation is that it can be extended to all the knowledge that can be represented by a neuro-symbolic method, and it allows the development of a novel class of loss functions, that the experimental results show to lead to faster convergence rates than other approaches previously proposed in the literature.

[1]  Artur S. d'Avila Garcez,et al.  Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge , 2016, NeSy@HLAI.

[2]  Guy Van den Broeck,et al.  A Semantic Loss Function for Deep Learning with Symbolic Knowledge , 2017, ICML.

[3]  Luc De Raedt,et al.  DeepProbLog: Neural Probabilistic Logic Programming , 2018, BNAIC/BENELEARN.

[4]  Nikhil Ketkar,et al.  Introduction to PyTorch , 2021, Deep Learning with Python.

[5]  Marco Gori,et al.  Image Classification Using Deep Learning and Prior Knowledge , 2018, AAAI Workshops.

[6]  Jennifer Neville,et al.  Iterative Classification in Relational Data , 2000 .

[7]  Petr Hájek,et al.  Metamathematics of Fuzzy Logic , 1998, Trends in Logic.

[8]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[9]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[10]  Stephen H. Bach,et al.  Hinge-Loss Markov Random Fields and Probabilistic Soft Logic , 2015, J. Mach. Learn. Res..

[11]  James R. Foulds,et al.  Collective Spammer Detection in Evolving Multi-Relational Social Networks , 2015, KDD.

[12]  Lise Getoor,et al.  Collective Classification in Network Data , 2008, AI Mag..

[13]  Lise Getoor,et al.  Link-Based Classification , 2003, Encyclopedia of Machine Learning and Data Mining.

[14]  Krysia Broda,et al.  Neural-symbolic learning systems - foundations and applications , 2012, Perspectives in neural computing.

[15]  Artur S. d'Avila Garcez,et al.  Learning and reasoning in logic tensor networks: theory and application to semantic image interpretation , 2017, SAC.

[16]  Marco Gori,et al.  Semantic-based regularization for learning and inference , 2017, Artif. Intell..

[17]  Luc De Raedt,et al.  ProbLog: A Probabilistic Prolog and its Application in Link Discovery , 2007, IJCAI.

[18]  Marco Gori,et al.  LYRICS: a General Interface Layer to Integrate AI and Deep Learning , 2019, ArXiv.

[19]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[20]  Marco Gori,et al.  Integrating Learning and Reasoning with Deep Logic Models , 2019, ECML/PKDD.

[21]  Luc De Raedt,et al.  Learning SMT(LRA) Constraints using SMT Solvers , 2018, IJCAI.

[22]  Matthew Richardson,et al.  Markov logic networks , 2006, Machine Learning.

[23]  P. Mostert,et al.  On the Structure of Semigroups on a Compact Manifold With Boundary , 1957 .

[24]  Marcello Sanguineti,et al.  Foundations of Support Constraint Machines , 2015, Neural Computation.

[25]  Artur S. d'Avila Garcez,et al.  Logic Tensor Networks for Semantic Image Interpretation , 2017, IJCAI.

[26]  Sándor Jenei,et al.  A note on the ordinal sum theorem and its consequence for the construction of triangular norms , 2002, Fuzzy Sets Syst..

[27]  Radko Mesiar,et al.  Triangular norms. Position paper I: basic analytical and algebraic properties , 2004, Fuzzy Sets Syst..

[28]  Marco Gori,et al.  On a Convex Logic Fragment for Learning and Reasoning , 2018, IEEE Transactions on Fuzzy Systems.

[29]  Radko Mesiar,et al.  Triangular norms. Position paper III: continuous t-norms , 2004, Fuzzy Sets Syst..

[30]  Radko Mesiar,et al.  Triangular norms. Position paper II: general constructions and parameterized families , 2004, Fuzzy Sets Syst..