Techniques for Symbol Grounding with SATNet

Many experts argue that the future of artificial intelligence is limited by the field’s ability to integrate symbolic logical reasoning into deep learning architectures. The recently proposed differentiable MAXSAT solver, SATNet, was a breakthrough in its capacity to integrate with a traditional neural network and solve visual reasoning problems. For instance, it can learn the rules of Sudoku purely from image examples. Despite its success, SATNet was shown to succumb to a key challenge in neurosymbolic systems known as the Symbol Grounding Problem: the inability to map visual inputs to symbolic variables without explicit supervision (“label leakage”). In this work, we present a self-supervised pre-training pipeline that enables SATNet to overcome this limitation, thus broadening the class of problems that SATNet architectures can solve to include datasets where no intermediary labels are available at all. We demonstrate that our method allows SATNet to attain full accuracy even with a harder problem setup that prevents any label leakage. We additionally introduce a proofreading method that further improves the performance of SATNet architectures, beating the state-of-the-art on Visual Sudoku.

[1]  Moshe Y. Vardi,et al.  Graph Neural Networks Meet Neural-Symbolic Computing: A Survey and Perspective , 2020, IJCAI.

[2]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[3]  David L. Dill,et al.  Learning a SAT Solver from Single-Bit Supervision , 2018, ICLR.

[4]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[5]  Geoffrey E. Hinton,et al.  Distilling the Knowledge in a Neural Network , 2015, ArXiv.

[6]  H. H. Mao,et al.  A Survey on Self-supervised Pre-training for Sequential Transfer Learning in Neural Networks , 2020, ArXiv.

[7]  Ole Winther,et al.  Recurrent Relational Networks , 2017, NeurIPS.

[8]  Yingjie Tian,et al.  A Comprehensive Survey of Clustering Algorithms , 2015, Annals of Data Science.

[9]  Pieter Abbeel,et al.  InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.

[10]  Yoshua Bengio,et al.  Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.

[11]  Priya L. Donti,et al.  SATNet: Bridging deep learning and logical reasoning using a differentiable satisfiability solver , 2019, ICML.

[12]  Sebastian Nowozin,et al.  Structured Learning and Prediction in Computer Vision , 2011, Found. Trends Comput. Graph. Vis..

[13]  Hod Lipson,et al.  Assessing SATNet's Ability to Solve the Symbol Grounding Problem , 2023, NeurIPS.

[14]  Richard Evans,et al.  Learning Explanatory Rules from Noisy Data , 2017, J. Artif. Intell. Res..

[15]  A. Freitas,et al.  ∂-Explainer: Abductive Natural Language Inference via Differentiable Convex Optimization , 2021, ArXiv.

[16]  Luc De Raedt,et al.  DeepProbLog: Neural Probabilistic Logic Programming , 2018, BNAIC/BENELEARN.

[17]  Kai-Wei Chang,et al.  An Integer Linear Programming Framework for Mining Constraints from Data , 2020, ICML.

[18]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[19]  Ferdinando Fioretto,et al.  End-to-End Constrained Optimization Learning: A Survey , 2021, IJCAI.

[20]  Toby Walsh,et al.  New Approaches to Constraint Acquisition , 2016, Data Mining and Constraint Programming.

[21]  Aleksander Madry,et al.  On Evaluating Adversarial Robustness , 2019, ArXiv.

[22]  Samuel Henrique Silva,et al.  Opportunities and Challenges in Deep Learning Adversarial Robustness: A Survey , 2020, ArXiv.

[23]  Olivier Teboul,et al.  Fast Differentiable Sorting and Ranking , 2020, ICML.

[24]  Georg Martius,et al.  Differentiation of Blackbox Combinatorial Solvers , 2020, ICLR.

[25]  Yang Yu,et al.  Tunneling Neural Perception and Logic Reasoning through Abductive Learning , 2018, ArXiv.

[26]  Julien Mairal,et al.  Unsupervised Pre-Training of Image Features on Non-Curated Data , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[27]  Cuntai Guan,et al.  A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[28]  Rich Caruana,et al.  Do Deep Nets Really Need to be Deep? , 2013, NIPS.

[29]  Ohad Shamir,et al.  Failures of Gradient-Based Deep Learning , 2017, ICML.

[30]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[31]  J. Zico Kolter,et al.  OptNet: Differentiable Optimization as a Layer in Neural Networks , 2017, ICML.

[32]  Razvan Pascanu,et al.  Policy Distillation , 2015, ICLR.

[33]  Matthew Richardson,et al.  Do Deep Convolutional Nets Really Need to be Deep and Convolutional? , 2016, ICLR.

[34]  Luis C. Lamb,et al.  Neurosymbolic AI: the 3rd wave , 2020, Artificial Intelligence Review.