Mathematical Reasoning via Self-supervised Skip-tree Training

We examine whether self-supervised language modeling applied to mathematical formulas enables logical reasoning. We suggest several logical reasoning tasks that can be used to evaluate language models trained on formal mathematical statements, such as type inference, suggesting missing assumptions and completing equalities. To train language models for formal mathematics, we propose a novel skip-tree task. We find that models trained on the skip-tree task show surprisingly strong mathematical reasoning abilities, and outperform models trained on standard skip-sequence tasks. We also analyze the models' ability to formulate new conjectures by measuring how often the predictions are provable and useful in other proofs.

[1]  Guillaume Lample,et al.  Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.

[2]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[3]  Xiaodong Liu,et al.  Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.

[4]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[5]  Paul A. Cairns Informalising Formal Mathematics: Searching the Mizar Library with Latent Semantics , 2004, MKM.

[6]  Guillaume Lample,et al.  Deep Learning for Symbolic Mathematics , 2019, ICLR.

[7]  Thibault Gauthier,et al.  Initial Experiments with Statistical Conjecturing over Large Formal Corpora , 2016, FM4M/MathUI/ThEdu/DP/WIP@CIKM.

[8]  Jia Deng,et al.  Learning to Prove Theorems via Interacting with Proof Assistants , 2019, ICML.

[9]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[10]  Yao Zhao,et al.  PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.

[11]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[12]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[13]  Dawn Xiaodong Song,et al.  GamePad: A Learning Environment for Theorem Proving , 2018, ICLR.

[14]  Jan Jakubuv,et al.  First Neural Conjecturing Datasets and Experiments , 2020, CICM.

[15]  Sarah M. Loos,et al.  HOList: An Environment for Machine Learning of Higher-Order Theorem Proving (extended version) , 2019, ArXiv.

[16]  Rishabh Singh,et al.  Global Relational Models of Source Code , 2020, ICLR.

[17]  Lei Yu,et al.  Modelling High-Level Mathematical Reasoning in Mechanised Declarative Proofs , 2020, ArXiv.

[18]  Sorin Lerner,et al.  Generating correctness proofs with neural networks , 2019, MAPL@PLDI.

[19]  Thibault Gauthier,et al.  TacticToe: Learning to Reason with HOL4 Tactics , 2017, LPAR.

[20]  Qingxiang Wang,et al.  Exploration of neural machine translation in autoformalization of mathematics in Mizar , 2019, CPP.

[21]  Bernd Finkbeiner,et al.  Teaching Temporal Logics to Neural Networks , 2020, ICLR.

[22]  Omer Levy,et al.  SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.

[23]  Chin-Yew Lin,et al.  ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.

[24]  Josef Urban MPTP – Motivation, Implementation, First Experiments , 2004, Journal of Automated Reasoning.

[25]  Xu Tan,et al.  MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.

[26]  John Harrison,et al.  HOL Light: A Tutorial Introduction , 1996, FMCAD.

[27]  Cezary Kaliszyk,et al.  Can Neural Networks Learn Symbolic Rewriting? , 2019, ArXiv.

[28]  Qingxiang Wang,et al.  First Experiments with Neural Translation of Informal to Formal Mathematics , 2018, CICM.

[29]  Cezary Kaliszyk,et al.  Learning-Assisted Automated Reasoning with Flyspeck , 2012, Journal of Automated Reasoning.

[30]  Sarah M. Loos,et al.  HOList: An Environment for Machine Learning of Higher Order Logic Theorem Proving , 2019, ICML.

[31]  Guillaume Lample,et al.  Cross-lingual Language Model Pretraining , 2019, NeurIPS.

[32]  Chris Quirk,et al.  Novel positional encodings to enable tree-based transformers , 2019, NeurIPS.