Mathematical Reasoning via Self-supervised Skip-tree Training

We examine whether self-supervised language modeling applied to mathematical formulas enables logical reasoning. We suggest several logical reasoning tasks that can be used to evaluate language models trained on formal mathematical statements, such as type inference, suggesting missing assumptions and completing equalities. To train language models for formal mathematics, we propose a novel skip-tree task. We find that models trained on the skip-tree task show surprisingly strong mathematical reasoning abilities, and outperform models trained on standard skip-sequence tasks. We also analyze the models' ability to formulate new conjectures by measuring how often the predictions are provable and useful in other proofs.

[1]  Thibault Gauthier,et al.  TacticToe: Learning to Reason with HOL4 Tactics , 2017, LPAR.

[2]  Qingxiang Wang,et al.  Exploration of neural machine translation in autoformalization of mathematics in Mizar , 2019, CPP.

[3]  Chin-Yew Lin,et al.  ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.

[4]  Guillaume Lample,et al.  Deep Learning for Symbolic Mathematics , 2019, ICLR.

[5]  Thibault Gauthier,et al.  Initial Experiments with Statistical Conjecturing over Large Formal Corpora , 2016, FM4M/MathUI/ThEdu/DP/WIP@CIKM.

[6]  Dawn Xiaodong Song,et al.  GamePad: A Learning Environment for Theorem Proving , 2018, ICLR.

[7]  Cezary Kaliszyk,et al.  Can Neural Networks Learn Symbolic Rewriting? , 2019, ArXiv.

[8]  Omer Levy,et al.  SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.

[9]  Guillaume Lample,et al.  Cross-lingual Language Model Pretraining , 2019, NeurIPS.

[10]  Lei Yu,et al.  Modelling High-Level Mathematical Reasoning in Mechanised Declarative Proofs , 2020, ArXiv.

[11]  Qingxiang Wang,et al.  First Experiments with Neural Translation of Informal to Formal Mathematics , 2018, CICM.

[12]  Bernd Finkbeiner,et al.  Teaching Temporal Logics to Neural Networks , 2020, ArXiv.

[13]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[14]  John Harrison,et al.  HOL Light: A Tutorial Introduction , 1996, FMCAD.

[15]  Guillaume Lample,et al.  Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.

[16]  Cezary Kaliszyk,et al.  Learning-Assisted Automated Reasoning with Flyspeck , 2012, Journal of Automated Reasoning.

[17]  Yao Zhao,et al.  PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.

[18]  Josef Urban MPTP – Motivation, Implementation, First Experiments , 2004, Journal of Automated Reasoning.

[19]  Xu Tan,et al.  MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.

[20]  Jan Jakubuv,et al.  First Neural Conjecturing Datasets and Experiments , 2020, CICM.

[21]  Xiaodong Liu,et al.  Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.

[22]  Chris Quirk,et al.  Novel positional encodings to enable tree-based transformers , 2019, NeurIPS.

[23]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[24]  Paul A. Cairns Informalising Formal Mathematics: Searching the Mizar Library with Latent Semantics , 2004, MKM.

[25]  Sarah M. Loos,et al.  HOList: An Environment for Machine Learning of Higher Order Logic Theorem Proving , 2019, ICML.

[26]  Rishabh Singh,et al.  Global Relational Models of Source Code , 2020, ICLR.

[27]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[28]  Jia Deng,et al.  Learning to Prove Theorems via Interacting with Proof Assistants , 2019, ICML.