Mathematical Reasoning via Self-supervised Skip-tree Training
暂无分享,去创建一个
Christian Szegedy | Kshitij Bansal | Markus N. Rabe | Dennis Lee | Markus Norman Rabe | Christian Szegedy | Dennis Lee | Kshitij Bansal
[1] Guillaume Lample,et al. Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.
[2] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[3] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[4] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[5] Paul A. Cairns. Informalising Formal Mathematics: Searching the Mizar Library with Latent Semantics , 2004, MKM.
[6] Guillaume Lample,et al. Deep Learning for Symbolic Mathematics , 2019, ICLR.
[7] Thibault Gauthier,et al. Initial Experiments with Statistical Conjecturing over Large Formal Corpora , 2016, FM4M/MathUI/ThEdu/DP/WIP@CIKM.
[8] Jia Deng,et al. Learning to Prove Theorems via Interacting with Proof Assistants , 2019, ICML.
[9] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[10] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[11] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[12] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[13] Dawn Xiaodong Song,et al. GamePad: A Learning Environment for Theorem Proving , 2018, ICLR.
[14] Jan Jakubuv,et al. First Neural Conjecturing Datasets and Experiments , 2020, CICM.
[15] Sarah M. Loos,et al. HOList: An Environment for Machine Learning of Higher-Order Theorem Proving (extended version) , 2019, ArXiv.
[16] Rishabh Singh,et al. Global Relational Models of Source Code , 2020, ICLR.
[17] Lei Yu,et al. Modelling High-Level Mathematical Reasoning in Mechanised Declarative Proofs , 2020, ArXiv.
[18] Sorin Lerner,et al. Generating correctness proofs with neural networks , 2019, MAPL@PLDI.
[19] Thibault Gauthier,et al. TacticToe: Learning to Reason with HOL4 Tactics , 2017, LPAR.
[20] Qingxiang Wang,et al. Exploration of neural machine translation in autoformalization of mathematics in Mizar , 2019, CPP.
[21] Bernd Finkbeiner,et al. Teaching Temporal Logics to Neural Networks , 2020, ICLR.
[22] Omer Levy,et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.
[23] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[24] Josef Urban. MPTP – Motivation, Implementation, First Experiments , 2004, Journal of Automated Reasoning.
[25] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[26] John Harrison,et al. HOL Light: A Tutorial Introduction , 1996, FMCAD.
[27] Cezary Kaliszyk,et al. Can Neural Networks Learn Symbolic Rewriting? , 2019, ArXiv.
[28] Qingxiang Wang,et al. First Experiments with Neural Translation of Informal to Formal Mathematics , 2018, CICM.
[29] Cezary Kaliszyk,et al. Learning-Assisted Automated Reasoning with Flyspeck , 2012, Journal of Automated Reasoning.
[30] Sarah M. Loos,et al. HOList: An Environment for Machine Learning of Higher Order Logic Theorem Proving , 2019, ICML.
[31] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[32] Chris Quirk,et al. Novel positional encodings to enable tree-based transformers , 2019, NeurIPS.