Mathematical Reasoning via Self-supervised Skip-tree Training
暂无分享,去创建一个
[1] Thibault Gauthier,et al. TacticToe: Learning to Reason with HOL4 Tactics , 2017, LPAR.
[2] Qingxiang Wang,et al. Exploration of neural machine translation in autoformalization of mathematics in Mizar , 2019, CPP.
[3] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[4] Guillaume Lample,et al. Deep Learning for Symbolic Mathematics , 2019, ICLR.
[5] Thibault Gauthier,et al. Initial Experiments with Statistical Conjecturing over Large Formal Corpora , 2016, FM4M/MathUI/ThEdu/DP/WIP@CIKM.
[6] Dawn Xiaodong Song,et al. GamePad: A Learning Environment for Theorem Proving , 2018, ICLR.
[7] Cezary Kaliszyk,et al. Can Neural Networks Learn Symbolic Rewriting? , 2019, ArXiv.
[8] Omer Levy,et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.
[9] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[10] Lei Yu,et al. Modelling High-Level Mathematical Reasoning in Mechanised Declarative Proofs , 2020, ArXiv.
[11] Qingxiang Wang,et al. First Experiments with Neural Translation of Informal to Formal Mathematics , 2018, CICM.
[12] Bernd Finkbeiner,et al. Teaching Temporal Logics to Neural Networks , 2020, ArXiv.
[13] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[14] John Harrison,et al. HOL Light: A Tutorial Introduction , 1996, FMCAD.
[15] Guillaume Lample,et al. Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.
[16] Cezary Kaliszyk,et al. Learning-Assisted Automated Reasoning with Flyspeck , 2012, Journal of Automated Reasoning.
[17] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[18] Josef Urban. MPTP – Motivation, Implementation, First Experiments , 2004, Journal of Automated Reasoning.
[19] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[20] Jan Jakubuv,et al. First Neural Conjecturing Datasets and Experiments , 2020, CICM.
[21] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[22] Chris Quirk,et al. Novel positional encodings to enable tree-based transformers , 2019, NeurIPS.
[23] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[24] Paul A. Cairns. Informalising Formal Mathematics: Searching the Mizar Library with Latent Semantics , 2004, MKM.
[25] Sarah M. Loos,et al. HOList: An Environment for Machine Learning of Higher Order Logic Theorem Proving , 2019, ICML.
[26] Rishabh Singh,et al. Global Relational Models of Source Code , 2020, ICLR.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] Jia Deng,et al. Learning to Prove Theorems via Interacting with Proof Assistants , 2019, ICML.