Investigating Numeracy Learning Ability of a Text-to-Text Transfer Model

The transformer-based pre-trained language models have been tremendously successful in most of the conventional NLP tasks. But they often struggle in those tasks where numerical understanding is required. Some possible reasons can be the tokenizers and pre-training objectives which are not specifically designed to learn and preserve numeracy. Here we investigate the ability of text-to-text transfer learning model (T5), which has outperformed its predecessors in the conventional NLP tasks, to learn numeracy. We consider four numeracy tasks : numeration, magnitude order prediction, finding minimum and maximum in a series, and sorting. We find that, although T5 models perform reasonably well in the interpolation setting, they struggle considerably in the extrapolation setting across all four tasks.

[1]  Rahul Khanna,et al.  Birds have four legs?! NumerSense: Probing Numerical Commonsense Knowledge of Pre-trained Language Models , 2020, EMNLP.

[2]  Alexander Rosenberg Johansen,et al.  Neural Arithmetic Units , 2020, ICLR.

[3]  Hiroya Takamura,et al.  Numeracy-600K: Learning Numeracy for Detecting Exaggerated Information in Market Comments , 2019, ACL.

[4]  D. Roth,et al.  Do Language Embeddings capture Scales? , 2020, FINDINGS.

[5]  Sameer Singh,et al.  Do NLP Models Know Numbers? Probing Numeracy in Embeddings , 2019, EMNLP.

[6]  Jonathan Berant,et al.  Injecting Numerical Reasoning Skills into Language Models , 2020, ACL.

[7]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[8]  Carolyn Penstein Rosé,et al.  Exploring Numeracy in Word Embeddings , 2019, ACL.

[9]  Jian Zhang,et al.  SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.

[10]  Neeraj Varshney,et al.  Towards Question Format Independent Numerical Reasoning: A Set of Prerequisite Tasks , 2020, ArXiv.

[11]  Pedro A. Szekely,et al.  Representing Numbers in NLP: a Survey and a Vision , 2021, NAACL.

[12]  Chris Dyer,et al.  Neural Arithmetic Logic Units , 2018, NeurIPS.

[13]  Gabriel Stanovsky,et al.  DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs , 2019, NAACL.

[14]  Kewei Tu,et al.  Learning Numeral Embedding , 2019, FINDINGS.

[15]  Drew Barker,et al.  Probing for Multilingual Numerical Understanding in Transformer-Based Language Models , 2020, BLACKBOXNLP.

[16]  Zhiying Jiang,et al.  Investigating the Limitations of the Transformers with Simple Arithmetic Tasks , 2021, ArXiv.

[17]  Daniel Cer,et al.  NT5?! Training T5 to Perform Numerical Reasoning , 2021, ArXiv.