Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond

A fundamental goal of scientific research is to learn about causal relationships. However, despite its critical role in the life and social sciences, causality has not had the same importance in Natural Language Processing (NLP), which has traditionally placed more emphasis on predictive tasks. This distinction is beginning to fade, with an emerging area of interdisciplinary research at the convergence of causal inference and language processing. Still, research on causality in NLP remains scattered across domains without unified definitions, benchmark datasets and clear articulations of the remaining challenges. In this survey, we consolidate research across academic areas and situate it in the broader NLP landscape. We introduce the statistical challenge of estimating causal effects, encompassing settings where text is used as an outcome, treatment, or as a means to address confounding. In addition, we explore potential uses of causal inference to improve the performance, robustness, fairness, and interpretability of NLP models. We thus provide a unified overview of causal inference for the computational linguistics community. 1 ∗All authors equally contributed to this paper. Author names are organized alphabetically in two clusters: First students and post-docs and then faculty members. The email address of the corresponding (first) author is: feder@campus.technion.ac.il. An online repository containing existing research on causal inference and language processing is available here: https://github.com/causaltext/

[1]  Yonatan Belinkov,et al.  Investigating Gender Bias in Language Models Using Causal Mediation Analysis , 2020, NeurIPS.

[2]  David G. Rand,et al.  Structural Topic Models for Open‐Ended Survey Responses , 2014, American Journal of Political Science.

[3]  Zhao Wang,et al.  Identifying spurious correlations for robust text classification , 2020, FINDINGS.

[4]  Po-Sen Huang,et al.  Reducing Sentiment Bias in Language Models via Counterfactual Evaluation , 2019, FINDINGS.

[5]  Bernhard Scholkopf Causality for Machine Learning , 2019 .

[6]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[7]  Margaret E. Roberts,et al.  Adjusting for Confounding with Text Matching , 2020 .

[8]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[9]  Lise Getoor,et al.  Estimating Causal Effects of Tone in Online Debates , 2019, IJCAI.

[10]  Omer Levy,et al.  Annotation Artifacts in Natural Language Inference Data , 2018, NAACL.

[11]  Martin Wattenberg,et al.  Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.

[12]  Margaret E. Roberts,et al.  How to make causal inferences using texts , 2018, Science advances.

[13]  Franco Turini,et al.  A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..

[14]  Judea Pearl,et al.  A Probabilistic Calculus of Actions , 1994, UAI.

[15]  Jieyu Zhao,et al.  Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints , 2017, EMNLP.

[16]  Virgile Landeiro,et al.  Robust Text Classification under Confounding Shift , 2018, J. Artif. Intell. Res..

[17]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[18]  Luke S. Zettlemoyer,et al.  Deep Contextualized Word Representations , 2018, NAACL.

[19]  Sameer Singh,et al.  Beyond Accuracy: Behavioral Testing of NLP Models with CheckList , 2020, ACL.

[20]  Yoav Goldberg,et al.  Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals , 2021, Transactions of the Association for Computational Linguistics.

[21]  Jonas Peters,et al.  Causal inference by using invariant prediction: identification and confidence intervals , 2015, 1501.01332.

[22]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[23]  Dhanya Sridhar,et al.  Adapting Text Embeddings for Causal Inference , 2020, UAI.

[24]  David C. Uthus,et al.  TextSETTR: Label-Free Text Style Extraction and Tunable Targeted Restyling , 2021, ArXiv.

[25]  Yonatan Belinkov,et al.  Causal Analysis of Syntactic Agreement Mechanisms in Neural Language Models , 2021, ACL.

[26]  Barbara F. Walter,et al.  The Gender Citation Gap in International Relations , 2013, International Organization.

[27]  Daniel Jurafsky,et al.  Deconfounded Lexicon Induction for Interpretable Social Science , 2018, NAACL.

[28]  Ryan Cotterell,et al.  It’s All in the Name: Mitigating Gender Bias with Name-Based Counterfactual Data Substitution , 2019, EMNLP.

[29]  Katherine A. Keith,et al.  Text and Causal Inference: A Review of Using Text to Remove Confounding from Causal Estimates , 2020, ACL.

[30]  Rotem Dror,et al.  Replicability Analysis for Natural Language Processing: Testing Significance with Multiple Datasets , 2017, TACL.

[31]  Amit Sharma,et al.  Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.

[32]  Alexander D'Amour,et al.  Counterfactual Invariance to Spurious Correlations: Why and How to Pass Stress Tests , 2021, ArXiv.

[33]  Justin Grimmer,et al.  Discovery of Treatments from Text Corpora , 2016, ACL.

[34]  Uri Shalit,et al.  CausaLM: Causal Model Explanation Through Counterfactual Language Models , 2020, CL.

[35]  JUSTINE ZHANG,et al.  Quantifying the Causal Effects of Conversational Tendencies , 2020, Proc. ACM Hum. Comput. Interact..

[36]  Li Zhao,et al.  Attention-based LSTM for Aspect-level Sentiment Classification , 2016, EMNLP.

[37]  Mark Dredze,et al.  Challenges of Using Text Classifiers for Causal Inference , 2018, EMNLP.

[38]  Alexander D'Amour,et al.  Overlap in observational studies with high-dimensional covariates , 2017, Journal of Econometrics.

[39]  Issa Kohler-Hausmann Eddie Murphy and the Dangers of Counterfactual Causal Thinking About Detecting Racial Discrimination , 2019 .

[40]  Dhanya Sridhar,et al.  Causal Effects of Linguistic Properties , 2020, NAACL.

[41]  Ryan Cotterell,et al.  Counterfactual Data Augmentation for Mitigating Gender Stereotypes in Languages with Rich Morphology , 2019, ACL.

[42]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[43]  Daniel Jurafsky,et al.  Predicting Sales from the Language of Product Descriptions , 2017, eCOM@SIGIR.

[44]  Chris Russell,et al.  Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.

[45]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[46]  Jeffrey Heer,et al.  Polyjuice: Automated, General-purpose Counterfactual Generation , 2021, ArXiv.

[47]  Sandro Pezzelle,et al.  FOIL it! Find One mismatch between Image and Language caption , 2017, ACL.

[48]  Jieyu Zhao,et al.  Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods , 2018, NAACL.

[49]  Junmo Kim,et al.  Learning Not to Learn: Training Deep Neural Networks With Biased Data , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[50]  Bernhard Schölkopf,et al.  On causal and anticausal learning , 2012, ICML.

[51]  Christopher W. Larimer,et al.  Social Pressure and Voter Turnout: Evidence from a Large-Scale Field Experiment , 2008, American Political Science Review.

[52]  Luke Miratrix,et al.  Matching with Text Data: An Experimental Evaluation of Methods for Matching Documents and of Measuring Match Quality , 2018, Political Analysis.

[53]  Bernhard Schölkopf,et al.  Domain Generalization via Invariant Feature Representation , 2013, ICML.

[54]  Mengjie Zhang,et al.  Domain Generalization for Object Recognition with Multi-task Autoencoders , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[55]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[56]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[57]  Richard Zemel,et al.  Fairness and Robustness in Invariant Learning: A Case Study in Toxicity Classification , 2020, ArXiv.

[58]  Carolyn Penstein Rosé,et al.  Stress Test Evaluation for Natural Language Inference , 2018, COLING.

[59]  Yoav Goldberg,et al.  Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection , 2020, ACL.

[60]  Noah A. Smith,et al.  Evaluating Models’ Local Decision Boundaries via Contrast Sets , 2020, FINDINGS.