Understanding from Machine Learning Models

Simple idealized models seem to provide more understanding than opaque, complex, and hyper-realistic models. However, an increasing number of scientists are going in the opposite direction by utilizing opaque machine learning models to make predictions and draw inferences, suggesting that scientists are opting for models that have less potential for understanding. Are scientists trading understanding for some other epistemic or pragmatic good when they choose a machine learning model? Or are the assumptions behind why minimal models provide understanding misguided? In this paper, using the case of deep neural networks, I argue that it is not the complexity or black box nature of a model that limits how much understanding the model provides. Instead, it is a lack of scientific and empirical evidence supporting the link that connects a model to the target phenomenon that primarily prohibits understanding.

[1]  Joshua B. Tenenbaum,et al.  Human-level concept learning through probabilistic program induction , 2015, Science.

[2]  Zach Blas,et al.  “Escaping the Face: Biometric Facial Recognition and the Facial Weaponization Suite” , 2013 .

[3]  Collin Rice,et al.  Hypothetical Pattern Idealization and Explanatory Models , 2013, Philosophy of Science.

[4]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[5]  Kareem Khalifa,et al.  Understanding Without Explanation , 2017 .

[6]  Kareem Khalifa,et al.  Idealizations and Understanding: Much Ado About Nothing? , 2019, Australasian Journal of Philosophy.

[7]  Robert W. Batterman,et al.  Minimal Model Explanations , 2014, Philosophy of Science.

[8]  Jan Havlíček,et al.  Shape Differences Between the Faces of Homosexual and Heterosexual Men , 2013, Archives of Sexual Behavior.

[9]  B. Mustanski,et al.  A Critical Review of Recent Biological Research on Human Sexual Orientation , 2002, Annual review of sex research.

[10]  H. W. de Regt,et al.  Understanding Scientific Understanding , 2017 .

[11]  W. Clark,et al.  Residential preferences and neighborhood racial segregation: A test of the schelling segregation model , 1991, Demography.

[12]  Eric Winsberg Values and Uncertainties in the Predictions of Global Climate Models , 2012, Kennedy Institute of Ethics journal.

[13]  Coryn A. L. Bailer-Jones,et al.  Modeling Data: Analogies in Neural Networks, Simulated Annealing and Genetic Algorithms , 2002 .

[14]  Paul Humphreys,et al.  Extending Ourselves: Computational Science, Empiricism, and Scientific Method , 2004 .

[15]  Naftali Tishby,et al.  Deep learning and the information bottleneck principle , 2015, 2015 IEEE Information Theory Workshop (ITW).

[16]  Stephan Hartmann,et al.  Understanding (with) Toy Models , 2018, The British Journal for the Philosophy of Science.

[17]  Cameron Buckner,et al.  Empiricism without magic: transformational abstraction in deep convolutional neural networks , 2018, Synthese.

[18]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[19]  Michael S. Lew,et al.  Deep learning for visual understanding: A review , 2016, Neurocomputing.

[20]  Heather Douglas Inductive Risk and Values in Science , 2000, Philosophy of Science.

[21]  Sebastian Thrun,et al.  Dermatologist-level classification of skin cancer with deep neural networks , 2017, Nature.

[22]  Kareem Khalifa Understanding, Explanation, and Scientific Knowledge , 2017 .

[23]  Ryan Muldoon,et al.  Segregation That No One Seeks* , 2012, Philosophy of Science.

[24]  Li Li,et al.  Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records , 2016, Scientific Reports.

[25]  Parisa Rashidi,et al.  Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis , 2017, IEEE Journal of Biomedical and Health Informatics.

[26]  K. Elliott Douglas on values: From indirect roles to multiple goals , 2013 .

[27]  M.N. Sastry,et al.  Structure and interpretation of computer programs , 1986, Proceedings of the IEEE.

[28]  Till Grüne-Yanoff,et al.  Learning from Minimal Economic Models , 2009 .

[29]  Nello Cristianini Are We There Yet? , 2009, ECML/PKDD.

[30]  Insa Lawler,et al.  Understanding why, knowing why, and cognitive achievements , 2019, Synthese.

[31]  Explanation and understanding , 2011 .

[32]  Stephen R. Grimm The goal of explanation , 2010 .

[33]  Jenna Burrell,et al.  How the machine ‘thinks’: Understanding opacity in machine learning algorithms , 2016 .

[34]  L. Bobo,et al.  Attitudes on Residential Integration: Perceived Status Differences, Mere In-Group Preference, or Racial Prejudice? , 1996 .

[35]  Arianne E. Miller Searching for gaydar: Blind spots in the study of sexual orientation perception , 2018 .

[36]  M. Kosinski,et al.  Deep Neural Networks Are More Accurate Than Humans at Detecting Sexual Orientation From Facial Images , 2018, Journal of personality and social psychology.

[37]  Charles Weijer,et al.  Queer Science: The Use and Abuse of Research Into Homosexuality , 1996 .

[38]  M. Strevens Depth: An Account of Scientific Explanation , 2008 .

[39]  Jaakko Kuorikoski,et al.  External representations and scientific understanding , 2014, Synthese.