Gaining Insights Into Patient Satisfaction Through Interpretable Machine Learning

Patient satisfaction is a key performance indicator of patient-centered care and hospital reimbursement. To discover the major factors that affect patient experiences is considered as an effective way to formulate corrective actions. A patient during his/her healthcare journey interacts with multiple health professionals across different service units. The health-related data generated at each step of the journey is a valuable resource for extracting actionable insights. In particular, self-reported satisfaction survey and the associated patient electronic health records play an important role in the hospital-patient interaction analysis. In this paper, we propose an interpretable machine learning framework to formulate the patient satisfaction problem as a supervised learning task and utilize a mixed-integer programming model to identify the most influential factors. The proposed framework transforms heterogeneous data into human-understandable features and integrates feature transformation, variable selection, and coefficient learning into the optimization process. Therefore, it can achieve desirable model performance while maintaining excellent model interpretability, which paves the way for successful real-world applications.

[1]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[2]  J. LaFountain Inc. , 2013, American Art.

[3]  J. Jamison PATIENT SATISFACTION , 1996, Australasian Chiropractic & Osteopathy.

[4]  W. Marsden I and J , 2012 .

[5]  Cynthia Rudin,et al.  Supersparse linear integer models for optimized medical scoring systems , 2015, Machine Learning.

[6]  Matthew P. Manary,et al.  Relationship between patient satisfaction with inpatient care and hospital readmission within 30 days. , 2011, The American journal of managed care.

[7]  Matteo Mario Savino,et al.  Determinants of patient satisfaction: a systematic review , 2017, Perspectives in public health.

[8]  A. Vozikis,et al.  Patient satisfaction with the healthcare system: Assessing the impact of socio-economic and healthcare provision factors , 2016, BMC Health Services Research.

[9]  Ari Voutilainen,et al.  A New Look at Patient Satisfaction: Learning From Self-Organizing Maps , 2014, Nursing research.

[10]  J. Kugler,et al.  Determinants of patient satisfaction: a study among 39 hospitals in an in-patient setting in Germany. , 2011, International journal for quality in health care : journal of the International Society for Quality in Health Care.

[11]  J. Graham,et al.  Missing data analysis: making it work in the real world. , 2009, Annual review of psychology.

[12]  Brandon M. Greenwell,et al.  Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.

[13]  D. Rabago,et al.  Impact of patient satisfaction ratings on physicians and clinical care , 2014, Patient preference and adherence.

[14]  Usama M. Fayyad,et al.  Multi-Interval Discretization of Continuous-Valued Attributes for Classification Learning , 1993, IJCAI.

[15]  A. Jena,et al.  Hospital value-based purchasing. , 2013, Journal of hospital medicine.

[16]  Gerhard Tutz,et al.  Regularization and Model Selection with Categorial Effect Modifiers , 2010 .

[17]  Ye Zhang,et al.  A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification , 2015, IJCNLP.

[18]  Harry N. Boone,et al.  Analyzing Likert Data , 2012, Journal of Extension.

[19]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[20]  G. Fonarow,et al.  Improving patient satisfaction through physician education, feedback, and incentives. , 2015, Journal of hospital medicine.

[21]  Cynthia Rudin,et al.  Optimized Scoring Systems: Toward Trust in Machine Learning for Healthcare and Criminal Justice , 2018, Interfaces.

[22]  Cynthia Rudin,et al.  Interpretable classification models for recidivism prediction , 2015, 1503.07810.

[23]  Ning Liu,et al.  Explainable data-driven modeling of patient satisfaction survey data , 2017, 2017 IEEE International Conference on Big Data (Big Data).

[24]  Max Kuhn,et al.  Building Predictive Models in R Using the caret Package , 2008 .

[25]  Scott Sanner,et al.  Algorithms for Direct 0-1 Loss Optimization in Binary Classification , 2013, ICML.

[26]  B. McNeil,et al.  Patient satisfaction as an indicator of quality care. , 1988, Inquiry : a journal of medical care organization, provision and financing.

[27]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[28]  Martin McKee,et al.  Patient safety, satisfaction, and quality of hospital care: cross sectional surveys of nurses and patients in 12 countries in Europe and the United States , 2012, BMJ : British Medical Journal.

[29]  Motoaki Kawanabe,et al.  How to Explain Individual Classification Decisions , 2009, J. Mach. Learn. Res..

[30]  Cynthia Rudin,et al.  Methods and Models for Interpretable Linear Classification , 2014, ArXiv.

[31]  Joel T. Dudley,et al.  Data-Driven Identification of Risk Factors of Patient Satisfaction at a Large Urban Academic Medical Center , 2016, PloS one.

[32]  Kenney Ng,et al.  Interacting with Predictions: Visual Inspection of Black-box Machine Learning Models , 2016, CHI.

[33]  Verónica Bolón-Canedo,et al.  Data discretization: taxonomy and big data challenge , 2016, WIREs Data Mining Knowl. Discov..

[34]  Cynthia Rudin,et al.  A Bayesian Framework for Learning Rule Sets for Interpretable Classification , 2017, J. Mach. Learn. Res..

[35]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[36]  Cynthia Rudin,et al.  Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model , 2015, ArXiv.