Predicting Paper Acceptance via Interpretable Decision Sets

Measuring the quality of research work is an essential component of the scientific process. With the ever-growing rates of articles being submitted to top-tier conferences, and the potential consistency and bias issues in the peer review process identified by scientific community, it is thus of great necessary and challenge to automatically evaluate submissions. Existing works mainly focus on exploring relevant factors and applying machine learning models to simply be accurate at predicting the acceptance of a given academic paper, while ignoring the interpretability power which is required by a wide range of applications. In this paper, we propose a framework to construct decision sets that consist of unordered if-then rules for predicting paper acceptance. We formalize decision set learning problem via a joint objective function that simultaneously optimize accuracy and interpretability of the rules, rather than organizing them in a hierarchy. We evaluate the effectiveness of the proposed framework by applying it on a public scientific peer reviews dataset. Experimental results demonstrate that the learned interpretable decision sets by our framework performs on par with state-of-the-art classification algorithms which optimize exclusively for predictive accuracy and much more interpretable than rule-based methods.

[1]  Xiaoxia Zhang,et al.  Uncovering and Predicting the Dynamic Process of Collective Attention with Survival Theory , 2017, Scientific Reports.

[2]  Jie Tang,et al.  ArnetMiner: extraction and mining of academic social networks , 2008, KDD.

[3]  Bhavana Dalvi,et al.  A Dataset of Peer Reviews (PeerRead): Collection, Insights and NLP Applications , 2018, NAACL.

[4]  Peter Clark,et al.  The CN2 Induction Algorithm , 1989, Machine Learning.

[5]  Cynthia Rudin,et al.  Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model , 2015, ArXiv.

[6]  Vipin Balachandran,et al.  Reducing human effort and improving quality in peer code reviews using automatic static analysis and reviewer recommendation , 2013, 2013 35th International Conference on Software Engineering (ICSE).

[7]  Ramakrishnan Srikant,et al.  Fast Algorithms for Mining Association Rules in Large Databases , 1994, VLDB.

[8]  Katsiaryna Mirylenka,et al.  Alternatives to Peer Review: Novel Approaches for Research Evaluation , 2011, Front. Comput. Neurosci..

[9]  Samir Khuller,et al.  The Budgeted Maximum Coverage Problem , 1999, Inf. Process. Lett..

[10]  Albert-László Barabási,et al.  Collective credit allocation in science , 2014, Proceedings of the National Academy of Sciences.

[11]  A. Barabasi,et al.  Quantifying the evolution of individual scientific impact , 2016, Science.

[12]  Yizhou Sun,et al.  Task-Guided and Path-Augmented Heterogeneous Network Embedding for Author Identification , 2016, WSDM.

[13]  Albert-László Barabási,et al.  Quantifying Long-Term Scientific Impact , 2013, Science.

[14]  Foster J. Provost,et al.  The myth of the double-blind review?: author identification using only citations , 2003, SKDD.

[15]  Jiahui Wang,et al.  Metapath-Guided Credit Allocation for Identifying Representative Works , 2020, WWW.

[16]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[17]  Albert-László Barabási,et al.  The universal decay of collective memory and attention , 2018, Nature Human Behaviour.

[18]  Vahab S. Mirrokni,et al.  Maximizing Non-Monotone Submodular Functions , 2011, 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS'07).

[19]  Jure Leskovec,et al.  Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.

[20]  Jiahui Wang,et al.  Identifying Your Representative Work Based on Credit Allocation , 2018, WWW.

[21]  Tomasz Imielinski,et al.  Mining association rules between sets of items in large databases , 1993, SIGMOD Conference.

[22]  ChengXiang Zhai,et al.  Dynamic credit allocation in scientific literature , 2017, Scientometrics.

[23]  Wynne Hsu,et al.  Integrating Classification and Association Rule Mining , 1998, KDD.