Parameter Selection: Why We Should Pay More Attention to It

The importance of parameter selection in supervised learning is well known. However, due to the many parameter combinations, an incomplete or an insufficient procedure is often applied. This situation may cause misleading or confusing conclusions. In this opinion paper, through an intriguing example we point out that the seriousness goes beyond what is generally recognized. In the topic of multilabel classification for medical code prediction, one influential paper conducted a proper parameter selection on a set, but when moving to a subset of frequently occurring labels, the authors used the same parameters without a separate tuning. The set of frequent labels became a popular benchmark in subsequent studies, which kept pushing the state of the art. However, we discovered that most of the results in these studies cannot surpass the approach in the original paper if a parameter tuning had been conducted at the time. Thus it is unclear how much progress the subsequent developments have actually brought. The lesson clearly indicates that without enough attention on parameter selection, the research progress in our field can be uncertain or even illusive.

[1]  Juri Opitz,et al.  Macro F1 and Macro F1 , 2019, ArXiv.

[2]  Shirui Pan,et al.  Medical Code Assignment with Gated Convolution and Note-Code Interaction , 2021, FINDINGS.

[3]  Fei Li,et al.  ICD Coding from Clinical Text Using Multi-Filter Residual Convolutional Neural Network , 2019, AAAI.

[4]  Guoyin Wang,et al.  Joint Embedding of Words and Labels for Text Classification , 2018, ACL.

[5]  Yubo Chen,et al.  HyperCore: Hyperbolic and Co-graph Representation for Automatic ICD Coding , 2020, ACL.

[6]  Fei Teng,et al.  Explainable Prediction of Medical Codes With Knowledge Graphs , 2020, Frontiers in Bioengineering and Biotechnology.

[8]  Yun-Nung Chen,et al.  Leveraging Hierarchical Category Knowledge for Data-Imbalanced Multi-Label Diagnostic Text Understanding , 2019, EMNLP.

[9]  Pengtao Xie,et al.  A Neural Architecture for Automated ICD Coding , 2017, ACL.

[10]  Jasper Snoek,et al.  Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.

[11]  Philip S. Yu,et al.  EHR Coding with Multi-scale Feature Attention and Structured Knowledge Graph Propagation , 2019, CIKM.

[12]  Anthony N. Nguyen,et al.  A Label Attention Model for ICD Coding from Clinical Text , 2020, IJCAI.

[13]  Jimeng Sun,et al.  Explainable Prediction of Medical Codes from Clinical Text , 2018, NAACL.

[14]  Xiangling Fu,et al.  Clinical-Coder: Assigning Interpretable ICD-10 Codes to Chinese Clinical Notes , 2020, ACL.

[15]  Peter Szolovits,et al.  MIMIC-III, a freely accessible critical care database , 2016, Scientific Data.

[16]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[17]  Honghan Wu,et al.  Explainable Automated Coding of Clinical Notes using Hierarchical Label-wise Attention Networks and Label Embedding Initialisation , 2021, J. Biomed. Informatics.

[18]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[19]  Erik Cambria,et al.  Dilated Convolutional Attention Network for Medical Code Assignment from Clinical Text , 2020, CLINICALNLP.

[20]  David Suendermann-Oeft,et al.  Medical code prediction with multi-view convolution and description-regularized label-dependent attention , 2018, ArXiv.