Generating rules from examples of human multiattribute decision making should be simple

Abstract How many prototypes or clusters are needed to predict real world human multiattribute subjective decision making? Although subjective decision making problems occur daily in our life, they have received relatively little attention in artificial intelligence, machine learning and data mining communities. We claim that for most problems, a simple set of rules derived by a nearest neighbor algorithm is the appropriate approach. A simple version of a nearest neighbor model is tested and compared with two other well-established classification methods: neural networks and classifications and regression trees (CART). The results of the experiments show that the simple nearest neighbor method provides very accurate predictions while using very few prototypes or clusters. Although not always the best in accuracy, the differences are sufficiently slight to not warrant greater complexity in deriving rules. Our research on the effectiveness of parsimonious rule sets suggests that decision trees with more than 7–10 branches are not needed for capturing most human multiattribute decision-making problems, and minimal time or memory resources should be used to generate decision making rules.

[1]  Gerhard Widmer,et al.  Prediction of Ordinal Classes Using Regression Trees , 2001, Fundam. Informaticae.

[2]  Herbert A. Simon,et al.  Information-Processing Theory of Human Problem Solving , 1978 .

[3]  Bernard De Baets,et al.  On the Definition and Representation of a Ranking , 2001, RelMiCS.

[4]  Mark Last,et al.  A compact and accurate model for classification , 2004, IEEE Transactions on Knowledge and Data Engineering.

[5]  Jan C. Bioch,et al.  Decision trees for ordinal classification , 2000, Intell. Data Anal..

[6]  Karl Rihaczek,et al.  1. WHAT IS DATA MINING? , 2019, Data Mining for the Social Sciences.

[7]  Ian Witten,et al.  Data Mining , 2000 .

[8]  P. McCullagh,et al.  Generalized Linear Models , 1992 .

[9]  Yoav Ganzach Goals as Determinants of Nonlinear Noncompensatory Judgment Strategies: Leniency vs Strictness , 1993 .

[10]  Robert C. Holte,et al.  Very Simple Classification Rules Perform Well on Most Commonly Used Datasets , 1993, Machine Learning.

[11]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[12]  Leon Sterling,et al.  Learning and classification of monotonic ordinal concepts , 1989, Comput. Intell..

[13]  Yoh-Han Pao,et al.  Adaptive pattern recognition and neural networks , 1989 .

[14]  Toshihide Ibaraki,et al.  Data Analysis by Positive Decision Trees , 1999, CODAS.

[15]  A. Tversky Intransitivity of preferences. , 1969 .

[16]  Gian Luigi Ferrari,et al.  Parameterized Structured Operational Semantics , 1998, Fundam. Informaticae.

[17]  Berndt Brehmer,et al.  New directions in research on decision making , 1986 .

[18]  Sotiris B. Kotsiantis,et al.  A Cost Sensitive Technique for Ordinal Classification Problems , 2004, SETN.

[19]  Pedro M. Domingos The Role of Occam's Razor in Knowledge Discovery , 1999, Data Mining and Knowledge Discovery.

[20]  G. A. Miller THE PSYCHOLOGICAL REVIEW THE MAGICAL NUMBER SEVEN, PLUS OR MINUS TWO: SOME LIMITS ON OUR CAPACITY FOR PROCESSING INFORMATION 1 , 1956 .

[21]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[22]  Wei-Yin Loh,et al.  A Comparison of Prediction Accuracy, Complexity, and Training Time of Thirty-Three Old and New Classification Algorithms , 2000, Machine Learning.

[23]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[24]  G. A. Miller The magical number seven plus or minus two: some limits on our capacity for processing information. , 1956, Psychological review.

[25]  A. Ben-David Monotonicity Maintenance in Information-Theoretic Machine Learning Algorithms , 1995, Machine Learning.

[26]  Sholom M. Weiss,et al.  Computer Systems That Learn , 1990 .

[27]  Gary G. R. Green,et al.  Neural networks, approximation theory, and finite precision computation , 1995, Neural Networks.

[28]  Allan Pinkus,et al.  Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.