Credit scoring models are usually developed using the accepted Known Good-Bad applicants, called KGB model. Yet, the KGB model does not represent the entire Through-The-Door population. Reject inference attempts to correct this inherent flaw by using information of the rejected accounts. Augmentation methods are widely used methods of reject inference, among which Fuzzy Augmentation is the most accurate one. In this paper, we first establish an important property of Fuzzy Augmentation: If Fuzzy Augmentation is not incorporated with variable re-selection, it will produce the same results as the KGB model. We then propose a rule of thumb for Augmentation methods. Based on this rule of thumb, we present a two-phase Augmentation. This two-phase method works not only for Machine Learning in Python but also for the traditional approach using SAS. Moreover, it is user friendly in that the user can specify a factor to increase the bad rate of rejected accounts.
[1]
Jonathan Crook,et al.
Reject inference, augmentation, and sample selection
,
2007,
Eur. J. Oper. Res..
[2]
G. G. Stokes.
"J."
,
1890,
The New Yale Book of Quotations.
[3]
Metric Divergence Measures and Information Value in Credit Scoring
,
2013
.
[4]
A. Albert,et al.
On the existence of maximum likelihood estimates in logistic regression models
,
1984
.
[5]
David J. Hand,et al.
Can reject inference ever work
,
1993
.
[6]
Jonathan Crook,et al.
Does reject inference really improve the performance of application scoring models
,
2004
.
[7]
Jonathan Crook,et al.
Sample selection bias in credit scoring models
,
2003,
J. Oper. Res. Soc..
[8]
David W. Hosmer,et al.
Applied Logistic Regression
,
1991
.