2020 IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA) 2020
DOI: 10.1109/dsaa49011.2020.00058
|View full text |Cite
|
Sign up to set email alerts
|

Probability of default estimation, with a reject option

Abstract: Many companies, such as credit granting companies, have to decide on granting or denying customer or invoice loans on a daily basis. Increasingly, machine learning is used to learn probability-of-default models from previously granted cases and, thus, whether the outcome was positive or negative for the company, i.e. whether the client paid back or defaulted. However, as the outcome can only be observed for the granted cases, the data inherently has sample selection bias and caution should be taken when applyi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 22 publications
0
1
0
Order By: Relevance
“…Instead, we simply assume that the predictor is likely to be inaccurate on data points that are highly dissimilar to those samples in the training data. This yields a model-agnostic approach where r models the training data using, for example, a one-class model such as a Gaussian-mixture [20] or a One-Class Support Vector Machine (OCSVM) [3,18]. During deployment, r only passes samples to h that are similar to those found in the training data.…”
Section: Related Work On Machine Learning With a Reject Optionmentioning
confidence: 99%
“…Instead, we simply assume that the predictor is likely to be inaccurate on data points that are highly dissimilar to those samples in the training data. This yields a model-agnostic approach where r models the training data using, for example, a one-class model such as a Gaussian-mixture [20] or a One-Class Support Vector Machine (OCSVM) [3,18]. During deployment, r only passes samples to h that are similar to those found in the training data.…”
Section: Related Work On Machine Learning With a Reject Optionmentioning
confidence: 99%
“…The novel observations can be identified as lying in low density areas [3] or outside a boundary encapsulating the training data [10,13]. Alternatively, an anomaly detector can be used to identify the deviating data such as the k-nearest neighbor outlier detector [1] or isolation forest [6]. All these methods only look at the independent variable, without assessing the accuracy of the classifier.…”
Section: Introductionmentioning
confidence: 99%