Hostname: page-component-89b8bd64d-rbxfs Total loading time: 0 Render date: 2026-05-06T23:54:46.417Z Has data issue: false hasContentIssue false

Combining white box models, black box machines and humaninterventions for interpretable decision strategies

Published online by Cambridge University Press:  01 January 2023

Gregory Gadzinski*
Affiliation:
International University of Monaco – Omnes Education, 14, rue Hubert Clerissi, MC98000 – Monaco
Alessio Castello*
Affiliation:
International University of Monaco – Omnes Education, 14, rue Hubert Clerissi, MC98000 – Monaco
Rights & Permissions [Opens in a new window]

Abstract

Granting a short-term loan is a critical decision. A great deal of research hasconcerned the prediction of credit default, notably through Machine Learning(ML) algorithms. However, given that their black-box nature has sometimes led tounwanted outcomes, comprehensibility in ML guided decision-making strategies hasbecome more important. In many domains, transparency and accountability are nolonger optional. In this article, instead of opposing white-box againstblack-box models, we use a multi-step procedure that combines the Fast andFrugal Tree (FFT) methodology of Martignon et al. (2005) and Phillips et al.(2017) with the extraction of post-hoc explainable informationfrom ensemble ML models. New interpretable models are then built thanks to theinclusion of explainable ML outputs chosen by human intervention. Ourmethodology improves significantly the accuracy of the FFT predictions whilepreserving their explainable nature. We apply our approach to a dataset ofshort-term loans granted to borrowers in the UK, and show how complex machinelearning can challenge simpler machines and help decision makers.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
The authors license this article under the terms of the Creative Commons Attribution 3.0 License.
Copyright
Copyright © The Authors [2022] This is an Open Access article, distributed under the terms of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Figure 0

Figure 1: Human-machines process control framework.

Figure 1

Figure 2: Visualization of a restricted FFT object with two variables (Phillips et al., 2017). The top panel shows the frequencies of negative and positive criterion classes. The middle panel contains the FFT with icon arrays displaying the accuracy of cases classified at each node.

Figure 2

Figure 3: A 2x2 confusion matrix used to evaluate a decision algorithm from Phillips et al. (2017).

Figure 3

Figure 4: ifan algorithm, adapted from Phillips et al. (2017).

Figure 4

Figure 5: Visualization of an FFT with five variables (Phillips et al., 2017). The top panel shows the frequencies of negative and positive criterion classes. The middle panel contains the FFT with icon arrays displaying the accuracy of cases classified at each node. The bottom panel shows the confusion matrix, and the FFT’s ROC classification performance.

Figure 5

Figure 6: Visualization of the ensemble first-order Partial Dependence Plots. Each line represents the partial dependence function of one trained neural network model. The y-axis represents the average prediction of the probability of default (1 = default) and the x-axis represents the values of the original independent variables. For the continuous variables, the vertical line in red represents the threshold computed by the FFT in Figure 5.

Figure 6

Figure 7: Visualization of the ensemble first-order Accumulated Local Plots. Each line represents the Accumulated Local Effects (ALE) of one trained neural network model. The y-axis represents the average prediction of the probability of default (1 = default) and the x-axis represents the values of the original independent variables. For the continuous variables, the vertical line in red represents the threshold computed by the FFT in Figure 5.

Figure 7

Figure 8: Visualization of the ensemble Partial Dependence Plots for selected pairs of variables. The axes represent the values of the original independent variables. The outputs of the partial dependence function, i.e. the prediction of the probability of default, from the 10 neural network models have been averaged for each grid value.

Figure 8

Figure 9: Visualization of the augmented FFT-1.

Figure 9

Figure 10: Visualization of the augmented FFT-2.

Figure 10

Table 1: Model accuracy (%) by type of model, sorted from the lowest to the highest absolute prediction. Bold highlights maximum of each column.

Figure 11

Table 2: P-values of a test for the difference in two proportions (absolute correct predictions in Table 1) Values are expressed in percentages.