Explainers overview

Representatives
All explainers
explanation chart

Approximate model behavior

Surrogate Decision Tree

The surrogate decision tree is an approximate overall flow chart of the model, created by training a simple...

explanation chart

Original feature importance

Shapley Values for Original Features (Kernel SHAP Method)

Shapley explanations are a technique with credible theoretical support that presents consistent global and ...

explanation chart

Feature behavior

Shapley Summary Plot for Original Features (Kernel SHAP Method)

Shapley explanations are a technique with credible theoretical support that presents consistent global and ...

explanation chart

Fairness

Disparate Impact Analysis

Disparate Impact Analysis (DIA) is a technique that is used to evaluate fairness. Bias can be introduced to...

explanation chart

Model debugging

Residual Surrogate Decision Tree

The residual surrogate decision tree predicts which paths in the tree (paths explain approximate model beha...

Problems

Explainers identified the following problems:

Severity Type Problem Suggested actions Explainer Resources
 MEDIUM  bias The residual partial dependence plot of feature 'LIMIT_BAL' indicates the highest interaction of this feature with the error (residual abs(max-min) = 0.03134083655380704) from all the features used by the model (or features configured for PD calculation) Verify that feature `LIMIT_BAL' error interaction does not indicate a model bias or other problem. Residual Partial Dependence Plot PartialDependenceExplanation / application/json
 LOW  bias A path in the residual surrogate decision tree leading to the largest residual (2.0222235) may indicate a problem in the model. Verify that the following surrogate decision tree path does not indicate a model bias or other problem: IF (LIMIT_BAL >= 54956.0 OR LIMIT_BAL IS N/A) AND (AGE < 25.5) AND (LIMIT_BAL >= 65507.5 OR LIMIT_BAL IS N/A) THEN AVERAGE VALUE OF TARGET IS 2.0222235 Residual Surrogate Decision Tree GlobalDtExplanation / application/json

Explainers

100%

Scheduled explainers (10):

Finished explainers (10): Successful explainers (10):

Explainer: Disparate Impact Analysis

Explainer description

Disparate Impact Analysis (DIA) is a technique that is used to evaluate fairness. Bias can be introduced to models during the process of collecting, processing, and labeling data as a result, it is important to determine whether a model is harming certain users by making a significant number of biased decisions. DIA typically works by comparing aggregate measurements of unprivileged groups to a privileged group. For instance, the proportion of the unprivileged group that receives the potentially harmful outcome is divided by the proportion of the privileged group that receives the same outcome - the resulting proportion is then used to determine whether the model is biased.

Explanations

Fairness metrics for the feature: PAY_5

N for feature 'PAY_5'
Adverse Impact for feature 'PAY_5'
Accuracy for feature 'PAY_5'
True Positive Rate for feature 'PAY_5'
Precision for feature 'PAY_5'
Specificity for feature 'PAY_5'
Negative Predicted Value for feature 'PAY_5'
False Positive Rate for feature 'PAY_5'
False Discovery Rate for feature 'PAY_5'
False Negative Rate for feature 'PAY_5'
False Omissions Rate for feature 'PAY_5'
Fairness metrics for the feature: PAY_0
N for feature 'PAY_0'
Adverse Impact for feature 'PAY_0'
Accuracy for feature 'PAY_0'
True Positive Rate for feature 'PAY_0'
Precision for feature 'PAY_0'
Specificity for feature 'PAY_0'
Negative Predicted Value for feature 'PAY_0'
False Positive Rate for feature 'PAY_0'
False Discovery Rate for feature 'PAY_0'
False Negative Rate for feature 'PAY_0'
False Omissions Rate for feature 'PAY_0'
Fairness metrics for the feature: EDUCATION
N for feature 'EDUCATION'
Adverse Impact for feature 'EDUCATION'
Accuracy for feature 'EDUCATION'
True Positive Rate for feature 'EDUCATION'
Precision for feature 'EDUCATION'
Specificity for feature 'EDUCATION'
Negative Predicted Value for feature 'EDUCATION'
False Positive Rate for feature 'EDUCATION'
False Discovery Rate for feature 'EDUCATION'
False Negative Rate for feature 'EDUCATION'
False Omissions Rate for feature 'EDUCATION'
Fairness metrics for the feature: PAY_6
N for feature 'PAY_6'
Adverse Impact for feature 'PAY_6'
Accuracy for feature 'PAY_6'
True Positive Rate for feature 'PAY_6'
Precision for feature 'PAY_6'
Specificity for feature 'PAY_6'
Negative Predicted Value for feature 'PAY_6'
False Positive Rate for feature 'PAY_6'
False Discovery Rate for feature 'PAY_6'
False Negative Rate for feature 'PAY_6'
False Omissions Rate for feature 'PAY_6'
Fairness metrics for the feature: MARRIAGE
N for feature 'MARRIAGE'
Adverse Impact for feature 'MARRIAGE'
Accuracy for feature 'MARRIAGE'
True Positive Rate for feature 'MARRIAGE'
Precision for feature 'MARRIAGE'
Specificity for feature 'MARRIAGE'
Negative Predicted Value for feature 'MARRIAGE'
False Positive Rate for feature 'MARRIAGE'
False Discovery Rate for feature 'MARRIAGE'
False Negative Rate for feature 'MARRIAGE'
False Omissions Rate for feature 'MARRIAGE'
Fairness metrics for the feature: PAY_4
N for feature 'PAY_4'
Adverse Impact for feature 'PAY_4'
Accuracy for feature 'PAY_4'
True Positive Rate for feature 'PAY_4'
Precision for feature 'PAY_4'
Specificity for feature 'PAY_4'
Negative Predicted Value for feature 'PAY_4'
False Positive Rate for feature 'PAY_4'
False Discovery Rate for feature 'PAY_4'
False Negative Rate for feature 'PAY_4'
False Omissions Rate for feature 'PAY_4'
Fairness metrics for the feature: PAY_2
N for feature 'PAY_2'
Adverse Impact for feature 'PAY_2'
Accuracy for feature 'PAY_2'
True Positive Rate for feature 'PAY_2'
Precision for feature 'PAY_2'
Specificity for feature 'PAY_2'
Negative Predicted Value for feature 'PAY_2'
False Positive Rate for feature 'PAY_2'
False Discovery Rate for feature 'PAY_2'
False Negative Rate for feature 'PAY_2'
False Omissions Rate for feature 'PAY_2'
Fairness metrics for the feature: PAY_3
N for feature 'PAY_3'
Adverse Impact for feature 'PAY_3'
Accuracy for feature 'PAY_3'
True Positive Rate for feature 'PAY_3'
Precision for feature 'PAY_3'
Specificity for feature 'PAY_3'
Negative Predicted Value for feature 'PAY_3'
False Positive Rate for feature 'PAY_3'
False Discovery Rate for feature 'PAY_3'
False Negative Rate for feature 'PAY_3'
False Omissions Rate for feature 'PAY_3'

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
  • Disparate Impact Analysis
    global-disparate-impact-analysis
  • Disparate Impact Analysis
    global-html-fragment
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
dia_cols None List of features for which to compute DIA. list None
cut_off 0.0 Cut off. float 0.0
maximize_metric F1 Maximize metric. str F1
max_cardinality 10 Max cardinality for categorical variables. int 10
min_cardinality 2 Minimum cardinality for categorical variables. int 2
num_card 25 Max cardinality for numeric variables to be considered categorical. int 25
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
Explainer keywords:
  • run-by-default
  • explains-fairness
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:30:59 T+0100
  • Duration: 4.85s
Explainer log
Explainer log file:

Explainer: Residual Surrogate Decision Tree

Problems

Explainer identified the following problems:

Severity Type Problem Suggested actions Explainer Resources
 LOW  bias A path in the residual surrogate decision tree leading to the largest residual (2.0222235) may indicate a problem in the model. Verify that the following surrogate decision tree path does not indicate a model bias or other problem: IF (LIMIT_BAL >= 54956.0 OR LIMIT_BAL IS N/A) AND (AGE < 25.5) AND (LIMIT_BAL >= 65507.5 OR LIMIT_BAL IS N/A) THEN AVERAGE VALUE OF TARGET IS 2.0222235 Residual Surrogate Decision Tree GlobalDtExplanation / application/json
Explainer description

The residual surrogate decision tree predicts which paths in the tree (paths explain approximate model behavior) lead to highest or lowest error. The residual surrogate decision tree is created by training a simple decision tree on the residuals of the predictions of the model. Residuals are differences between observed and predicted values which can be used as targets in surrogate models for the purpose of model debugging. The method used to calculate residuals varies depending on the type of problem. For classification problems, logloss residuals are calculated for a specified class (only one residual surrogate decision is created by the explainer and it is built for this class). For regression problems, residuals are determined by calculating the square of the difference between targeted and predicted values.

Explanations

Approximate model behavior for the class '1':

Decision tree for class '1'

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
debug_residuals_class 1 Class for debugging classification model logloss residuals, empty string for debugging regression model residuals. str
dt_tree_depth 3 Decision tree depth. int 3
nfolds 3 Number of CV folds. int 3
qbin_cols None Quantile binning columns. list None
qbin_count 0 Quantile bins count. int 0
categorical_encoding onehotexplicit Categorical encoding. str onehotexplicit
debug_residuals True
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • run-by-default
  • requires-h2o3
  • explains-model-debugging
  • surrogate
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:04 T+0100
  • Duration: 2.131s
Explainer log
Explainer log file:

Explainer: Surrogate Decision Tree

Explainer description

The surrogate decision tree is an approximate overall flow chart of the model, created by training a simple decision tree on the original inputs and the predictions of the model.

Explanations

Approximate model behavior for the class '1':

Decision tree for class '1'

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
debug_residuals False Debug model residuals. bool False
debug_residuals_class Class for debugging classification model logloss residuals, empty string for debugging regression model residuals. str
dt_tree_depth 3 Decision tree depth. int 3
nfolds 3 Number of CV folds. int 3
qbin_cols None Quantile binning columns. list None
qbin_count 0 Quantile bins count. int 0
categorical_encoding onehotexplicit Categorical encoding. str onehotexplicit
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • run-by-default
  • requires-h2o3
  • surrogate
  • explains-approximate-behavior
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:06 T+0100
  • Duration: 2.179s
Explainer log
Explainer log file:

Explainer: Shapley Summary Plot for Original Features (Kernel SHAP Method)

Explainer description

Shapley explanations are a technique with credible theoretical support that presents consistent global and local feature contributions.

The Shapley Summary Plot shows original features versus their local Shapley values on a sample of the dataset. Feature values are binned by Shapley values and the average normalized feature value for each bin is plotted. The legend corresponds to numeric features and maps to their normalized value - yellow is the lowest value and deep orange is the highest. You can also get a scatter plot of the actual numeric features values versus their corresponding Shapley values. Categorical features are shown in grey and do not provide an actual-value scatter plot.

Notes:

  • The Shapley Summary Plot only shows original features that are used in the model.
  • The dataset sample size and the number of bins can be updated in the interpretation settings.

Explanations

Global Shapley values for original features of class 'None (Regression)':

Global Shapley values for class 'None (Regression)'
Local Shapley values for original features of class 'None (Regression)':
Local Shapley values for class None (Regression) and feature PAY_4
Local Shapley values for class None (Regression) and feature PAY_3
Local Shapley values for class None (Regression) and feature PAY_AMT1
Local Shapley values for class None (Regression) and feature PAY_AMT4
Local Shapley values for class None (Regression) and feature PAY_AMT3
Local Shapley values for class None (Regression) and feature BILL_AMT6
Local Shapley values for class None (Regression) and feature AGE
Local Shapley values for class None (Regression) and feature BILL_AMT3
Local Shapley values for class None (Regression) and feature PAY_6
Local Shapley values for class None (Regression) and feature PAY_AMT6
Local Shapley values for class None (Regression) and feature MARRIAGE
Local Shapley values for class None (Regression) and feature BILL_AMT1
Local Shapley values for class None (Regression) and feature EDUCATION
Local Shapley values for class None (Regression) and feature BILL_AMT5
Local Shapley values for class None (Regression) and feature BILL_AMT2
Local Shapley values for class None (Regression) and feature PAY_0
Local Shapley values for class None (Regression) and feature PAY_5
Local Shapley values for class None (Regression) and feature ID
Local Shapley values for class None (Regression) and feature BILL_AMT4
Local Shapley values for class None (Regression) and feature PAY_AMT2
Local Shapley values for class None (Regression) and feature PAY_AMT5
Local Shapley values for class None (Regression) and feature PAY_2
Local Shapley values for class None (Regression) and feature LIMIT_BAL

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
max_features 50 Maximum number of features to be shown in the plot. int 50
sample_size 20000 Sample size. int 20000
x_shapley_resolution 500 x-axis resolution (number of Shapley values bins). int 500
enable_drilldown_charts True Enable creation of per-feature Shapley/feature value scatter plots. bool True
fast_approx_contribs True Speed up predictions with fast predictions and contributions approximations. bool True
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • run-by-default
  • explains-feature-behavior
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:08 T+0100
  • Duration: 12.129s
Explainer log
Explainer log file:

Explainer: Partial Dependence Plot

Explainer description

Partial dependence plot (PDP) portrays the average prediction behavior of the model across the domain of an input variable along with +/- 1 standard deviation bands. Individual Conditional Expectations plot (ICE) displays the prediction behavior for an individual row of data when an input variable is toggled across its domain.

PD binning:

Integer feature:

  • bins in numeric mode:
    • bins are integers
    • (at most) grid_resolution integer values in between minimum and maximum of feature values
    • bin values are created as evenly as possible
    • minimum and maximum is included in bins (if grid_resolution is bigger or equal to 2)
  • bins in categorical mode:
    • bins are integers
    • top grid_resolution values from feature values ordered by frequency (int values are converted to strings and most frequent values are used as bins)
  • quantile bins in numeric mode:
    • bins are integers
    • bin values are created with the aim that there will be the same number of observations per bin
    • q-quantile used to created q bins where q is specified by PD parameter
  • quantile bins in categorical mode:
    • not supported

Float feature:

  • bins in numeric mode:
    • bins are floats
    • grid_resolution float values in between minimum and maximum of feature values
    • bin values are created as evenly as possible
    • minimum and maximum is included in bins (if grid_resolution is bigger or equal to 2)
  • bins in categorical mode:
    • bins are floats
    • top grid_resolution values from feature values ordered by frequency (float values are converted to strings and most frequent values are used as bins)
  • quantile bins in numeric mode:
    • bins are floats
    • bin values are created with the aim that there will be the same number of observations per bin
    • q-quantile used to created q bins where q is specified by PD parameter
  • quantile bins in categorical mode:
    • not supported

String feature:

  • bins in numeric mode:
    • not supported
  • bins in categorical mode:
    • bins are strings
    • top grid_resolution values from feature values ordered by frequency
  • quantile bins:
    • not supported

Date/datetime feature:

  • bins in numeric mode:
    • bins are dates
    • grid_resolution date values in between minimum and maximum of feature values
    • bin values are created as evenly as possible:
      1. dates are parsed and converted to epoch timestamps i.e integers
      2. bins are created as in case of numeric integer binning
      3. integer bins are converted back to original date format
    • minimum and maximum is included in bins (if grid_resolution is bigger or equal to 2)
  • bins in categorical mode:
    • bins are dates
    • top grid_resolution values from feature values ordered by frequency (dates are handled as opaque strings and most frequent values are used as bins)
  • quantile bins:
    • not supported

PD out of range binning:

Integer feature:

  • OOR bins in numeric mode:
    • OOR bins are integers
    • (at most) oor_grid_resolution integer values are added below minimum and above maximum
    • bin values are created by adding/substracting rounded standard deviation (of feature values) above and below maximum and minimum oor_grid_resolution times
      • 1 used used if rounded standard deviation would be 0
    • if feature is of unsigned integer type, then bins below 0 are not created
      • if rounded standard deviation and/or oor_grid_resolution is so high that it would cause lower OOR bins to be negative numbers, then standard deviation of size 1 is tried instead
  • OOR bins in categorical mode:
    • same as numeric mode

Float feature:

  • OOR bins in numeric mode:
    • OOR bins are floats
    • oor_grid_resolution float values are added below minimum and above maximum
    • bin values are created by adding/substracting standard deviation (of feature values) above and below maximum and minimum oor_grid_resolution times
  • OOR bins in categorical mode:
    • same as numeric mode

String feature:

  • bins in numeric mode:
    • not supported
  • bins in categorical mode:
    • OOR bins are strings
    • value UNSEEN is added as OOR bin

Date feature:

  • bins in numeric mode:
    • not supported
  • bins in categorical mode:
    • OOR bins are strings
    • value UNSEEN is added as OOR bin

Explanations

Partial Dependence Plot for the feature 'ID' and class 'None (Regression)':

PD for class 'None (Regression)' and feature 'ID
Partial Dependence Plot for the feature 'LIMIT_BAL' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'LIMIT_BAL
Partial Dependence Plot for the feature 'EDUCATION' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'EDUCATION
Partial Dependence Plot for the feature 'MARRIAGE' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'MARRIAGE
Partial Dependence Plot for the feature 'AGE' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'AGE
Partial Dependence Plot for the feature 'PAY_0' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_0
Partial Dependence Plot for the feature 'PAY_2' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_2
Partial Dependence Plot for the feature 'PAY_3' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_3
Partial Dependence Plot for the feature 'PAY_4' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_4
Partial Dependence Plot for the feature 'PAY_5' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_5

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
sample_size 25000 Sample size for Partial Dependence Plot. int 25000
max_features 10 Partial Dependence Plot number of features (to see all features used by model set to -1). int 10
features None Partial Dependence Plot feature list. list None
oor_grid_resolution 0 Partial Dependence Plot number of out of range bins. int 0
quantile-bin-grid-resolution 0 Partial Dependence Plot quantile binning (total quantile points used to create bins). int 0
grid_resolution 20 Partial Dependence Plot observations per bin (number of equally spaced points used to create bins). int 20
center False Center Partial Dependence Plot using ICE centered at 0. bool False
sort_bins True Ensure bin values sorting. bool True
histograms True Enable histograms. bool True
quantile-bins Per-feature quantile binning (Example: if choosing features F1 and F2, this parameter is '{"F1": 2,"F2": 5}'. Note, you can set all features to use the same quantile binning with the `Partial Dependence Plot quantile binning` parameter and then adjust the quantile binning for a subset of PDP features with this parameter). str
numcat_num_chart True Unique feature values count driven Partial Dependence Plot binning and chart selection. bool True
numcat_threshold 11 Threshold for Partial Dependence Plot binning and chart selection (<=threshold categorical, >threshold numeric). int 11
debug_residuals False Debug model residuals. bool False
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • run-by-default
  • can-add-feature
  • explains-feature-behavior
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:20 T+0100
  • Duration: 1.342s
Explainer log
Explainer log file:

Explainer: Residual Partial Dependence Plot

Problems

Explainer identified the following problems:

Severity Type Problem Suggested actions Explainer Resources
 MEDIUM  bias The residual partial dependence plot of feature 'LIMIT_BAL' indicates the highest interaction of this feature with the error (residual abs(max-min) = 0.03134083655380704) from all the features used by the model (or features configured for PD calculation) Verify that feature `LIMIT_BAL' error interaction does not indicate a model bias or other problem. Residual Partial Dependence Plot PartialDependenceExplanation / application/json
Explainer description

The residual partial dependence plot (PDP) indicates which variables interact most with the error. Residuals are transformed differences between observed and predicted values: the square of the difference between observed and predicted values is used in case of regression problems; -1 * log(p) is used in case of classification problems. The residual partial dependence is created using normal partial dependence algorithm, while instead of prediction is used the residual. Individual Conditional Expectations plot (ICE) displays the interaction with error for an individual row of data when an input variable is toggled across its domain.

Explanations

Partial Dependence Plot for the feature 'ID' and class 'None (Regression)':

PD for class 'None (Regression)' and feature 'ID
Partial Dependence Plot for the feature 'LIMIT_BAL' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'LIMIT_BAL
Partial Dependence Plot for the feature 'EDUCATION' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'EDUCATION
Partial Dependence Plot for the feature 'MARRIAGE' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'MARRIAGE
Partial Dependence Plot for the feature 'AGE' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'AGE
Partial Dependence Plot for the feature 'PAY_0' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_0
Partial Dependence Plot for the feature 'PAY_2' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_2
Partial Dependence Plot for the feature 'PAY_3' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_3
Partial Dependence Plot for the feature 'PAY_4' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_4
Partial Dependence Plot for the feature 'PAY_5' and class 'None (Regression)':
PD for class 'None (Regression)' and feature 'PAY_5

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
sample_size 25000 Sample size for Partial Dependence Plot. int 25000
max_features 10 Partial Dependence Plot number of features (to see all features used by model set to -1). int 10
features None Partial Dependence Plot feature list. list None
oor_grid_resolution 0 Partial Dependence Plot number of out of range bins. int 0
quantile-bin-grid-resolution 0 Partial Dependence Plot quantile binning (total quantile points used to create bins). int 0
grid_resolution 20 Partial Dependence Plot observations per bin (number of equally spaced points used to create bins). int 20
center False Center Partial Dependence Plot using ICE centered at 0. bool False
sort_bins True Ensure bin values sorting. bool True
histograms True Enable histograms. bool True
quantile-bins Per-feature quantile binning (Example: if choosing features F1 and F2, this parameter is '{"F1": 2,"F2": 5}'. Note, you can set all features to use the same quantile binning with the `Partial Dependence Plot quantile binning` parameter and then adjust the quantile binning for a subset of PDP features with this parameter). str
numcat_num_chart True Unique feature values count driven Partial Dependence Plot binning and chart selection. bool True
numcat_threshold 11 Threshold for Partial Dependence Plot binning and chart selection (<=threshold categorical, >threshold numeric). int 11
debug_residuals True
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • can-add-feature
  • explains-model-debugging
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:21 T+0100
  • Duration: 0.86s
Explainer log
Explainer log file:

Explainer: Shapley Values for Original Features (Kernel SHAP Method)

Explainer description

Shapley explanations are a technique with credible theoretical support that presents consistent global and local variable contributions. Local numeric Shapley values are calculated by tracing single rows of data through a trained tree ensemble and aggregating the contribution of each input variable as the row of data moves through the trained ensemble. For regression tasks, Shapley values sum to the prediction of the Driverless AI model. For classification problems, Shapley values sum to the prediction of the Driverless AI model before applying the link function. Global Shapley values are the average of the absolute Shapley values over every row of a dataset. Shapley values for original features are calculated with the Kernel Explainer method, which uses a special weighted linear regression to compute the importance of each feature. More information about Kernel SHAP is available at http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf.

Explanations

Feature importance for the class '1':

Feature importance for class '1'

The most important original feature of the class 1 is MARRIAGE .

Original feature importances for the class '1':

  • 1. MARRIAGE feature with importance 0.5048248714839003
  • 2. BILL_AMT6 feature with importance 0.5032058119254781
  • 3. PAY_AMT1 feature with importance 0.5029552167108501
  • 4. PAY_4 feature with importance 0.5027351920100945
  • 5. PAY_AMT6 feature with importance 0.5010883391722856
  • 6. PAY_AMT3 feature with importance 0.5004985354974225
  • 7. PAY_0 feature with importance 0.500284473535244
  • 8. BILL_AMT1 feature with importance 0.5002006452352199
  • 9. LIMIT_BAL feature with importance 0.4997530831820978
  • 10. PAY_3 feature with importance 0.49947839694224516
  • 11. AGE feature with importance 0.4993569130311947
  • 12. PAY_AMT4 feature with importance 0.49910010407530536
  • 13. BILL_AMT5 feature with importance 0.4990933644743017
  • 14. ID feature with importance 0.49842250047515246
  • 15. BILL_AMT4 feature with importance 0.4982346212625947
  • 16. default payment next month feature with importance 0.497985070650521
  • 17. PAY_AMT2 feature with importance 0.49785012302559956
  • 18. BILL_AMT2 feature with importance 0.4974761425230914
  • 19. PAY_5 feature with importance 0.49722928324889265
  • 20. PAY_6 feature with importance 0.4971471089510063
  • 21. BILL_AMT3 feature with importance 0.4969625066178144
  • 22. PAY_AMT5 feature with importance 0.49647477937225026
  • 23. EDUCATION feature with importance 0.49621047003623286
  • 24. PAY_2 feature with importance 0.4932958341860075

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
sample_size 100000 Sample size. int 100000
sample True Sample Kernel Shapley. bool True
nsample Number of times to re-evaluate the model when explaining each prediction with Kernel Explainer. Default is determined internally.'auto' or int. Number of times to re-evaluate the model when explaining each prediction. More samples lead to lower variance estimates of the SHAP values. The 'auto' setting uses nsamples = 2 * X.shape[1] + 2048. This setting is disabled by default and runtime determines the right number internally. int
L1 auto L1 regularization for Kernel Explainer. 'num_features(int)', 'auto' (default for now, but deprecated), 'aic', 'bic', or float. The L1 regularization to use for feature selection (the estimation procedure is based on a debiased lasso). The 'auto' option currently uses aic when less that 20% of the possible sample space is enumerated, otherwise it uses no regularization. The aic and bic options use the AIC and BIC rules for regularization. Using 'num_features(int)' selects a fix number of top features. Passing a float directly sets the alpha parameter of the sklearn.linear_model.Lasso model used for feature selection. str auto
max runtime 900 Max runtime for Kernel explainer in seconds. int 900
fast_approx True Speed up predictions with fast predictions approximation. bool True
leakage_warning_threshold 0.95 The threshold above which to report a potentially detected feature importance leak problem. float 0.95
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer keywords:
  • explains-original-feature-importance
  • is_slow
  • h2o-sonar
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:22 T+0100
  • Duration: 1.684s
Explainer log
Explainer log file:

Explainer: Partial Dependence Plot for Two Features

Explainer description

Partial dependence for 2 features portrays the average prediction behavior of a model across the domains of two input variables i.e. interaction of feature tuples with the prediction. While PD for one feature produces 2D plot, PD for two features produces 3D plots. This explainer plots PD for two features using heatmap, contour 3D or surface 3D.

Explanations

Partial dependence plot for features 'ID' and 'LIMIT_BAL':

Partial dependence plot for features 'ID' and 'LIMIT_BAL'
Partial dependence plot for features 'ID' and 'EDUCATION':
Partial dependence plot for features 'ID' and 'EDUCATION'
Partial dependence plot for features 'LIMIT_BAL' and 'EDUCATION':
Partial dependence plot for features 'LIMIT_BAL' and 'EDUCATION'

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
sample_size 25000 Sample size for Partial Dependence Plot of 2 features. int 25000
max_features 3 Partial Dependence Plot number of features. int 3
features None List of features from which to choose pairs to compute PD for two features. list None
grid_resolution 10 Partial Dependence Plot observations per bin (number of equally spaced points used to create bins). int 10
oor_grid_resolution 0 Partial Dependence Plot number of out of range bins. int 0
quantile-bin-grid-resolution 0 Partial Dependence Plot quantile binning (total quantile points used to create bins). int 0
plot_type heatmap Plot type. str heatmap
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
Explainer keywords:
  • is_slow
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:24 T+0100
  • Duration: 2.225s
Explainer log
Explainer log file:

Explainer: Friedman's H-statistic

Explainer description

Friedman's H-statistic describes the amount of variance explained by the feature pair. It's expressed with a graph where most important original features are nodes and the interaction scores are edges. When features interact with each other, then the influence of the features on the prediction does not have be additive, but more complex. For instance the contribution might be greater than the sum of contributions. Friedman's H-statistic calculation is computationally intensive and typically requires long time to finish - calculation duration grows with the number of features and bins.

Explanations

Feature importance for the class 'None (Regression)':

Feature importance for class 'None (Regression)'

Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
Explainer parameters
Explainer was run with the following parameters:
Parameter Value Description Type Default value
features_number 4 Number of features for which to calculate H-Statistic. int 4
grid_resolution 3 Observations per bin (number of equally spaced points used to create bins). int 3
features None Feature list - at least 2 features must be selected. multilist None
sample_size 25000 Sample size for Partial Dependence Plot int 25000
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
Explainer keywords:
  • explains-feature-behavior
  • h2o-sonar
  • is_slow
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:26 T+0100
  • Duration: 0.486s
Explainer log
Explainer log file:

Explainer: Dataset and model insights explainer

Explainer description

The explainer checks the dataset and model for various issues. For example, it provides problems and actions for missing values in the target column and a low number of unique values across columns of a dataset.

Explanations
Explanations
Model explanations created by the explainer organized by explanation types with its formats (representations) identified by media types :
  • Dataset and model insights explainer
    global-text-explanation
Explainer parameters
Explainer was run with the following parameters:
Explainer metadata
The explainer can explain model types:
  • regression
  • binomial
  • multinomial
Explainer run
Explainer run details:
  • Status code:  SUCCESS 
  • Progress: 100%
  • Started: 2026-01-30 17:31:27 T+0100
  • Duration: 0.212s
Explainer log
Explainer log file:

Dataset

Dataset description:

Model

Model description:

Configuration and parameters

Interpretation parameters:

H2O Sonar library configuration:

Directories, files and logs

Directories and files: