This vignette demonstrates using
the ale
package on small datasets, as is often the case
with statistical inference. You should first read the introductory vignette that
explains general functionality of the package; this vignette goes beyond
those details to discuss issues unique to small datasets.
An obvious question is, “How small is ‘small’?” This is a very complex question and it is way beyond the scope of this vignette to try to answer it rigorously. But we can simply say that the key issue at stake is that applying the training-test split that is common in machine learning is a crucial technique for increasing the generalizability of data analysis. So, the question becomes focused to, “How small is too small for a training-test split for machine learning analysis?” The rule of thumb I am familiar with is that machine learning requires at least 200 rows of data for each predictor variable. So, for example, if you have five input variables, you would need at least 1000 rows of data. But note that this does not refer to the size of the entire dataset but to the minimum size of the training subset. So, if you carry out an 80-20 split on the full dataset (that is, 80% training set), then you would need at least 1000 rows for the training set and another 250 rows for the test set, for a minimum of 1250 rows. (And if you carry out hyperparameter tuning with cross validation on that training set, then you need even more data.) If you see where this is headed, you might quickly realize that most datasets of less than 2000 rows are probably “small”. You can see that even many datasets that are more than 2000 rows are nonetheless “small”, and so probably need the techniques mentioned in this vignette.
We begin by loading the necessary libraries.
Most analyses use the attitude
dataset, built-in with R:
“From a survey of the clerical employees of a large financial
organization, the data are aggregated from the questionnaires of the
approximately 35 employees for each of 30 (randomly selected)
departments.” Since we’re talking about “small” datasets, we figure that
we might as well demonstrate principles with extremely small
examples.
From a survey of the clerical employees of a large financial organization, the data are aggregated from the questionnaires of the approximately 35 employees for each of 30 (randomly selected) departments. The numbers give the percent proportion of favourable responses to seven questions in each department.
A data frame with 30 observations on 7 variables. The first column are the short names from the reference, the second one the variable names in the data frame:
Variable | Type | Description | |
---|---|---|---|
Y | rating | numeric | Overall rating |
X[1] | complaints | numeric | Handling of employee complaints |
X[2] | privileges | numeric | Does not allow special privileges |
X[3] | learning | numeric | Opportunity to learn |
X[4] | raises | numeric | Raises based on performance |
X[5] | critical | numeric | Too critical |
X[6] | advance | numeric | Advancement |
Chatterjee, S. and Price, B. (1977) Regression Analysis by Example. New York: Wiley. (Section 3.7, p.68ff of 2nd ed.(1991).)
str(attitude)
#> 'data.frame': 30 obs. of 7 variables:
#> $ rating : num 43 63 71 61 81 43 58 71 72 67 ...
#> $ complaints: num 51 64 70 63 78 55 67 75 82 61 ...
#> $ privileges: num 30 51 68 45 56 49 42 50 72 45 ...
#> $ learning : num 39 54 69 47 66 44 56 55 67 47 ...
#> $ raises : num 61 63 76 54 71 54 66 70 71 62 ...
#> $ critical : num 92 73 86 84 83 49 68 66 83 80 ...
#> $ advance : num 45 47 48 35 47 34 35 41 31 41 ...
summary(attitude)
#> rating complaints privileges learning raises
#> Min. :40.00 Min. :37.0 Min. :30.00 Min. :34.00 Min. :43.00
#> 1st Qu.:58.75 1st Qu.:58.5 1st Qu.:45.00 1st Qu.:47.00 1st Qu.:58.25
#> Median :65.50 Median :65.0 Median :51.50 Median :56.50 Median :63.50
#> Mean :64.63 Mean :66.6 Mean :53.13 Mean :56.37 Mean :64.63
#> 3rd Qu.:71.75 3rd Qu.:77.0 3rd Qu.:62.50 3rd Qu.:66.75 3rd Qu.:71.00
#> Max. :85.00 Max. :90.0 Max. :83.00 Max. :75.00 Max. :88.00
#> critical advance
#> Min. :49.00 Min. :25.00
#> 1st Qu.:69.25 1st Qu.:35.00
#> Median :77.50 Median :41.00
#> Mean :74.77 Mean :42.93
#> 3rd Qu.:80.00 3rd Qu.:47.75
#> Max. :92.00 Max. :72.00
We first run ALE analysis on this dataset as if it were a valid
regular dataset, even though it is too small for a proper training-test
split. This is a small-scale demonstration mainly to demonstrate that
ale
package is valid for analyzing even small datasets, not
just the large datasets typically used for machine learning.
Ordinary least squares (OLS) regression is the most generic multivariate statistical technique. Thus, we use it as a baseline illustration to help motivate the value of ALE for interpreting the analysis of small data samples.
We train an OLS model to predict average rating:
lm_attitude <- lm(rating ~ ., data = attitude)
summary(lm_attitude)
#>
#> Call:
#> lm(formula = rating ~ ., data = attitude)
#>
#> Residuals:
#> Min 1Q Median 3Q Max
#> -10.9418 -4.3555 0.3158 5.5425 11.5990
#>
#> Coefficients:
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) 10.78708 11.58926 0.931 0.361634
#> complaints 0.61319 0.16098 3.809 0.000903 ***
#> privileges -0.07305 0.13572 -0.538 0.595594
#> learning 0.32033 0.16852 1.901 0.069925 .
#> raises 0.08173 0.22148 0.369 0.715480
#> critical 0.03838 0.14700 0.261 0.796334
#> advance -0.21706 0.17821 -1.218 0.235577
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> Residual standard error: 7.068 on 23 degrees of freedom
#> Multiple R-squared: 0.7326, Adjusted R-squared: 0.6628
#> F-statistic: 10.5 on 6 and 23 DF, p-value: 1.24e-05
At the very least, the ale
is useful for visualizing the
effects of model variables.
Before starting, we recommend that you enable progress bars to see how long procedures will take. Simply run the following code at the beginning of your R session:
# Run this in an R console; it will not work directly within an R Markdown or Quarto block
progressr::handlers(global = TRUE)
progressr::handlers('cli')
If you forget to do that, the {ale}
package will do it
automatically for you with a notification message.
Note that for now, we run ale
with no bootstrapping (the
default) because small samples require a special bootstrap approach, as
explained below. For now, all we are doing is using ALE to accurately
visualize what the model estimates.
ale_lm_attitude_simple <- ale(
attitude, lm_attitude,
parallel = 2 # CRAN limit (delete this line on your own computer)
)
# Print all plots
lm_attitude_simple_plots <- plot(ale_lm_attitude_simple)
lm_attitude_simple_1D_plots <- lm_attitude_simple_plots$distinct$rating$plots[[1]]
patchwork::wrap_plots(lm_attitude_simple_1D_plots, ncol = 2)
This visualization confirms what we see in the model coefficients above: complaints have a strong positive effect on ratings and learning has a more moderate effect. However, the ALE indicates a stronger effect of advance than the regression coefficients suggest. The other variables have relatively little effect on ratings. We will see shortly that proper bootstrapping of the model can shed some light on the discrepancies.
What is unique about ALE compared to other approaches is that it visualizes the effect of each variable irrespective of interactions that might or might not exist with other variables, whether these other interacting variables are included in the model or not.
We can also use ale()
to visualize the possible
existence of any interactions by specifying complete_d = 2
to calculate all 2D interactions:
ale_lm_attitude_2D <- ale(
attitude, lm_attitude,
complete_d = 2,
parallel = 2 # CRAN limit (delete this line on your own computer)
)
# Create ale_plots object
attitude_plots <- plot(ale_lm_attitude_2D)
attitude_2D_plots <- attitude_plots$distinct$rating$plots[[2]]
# Print plots
attitude_2D_plots |>
# extract list of x1 ALE outputs
purrr::walk(\(it.x1) {
# plot all x2 plots in each .x1 element
patchwork::wrap_plots(it.x1, ncol = 2) |>
print()
})
This is a powerful use-case for the ale
package: it can
be used to explore the existence of interactions after the fact; they do
not need to be hypothesized beforehand. However, without bootstrapping,
such findings cannot be considered reliable. In this case, there are no
interactions in this dataset, so we will not explore them further.
We have referred frequently to the importance of bootstrapping. None
of our model results, with or without ALE, should be considered reliable
without being bootstrapped. For large datasets whose models have been
properly trained and evaluated on separate subsets before ALE analysis,
ale()
bootstraps the ALE results of the final deployment
model on the full dataset. However, when a dataset is too small to be
subdivided into training and test sets, then the entire model should be
bootstrapped, not just the ALE data from a single deployment model. That
is, multiple models should be trained, one on each bootstrap sample. The
reliable results are the average results of all the bootstrap models,
however many there are.
The model_bootstrap()
function automatically carries out
full-model bootstrapping suitable for small datasets. Specifically,
it:
model_bootstrap()
has two required arguments. Consistent
with tidyverse conventions, its first argument is a dataset,
data.
The second argument is the model object to be
analyzed. For objects that follow standard R modelling conventions,
model_bootstrap()
should be able to automatically recognize
and parse the model object. So, here is the call to
model_bootstrap()
:
mb_lm <- model_bootstrap(
attitude,
lm_attitude,
boot_it = 10, # 100 by default but reduced here for a faster demonstration
parallel = 2 # CRAN limit (delete this line on your own computer)
)
By default, model_bootstrap()
creates 100 bootstrap
samples of the provided dataset and creates 100 + 1 models on the data
(one for each bootstrap sample and then once for the original dataset).
(However, so that this illustration runs faster, we demonstrate it here
with only 10 iterations.) Beyond ALE data, it also provides bootstrapped
overall model statistics (provided through broom::glance()
)
and bootstrapped model coefficients (provided through
broom::tidy()
). Any of the default options for
broom::glance()
, broom::tidy()
, and
ale()
can be customized, along with defaults for
model_bootstrap()
, such as the number of bootstrap
iterations. You can consult the help file for these details with
help(model_bootstrap)
.
model_bootstrap()
returns a list with the following
elements (depending on values requested in the output
argument:
model_stats
: bootstrapped results from
broom::glance()
model_coefs
: bootstrapped results from
broom::tidy()
ale_data
: bootstrapped ALE data and plotsboot_data
: full bootstrap data (not returned by
default)Here are the bootstrapped overall model statistics:
mb_lm$model_stats
#> # A tibble: 12 × 7
#> name boot_valid conf.low median mean conf.high sd
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 r.squared NA 6.78e-1 0.822 0.793 0.874 7.58e-2
#> 2 adj.r.squared NA 5.94e-1 0.775 0.739 0.841 9.56e-2
#> 3 sigma NA 4.62e+0 5.91 6.03 7.65 1.05e+0
#> 4 statistic NA 8.07e+0 17.7 16.9 26.7 6.86e+0
#> 5 p.value NA 3.53e-9 0.000000159 0.0000203 0.0000922 3.62e-5
#> 6 df NA 6 e+0 6 6 6 0
#> 7 df.residual NA 2.3 e+1 23 23 23 0
#> 8 nobs NA 3 e+1 30 30 30 0
#> 9 mae 7.08 5.70e+0 NA NA 10.2 1.62e+0
#> 10 sa_mae_mad 0.597 3.82e-1 NA NA 0.709 1.21e-1
#> 11 rmse 8.34 6.47e+0 NA NA 11.9 1.85e+0
#> 12 sa_rmse_sd 0.638 4.59e-1 NA NA 0.748 9.47e-2
Here are the bootstrapped model coefficients:
mb_lm$model_coefs
#> # A tibble: 7 × 6
#> term conf.low median mean conf.high std.error
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 (Intercept) -15.4 6.57 8.40 37.1 19.9
#> 2 complaints 0.370 0.556 0.561 0.772 0.144
#> 3 privileges -0.325 0.0323 -0.0575 0.187 0.199
#> 4 learning 0.0385 0.233 0.227 0.434 0.131
#> 5 raises -0.0105 0.169 0.215 0.472 0.179
#> 6 critical -0.303 0.120 0.0300 0.302 0.235
#> 7 advance -0.509 -0.0816 -0.173 0.133 0.239
Here we can visualize the results of the ALE plots.
mb_lm_plots <- plot(mb_lm)
mb_lm_1D_plots <- mb_lm_plots$distinct$rating$plots[[1]]
patchwork::wrap_plots(mb_lm_1D_plots, ncol = 2)
The key to interpreting the effects in these models is contrasting the grey bootstrapped confidence bands surrounding the average (median) ALE effect with the thin horizontal grey band labelled ‘median ± 2.5%’. Anything within ± 2.5% of the median is in the 5% middle of the data. Only bootstrapped effects that are clearly beyond this middle band may be considered significant. By this criteria, considering that the median rating was 65.5%, we can conclude that:
Although this is a basic demonstration, it readily shows how crucial proper bootstrapping is to make meaningful inferences from data analysis.
A major limitation of OLS regression is that it models all relationships between the x variables and y as straight lines. But it is unlikely that all relationships are truly linear. OLS cannot accurately capture non-linear relationships.
Because the samples here are relatively small, we will use general additive models (GAM) for the modelling. To grossly oversimplify things, GAM is an extension of statistical regression analysis that lets the model fit flexible patterns in the data instead of being restricted to the best-fitting straight line. It is an ideal approach for samples that are too small for machine learning because it provides flexible curves unlike ordinary least squares regression yet will not overfit excessively as would most machine learning techniques when working with such small samples.
With GAM, the variables that we want to become flexible need to be
wrapped in the s
(smooth) function, e.g.,
s(complaints)
. For this example, we will smooth all our
numerical input variables:
gam_attitude <- mgcv::gam(rating ~ complaints + privileges + s(learning) +
raises + s(critical) + advance,
data = attitude)
summary(gam_attitude)
#>
#> Family: gaussian
#> Link function: identity
#>
#> Formula:
#> rating ~ complaints + privileges + s(learning) + raises + s(critical) +
#> advance
#>
#> Parametric coefficients:
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) 36.97245 11.60967 3.185 0.004501 **
#> complaints 0.60933 0.13297 4.582 0.000165 ***
#> privileges -0.12662 0.11432 -1.108 0.280715
#> raises 0.06222 0.18900 0.329 0.745314
#> advance -0.23790 0.14807 -1.607 0.123198
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> Approximate significance of smooth terms:
#> edf Ref.df F p-value
#> s(learning) 1.923 2.369 3.761 0.0312 *
#> s(critical) 2.296 2.862 3.272 0.0565 .
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> R-sq.(adj) = 0.776 Deviance explained = 83.9%
#> GCV = 47.947 Scale est. = 33.213 n = 30
By comparing the adjusted R2 of the OLS model (0.663) with that of the GAM model (0.776), we can readily see that the GAM model provides a superior fit to the data.
To understand which variables were responsible for this relationship, the results for the smooth terms in GAM are not readily interpretable. They need to be visualized for effective interpretation—ALE is perfect for such purposes.
ale_gam_attitude_simple <- ale(
attitude, gam_attitude,
parallel = 2 # CRAN limit (delete this line on your own computer)
)
gam_attitude_simple_plots <- plot(ale_gam_attitude_simple)
gam_attitude_simple_1D_plots <- gam_attitude_simple_plots$distinct$rating$plots[[1]]
patchwork::wrap_plots(gam_attitude_simple_1D_plots, ncol = 2)
Compared to the OLS results above, the GAM results provide quite a surprise concerning the shape of the effect of employees’ perceptions that their department is too critical–it seems that both low criticism and very high criticism negatively affect ratings. However, before trying to interpret these results, we must remember that results that are not bootstrapped are simply not reliable. So, let us see what bootstrapping will give us.
mb_gam <- model_bootstrap(
attitude,
gam_attitude,
boot_it = 10, # 100 by default but reduced here for a faster demonstration
parallel = 2 # CRAN limit (delete this line on your own computer)
)
mb_gam$model_stats
#> # A tibble: 9 × 7
#> name boot_valid conf.low median mean conf.high sd
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 df NA 8.18 14.5 14.4 20.6 4.62
#> 2 df.residual NA 9.45 15.5 15.6 21.8 4.62
#> 3 nobs NA 30 30 30 30 0
#> 4 adj.r.squared NA 0.851 0.981 0.943 1.00 0.0675
#> 5 npar NA 23 23 23 23 0
#> 6 mae 12.1 6.20 NA NA 53.0 16.6
#> 7 sa_mae_mad 0.364 -1.25 NA NA 0.681 0.653
#> 8 rmse 15.1 7.49 NA NA 68.8 22.1
#> 9 sa_rmse_sd 0.383 -1.41 NA NA 0.699 0.730
mb_gam$model_coefs
#> # A tibble: 2 × 6
#> term conf.low median mean conf.high std.error
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 s(learning) 1.20 4.55 4.92 8.98 3.34
#> 2 s(critical) 1.97 4.56 4.49 7.12 2.19
mb_gam_plots <- plot(mb_gam)
mb_gam_1D_plots <- mb_gam_plots$distinct$rating$plots[[1]]
patchwork::wrap_plots(mb_gam_1D_plots, ncol = 2)
The bootstrapped GAM results tell a rather different story from the OLS results. In this case, the bootstrap confidence bands of all the variables (even of complaints) fully overlap with the entirety of the median non-significance region. Even the average slopes have vanished from all variables except for complaint, where it remains positive, yet insignificant because of the wide confidence interval.
So, what should we conclude? First, it is tempting to retain the OLS results because they tell a more interesting story. But we consider that this would be irresponsible since the GAM model is clearly superior in terms of adjusted R2: it is the model that far more reliably tells us what is really going on. And what does it tell us?
No doubt, the inconclusive results are because the dataset is so small (only 30 rows). A dataset even double that size might show significant effects at least for complaints, if not for other variables.
model_call_string
argument for non-standard modelsmodel_bootstrap()
accesses the model object and
internally modifies it to retrain the model on bootstrapped datasets. It
should be able to automatically manipulate most R model objects that are
used for statistical analysis. However, if an object does not follow
standard conventions for R model objects, model_bootstrap()
might not be able to manipulate it. If so, the function will fail early
with an appropriate error message. In that case, the user must specify
the model_call_string
argument with a character string of
the full call for the model with boot_data
as the data
argument for the call. (boot_data
is a placeholder for the
bootstrap datasets that model_bootstrap()
will internally
work with.)
To show how this works, let’s pretend that the mgcv::gam
object needs such special treatment. To construct, the
model_call_string
, we must first execute the model and make
sure that it works. We did that earlier but we repeat it here for this
demonstration
gam_attitude_again <- mgcv::gam(rating ~ complaints + privileges + s(learning) +
raises + s(critical) + advance,
data = attitude)
summary(gam_attitude_again)
#>
#> Family: gaussian
#> Link function: identity
#>
#> Formula:
#> rating ~ complaints + privileges + s(learning) + raises + s(critical) +
#> advance
#>
#> Parametric coefficients:
#> Estimate Std. Error t value Pr(>|t|)
#> (Intercept) 36.97245 11.60967 3.185 0.004501 **
#> complaints 0.60933 0.13297 4.582 0.000165 ***
#> privileges -0.12662 0.11432 -1.108 0.280715
#> raises 0.06222 0.18900 0.329 0.745314
#> advance -0.23790 0.14807 -1.607 0.123198
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> Approximate significance of smooth terms:
#> edf Ref.df F p-value
#> s(learning) 1.923 2.369 3.761 0.0312 *
#> s(critical) 2.296 2.862 3.272 0.0565 .
#> ---
#> Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
#>
#> R-sq.(adj) = 0.776 Deviance explained = 83.9%
#> GCV = 47.947 Scale est. = 33.213 n = 30
Once we’re sure that the model call works, then the
model_call_string
is constructed with three simple
steps:
<-
) in quotes.boot_data
.model_bootstrap()
as the
model_call_string
argument (the argument must be explicitly
named).So, here is the form of the call to model_bootstrap()
for a non-standard model object type:
mb_gam_non_standard <- model_bootstrap(
attitude,
gam_attitude_again,
model_call_string = 'mgcv::gam(rating ~ complaints + privileges + s(learning) +
raises + s(critical) + advance,
data = boot_data)',
boot_it = 10, # 100 by default but reduced here for a faster demonstration
parallel = 2 # CRAN limit (delete this line on your own computer)
)
mb_gam_non_standard$model_stats
#> # A tibble: 9 × 7
#> name boot_valid conf.low median mean conf.high sd
#> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
#> 1 df NA 8.18 14.5 14.4 20.6 4.62
#> 2 df.residual NA 9.45 15.5 15.6 21.8 4.62
#> 3 nobs NA 30 30 30 30 0
#> 4 adj.r.squared NA 0.851 0.981 0.943 1.00 0.0675
#> 5 npar NA 23 23 23 23 0
#> 6 mae 12.1 6.20 NA NA 53.0 16.6
#> 7 sa_mae_mad 0.364 -1.25 NA NA 0.681 0.653
#> 8 rmse 15.1 7.49 NA NA 68.8 22.1
#> 9 sa_rmse_sd 0.383 -1.41 NA NA 0.699 0.730
Everything else works as usual.