Techniques for Model Interpretability (Feature Importance, Partial Dependence)

When it comes to machine learning models, it is not enough to have accurate predictions. It is equally important to understand how these predictions are being made. This is where model interpretability techniques come into play. They allow us to gain insights into the inner workings of the model and understand the importance of different features.

Feature Importance

One of the most common techniques for model interpretability is feature importance. Feature importance is a way to determine which features have the most influence on the predictions made by the model. In scikit-learn, there are different ways to calculate feature importance depending on the type of model used.

For tree-based models such as the decision tree or random forest, feature importance is typically calculated based on the Gini importance or mean decrease impurity. Gini importance measures the total reduction of the Gini impurity criterion brought by a particular feature, while mean decrease impurity calculates the average reduction in impurity for each feature over all the trees in the forest.

To obtain the feature importances in scikit-learn, you can use the feature_importances_ attribute of the trained model. For example:

from sklearn.ensemble import RandomForestClassifier

model = RandomForestClassifier(), y)

importance = model.feature_importances_

The importance variable will then contain an array of values representing the importance of each feature, which can be further analyzed or visualized to gain insights.

Partial Dependence

While feature importance tells us the relative importance of each feature, it does not provide a clear understanding of how a particular feature affects the model's predictions. This is where partial dependence comes into play. Partial dependence allows us to understand the relationship between a feature and the predicted outcome while holding other features constant.

Scikit-learn provides a useful method called plot_partial_dependence() that enables us to visualize the partial dependence of a feature. This function takes a trained model, the input data, and the indices of the features of interest. It then generates plots showing how the predicted outcome varies as we change the values of the selected features, while keeping the other features fixed.

Here's an example of how to use the plot_partial_dependence() function with scikit-learn:

from sklearn.ensemble import GradientBoostingRegressor
from sklearn.inspection import plot_partial_dependence

model = GradientBoostingRegressor(), y)

plot_partial_dependence(model, X, [0, 1])  # plots partial dependence of features at indices 0 and 1

The plots generated will provide insights into the relationships between the selected features and the predicted outcome. This can be particularly useful in understanding how changing different features might impact the model's predictions.


Techniques for model interpretability such as feature importance and partial dependence are essential for understanding and trusting the predictions made by machine learning models. Scikit-learn provides built-in functionality to calculate feature importance and visualize partial dependence, making it easier for data scientists and researchers to gain insights into their models. By utilizing these techniques, we can better understand the inner workings of our models and make more informed decisions based on the predictions they provide.

noob to master © copyleft