# 6 Introduction to Instance-level Exploration

Instance-level exploration methods help to understand how does a model yield a prediction for a particular single observation. We can think about the following situations as examples:

- We may want to evaluate effects of explanatory variables on the model’s predictions. For instance, we may be interested in predicting the risk of heart attack based on a person’s age, sex, and smoking habits. A model may be used to construct a score (for instance, a linear combination of the explanatory variables representing age, sex, and smoking habits) that could be used for the purposes of prediction. For a particular patient, we may want to learn how much do the different variables contribute to the score?
- We may want to understand how would the model’s predictions change if values of some of the explanatory variables changed? For instance, what would be the predicted risk of heart attack if the patient cut the number of cigarettes smoked per day by half?
- We may discover that the model is providing incorrect predictions and we may want to find the reason. For instance, a patient with a very low risk-score experienced a heart attack. What has driven the wrong prediction?

In this part of the book, we describe the most popular approaches to instance-level exploration. They can be divided into three classes:

- One approach is to analyze how does the model’s prediction for a particular instance differ from the average prediction and how can the difference be distributed among explanatory variables? This method is often called the “variable attributions” approach. An example is provided in panel A of Figure 6.1. Chapters 7-9 present various methods for implementing this approach.
- Another approach uses the interpretation of the model as a function and investigates the local behaviour of this function around the point (observation) of interest \(\underline{x}_*\). In particular, we analyze the curvature of the model response (prediction) surface around \(\underline{x}_*\). In case of a black-box model, we may approximate it with a simpler glass-box model around \(\underline{x}_*\). An example is provided in panel B of Figure 6.1. Chapter 10 presents the Local Interpretable Model-agnostic Explanations (LIME) method that exploits the concept of a “local model”.
- Yet another approach is to investigate how does the model’s prediction change if the value of a single explanatory variable changes? The approach is useful in the so-called “What-if” analyses. In particular, we can construct plots presenting the change in model-based predictions induced by a change of a single explanatory variable. Such plots are usually called ceteris-paribus (CP) profiles. An example is provided in panel C in Figure 6.1. Chapters 11-13 introduce the CP profiles and methods based on them.

Each method has its own merits and limitations. They are briefly discussed in the corresponding chapters. Chapter 14 offers a comparison of the methods.