Chapter 9 of Kahneman et al. (2021) discusses how predictions made by humans can be less accurate than predictions made using statistical models. Part of the chapter describes research by Goldberg (1970) and subsequent authors showing that models of human predictions can out-perform the humans on which those models are based.
For example, suppose I’m asked to make predictions in a range of contexts . My goal is to use some contextual data to predict the value of a context-specific outcome . I generate predictions where the are context-specific errors. The accuracy of my predictions can be measured via their mean squared error (MSE) where a lower MSE implies higher accuracy. Another way to generate predictions could be to posit a linear model where is a row vector of coefficients and the are random errors. But I don’t know the true outcomes —hence needing to predict them—and so I can’t just use ordinary least squares (OLS) to estimate . Instead, Goldberg (1970) suggests replacing this linear model with where is a (possibly different) vector of coefficients and the are (possibly different) random errors. This second model describes the linearized relationship between my (possibly incorrect) predictions and the data on which those predictions are based. Since I know my predictions , I can use OLS to obtain an estimate of and produce a set of “modeled predictions” The difference between the and is that the latter ignore the non-linearities in my method for generating predictions. Intuitively, the represent what I would predict using a simple, linear formula; my predictions may be generated using a formula that is much more complex, or may not be generated using a formula at all.
So, how do my raw predictions and their modeled counterparts compare? The chart below plots the and against the true values when
- the and are iid standard normal, and
- with iid standard normal.
The modeled predictions are far more accurate: they have an MSE of 0.22, whereas my raw predictions have an MSE of 0.76. In this case, the true relationship between the and is linear, and so a linear model of my predictions is well-placed to out-perform those predictions.
However, modeling predictions does not always improve their accuracy. For example, suppose the contextual data are scalars, and the , , and have zero means. Then the MSE of the modeled predictions turns out to be where and are the variances of the and , where is the correlation of the and , and where is the correlation of the and . Consequently, replacing my raw predictions with their modeled counterparts leads to an accuracy improvement if and only if This condition holds in the example plotted above: both and equal unity, but is much larger in absolute value than . In general, the condition is most likely to hold when
- is larger than (i.e., my raw predictions are relatively noisy);
- is large (i.e., the relationship between the and is approximately linear and deterministic); and
- is small (i.e., the errors in my raw predictions are relatively uncorrelated with the ).
Intuitively, if the outcomes are a linear function of the (i.e., if ) then linearizing my predictions improves their accuracy by removing non-linear errors. On the other hand, if my prediction errors are a linear function of the (i.e., if ) then linearizing my predictions cannot improve their accuracy because there are no non-linear errors to remove.