This post describes a continuous-time model of Bayesian learning about a binary state. It complements the discrete-time models discussed in previous posts (see, e.g., here or here). I present the model, discuss its learning dynamics, and derive these dynamics analytically.
The model has been used to study decision times (Fudenberg et al., 2018), experimentation (Bolton and Harris, 1999; Moscarini and Smith, 2001), information acquisition (Morris and Strack, 2019), and persuasion (Liao, 2021). It also underlies the drift-diffusion model of reaction times used by psychologists—see Ratcliff (1978) for an early example, and Hébert and Woodford (2023) or Smith (2000) for related discussions.
Model
Suppose I want to learn about a state \(\mu\)
that may be high (equal to \(H\)
) or low (equal to \(L<H\)
).
I observe a continuous sample path \((X_t)_{t\ge0}\)
with random, instantaneous increments
$$\DeclareMathOperator{\E}{E} \newcommand{\der}{\mathrm{d}} \newcommand{\R}{\mathbb{R}} \der X_t=\mu\der t+\sigma \der W_t,$$
where \(\sigma>0\)
amplifies the noise generated by the standard Wiener process \((W_t)_{t\ge0}\)
.
These increments provide noisy signals of the state \(\mu\)
.
I use these signals, my prior belief \(p_0=\Pr(\mu=H)\)
, and Bayes’ rule to form a posterior belief
$$p_t\equiv \Pr\left(\mu=H\mid (X_s)_{s<t}\right)$$
about \(\mu\)
given the sample path observed up to time \(t\)
.
As shown below, this posterior belief has increments
$$\der p_t=p_t(1-p_t)\frac{(H-L)}{\sigma}\der Z_t,$$
where \((Z_t)_{t\ge0}\)
is a Wiener process with respect to my information at time \(t\)
.
Its increments
$$\der Z_t=\frac{1}{\sigma}\left(\der X_t-\hat\mu_t\der t\right)$$
exceed zero precisely when the corresponding increments \(\der X_t\)
in the sample path exceed my posterior estimates
$$\begin{align} \hat\mu_t &\equiv \E\left[\mu\mid (X_s)_{s<t}\right] \\ &= p_tH+(1-p_t)L. \end{align}$$
Learning dynamics
My belief increments \(\der p_t\)
get smaller as \(p_t\)
approaches zero or one.
The ratio \((H-L)/\sigma\)
controls how quickly this happens.
Intuitively, if \((H-L)\)
is large then the high and low states are easy to tell apart from the trends in \((X_t)_{t\ge0}\)
they imply.
But if \(\sigma\)
is large then these trends are blurred by the random fluctuations \(\sigma\der W_t\)
.
I illustrate these dynamics in the chart below.
It shows the sample paths \((X_t)_{t\ge0}\)
and corresponding beliefs \((p_t)_{t\ge0}\)
when \((H,L,\mu,p_0)=(1,0,H,0.5)\)
and \(\sigma\in\{1,2\}\)
.
I use the same realization of the underlying Wiener process \((W_t)_{t\ge0}\)
for each value of \(\sigma\)
.
Increasing this value slows my convergence to the correct belief \(p_t=1\)
because it makes the signals \(\der X_t\)
less informative about \(\mu=H\)
.
Deriving the belief increments
The increments \(\der W_t\)
of the Wiener process \((W_t)_{t\ge0}\)
are iid normally distributed with mean zero and variance \(\der t\)
:
$$\der W_t\sim N(0,\der t).$$
Thus, given \(\mu\)
, the increments \(\der X_t\)
of the sample path \((X_t)_{t\ge0}\)
are iid normal with mean \(\mu\der t\)
and variance \(\sigma^2\der t\)
:
$$\der X_t\mid\mu\sim N(\mu\der t,\sigma^2\der t).$$
So these increments have conditional PDF
$$\begin{align} f_\mu(\der X_t) &= \frac{1}{\sigma\sqrt{2\pi\der t}}\exp\left(-\frac{(\der X_t-\mu\der t)^2}{2\sigma^2\der t}\right) \\ &= \frac{1}{\sigma\sqrt{2\pi\der t}}\exp\left(-\frac{(\der X_t)^2}{2\sigma^2\der t}\right)\exp\left(\frac{\mu\der X_t}{\sigma^2}-\frac{\mu^2\der t}{2\sigma^2}\right). \end{align}$$
But the rules of Itô calculus imply \((\der X_t)^2=\sigma^2\der t\)
and
$$\begin{align} \exp\left(\frac{\der X_t\mu}{\sigma^2}-\frac{\mu^2\der t}{2\sigma^2}\right) &= \sum_{k\ge0}\frac{1}{k!}\left(\frac{\mu\der X_t}{\sigma^2}-\frac{\mu^2\der t}{2\sigma^2}\right)^k \\ &= 1+\frac{\mu\der X_t}{\sigma^2} \end{align}$$
because these rules treat terms of order \((\der t)^2\)
or smaller as equal to zero.
Thus
$$f_\mu(\der X_t)=\frac{1}{\sigma^3\sqrt{2\pi\der t}}\exp\left(-\frac{1}{2}\right)\left(\mu\der X_t+\sigma^2\right)$$
for each \(\mu\in\{H,L\}\)
.
Applying Bayes’ rule then gives
$$\begin{align} p_{t+\der t} &= \frac{p_tf_H(\der X_t)}{p_tf_H(\der X_t)+(1-p_t)f_L(\der X_t)} \\ &= \frac{p_t\left(H\der X_t+\sigma^2\right)}{\hat\mu_t\der X_t+\sigma^2}, \end{align}$$
where \(\hat\mu_t=\E[\mu\mid (X_s)_{s<t}]\)
is my posterior estimate of \(\mu\)
.
So the belief process \((p_t)_{t\ge0}\)
has increments
$$\begin{align} \der p_t &\equiv p_{t+\der t}-p_t \\ &= \frac{p_t(1-p_t)(H-L)\der X_t}{\hat\mu_t\der X_t+\sigma^2}. \end{align}$$
Finally, taking a Maclaurin series expansion and applying the rules of Itô calculus gives
$$\begin{align} \frac{\der X_t}{\hat\mu_t\der X_t+\sigma^2} &= \der X_t\sum_{k\ge0}\frac{(-1)^kk!}{(\sigma^2)^{k+1}}(\der X_t)^k \\ &= \der X_t\left(\frac{1}{\sigma^2}-\frac{1}{\sigma^4}\der X_t\right) \\ &= \frac{1}{\sigma^2}\left(\der X_t-\hat\mu_t\der t\right), \end{align}$$
from which we obtain the expressions for \(\der p_t\)
and \(\der Z_t\)
provided above.