Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Fair value estimation

Introduction

In this chapter we introduce techniques to estimate the fair value of a financial instrument exploiting the pricing information available in financial markets.

Fair value is the price at which a financial instrument is economically equivalent, at a given time and under a given information set, to its future cash flows or payoffs, once transactions costs, opportunity costs and risk are appropriately accounted for

In this chapter we will focus on two conceptually different ways to estimate fair value. The first one assumes markets are efficient enough so that prices observed in the market are our best estimation of fair value, bar some idiosyncratic components coming from trading fictions like liquidity premiums, dealer spreads, and transaction costs. In this case, fair value estimation can be seen as a filtering problem, and we will study a simple albeit powerful model to carry out this task: the Kalman filter.

When financial instruments are highly illiquid or do not trade directly in organized markets, fair value estimation must rely on economic—often referred to as fundamental—valuation models. These models infer the value of a financial instrument from the future cash flows specified by its contractual structure. Central to this approach is the concept of the time value of money, which states that a payment received in the future is not economically equivalent to the same payment received today, due to opportunity costs. As a result, future cash flows must be converted into present values through the application of a discount factor, which renders cash flows occurring at different points in time comparable.

Time discounting, however, is not the only challenge faced by fundamental valuation models. Future cash flows are frequently contingent on information that is not known at the valuation date, such as future prices of financial assets or macroeconomic variables. Examples include dividends paid by a company or the value of the underlying asset referenced by a derivative contract. A simple, albeit theoretically naive, approach consists in valuing the instrument as the expected value of its future cash flows, treated as random variables. This approach, however, fails to account for heterogeneity in investors’ risk preferences. Once cash flows are stochastic, the realized return on the investment becomes uncertain, and uncertainty is not valued equally by all investors. To address this limitation, one can adopt a utility indifference pricing framework, in which risk preferences explicitly enter the valuation.

In many situations, particularly for illiquid flow instruments, this is the most refined valuation approach available. However, for the specific case of derivative instruments, stronger theoretical results can be obtained under additional assumptions. As shown by Fischer Black, Myron Scholes, and Robert C. Merton in the 1970s, it is possible to construct dynamic replication portfolios that reproduce the payoffs of a derivative using traded instruments. In such settings, the fair value of the derivative becomes independent of investors’ risk aversion, since any deviation from this price would give rise to risk-free arbitrage opportunities. This insight leads to the arbitrage-free pricing framework, which will be introduced in the final section of this chapter.

Finally, a unifying pricing framework that can accommodate both the risk-aversion profile of the investor and the arbitrage-free constraints is the stochastic discount factor pricing framework Cochrane, 2005, which we will briefly describe at the end of the chapter.

Filtering models for fair value estimation

As mentioned in the introduction, for financial instruments that are relatively liquid, we can aim at extracting all the pricing information from price indications and trades in the market, without having to resort to economic theories of fair value. In this setup, we consider as their fair value the one that market participants are willing to pay for.

The issue, though, is that price indications and trades cannot be considered themselves pure observations of fair value, since they might be affected by market frictions: bid ask spreads, particularities of the negotiation mechanism, liquidity fluctuations, specific needs of market participants at a given time, etc. When instruments trade in limit order books, a popular estimation of the fair value is using the mid-price, the arithmetic average of the best bid and ask. However, if bid-ask spreads are wide of liquidity is thin in the first levels, such estimation is not necessary very precise. Trades provide a lot of information, since they are real transaction and not indications of interests, the larger they are in principle the more information. Still, they are subject to the aforementioned market frictions that reduce their reliability.

These makes all these price observations noisy estimates of the fair value, so if we want to estimate a fair value out of them we need to be able to separate the signal from the noise, or in other words, filter those observations. This is precisely what, under certain model assumptions, a Kalman filter does.

The Kalman filter was introduced in the chapter on Bayesian Theory. It is a Bayesian filtering algorithm that allows to perform exact inference, i.e. compute the closed-form distribution, of the latent state vector in a Linear Gaussian State Space Model (LG-SSM).

Recall that a State Space Model (SSM) is a model to describe dynamic systems where we have a non or partially observable state, a vector x\mathbf{x}, whose dynamics in time is described by a so-called transition equation:

xt+Δt=f(xt,ut)+ϵt\mathbf{x}_{t+\Delta t} = f(\mathbf{x}_t, \mathbf{u}_t) + \mathbf{\epsilon}_t

where f(xt,ut)f(\mathbf{x}_t, \mathbf{u}_t) is a general function, ut\mathbf{u}_t are inputs (or controls) that affect the dynamics, Δt\Delta t is the time-step between observations, and ϵt\mathbf{ \epsilon}_t is a transition noise with a given distribution. The state is observed indirectly via a proxy vector y\mathbf{y} via the observation equation:

yt=g(xt,ut)+η\mathbf{y}_t = g(\mathbf{x}_t, \mathbf{u}_t) + \mathbf{\eta}

where g(xt,ut)g(\mathbf{x}_t, \mathbf{u}_t) is another general function and ηt\mathbf{\eta}_t the observation noise, meaning that observations have a degree of uncertainty with respect to the latent space.

A Linear Gaussian Model (LGM) is a specific case of the SSM were both the transition and observation functions are linear and the noise terms are Gaussian. In this case, we can use the Kalman filter algorithm to compute the distribution of the state vector at any time, given the observations and the transition and observation model. If some or all the parameters of these models are not known, they can be estimated using standard techniques like Maximum Likelihood Estimation (MLE) or Expectation Maximization (EM) when the former becomes computationally intractable due to the latent state vector.

For non-Linear Gaussian Models, there are extensions of the Kalman filter that can be used:

A simple pricing model

Let us consider a simple setup where we aim to infer the distribution of the fair value mtm_t of a financial instrument that follows a random walk:

mt+Δt=mt+ϵt,ϵtN(0,σϵ2Δt)m_{t+\Delta t} = m_t + \epsilon_t, \epsilon_t \sim N(0, \sigma_\epsilon^2 \Delta t)

We don’t observe this fair value, only trades which we consider noisy observation of the mid since they include transaction costs and potentially other external factors like dealer inventory positions, etc:

pt=mt+νt,νtN(0,σν2)p_t = m_t + \nu_t, \nu_t \sim N(0, \sigma_\nu^2)

Readers will recognize that this is the local level model discussed extensively in the Chapter on Bayesian Modelling. For the observation noise we can introduce prior business knowledge about the confidence we have on trade observations as a source of pricing information. In his Option Trading’s book Sinclair, 2010, Euan Sinclair describes a simple model that quantifies the information provided by trades based on the size of the trade, vv:

σν(v)=σp(vmaxv1))+\sigma_\nu (v)= \sigma_p \left(\frac{v_\text{max}}{v}-1)\right)^+

where σp\sigma_p is a baseline observation noise and vmaxv_\text{max} is an input to the model, the trade size we believe saturates information the information provided in the sense that our mid estimation will essentially move the price of the trade. In contrast, trades of small size, vvmaxv \ll v_\text{max}, will have σp\sigma_p \rightarrow \infty and will provide a negligible pricing information. An alternative simple model is:

σν(v)=σpv0v\sigma_\nu(v) = \sigma_p \frac{v_0}{v}

where v0v_0 in this case is a size scale that separates the regimes where the information provided by the trade is negligible, vv0v \ll v_0, or relevant vv0v \gg v_0, but it does not saturate for a specific trading size, as in Sinclair’s model. Of course nothing prevents to use more business prior knowledge to enrich the observation model with other observable characteristics of the trade or the wider market context.

Estimation of the simple pricing model

As discussed in Bayesian Modelling, the standard way to estimate the parameters of a Kalman Filter is using the Expectation Maximization (EM) algorithm, suitable for probabilistic models with latent variables. However, the properties of the simple pricing model can be exploited to obtain closed-form estimators for its parameters using moment matching.

Let us start by working with a model where observation errors have no dependency on the volume: σν(v)=σν\sigma_\nu (v) = \sigma_\nu. The key is to compute statistics of:

dtpt+Δtpt=(mt+Δtmt)+(νt+Δtνt)=ϵt+(νt+Δtνt)d_t \equiv p_{t+\Delta t} - p_t = (m_{t+\Delta t}- m_t) + (\nu_{t+\Delta t} - \nu_t) = \epsilon_t + (\nu_{t+\Delta t} - \nu_t)

which depend only on observed trades. First we compute the variance:

Var[dt]=Var[ϵt]+Var[(νt+Δtνt)]=σϵ2Δt+2σν2Var[d_t]= Var[\epsilon_t] + Var[(\nu_{t+\Delta t} - \nu_t)]= \sigma_\epsilon^2 \Delta t + 2 \sigma_\nu^2

where we have used that ϵt\epsilon_t, νt+Δt\nu_{t+\Delta t} and νt\nu_t are independent random variables. This expression links the variance of the first differences in trade prices with the parameters to estimate. We need though a second expression to solve for each parameters separately. For that we compute the lag-1 auto-covariance of dtd_t:

Cov[dt,dtΔt]=Cov[ϵt+(νt+Δtνt),ϵtΔt+(νtνtΔt)]=Var[νt]=σν2Cov[d_t,d_{t-\Delta t}]=Cov[\epsilon_t + (\nu_{t+\Delta t} - \nu_t), \epsilon_{t-\Delta t} + (\nu_t - \nu_{t-\Delta t})] = -Var[\nu_t] = -\sigma_\nu^2

where again we have used the independence between the noise terms. Our estimators read then:

σ^ϵ2Δt=1N1i=1N1dti22σ^ν2\hat{\sigma}_\epsilon^2 \Delta t = \frac{1}{N-1} \sum_{i=1}^{N-1} d_{t_i}^2 - 2 \hat{\sigma}_\nu^2
σ^ν2=1N2i=2N1dtidtiΔt\hat{\sigma}_\nu^2 = -\frac{1}{N-2}\sum_{i=2}^{N-1} d_{t_i} d_{t_{i}-\Delta t}

The results have an interesting interpretation:

In the case of volume dependent observation errors, we can still compute these statistics, which now read:

Var[dt]=σϵ2Δt+σν2(vt+Δt)+σν2(vt)Var[d_t]= \sigma_\epsilon^2 \Delta t + \sigma_\nu^2(v_{t+\Delta t}) + \sigma_\nu^2(v_t)
Cov[dt,dtΔt]=σν2(vt)Cov[d_t,d_{t-\Delta t}]= -\sigma_\nu^2 (v_t)

The statistical estimator of the variance of the first differences can still be used, by accounting by the variability of the error with volume (heteroskedasticity):

E[1N1i=1N1dti2]=1N1i=1N1(σϵ2Δt+σν2(vt+Δt)+σν2(vt))=σϵ2Δt+1N1i=1N1(σν2(vt+Δt)+σν2(vt))E[\frac{1}{N-1} \sum_{i=1}^{N-1} d_{t_i}^2]= \frac{1}{N-1} \sum_{i=1}^{N-1} \left(\sigma_\epsilon^2 \Delta t + \sigma_\nu^2(v_{t+\Delta t}) + \sigma_\nu^2(v_t)\right)=\sigma_\epsilon^2 \Delta t+\frac{1}{N-1} \sum_{i=1}^{N-1} \left( \sigma_\nu^2(v_{t+\Delta t}) + \sigma_\nu^2(v_t)\right)

Moving into the 1-lag covariance, we have:

E[1N2i=2N1dtidtiΔt]=1N2i=2N1σv2(vti)E[\frac{1}{N-2}\sum_{i=2}^{N-1} d_{t_i} d_{t_{i}-\Delta t}] = - \frac{1}{N-2}\sum_{i=2}^{N-1} \sigma_v^2(v_{t_i})

As far as we the volume dependency has a single parameter to fit, we can still use these two equations to solve for the parameters. If we use, for instance, the simple model σν(v)=σpv0v\sigma_\nu(v) = \sigma_p \frac{v_0}{v}, where v0v_0 is given and σp\sigma_p is to be estimated from data (notice that we could simply estimate σpv0\sigma_p v_0, the factorization is useful for business interpretation):

E[1N1i=1N1dti2]=σϵ2Δt+σp2N1i=1N1(v02vti+Δt2+v02vti2)E[\frac{1}{N-1} \sum_{i=1}^{N-1} d_{t_i}^2]=\sigma_\epsilon^2 \Delta t+ \frac{\sigma_p^2}{N-1} \sum_{i=1}^{N-1} \left( \frac{v^2_0}{v^2_{t_i+\Delta t}} + \frac{v^2_0}{v^2_{t_i}} \right)
E[1N2i=2N1dtidtiΔt]=σp2N2i=2N1v02vti2E[\frac{1}{N-2}\sum_{i=2}^{N-1} d_{t_i} d_{t_{i}-\Delta t}] = - \frac{\sigma_p^2}{N-2}\sum_{i=2}^{N-1} \frac{v^2_0}{v^2_{t_i}}

In this simple case, the estimation of the parameters σp\sigma_p and σϵ\sigma_\epsilon is straightforward. More complex functions like the one from Sinclair cannot be estimated with only two moments. Further moments can be computed to provide extra equations, although at this point it might be worthy to resort to standard estimation techniques like EM if available.

Inference on the simple pricing model

We can use the general Kalman filter equations described in Bayesian Modelling to derive the distribution of our mid - price at the next time t+Δtt + \Delta t where a trade happens.

The Kalman filter algorithm operates sequentially over observation steps applying two steps, the predict step, where we compute the distribution of the fair value based purely on the random walk model, and the update step in which we incorporate the information provided by the observation of a new trade. We define mt+Δttm_{t+\Delta t}^t as the distribution of the fair value at t+Δtt+\Delta t before observing the trade, and mt+Δt+Δtm_{t+\Delta}^{t+\Delta t} afterwards.

Let us apply first the predict step. The distribution of mt+Δttm_{t+\Delta t}^t is Gaussian with mean and variance given by:

mˉt+Δtt=mˉtt\bar{m}_{t+\Delta t}^t = \bar{m}_{t}^t
(σm,t+Δtt)2=(σm,tt)2+σϵ2Δt(\sigma_{m,t+\Delta t}^t)^2 = (\sigma_{m,t}^t)^2 + \sigma_\epsilon^2 \Delta t

Since our model uses a drift-less random walk dynamics for the evolution of the fair value, the updated mean does not change and the variance increases proportionally to the time step Δt\Delta t.

Now we use the update step to incorporate the information from a trade happening at t+Δtt + \Delta t:

mˉt+Δtt+Δt=mˉtt+Δt+Kt(pt+Δtmˉtt+Δt)\bar{m}_{t+\Delta t}^{t+\Delta t} = \bar{m}_{t}^{t+\Delta t} + K_t (p_{t +\Delta t} - \bar{m}_{t}^{t+\Delta t})
(σm,t+Δtt+Δt)2=(σm,t+Δtt)2ση2(σm,t+Δtt)2+ση2(\sigma_{m,t+\Delta t}^{t+\Delta t})^2 = \frac{(\sigma_{m,t+\Delta t}^t)^2 \sigma_\eta^2}{(\sigma_{m,t+\Delta t}^t)^2 + \sigma_\eta^2}

where KtK_t is the Kalman gain, given by:

Kt=(σm,t+Δtt)2(σm,t+Δtt)2+ση2K_t = \frac{(\sigma_{m,t+\Delta t}^t)^2}{(\sigma_{m,t+\Delta t}^t)^2 + \sigma_\eta^2}

The updated mean is an interpolation between the predicted mean and the trade observation, weighted by the Kalman gain

If we look at the new standard deviation, we also find similar limiting behaviors:

One interesting consequence of the optimality of the Kalman filter is that the updated standard deviation cannot be larger than the predicted one, and for any finite ση\sigma_\eta is always smaller: the information from the trade always contributes to improve our estimation of the fair value. This is easily seen writing:

σm,t+Δtt+Δt=σm,t+Δtt1(σm,t+Δttση)2+1\sigma_{m,t+\Delta t}^{t+\Delta t} =\sigma_{m,t+\Delta t}^t \frac{1}{\sqrt{(\frac{\sigma_{m,t+\Delta t}^t}{\sigma_\eta})^2 + 1}}

Since σm,t+Δttση\frac{\sigma_{m,t+\Delta t}^t}{\sigma_\eta} is non-negative, then the denominator is never lower than 1.

In some applications of the local level model to pricing we might also be interested in the Kalman smoothing algorithm. Recall that the difference with the Kalman filtering we have just seen is that in smoothing we estimate the latent variable using all the available data, including the future. Of course this means Kalman smoothing does not make sense for online price inference, but there are other applications of this pricing model where using the best estimation of the latent fair value is relevant:

Multiple observations of the same instrument

In many real pricing situations, we might have different sources that reveal information about the fair value of an instrument, for example trades in different platforms, information from composites, or pricing information derived indirectly from trading indicators like the hit&miss. We will explore later those sources in detail. If they happen asynchronously, we can just use the simple pricing model introduced in the previous section, adjusting the observation error depending on the pricing source.

If they happen synchronously, though, we need to expand the Kalman filter to cope with simultaneous observations. This requires to change the observation model to a system of equations:

pt,i=mt+νt,i,νt,iN(0,σν,i2)p_{t,i} = m_t + \nu_{t,i}, \vec{\nu}_{t,i} \sim N(0, \sigma_{\nu, i}^2)

We can compute the Kalman gain in this case, which is a vector matrix:

Kt,i=(σm,t+Δtt)2/σν,i21+(σm,t+Δtt)2/ΛK_{t,i} = \frac{(\sigma_{m,t+\Delta t}^{\,t})^2/\sigma_{\nu,i}^2}{1 + (\sigma_{m,t+\Delta t}^{\,t})^2/\Lambda}

where:

Λ=i=1n1σν,i2\Lambda = \sum_{i=1}^n \frac{1}{\sigma_{\nu,i}^2}

is the total observation precision. Notice that, as a sanity check, in the case of a single observation we recover the Kalman gain derived in the previous section. For multiple observations, the update equation then reads:

mˉt+Δtt+Δt=mˉtt+Δt+i=1nKt,i(pt+Δt,imˉtt+Δt)\bar{m}_{t+\Delta t}^{t+\Delta t} = \bar{m}_{t}^{t+\Delta t} + \sum_{i=1}^n K_{t,i} (p_{t + \Delta t ,i} - \bar{m}_{t}^{t+\Delta t})

The relative weight of influence of each observation depends on the fraction of the total variance that the observation variance represents, with more noisy observations having a smaller effect in the update.

Multiple correlated instruments

The Kalman filter model for pricing becomes even more relevant when we include information from other financial instruments that are historically correlated with the one whose fair value we are estimated. Typical situations are:

The simple pricing model we have analyzed so far can be easily extended to include information from a set of N instruments. Notice that in this case what we are actually doing is estimating the fair values of all the instruments in the set, not necessarily only the one of interest. The evolution of the fair values is now modelled using:

mt+Δt=mt+ϵt,ϵtN(0,ΣϵΔt)\vec{m}_{t+\Delta t} = \vec{m}_t + \vec{\epsilon}_t, \vec{\epsilon}_t \sim N(\vec{0}, \Sigma_\epsilon \Delta t)

where Σϵ\Sigma_\epsilon is now a covariance matrix that takes into account the effect of correlations in the pricing movements. Price observations follow the model:

pt=mt+νt,νtN(0,Σν)\vec{p}_t = \vec{m}_t + \vec{\nu}_t, \vec{\nu}_t \sim N(0, \Sigma_\nu)

In this case, since we are already modelling correlations at fair value level, a typical choice is to take Σν\Sigma_\nu diagonal, i.e. Σν=diag(σν,12,...,σνN2)\Sigma_\nu = \text{diag}(\sigma_{\nu,1}^2, ..., \sigma_{\nu_N}^2), although in certain setups one might want to include some of form of bid-ask spread correlation between instruments.

With this model specification, we can directly use the filtering, smoothing and EM equations discussed in the Bayesian Modelling chapter. Let us though specifically focus on the case of N=2N=2 instruments, where we can work out in detail the Kalman filter equations to get further insights into the model’s inner workings.

The predict step in the Kalman filter is given by the following equations:

mtt1=mt1t1\vec{m}_{t|t-1} = \vec{m}_{t-1|t-1}
Σtt1=Σt1t1+ΣϵΔt\Sigma_{t|t-1} = \Sigma_{t-1|t-1} + \Sigma_\epsilon \Delta t

which are relatively simple, as expected. The update step equations are more interesting, since they include the effect of observations, and critically the impact of one instrument’s trades into the fair value or the other:

mtt=mtt1+Kt(Pt+Δtmtt1)\vec{m}_{t|t} = \vec{m}_{t|t-1} + K_t ( \vec{P}_{t+\Delta t}-\vec{m}_{t|t-1})
Σtt=(1Kt)Σtt1\Sigma_{t|t} = (1-K_t) \Sigma_{t|t-1}

with the Kalman gain being:

Kt=Σtt1(Σtt1+Σν)1K_t = \Sigma_{t|t-1} ( \Sigma_{t|t-1} + \Sigma_\nu)^{-1}

For the two instrument case, the Kalman gain can be expanded into:

Kt=1(σν,12+(σ1,tt1)2)(σν,22+(σ2,tt1)2)(ρtt1σ1,tt1σ2,tt1)2((σ1,tt1)2σν,22+(σ1,tt1)2(σ2,tt1)2(1(ρtt1)2)ρtt1σ1,tt1σ2,tt1σν,12ρtt1σ1,tt1σ2,tt1σν,22(σ2,tt1)2σν,12+(σ1,tt1)2(σ2,tt1)2(1(ρtt1)2))\begin{aligned} K_t = \frac{1}{(\sigma_{\nu,1}^2 + (\sigma_{1,t}^{t-1})^2)(\sigma_{\nu,2}^2 + (\sigma_{2,t}^{t-1})^2) - (\rho_t^{t-1}\sigma_{1,t}^{t-1}\sigma_{2,t}^{t-1})^2} \nonumber \\ \begin{pmatrix} (\sigma_{1,t}^{t-1})^2 \sigma_{\nu,2}^2 + (\sigma_{1,t}^{t-1})^2(\sigma_{2,t}^{t-1})^2(1- (\rho_t^{t-1})^2) & \rho_t^{t-1} \sigma_{1,t}^{t-1}\sigma_{2,t}^{t-1}\sigma_{\nu,1}^2 \\ \rho_t^{t-1} \sigma_{1,t}^{t-1}\sigma_{2,t}^{t-1}\sigma_{\nu,2}^2 & (\sigma_{2,t}^{t-1})^2\sigma_{\nu,1}^2 + (\sigma_{1,t}^{t-1})^2 (\sigma_{2,t}^{t-1})^2(1-(\rho_t^{t-1})^2) \nonumber \\ \end{pmatrix} \end{aligned}

Let us analyze some particular cases:

priceLet’s evaluate this model in one of the typical scenarios for fair value discovery discussed above: two correlated instruments traded in markets with some periods of non-overlapping trading. The objective is to leverage their correlation to estimate fair values for the instrument whose market is closed. The underlying principle is that new information affecting the price of the actively traded instrument during its market hours would similarly impact the closed-market instrument, if it were tradable.

For that, we first generate synthetic fair values using a correlated Brownian motion with ρ=0.9\rho = 0.9, σ1=5e4\sigma_1 = 5e-4 and σ2=4e4\sigma_2 = 4e-4. Then we generate trades over 22 days but for each day, each day consisting on 60 time-steps to make the simulation efficient. We consider three situations: one in which only the first instrument is traded, on in which only the second instrument is traded, and a third one in which both are simultaneously traded. We use a diagonal observation covariance to generate the trades, i.e. we assume that there is no correlation between the spreads with respect to the fair value, so correlation is driven exclusively by fair value correlations. To generate the trades, we use standard deviations in the observation covariance of 0.032 and 0.045, respectively. Then we use Expectation Maximization (EM) over the first half of the synthetic trade data to estimate the parameters of the model, and run the Kalman filter over the second half of the data to compare the estimations of the fair value to the real simulated values. The results can be seen in the following figure:

Estimation of the fair value of instruments when their market is closed, using information from correlated instrument that trade at those times. The results are based on a simulation in which first the fair values are generated (blue lines) and trades (blue dots) are simulated when the market is open, which. happens half of the day. Notice that a third of the day both instruments trade simultaneously. The orange line are the fair values estimated using the Kalman filter, which is trained with half of the data using EM and the run over the second half of the data. The figure focus on four days of test data.

Figure 1:Estimation of the fair value of instruments when their market is closed, using information from correlated instrument that trade at those times. The results are based on a simulation in which first the fair values are generated (blue lines) and trades (blue dots) are simulated when the market is open, which. happens half of the day. Notice that a third of the day both instruments trade simultaneously. The orange line are the fair values estimated using the Kalman filter, which is trained with half of the data using EM and the run over the second half of the data. The figure focus on four days of test data.

As we see, the Kalman filter successfully exploits the correlation between instruments to update the fair value of the instruments when the market is closed. The updates are not perfect but they capture the overnight trends, improving over typical baselines like the closing price of the instrument. In general, the estimated fair values include the true fair values within one standard deviation, depicted as the shaded grey area in the figure.

Pricing sources and models

When it comes to feeding the Kalman filter with information to improve the estimations of the fair value, there is a variety of sources that is typically used. Let us discuss the most typical sources, those based on Limit Order Book (LOB) data and those based on Request for Quote (RfQ) data. As we have seen, these sources can be used both if they belong to the instrument of interest or correlated ones. However, there are some considerations to take into account when using correlated instruments, which we discuss in the end.

One point that will become a common theme in this section is the information content of the observations. Intuitively, not every source of pricing is equally informative. For example, as we discussed above, we expect that trades with larger volumes are more informative than small trades. In the same way, a cancellation of a limit order deep in a Limit Order Book might not carry meaningful new pricing information. These effects need to be incorporated case by case into the pricing model.

LOB traded

Limit Order Books (LOBs) are complex structures that contain extensive pricing information. Nevertheless, as discussed in Chapter Market microstructure, the primary references for price discovery are the best bid and ask quotes, the mid-price —defined as the average of these two—and the prices of executed trades. To improve robustness and reduce the impact of potential market manipulation, it is standard practice to compute volume-weighted averages of the first K levels on both the bid and ask sides, and to define a robust mid-price based on these averages.

Mid-price information

As mentioned above, a robust mid-price indicator in a LOB is:

Pmid,tLOB=12(i=1Kvb,iPb,ii=1Kvb,i+i=1Kva,iPa,ii=1Kva,i)P_{\text{mid},t}^{\text{LOB}} = \frac{1}{2} \left(\frac{\sum_{i=1}^K v_{b,i} P_{b,i}}{\sum_{i=1}^K v_{b,i}} + \frac{\sum_{i=1}^K v_{a,i} P_{a,i}}{\sum_{i=1}^K v_{a,i}}\right)

where vb/a,iv_{b/a, i} and Pb/a,iP_{b/a, i} are volumes and prices of bid and ask, respectively, and KK is the number of levels that are taken into account, for instance K=4K = 4. We have omitted the time subscript in volumes and prices but they are also time dependent. Similarly, we can compute a robust bid-ask spread in the form:

StLOB=i=1Kva,iPa,ii=1Kva,ii=1Kvb,iPb,ii=1Kvb,iS_t^{\text{LOB}} = \frac{\sum_{i=1}^K v_{a,i} P_{a,i}}{\sum_{i=1}^K v_{a,i}} - \frac{\sum_{i=1}^K v_{b,i} P_{b,i}}{\sum_{i=1}^K v_{b,i}}

which has to be positive since any bid and ask limit orders with the same price are automatically matched in the LOB. With this information, we can compute a simple model of fair value based on LOB information:

mtLOBN(Pmid,tLOB,(StLOB)2)m_t^{\text{LOB}} \sim N(P_{\text{mid},t}^\text{LOB}, (S_t^{\text{LOB}})^2 )

The choice of a Gaussian distribution is for simplicity, since it captures well our subjective view of the pricing information that the LOB contains, i.e. as we saw in chapter Bayesian Modelling, since we only identify the mean and variance as the constraints, the maximum entropy distribution associated with these constraints in the Gaussian distribution --admittedly, there is an extra constraint in the form of positivity of the fair value, but for prices far from the zero boundary we can safely omit this constraint. A more grounded criticism of this model could be that the Gaussian distribution allows the fair value to be beyond the best bid and ask prices, and those prices are tradeable: if the market consensus of fair value was beyond those prices, market participants would be willing to trade at those prices until they the fair value lies between the bid and ask spread. However, since liquidity might be thin at the first levels, and therefore those prices might not be available for bulk transactions, we consider a soft constraint as more appropriate, even when we use the robust bid-ask spread instead of the best bid and ask spread.

Of course, in many situations this pricing source might not be sufficient for actual applications, since in illiquid markets the bid-ask spreads are large and therefore the pricing source has a large uncertainty. In those situations is precisely where the Kalman filter model plays a part.

Using directly mtLOBm_t^{\text{LOB}} as an observation in the Kalman filter is problematic, since the feed is available in streaming and therefore, even if we limit the updates to the Kalman filter to the moments in which the LOB gets updates (i.e. arriving of new orders, cancellations, modifications), the pricing observations might jam the Kalman filter estimation, making the model to consider that mtLOBm_t^{\text{LOB}} is a perfect source of pricing information. To see this, let us see the effect of a second observation following an initial one. Recall that the first predict plus update gave:

mˉt+Δtt+Δt=mˉtt+Kt(pt+Δtmˉtt)\bar{m}_{t+\Delta t}^{t+\Delta t} = \bar{m}_{t}^{t} + K_t (p_{t+\Delta t} - \bar{m}_{t}^{t})
(σm,t+Δtt+Δt)2=((σm,tt)2+σϵ2Δt)ση2(σm,tt)2+σϵ2Δt+ση2(\sigma_{m,t+\Delta t}^{t+\Delta t})^2 = \frac{((\sigma_{m,t}^t)^2 + \sigma_\epsilon^2 \Delta t) \sigma_\eta^2}{(\sigma_{m,t}^t)^2 + \sigma_\epsilon^2 \Delta t + \sigma_\eta^2}

with:

Kt=(σm,tt)2+σϵ2Δt(σm,tt)2+σϵ2Δt+ση2K_t = \frac{(\sigma_{m,t}^t)^2 + \sigma_\epsilon^2 \Delta t }{(\sigma_{m,t}^t)^2 + \sigma_\epsilon^2 \Delta t + \sigma_\eta^2}

If we approach the limit Δt0\Delta t \rightarrow 0 this simplifies to:

mˉt+Δtt+Δt=mˉtt+Kt(pt+Δtmˉtt)\bar{m}_{t+\Delta t}^{t+\Delta t} = \bar{m}_{t}^{t} + K_t (p_{t+\Delta t} - \bar{m}_{t}^{t})
(σm,t+Δtt+Δt)2=(σm,tt)2ση2(σm,tt)2+ση2(\sigma_{m,t+\Delta t}^{t+\Delta t})^2 = \frac{(\sigma_{m,t}^t)^2 \sigma_\eta^2}{(\sigma_{m,t}^t)^2 + \sigma_\eta^2}
Kt=(σm,tt)2(σm,tt)2+ση2K_t = \frac{(\sigma_{m,t}^t)^2 }{(\sigma_{m,t}^t)^2 + \sigma_\eta^2}

Let us apply a second predict plus update steps on top of this, keeping the Δt0\Delta t \rightarrow 0 limit:

mˉt+2Δtt+2Δt=mˉtt+Kt(pt+Δtmˉtt)+Kt+Δt(pt+2Δtmˉt+Δtt+Δt)=mˉtt(1Kt(1Kt+Δt)Kt+Δt)+Kt(1Kt+Δt)pt+Δt+Kt+Δtpt+2Δt\bar{m}_{t+2\Delta t}^{t+2\Delta t} = \bar{m}_{t}^{t} + K_t (p_{t+\Delta t} - \bar{m}_{t}^{t}) + K_{t+\Delta t} (p_{t+2\Delta t} - \bar{m}_{t+\Delta t}^{t +\Delta t}) \nonumber \\ = \bar{m}_{t}^{t} (1- K_t (1-K_{t+\Delta t}) - K_{t+\Delta t}) + K_t(1-K_{t+\Delta t}) p_{t+\Delta t}+ K_{t+\Delta t} p_{t+2\Delta t}
(σm,t+2Δtt+2Δt)2=(σm,tt)2ση22(σm,tt)2+ση2(\sigma_{m,t+2\Delta t}^{t+2\Delta t})^2 = \frac{(\sigma_{m,t}^{t})^2 \sigma_\eta^2}{2 (\sigma_{m,t}^{t})^2 + \sigma_\eta^2}

Since:

Kt+Δt=(σm,t+Δtt+Δt)2(σm,t+Δtt+Δt)2+ση2=(σm,tt)22(σm,tt)2+ση2K_{t+\Delta t} = \frac{(\sigma_{m,t+\Delta t}^{t+\Delta t})^2}{(\sigma_{m,t+\Delta t}^{t+\Delta t})^2 + \sigma_\eta^2} = \frac{(\sigma_{m,t}^t)^2 }{2 (\sigma_{m,t}^t)^2 + \sigma_\eta^2}
1Kt+Δt=(σm,tt)2+ση22(σm,tt)2+ση21-K_{t+\Delta t} = \frac{(\sigma_{m,t}^t)^2 +\sigma_\eta^2}{2 (\sigma_{m,t}^t)^2 + \sigma_\eta^2}
Kt(1Kt+Δt)=(σm,tt)22(σm,tt)2+ση2=Kt+ΔtK_t (1-K_{t+\Delta t}) = \frac{(\sigma_{m,t}^t)^2 }{2(\sigma_{m,t}^t)^2 + \sigma_\eta^2} = K_{t+\Delta t}

then:

mˉt+2Δtt+2Δt=(12Kt+Δt)mˉtt+Kt+Δt(pt+Δt+pt+2Δt)=ση22(σm,tt)2+ση2mˉtt+(σm,tt)22(σm,tt)2+ση2(pt+Δt+pt+2Δt)\bar{m}_{t+2\Delta t}^{t+2\Delta t} = (1-2K_{t+\Delta t}) \bar{m}_{t}^{t} + K_{t+\Delta t}(p_{t+\Delta t} + p_{t+2\Delta t}) \nonumber \\ = \frac{\sigma_\eta^2}{2 (\sigma_{m,t}^t)^2 + \sigma_\eta^2} \bar{m}_{t}^{t} + \frac{(\sigma_{m,t}^t)^2 }{2 (\sigma_{m,t}^t)^2 + \sigma_\eta^2} (p_{t+\Delta t} + p_{t+2\Delta t})

If we continue applying nn predict plus update steps in the Δ0\Delta \rightarrow 0 limit, by using the induction principle we arrive at the following result:

mˉt+nΔtt+nΔt=ση2n(σm,tt)2+ση2mˉtt+(σm,tt)2n(σm,tt)2+ση2i=1npt+nΔt\bar{m}_{t+n\Delta t}^{t+n\Delta t} = \frac{\sigma_\eta^2}{n (\sigma_{m,t}^t)^2 + \sigma_\eta^2} \bar{m}_{t}^{t} + \frac{(\sigma_{m,t}^t)^2 }{n (\sigma_{m,t}^t)^2 + \sigma_\eta^2} \sum_{i=1}^n p_{t+n \Delta t}
(σm,t+nΔtt+nΔt)2=(σm,tt)2ση2n(σm,tt)2+ση2(\sigma_{m,t+n\Delta t}^{t+n\Delta t})^2 = \frac{(\sigma_{m,t}^{t})^2 \sigma_\eta^2}{n (\sigma_{m,t}^{t})^2 + \sigma_\eta^2}

If we now take the limit nn\rightarrow \infty we converge to:

mˉt+nΔtt+nΔt1ni=1npt+nΔt\bar{m}_{t+n\Delta t}^{t+n\Delta t} \rightarrow \frac{1}{n} \sum_{i=1}^n p_{t+n \Delta t}
(σm,t+nΔtt+nΔt)20(\sigma_{m,t+n\Delta t}^{t+n\Delta t})^2 \rightarrow 0

Therefore, in the case of LOB updates that don’t significantly provide new pricing information but happen with a high frequency, as in the case in LOBs for liquid instruments, the Kalman filter will converge to the mid-price with zero uncertainty.

A simple way to fix this issue is to estimate the Kalman filter with other pricing information (trades, RfQs, see next section) and combine it as two separate fair value estimations at any time. The best linear combination of two estimators is the one that minimizes the variance:

m^t=αmtLOB+(1α)mtkalman\hat{m}_t = \alpha m_t^{\text{LOB}} + (1-\alpha) m_t^{\text{kalman}}
Var(m^t)=α2(StLOB)2+(1α)2(σm,tt)2\text{Var}(\hat{m}_t) = \alpha^2 (S_t^{\text{LOB}})^2 + (1-\alpha)^2 (\sigma_{m,t}^t)^2

which is achieved for:

α^=(σm,tt)2(StLOB)2+(σm,tt)2\hat{\alpha} = \frac{(\sigma_{m,t}^t)^2}{(S_t^{\text{LOB}})^2 + (\sigma_{m,t}^t)^2}

Notice that we have considered them independent estimators, otherwise there would be an extra term accounting for the correlation. Therefore, the best linear estimator is:

m^t=(σm,tt)2(StLOB)2+(σm,tt)2mtLOB+(StLOB)2(StLOB)2+(σm,tt)2mtkalman\hat{m}_t = \frac{(\sigma_{m,t}^t)^2}{(S_t^{\text{LOB}})^2 + (\sigma_{m,t}^t)^2} m_t^{\text{LOB}} + \frac{(S_t^{\text{LOB}})^2 }{(S_t^{\text{LOB}})^2 + (\sigma_{m,t}^t)^2} m_t^{\text{kalman}}

This makes intuitive sense: the smaller the relative error of one estimator compared to the other, the more weight the combined estimate assigns to it. Importantly, the resulting variance is lower than that of either individual estimator.

Var(m^t)=(StLOB)2(σm,tt)2(StLOB)2+(σm,tt)2\text{Var}(\hat{m}_t) = \frac{(S_t^{\text{LOB}})^2(\sigma_{m,t}^t)^2}{(S_t^{\text{LOB}})^2 + (\sigma_{m,t}^t)^2}

To check it, simply take the ratio of the final variance with any of the variances of the independent predictors, for example:

Var(m^t)(StLOB)2=(σm,tt)2(StLOB)2+(σm,tt)21\frac{\text{Var}(\hat{m}_t)}{(S_t^{\text{LOB}})^2} = \frac{(\sigma_{m,t}^t)^2}{(S_t^{\text{LOB}})^2 + (\sigma_{m,t}^t)^2} \leq 1

the equality happening when StLOB=0S_t^{\text{LOB}} = 0 i.e. it was already a perfect predictor and, therefore, there is no way to improve the prediction.

Notice that, since the combined fair value estimator is reconstructed independently at each point in time —without relying on past estimates—it is not affected by the high-frequency sampling issue observed when using the LOB mid-price as an observation.

Trades

Trades happening in the LOB are a valuable source of pricing information, since they correspond to real transaction prices and not only interests to trade as limit orders. When a trade happens, the exchange reports publicly the time, the size and the price, but not the parties or the orders involved. The latter is particularly relevant since a relevant pricing information is the side (buy or sell) of the order that was aggressive, meaning the one that consumed the liquidity in the order book. As we discussed in Market microstructure, this can typically a market order or a limit order at a price that is equal or better that prices available in the opposite side. Reverse engineering the side from a trade is an inference problem, and requires a model. A simple one widely used is the so-called tick-rule model which consists on comparing the price of the trade with the mid-price available in the order book just before the trade:

This model is not perfect, however. It does not account for hidden liquidity that might exist at more favorable prices than those displayed, which would alter the reference mid-price and, consequently, the trade classification logic. Moreover, it assumes a sequential processing of orders, whereas in practice, multiple orders may arrive simultaneously. In such cases, the exchange’s internal matching engine determines execution priority and order pairing through mechanisms that are not observable externally, meaning the apparent sequence of trades and quote updates in public data may not reflect the true matching process. This lack of transparency can lead to misclassification when applying the tick rule or similar models.

Once we have the relevant pricing information for the trade, namely time, side, size and or course price, it can be used to update the current fair value estimation of the financial instrument. The size information is useful to include some measure of information content in the order, as discussed in the simple pricing model section when introducing the Sinclair model: intuitively, a very small size trade should not be as relevant as a large size trade when updating our fair value estimation. Sinclair proposes to model a trade observation as a Gaussian random variable N(Pttrade,σ2(v)){\mathcal N}(P_t^{\text{trade}}, \sigma^2(v)), where PttradeP_t^{\text{trade}} is the observed trade price at time tt, and σ(v)\sigma(v) is given by:

σ(v)=σp(vmaxv1))+\sigma (v)= \sigma_p \left(\frac{v_\text{max}}{v}-1)\right)^+

where σp\sigma_p is a constant to be estimated, and vmaxv_{\text{max}} is an exogenous parameters that provides a typical scale for which trades are considered informative. It can be given by business prior knowledge or estimated from the statistical distribution of trade sizes. Notice that this error has the desirable properties of becoming zero at the scale vmaxv_{\text{max}} and above, i.e. trades above this scale are considered maximally informative and the fair value is instantaneously update to this value. It also becomes infinitely large as the volume tends to zero, which makes the Kalman filter model to essentially ignore those observations.

Sinclair’s model is not the only way to introduce this behavior into the model. Other choices of σ(v)\sigma(v) are also valid. For example, the model:

σ(v)=σ0exp(vv0)1+exp(vv0)\sigma (v)= \sigma_0 \frac{\exp(-\frac{v}{v_0})}{1+ \exp(-\frac{v}{v_0})}

might be more realistic in the sense that volume adjusts the degree of information but it never completely ignores small trades, for which the error tends to σ0/2\sigma_0/2. Another alternative that does not completely agree with trades or large size is the following:

σ(v)=σmin+(σ0σmin)evv0\sigma (v)= \sigma_{\min} + (\sigma_0 - \sigma_{\min}) e^{-\frac{v}{v_0}}

which tends to σmin\sigma_{\min} as vv \rightarrow \infty.

So far we have not used the side information inferred from the tick-rule model. There are different ways this information can be factored into the pricing. In markets that are quite unbalanced, the observation of a trade in the opposite side where the market is prevalently trading might be considered more informative. This means potentially adjusting the error function with the side information. Another alternative is directly building separate Kalman filters for buy and sell information, which are then combined in a final fair value estimation using the model discussed in the previous section for optimally combining two predictors, although in this case neglecting correlation between bid and ask estimations might not be a solid modelling choice. We leave as an exercise to the reader to derive the optimal linear predictor with correlation.

RfQ traded

We have discussed the Request for Quote (RfQ) protocol in the chapter on Market Microstructure. In terms of pricing information, the main difference with LOBs is the asymmetry in information between the different participants in the process: dealers and clients. Let us focus on the case of negotiation via Multi-Dealer-to-Client platforms, which has the richer casuistic, from the point of view of the dealer, who is typically the one actively trying to calculate the fair value of the instruments. The pricing information that the dealer receives is the following:

Apart from the pricing information gained via the trading platform, in order to improve transparency and price formation, there are private and public initiatives to share pricing information post-trade, particularly in the bond markets:

Composites

Modelling composites is similar to modelling mid-prices from order books, since they share in common the structure of information in terms of a continuous feed of bid and ask prices, only that for composites are indicative prices. Again, given that the frequent updates in the feed might not carry new pricing information, it makes sense to model it as an independent fair value source, that can be then aggregated with other estimations, like those using Kalman filters on trades and other information.

The model is therefore:

mtcompN(Pmid,tcomp,(Stcomp)2)m_t^{\text{comp}} \sim N(P_{\text{mid},t}^\text{comp}, (S_t^{\text{comp}})^2 )

with:

Pmid,tcomp=12(atcomp+btcomp)P_{\text{mid},t}^\text{comp} = \frac{1}{2}(a_t^\text{comp} +b_t^\text{comp})
Smid,tcomp=12(atcompbtcomp)S_{\text{mid},t}^\text{comp} = \frac{1}{2}(a_t^\text{comp} -b_t^\text{comp})

where btcompb_t^\text{comp}, atcompa_t^\text{comp} are, respectively, the bid and ask composite prices published by the platform.

RfQs

The information from RfQs depends, as discussed above, on the final status of the dealer in the process. If the dealer wins the RfQ, the observation corresponds to a trade, and the modelization is overall similar to the one discussed in the context of trades in the LOB. There is, though, one key difference with the case of LOBs, in that the cover price is also informed to the dealer.

Intuitively, the closer the cover and the trade price, the more confidence we might put on the trade price as a pricing source, since we have the agreement from a second dealer quoting a similar level. Or put it in a different way, if the dealer wins the trade at a price much further than the cover price, it might imply a clear miss-pricing that needs to be adjusted, not a reliable pricing source. One way to incorporate this into the model is to make the observation error a function of the distance to the cover:

σ(v,dtcover)=σ(v)g(dtcover)\sigma (v, d_t^{\text{cover}}) = \sigma(v) g(d_t^{\text{cover}})

where dtcover=PttradePtcoverd_t^{\text{cover}} = |P_t^\text{trade} - P_t^\text{cover}| is the distance to the cover, and g()g(\cdot) a modulating function with a minimum at zero.

A second case is the one in which the dealer misses the RfQ, but the price was the second best quoted. The fact that among the number of dealers quoting competitively the price quoted was the second best carries some significant information that can be used to adjust the fair value. One way to do this is to fit a probabilistic model that predicts the distance to cover based on features fif_i available in the negotiation, for example, a simple linear regression model on dtcoverd_t^\text{cover}:

dtcover=iwifi+ϵd_t^\text{cover} = \sum_i w_i f_i + \epsilon

where wiw_i are the regression weights and ϵN(0,σd2)\epsilon \sim N(0, \sigma_d^2). Alternatively, if we want to explicitly model that the distance to cover cannot be negative, we can use a log transformation, although then we will have to use a non-linear Kalman filter to account for this observation. Assuming the simple linear regression model, we can add the observation to the Kalman filter:

Ptinf, coverN(Ptcover+sideE[dtcover{fi}],σd2)P_t^{\text{inf, cover}} \sim N(P_t^\text{cover} + \text{side} \cdot E[d_t^\text{cover}|\{f_i\}], \sigma_d^2)

where side=±1\text{side} = \pm 1 depending on the side of the RfQ (buy or sell).

A third case happens when the client trades, the dealer misses the RfQ but her quote was not even the second best (cover). The information in this setup is much weaker, in the form of a bound to the traded price (the quoted price). Although, in principle, a probabilistic inference on the traded price could be performed using this information, the potential model risk coming from the plethora of model assumptions typically outweighs the information gains, so we will not dive deeper here.

Hit & Miss

Another source of pricing information comes from analyzing patterns of trading information from the dealer. In particular, it is useful to analyze the hit & miss, i.e. the ratio between won RfQs over the total traded by the client with any dealer (i.e. excluding those where the client did not trade at all), over a certain window of time. Intuitively, a dealer that has an abnormally high or low hit & miss might be using an incorrect fair value estimation from the quoting. Of course, the tricky question here is to assess what are the normal levels of hit & miss, which have to take into account the influence of factors like, for example, inventory levels in the spreads quoted: if a dealer has relatively high inventory holdings, it is likely that she will skew quotes to reduce inventory risk, hence the hit & miss will be higher if there are more quotes overall in the direction of risk reduction.

If the dealer has estimated a model that estimates the probability of winning the RfQs, and the model is well calibrated, it can be evaluated for the same windows of RfQs as the empirical hit & miss (H&M). If we compute the latter at time tt using the last nn RfQs, it is given by:

H&Mt=1ni=1n1wini\text{H\&M}_t = \frac{1}{n} \sum_{i=1}^n 1_{\text{win}_i}

where wini\text{win}_i is an abbreviation of the condition status(i)=win\text{status}(i) = \text{win}. For a well calibrated model, such empirical hit & miss should be close to the expected by the model:

E[H&Mt]=1ni=1nP(winiFti)E[\text{H\&M}_t] = \frac{1}{n} \sum_{i=1}^n P(\text{win}_i|\mathcal{F}_{t_i})

Here, Fti\mathcal{F}_{t_i} are the filtrations at the time tit_i of request of the i-th RfQ, which include the spreads δi\delta_i quoted by the dealer:

δi=side(Pimti)\delta_i = \text{side} (P_{i} - m_{t_i})

with PiP_{i} the price quoted for the i-th RfQ and mtim_{t_i} the estimation of the fair value at time tit_i.

If we consider each RfQ independent of each other, the variance of the hit & miss provides us with a scale of natural variability of our estimation versus the empirical hit & miss:

Var[H&M]=1ni=1nP(winiFti)(1P(winiFti))Var[\text{H\&M}] = \frac{1}{n} \sum_{i=1}^n P(\text{win}_i|\mathcal{F}_{t_i}) (1-P(\text{win}_i|\mathcal{F}_{t_i}))

Intuitively, then, if we get a persistent deviation between H&Mt\text{H\&M}_t and E[H&Mt]E[\text{H\&M}_{t}] that cannot be explained by the expected variability from the randomness of the RfQ process, this could be attributed to a potential bias in the estimation of mtm_t, which becomes a further pricing source for our fair value estimation model. Further modelization is required, though, to inject this information into the Kalman filter, which goes beyond the scope of this book. For us, it suffices to point out how we can potentially convert hit & miss deviations from the target into pricing information.

Correlated instruments

As we discussed above, if a set of instruments exhibit historical price correlations and we have reasons to believe those are structural correlations, i.e. they will continue existing in the present, we can do a joint estimation of the fair values using a multivariate Kalman filter. This way, information about the price of one instrument can be used to improve the estimation of the others. The pricing sources for these instruments can be any of those discussed previously.

There are some caveats though to take into account when using this source of pricing information:

Fundamental models for fair value estimation

Fundamental models estimate the fair value of a financial instrument by analysing the value of their future cash-flows. Recall from our introductory chapter on financial markets Financial Markets, that financial instruments are essentially contracts that promise to pay back funds to the investor that purchases it under conditions specified in the contract.

The present value theory of fair value

Even the most simple financial instrument, a promise to pay back a deterministic amount of money in a future fixed date, requires some theoretical hypothesis to estimate its fair value. The basic idea is that a unit of currency received today is worth more than the same unit received in the future, because it can be invested in the interim. Consider a risk-free deposit that pays a deterministic interest rate rr. An amount of one unit invested today grows to (1+r)T(1+r)^T units after TT periods, as far as interest rate payments are reinvested at the same rate (compounded interest). Conversely, receiving one unit in tt periods is economically equivalent to receiving 1/(1+r)T1/(1+r)^T units today. This opportunity cost argument implies that any future cash flow must be discounted by the factor (1+r)T(1+r)^T to make it comparable with cash today. Such economic consideration is referred as the time value of money.

Formally, for an investment that delivers a future cash flow CTC_T at time TT, the present value PVPV satisfies

PV=CT(1+r)TPV = \frac{C_T}{(1+r)^T}

where we have assumed that an interest payment of rCTr C_T is paid each unit of time. The factor 1/(1+r)T1/(1+r)^T is called the discount factor, since it is used to discount future cash-flows. Present values becomes our fair value estimation within this framework.

A typical hypothesis that provides useful mathematical simplifications is that of continuously accrued interest rates. This is also a good approximation for real situations where interests are paid daily, for instance in money market funds. Consider and account that pays interests each time period of size Δ\Delta. The interest paid is rΔr \Delta over a unitary notional. If we reinvest the interests, at time tt we have accumulated (1+rΔ)T/Δ(1+ r \Delta)^{T/\Delta}. If we now take the limit Δ0\Delta \rightarrow 0:

limΔ0(1+rΔ)T/Δ=limΔ0eTΔlog(1+rΔ)=erT\lim_{\Delta \rightarrow 0} (1+ r \Delta)^{T/\Delta} = \lim_{\Delta \rightarrow 0} e^{\frac{T}{\Delta} \log (1 + r \Delta)} = e^{rT}

This gives us the expression of the discount factor for continuously paying interest rates, given by the inverse erTe^{-rT}. Under this approximation, the preset value of our simple financial instrument becomes:

PV=erTCTPV = e^{-rT} C_T

Exponential functions provide a lot of mathematical simplifications, hence the usefulness of this limit, particularly when applied to the computation of fair value for more complex financial instruments.

If we also include the price paid for the financial instrument, let us say in this case we pay initially C0C_0, we define net present value (NPV) as:

NPV=erTCTC0NPV = e^{-rT} C_T - C_0

Notice that, therefore, a rational investor will only be willing to invest in this financial instrument if C0erTCTC_0 \leq e^{-rT} C_T, otherwise its NPV would be negative.

We can now generalize this expression for a financial instrument that pays a stream of deterministic cash-flows CtiC_{t_i} at times ti,i=1,...,Nt_i, i = 1, ..., N. This is the case for example of standard government bonds issued by most countries. The fair value given by present value becomes then:

PV=i=1NertiCtiPV = \sum_{i=1}^N e^{-r t_i} C_{t_i}

What happens if these future cash-flows are contingent to information not known at present day? For instance, we can have bonds that pay floating interest rates depending on reference rates that don’t get fixed until a future date. Shares pay dividends contingent to the financial results of the corporation that issues the shares. And derivatives are financial instruments whose value depends on the future price of an underlying instrument, hence the name “derivative”. A simple naive extension of the theory of present value would consider these future cash-flows as functions of random variables, replacing its future value by an expectation of future value:

PV0=E[i=1NertiCtiF0]E0[i=1NertiCti]PV_0 = {\mathbb E}\left[\sum_{i=1}^N e^{-r t_i} C_{t_i}|F_0\right] \equiv {\mathbb E}_0 \left[\sum_{i=1}^N e^{-r t_i} C_{t_i}\right]

where we have conditioned the expectation to the information available at the time of estimation, the filtration F0F_0. If discount factors are not stochastic themselves (an argument we will revisit in the last section of this chapter), this becomes:

PV0=i=1NertiE0[Cti]PV_0 = \sum_{i=1}^N e^{-r t_i} {\mathbb E}_0 \left[C_{t_i}\right]

The issue with this approach is that, once future cash-flows become uncertain, their expected value is a point estimation of their possible range of values that neglects the rest of potential scenarios that can happen. Recall from our discussion in chapter Bayesian Modelling, that choosing to represent a random variable by their expected value in terms of decision theory (and fair value estimation is in the end linked to the decision to buy or sell a financial instrument) makes sense when the investor penalizes errors in the estimation using a square loss function. The behavior of real rational investors, though, shows a more asymmetric loss function, in which generally potential extra gains are valued less than the equivalent potential losses. This kind of behavior is better capture by using utility functions, as we will discuss in the next section.

The utility indifference theory of fair value estimation

To ground the discussion in another example, let us consider a specific case of a future uncertain cash-flow whose value depends on the price of another instrument at the time of payment, STS_T, for example a stock. The cash-flow is therefore CT=f(ST)C_T = f(S_T). Notice that this is a specific case of a derivative’s contract. The function f(ST)f(S_T) is called the pay-off of the derivative, the instrument whose price is STS_T is called the underlying of the derivative, and the time TT is the expiry date of the derivative. Apart from the value STS_T, it can also depend on other parameters that are deterministic. For instance, for an European call option we have CT=(STK)+C_T = (S_T-K)^+, where KK is called the strike of the option. An european put options has a payoff CT=(KST)+C_T = (K-S_T)^+. There are also American options where the option can be exercised before the expiry date, which becomes itself a random variable.

This derivative pays in the future a quantity that is contingent to the future value of the underlying, whose value is known today but is uncertain in the future. To get an estimation, we need to use probability theory to put some bounds to our uncertainty, so we characterize STS_T by a random distribution function g(ST)g(S_T). As we saw in chapter Stochastic Calculus, a popular model that allows us to compute such future distribution is a random walk model or a geometric random walk, the latter being a natural choice for prices that cannot be negative. In those cases the future distribution can be computed, being a normal distribution in the first case, and a log-normal distribution in the second, e.g. in the case of stocks. Although for short expiries the random walk can also be a good model for stocks. These models allow us to get sometimes closed-form solutions, but more realistic models that capture better empirical distributions of prices can be used.

As mentioned in the previous section, we cannot just value this cash-flow using the expected value of the pay-off, since it would ignore the risk-profile of the investor. As discussed in more detail in chapterStochastic optimal control, utility functions provide a mathematical formalism that allows us to capture realistic risk behaviors. Utility provides a description of the value that the cash-flows derived from the financial instrument have for the investor. Typical utility functions show the notion of marginally decreasing utility for increasingly larger cash-flows. In situations where cash-flows are random variables, such behavior models investors that are risk averse, meaning that they need to be compensated increasingly more to take on extra risk.

To apply the utility function framework to the problem of fair pricing, we need to compute expected utilities to characterize the value that the investor places on the contract. Using a exponential utility function for simplicity, this means:

Et[U]=1Et[eγi(er(Tt)f(ST))]\mathbb{E}_t[U] = 1- \mathbb{E}_t[e^{-\gamma_i \left(e^{-r(T-t)}f(S_T)\right)}]

where γi\gamma_i is the risk aversion coefficient of the investor. We have discounted the payoff at TT by the discount factor er(Tt)e^{-r(T-t)} in order to consider the time value of money, as discussed in the previous section, although now we consider for generality a initial time tt.

The fair value in this formalism is the so-called premium of the derivative, denoted CtC_t, that the investors is willing to pay (or be paid, depending on the pay-off function) to enter into the derivative’s contract today. This changes the utility calculation, since it needs to take into account the premium:

Et[U]=1Et[eγi(er(Tt)f(ST)Ct)]\mathbb{E}_t[U] = 1- \mathbb{E}_t[e^{-\gamma_i \left(e^{-r(T-t)}f(S_T) -C_t\right)}]

When modelling rational risk-averse agents with utility functions, we model their decisions as those that maximize the expected utility. However, in this case this cannot be used to compute the premium, since naturally the premium that maximizes utility is Ct=C_t = -\infty!. The problem is, of course, that it does not take into account the utility maximization of the dealer selling the derivative, who would not enter into the contract at this premium. Of course, the same framework could be used to model the dealer’s payoff, which is the reverse from the investor, albeit with a different dealer’s risk aversion, γd\gamma_d:

Et[U]=1Et[eγd(Cter(Tt)f(ST))]\mathbb{E}_t[U] = 1- \mathbb{E}_t[e^{-\gamma_d \left(C_t - e^{-r(T-t)}f(S_T) \right)}]

but even introducing the dealer’s utility function, how could we compute the value of the premium?

For the answer, we need first to frame the problem in other terms: what is the maximum premium that the investor would be willing to pay to enter into the contract? Since the alternative to not entering into the contract implies a zero payoff with total certainty, whose expected utility in this framework is 0$0\$, we can argue that the investor would be willing to buy the derivative as far as the premium makes him/her better off, i.e. Et[U]>0\mathbb{E}_t[U] > 0. For a value of the premium such that Et[U]=0\mathbb{E}_t[U] = 0 , the investor is indifferent to buy or not buy. This value of the premium is called the reservation price or the utility indifference price of the investor. Of course the same computation could be done for the dealer, obtaining a different reservation price. An agreement will only happen if the maximum premium that the investor is willing to pay is above the minimum premium that the dealer is willing to receive.

Let us first see the problem from the dealer’s point of view. In real situations, it is typically the investor who comes to the dealer and request a price for the derivative. The minimum premium that the dealer would be willing to accept to provide the contract as a reference for derivatives pricing, i.e. the reservation price of the dealer, is the one that solves:

1E0[eγd(Cter(Tt)f(ST))]=01- \mathbb{E}_0[e^{-\gamma_d \left(C_t - e^{-r(T-t)}f(S_T) \right)}] = 0

We can now obtain a general expression for the premium:

Ctd=1γdlogE0[eγd(er(Tt)f(ST))]=1γdlogdSTg(ST)eγd(er(Tt)f(ST))C_t^d = \frac{1}{\gamma_d} \log \mathbb{E}_0[e^{\gamma_d \left(e^{-r(T-t)}f(S_T)\right)}] = \frac{1}{\gamma_d} \log \int dS_T g(S_T) e^{\gamma_d \left(e^{-r(T-t)}f(S_T)\right)}

For those dealers that have zero risk aversion, i.e. they are risk neutral, by taking the limit γd0\gamma_d \rightarrow 0 we get:

Ct(0)=E0[er(Tt)f(ST)]=dSTg(ST)er(Tt)f(ST)C_{t}(0) = \mathbb{E}_0[ e^{-r(T-t)}f(S_T)] = \int dS_T g(S_T) e^{-r(T-t)}f(S_T)

And for small, but positive risk aversion:

Ctd=Ct(0)+γd2dSTg(ST)e2r(Tt)f2(ST)+O(γd2)C_t^d = C_{t}(0) + \frac{\gamma_d}{2}\int dS_T g(S_T) e^{-2r(T-t)}f^2(S_T) + O(\gamma_d^2)

We can derive the same expression for a investor we get:

Cti=1γilogE0[eγi(er(Tt)f(ST))]=1γilogdSTg(ST)eγi(er(Tt)f(ST))C_t^i = -\frac{1}{\gamma_i} \log \mathbb{E}_0[e^{-\gamma_i \left(e^{-r(T-t)}f(S_T)\right)}] = -\frac{1}{\gamma_i} \log \int dS_T g(S_T) e^{-\gamma_i \left(e^{-r(T-t)}f(S_T)\right)}

If the investor has a small but positive risk aversion:

Cti=Ct(0)γi2dSTg(ST)e2r(Tt)f2(ST)+O(γi2)C_t^i = C_{t}(0)- \frac{\gamma_i}{2}\int dS_T g(S_T) e^{-2r(T-t)}f^2(S_T) + O(\gamma_i^2)

We see immediately that CtiCdiC_t^i \leq C_d^i, so there is only agreement if both investor and dealer are risk neutral, or at least one is risk prone, which is not a normal situation. Therefore, according to this theory of pricing, there would not be trading of derivatives! However, we know empirically that it is not the case. So what was wrong in our theory? We will see that the dealer is not simply taking the opposite bet than the investor, and therefore we need to modify this analysis. Before that, though, let us see particular examples of the computation of the premium for investors.

Example: pricing of a simple contingent claim

A contingent claim is a contract that pays off only under the realization of an uncertain event. Many derivatives contracts like options are contingent claims. The most simple contingent claim pays 1$ under the realization of a specific uncertain event, and zero in all other cases. These contingent claims are called Arrow-Debreu securities, and have a theoretical interest since we could in principle decompose any contingent claim as a linear combination of these securities. Therefore, if we know the prices (premiums) of Arrow-Debreu securities, we could price any contingent claim. We say that in this case we have a complete market, where we can trade instruments linked to any future state of the market.

For our purposes, though, we just want to discuss a simple example of reservation prices. Let us consider a contingent claim in which the dealer pays the investor 1$ if we get heads when tossing a fair coin in the present. In our framework, the underlying now is the side of the coin, heads or tails, with probabilities pH=pT=1/2p_H = p_T = 1/2. We also make T=tT = t since we toss the coin in the present. The value of the reservation price for the investor reads then:

Cti=1γilog(12eγi+12)=121γilogcosh(γi2)C_t^i = -\frac{1}{\gamma_i} \log \left(\frac{1}{2}e^{-\gamma_i} + \frac{1}{2} \right) = \frac{1}{2} - \frac{1}{\gamma_i}\log \cosh \left(\frac{\gamma_i}{2}\right)

For a risk-neutral investor, by making γi0\gamma_i \rightarrow 0, we get simply Cti=1/2C_t^i = 1/2, which makes sense: the investor is willing to pay 0.5$ to make the game fair. Or in other terms, to make the expected value of the game zero. A fully risk averse investor for whom γi\gamma_i \rightarrow \infty has Ct=0C_t = 0, i.e. only is willing to buy the contract when there is guarantee of no losses under any scenario. In the middle, the premium lies between those two values: the investor will be willing to pay more than 0$ to trade, as far as the payoff is skewed in its favor.

Example: Forward on a non-dividend paying stock

Let us now focus on a more realistic case and find the maximum premium that a risk averse investor would be willing to pay for a forward contract on a non-dividend paying stock [1]. The buyer of a forward has the obligation to buy a stock at the expire TT at a pre-agreed price KK. Therefore, the payoff function reads:

f(ST)=STKf(S_T) = S_T - K

The maximum premium that the investor is willing to pay reads then:

Cti=1γilogdSTg(ST)eγier(Tt)(STK)C_t^i = - \frac{1}{\gamma_i} \log \int dS_T g(S_T) e^{-\gamma_i e^{-r(T-t)}(S_T-K)}

which in the case of a risk-neutral investor reduces to:

Cti(0)=dSTg(ST)er(Tt)(STK)=er(Tt)(E[ST]K)C_{t}^i(0) = \int dS_T g(S_T) e^{-r(T-t)}(S_T-K) = e^{-r(T-t)}(E[S_T] - K)

i.e. the price is simply the discounted expected pay-off. The expectation represents the belief from the investor on the value of the stock at expiry. It is model-free, meaning that don’t need to specify a model for the evolution of the stock to compute the maximum premium, although of course an investor could use a model to compute it. The value of the premium has the following dependencies:

Example: European Call Options

The calculation for a forward was relatively tractable since the payoff of the derivative was linear on the stock. What about non-linear payoffs? This is the case for instance of an European Call option on a stock, which has the payoff:

f(ST)=(STK)+f(S_T) = (S_T - K)^+

where KK is the strike of the option. The risk-averse investor will be willing to pay the dealer a maximum premium of:

Cti=1γilogdSTg(ST)eγier(Tt)(STK)+C_t^i = -\frac{1}{\gamma_i} \log \int dS_T g(S_T) e^{-\gamma_i e^{-r(T-t)}(S_T-K)^+}

In the limit of a risk-neutral investor, the premium is:

Cti(0)=dSTg(ST)er(Tt)(STK)+C_{t}^i(0) = \int dS_T g(S_T) e^{-r(T-t)} (S_T-K)^+

Let us consider the case of a non-dividend paying stock, which we model as Geometrical Brownian Motion to ensure non-negative prices are allowed:

dSt=μStdt+σStdWtd S_t = \mu S_t dt + \sigma S_t d W_t

where μ\mu and σ\sigma are the drift and volatility of the stock, respectively, and WtW_t a Wiener process. Integrating this SDE up to T:

ST=Ste(μσ22)(Tt)+σTtZS_T = S_t e^{(\mu - \frac{\sigma^2}{2})(T-t) + \sigma \sqrt{T-t} Z}

where ZN(0,1)Z \sim N(0,1). We can further decompose the expression of the premimum as:

Cti(0)=0dSTg(ST)er(Tt)(STK)+=KdSTg(ST)er(Tt)(STK)C_{t}^i(0) = \int_0^\infty dS_T g(S_T) e^{-r(T-t)} (S_T-K)^+ = \int_K^\infty dS_T g(S_T) e^{-r(T-t)} (S_T - K)
=er(Tt)(KdSTg(ST)STKKdSTg(ST))= e^{-r(T-t)} \left(\int_K^\infty dS_T g(S_T) S_T - K \int_K^\infty dS_T g(S_T)\right)

Let us now change variables from STS_T to ZZ in the integration. The first integral becomes:

KdSTg(ST)ST=St1σTt(logKSt(μσ22)(Tt))dZ2πeZ22e(μσ22)(Tt)+σTtZ\int_K^\infty dS_T g(S_T) S_T = S_t \int_{\frac{1}{\sigma \sqrt{T-t}} \left(\log \frac{K}{S_t} - (\mu - \frac{\sigma^2}{2})(T-t)\right)}^\infty \frac{dZ}{\sqrt{2\pi}} e^{-\frac{Z^2}{2}} e^{(\mu - \frac{\sigma^2}{2})(T-t) + \sigma \sqrt{T-t} Z}
=Ste(μσ22)(Tt)d2(μ)dZ2πeZ22eσTtZ=Steμ(Tt)d2(μ)σTtdZ2πe(ZσTt)22= S_t e^{(\mu - \frac{\sigma^2}{2})(T-t)} \int_{-d_2(\mu)}^\infty \frac{dZ}{\sqrt{2\pi}} e^{-\frac{Z^2}{2}} e^{ \sigma \sqrt{T-t} Z} = S_t e^{\mu (T-t)} \int_{-d_2(\mu)-\sigma\sqrt{T-t}}^\infty \frac{dZ}{\sqrt{2\pi}} e^{-\frac{(Z-\sigma \sqrt{T-t})^2}{2}}
=Steμ(Tt)(1N(d1(μ)))= S_t e^{\mu (T-t)} \left(1-N(-d_1(\mu))\right)

where we have defined the functions:

d1(μ)=1σTt(logStK+(μσ22)(Tt))d_1(\mu) = \frac{1}{\sigma \sqrt{T-t}} \left(\log \frac{S_t}{K} + (\mu - \frac{\sigma^2}{2})(T-t)\right)
d2(μ)=1σTt(logStK+(μ+σ22)(Tt))d_2(\mu) = \frac{1}{\sigma \sqrt{T-t}} \left(\log \frac{S_t}{K} + (\mu + \frac{\sigma^2}{2})(T-t)\right)

and N(x)N(x) is the cumulative distribution function of the standard normal distribution. The second integral is then:

KdSTg(ST)=d2(μ)12πeZ22=1N(d2(μ))\int_K^\infty dS_T g(S_T) = \int_{-d_2(\mu)}^\infty \frac{1}{\sqrt{2 \pi}} e^{-\frac{Z^2}{2}} = 1-N(-d_2(\mu))

Wrapping up all we get:

Cti(0)=Ste(μr)(Tt)N(d1(μ))Ker(Tt)N(d2(μ))C_{t}^i(0)=S_t e^{(\mu-r)(T-t)}N(d_1(\mu))-K e^{-r(T-t)} N(d_2(\mu))

where we have used the property 1N(x)=N(x)1-N(-x) = N(x). Let us analyze this formula, which recall is the maximum premium that a investor is willing to pay for the call option. It depends on the following parameters:

We plot those dependencies in the following picture:

Dependencies of a call option premium for a risk neutral investor, derived as the maximum premium the investor is willing to pay (reservation or indifference price). We use the parameters S_t=100, K =100, T-t = 1 in years, r = 0.05, \mu = 0.1, \sigma = 0.2.

Figure 2:Dependencies of a call option premium for a risk neutral investor, derived as the maximum premium the investor is willing to pay (reservation or indifference price). We use the parameters St=100S_t=100, K=100K =100, Tt=1T-t = 1 in years, r=0.05r = 0.05, μ=0.1\mu = 0.1, σ=0.2\sigma = 0.2.

The arbitrage-free theory of derivatives pricing

When we were applying the utility indifference theory of derivatives pricing to both parties agreeing in the transaction, the investor and the dealer, we found the issue that according to this theory, there would be a trade only if both of them are risk averse, since they are taking opposite sides of the bet on the payoff result. In reality, both dealers are investors are generally risk-averse and we observe transactions, so what are we missing in our theory?

The answer, as we anticipated above, is that the dealer is not really simply taking the opposite of the bet. There are mainly two ways a dealer will conduct this business:

As we discussed in the first part of this chapter, if we are in the first situation, we might not need a derivatives pricing model since we can extract the prices of derivatives directly from observations of trades or request for quotes that are not closed. It is the second case where we need a theory of derivatives pricing that takes into account potential hedging strategies that mitigate the risk of the dealer. We anticipate then that in this framework, the minimum price or premium that the dealer will accept to sell the derivative will be one that compensates it for the costs of hedging plus the residual risk.

More interestingly, we will see that in some cases, under certain theoretical situations, a perfect hedging strategy might exist, so the minimum price will be exactly the cost of the hedging strategy, which in this setup is also called the perfect replication strategy (since a perfect hedging implies no risk, and therefore a replication of the payoff using other financial instruments). A consequence of the existence of such replication strategy is that dealers are forced to price derivatives consistently, otherwise they would be generating risk-free arbitrage opportunities where other dealers who price correctly the derivative trade with the one miss-pricing it, and pocket the difference without risk. Hence, this theory of derivatives pricing is also called the arbitrage-free theory of derivatives pricing.

Let us revisit the case of forwards and options under this optic.

Example: Forward on a non-dividend paying stock

Under the modelling hypothesis used in the previous sections to value the premium of a forward, namely, that 1) the interest risk is locked during the period of the forward, 2) there is no counter-party risk, i.e. no risk that the investor will not satisfy its obligations, then there is actually a simple replication strategy that hedges all the risk of the contract. If the dealer is selling the forward to the investor, therefore guaranteeing a price KK to buy a share at time TT, then:

If we consider that daily accruing of interest can be well approximated by continuous accruing, then the investor needs to repay at TT an amount Ster(Tt)S_t e^{r(T-t)}. The payoff for the dealer at the expiry is therefore:

KSter(Tt)K - S_t e^{r(T-t)}

which is deterministic under the hypotheses of the model. A rational dealer of course will not accept a determinist loss, so the minimum premium that will command for this contract is the discounted value of this payoff:

CF,t=Ker(Tt)StC_{F,t} = K e^{-r(T-t)} - S_t

In practice, forward markets work by quoting the strike such that the premium is zero, hence:

FtK=Ster(Tt)F_t \equiv K = S_t e^{r(T-t)}

This is the arbitrage-free price of a forward contract. As mentioned above, it is called arbitrage-free since any other price would represent a risk-free arbitrage opportunity for other dealer. For example, let’s assume this dealer quotes a Kˉt<Kt\bar{K}_t < K_t. Another dealer could buy the forward from the dealer, and use the opposite replication strategy:

The payoff for this dealer is then:

Ster(Tt)Kˉt>Ster(Tt)Ster(Tt)=0S_t e^{r(T-t)} - \bar{K}_t > S_t e^{r(T-t)} - S_t e^{r(T-t)} = 0

i.e. a risk-free profit!

Example: European Call Options

In the case of options there is no such obvious static replication strategy if we are only allowed to use the underlying stock and repo contracts. By static replication strategy we mean that we don’t need to modify the positions of the replication portfolio (the stock and the repo) during the life of the forward. If we are able to trade other derivatives there is actually a static replication strategy. If the dealer sells the call option to an investor, then immediately

The payoff at the expiry is:

(STK)++(KST)++(STK)=0-(S_T-K)^+ + (K-S_T)^+ + (S_T-K) = 0

meaning that the replication portfolio of a put and a forward replicates the call option. At inception, the dealer is paid for the call a premium CC,tC_{C,t} and pays CP,tC_{P,t} for the put and CF,tC_{F,t} for the forward, hence the payoff at inception is:

CC,tCF,tCF,tC_{C,t} - C_{F,t} - C_{F,t}

In order to avoid losses, the minimum premium that must command is therefore:

CC,t=CF,t+CF,tC_{C,t} = C_{F,t} + C_{F,t}

which is called the put-call parity relationship. The premium for the forward can be derived as discussed in the previous section, however we are left with a sort of chicken and the egg problem with regards to the call and put premiums: given one, we can determine the other, but we don’t have yet a replication strategy for the put to derive the call premium, and vice-versa.

If no static replication is available, is it possible to find a dynamical one that reproduces the payoff without uncertainty? Or are we left with strategies that, though they might minimize uncertainty, they don’t remove it and therefore we need to go back to our utility indifference theory?

The Black - Scholes - Merton Model

Fisher Black and Myron Scholes Black & Scholes, 1973, and separately Robert Merton Merton, 1973 provided an answer to this question: under certain theoretical conditions, we can indeed find a dynamic replication strategy based on the underlying stock and a risk free account, that reproduces the payoff with no uncertainty. The main conditions are the following:

dSt=μStdt+σStdWtd S_t = \mu S_t dt + \sigma S_t d W_t
We have considered the drift $\mu$ and volatility $\sigma$ constant, but the model can be generalized to time-dependent deterministic drifts and volatilities. Other dynamics can also be considered within the same framework. Also, we will consider a non-dividend paying stock, but the model can be adjusted to consider deterministic dividends.
Πt=ΔtSt+βt\Pi_t = \Delta_t S_t + \beta_t
The model can be generalized easily to time-dependent risk-free interest rates. The original model considers a generic cash account, although in practice is more realistic to assume a repo on the stock is used for funding.
dΠt=ΔtdSt+dβt=ΔtdSt+rβtdt=ΔtdSt+r(ΠtΔtSt)dtd\Pi_t = \Delta_t d S_t + d \beta_t = \Delta_t d S_t + r \beta_t dt = \Delta_t d S_t + r (\Pi_t - \Delta_t S_t) dt
Notice that the position in the stock depends on time, but is always adjusted ex-post, i.e. after the market moves.

If the portfolio Πt\Pi_t indeed replicates the option payoff at the maturity TT in any scenario, we must have ΠT=CT=(STK)+\Pi_T = C_T = (S_T - K)^+, where we have considered a Call option but the argument applies to any other payoff function. But since by construction the portfolio Πt\Pi_t is self-financing, then if such dynamic strategy exists we must have Πt=Ct\Pi_t = C_t for any time tTt \leq T. Otherwise, there would be a risk-free arbitrage opportunity, e.g. selling the option at price CtC_t in the case Ct>ΠtC_t > \Pi_t, buying with the cash from the premium the replicating portfolio Πt\Pi_t and making an instantaneous gain of the difference CtΠtC_t - \Pi_t. The same reasoning applies Ct<ΠtC_t < \Pi_t, only in this case we buy the option paying a discounted premium. Therefore, if we can find such strategy we immediately solve the problem of option pricing, since the premium is equal to the cost of replication by using the non-arbitrage opportunity argument.

The condition Πt=Ct\Pi_t = C_t is equivalent to ΠT=CT\Pi_T = C_T and dΠt=dCtd \Pi_t = dC_t. Additionally, this equality implies that Ct=C(t,St)C_t = C(t, S_t), i.e. the premium of the option has to be a function of time and the contemporary value of the stock. We can then use Ito’s lemma to further understand the requirements for such strategy to exist:

dCt=Cttdt+CtStdSt+122Ct2Stσ2St2dt=dΠt=ΔtdSt+rβtdtd C_t = \frac{\partial C_t}{\partial t} dt + \frac{\partial C_t}{\partial S_t} dS_t + \frac{1}{2}\frac{\partial^2 C_t}{\partial^2 S_t} \sigma^2 S_t^2 dt = d\Pi_t = \Delta_t d S_t + r\beta_t dt

Grouping the terms dependent on dtdt and dStd S_t separately we have:

(Ctt+12σ2St22Ct2St+rβ)dt+(CtStΔt)dSt=0\left(\frac{\partial C_t}{\partial t} + \frac{1}{2} \sigma^2 S_t^2 \frac{\partial^2 C_t}{\partial^2 S_t} +r\beta\right) dt + \left(\frac{\partial C_t}{\partial S_t} - \Delta_t\right) dS_t = 0

Since the equality must apply for any arbitrary dtdt and dStdS_t, each term in parenthesis must cancel separately. Starting with the right hand term we have:

Δt=CtSt\Delta_t = \frac{\partial C_t}{\partial S_t}

which is call the delta-hedging condition, given its obvious connection with linear hedging strategies where we try to neutralize the exposure of a financial portfolio to a risk factor like the stock price in this case. Delta hedging the portfolio we ensure that its value is independent on the random dynamics of the stock, at least during an infinitesimal time. However, a strategy that delta-hedges at every time the portfolio is not necessarily self-financing, i.e. we might need to add extra cash during the life of the portfolio. The problem is that a priori the cash required for delta-hedging would depend on the path of the stock, making it a random variable. In that case the premium would also be a random quantity with a certain distribution.

In order to have a deterministic premium the portfolio has to be self-financing, which requires that the first term of the equality above is also zero, namely:

Ctt+12σ2St22Ct2St+rStCtSt=rCt\frac{\partial C_t}{\partial t} + \frac{1}{2} \sigma^2 S_t^2 \frac{\partial^2 C_t}{\partial^2 S_t} + r S_t \frac{\partial C_t}{\partial S_t} = rC_t

where we have used the self-financing condition βt=ΠtΔtSt=CtCtStSt\beta_t = \Pi_t - \Delta_t S_t = C_t - \frac{\partial C_t}{\partial S_t} S_t. The resulting equation is the celebrated Black-Scholes-Merton (BSM) partial differential equation. Solving this equation with the terminal condition CT=(STK)+C_T = (S_T - K)^+ allows us to compute the value of the premium for any time tTt \leq T deterministically. Therefore, by virtue of the replicating portfolio and non-arbitrage opportunity arguments, if the Black-Scholes-Merton conditions are satisfied the price of an option is no longer a random quantity as in the utility indifference framework.

Before discussing the solution to this equation, we can get another insight simply by inspecting it: the option premium does not depend on the drift μ\mu of the stock, only the volatility. In the utility indifference framework, the estimation of the drift plays a big role in the price that the investor is willing to pay for the option, since it affects the probability of exercising or not the option. However, for a dealer pricing the option, as far as the BSM framework holds, the directionality of the market is irrelevant since the strategy guarantees a replication of the payoff in any scenario by investing the premium into the BSM dynamic portfolio and implementing the dynamic strategy.

Solving the Black-Scholes-Merton equation

There are different ways to solve the BSM equation. Most introductory textbooks on the topic (see for example Joshi, 2003, Wilmott, 2007) follow the derivation used in the seminal paper that uses an ansatz for the solution that transforms the equation into the heat-equation, whose analytical solution is well-known Evans, 2010. Here, we will take a different approach and use the Feynman - Kac theorem introduced in Chapter (#stochastic_calculus), section (#feynman_kac). Recall that the Feynman - Kac theorem provides a general solution to a family of partial differential equations in term of an expected value. In the interest of the reader we review it again here: the solution to the PDE

ut+μ(x,t)ux+12σ2(x,t)2ux2r(x,t)u=0\frac{\partial u}{\partial t} + \mu(x,t) \frac{\partial u}{\partial x} + \frac{1}{2} \sigma^2(x,t) \frac{\partial^2 u}{\partial x^2} - r(x,t)u = 0

with the boundary condition u(x,T)=f(x)u(x,T) = f(x), is the following expected value

u(x,t)=E[etTr(Xs,s)dsf(XT)Xt=x]u(x,t) = \mathbb{E}\left[ e^{-\int_t^T r(X_s,s) ds} f(X_T) \Big| X_t = x \right]

where XtX_t satisfies the general stochastic differential equation:

dXt=μ(Xt,t)dt+σ(Xt,t)dWtd X_t = \mu(X_t, t) dt + \sigma(X_t, t) dW_t

The BSM equation is a specific case of this general PDE where XtStX_t \rightarrow S_t, u(x,t)C(s,t)u(x,t) \rightarrow C(s, t), μ(x,t)r\mu(x,t) \rightarrow r, σ(x,t)σ\sigma(x,t) \rightarrow \sigma, r(x,t)rr(x,t) \rightarrow r. The solution of the BSM equation can be expressed as:

C(s,t)=E[er(Tt)(STK)+St=s]C(s, t) = \mathbb{E}\left[ e^{-r(T-t)} (S_T - K)^+ \Big| S_t = s \right]

where StS_t satisfies the SDE:

dSt=rStdt+σStdWtd S_t = r S_t dt + \sigma S_t dW_t

The first crucial observation is that the solution is remarkably close to the one analyzed in the context of utility indifference pricing for a risk-neutral investor, with one caveat: the expected value is taken with respect to a SDE for the stock price that has the drift μ\mu replaced by the risk-free interest rr. In the former, the investor was taking the risk of the contract, but was indifferent to risk. In the latter, the dealer might be risk averse, but since the BSM dynamic hedging strategy neutralizes the risk (if the hypotheses are correct), there is no risk taken, only a deterministic cost to execute the hedging strategy, which is charged to the investor as the premium (plus potentially a margin for the service). In the jargon, the dealer prices the derivative as a risk-neutral investor if it computes probabilities in a different probability measure, where the drift is rr. Such measure is usually called the risk-neutral measure. This framework for derivatives pricing can be generalized to other derivatives, allowing dealers to compute prices directly skipping the construction of the hedging portfolio and solving the PDE. It is appropriately referred as risk-neutral derivatives pricing.

From a financial point of view, though, it is convenient not to lose the connection to the dynamic hedging strategy, since the validity of this theory is linked to the validity of the hypotheses done to derive the PDE. If we are interested in introducing more realistic dynamics into the pricing, like non-continuous hedging and transaction costs, we need to start again from the point of view of the replication portfolio and the hedging strategy. When those hypotheses are relaxed, though, we come back to a probability distribution to characterize the premium of the option instead of a deterministic one, the latter being one of the main remarkable results of the BSM theory.

The computation of the expectation value for the option premium can be reused from the one done in the context of utility indifference pricing, simply by substituting μr\mu \rightarrow r in the expressions:

C(St,t)=StN(d1(μ))Ker(Tt)N(d2(μ))C(S_t,t)= S_t N(d_1(\mu))-K e^{-r(T-t)} N(d_2(\mu))

where:

d1(μ)=1σTt(logStK+(rσ22)(Tt))d_1(\mu) = \frac{1}{\sigma \sqrt{T-t}} \left(\log \frac{S_t}{K} + (r - \frac{\sigma^2}{2})(T-t)\right)
d2(μ)=1σTt(logStK+(r+σ22)(Tt))d_2(\mu) = \frac{1}{\sigma \sqrt{T-t}} \left(\log \frac{S_t}{K} + (r + \frac{\sigma^2}{2})(T-t)\right)

This is the infamous Black-Scholes-Merton option pricing formula. It provides dealers with option prices that depend on

It does not depend on the stock drift μ\mu, i.e. the expected trend of the market. This is because in the BSM framework, the dealer uses the dynamic hedging strategy to make the portfolio risk-free and therefore shielded from any market trend.

Again, we must emphasize that despite the similarity with the risk-neutral option premium computed in the utility indifference framework, in the BSM framework the dealer is not computing the premium based on real probability scenarios of stock prices. The connection between replicating strategies and expectation values provided by the Feynman - Kac theorem is a useful mathematical result that simplifies the computation of derivatives premia in the BSM framework. But they shall not be interpreted in the sense of real probabilities (or probabilities in the real probability measure, as they are also referred to).

The dependencies of the option formula for a European call option with respect to the parameters are similar to the ones seen in the context of utility indifference pricing, with the exception of the dependence with respect to the risk-free interest rate rr, whose role in the formula goes now beyond cash-flow discounting. In the following picture we compare the premia derived with BSM versus utility indifference pricing:

Option price premium calculated using the utility indifference pricing versus arbitrage-free theory. The results are shown for an european call option with parameters S_t=100, K =100, T-t = 1 in years, r = 0.05, \mu = 0.1, \sigma = 0.2.

Figure 3:Option price premium calculated using the utility indifference pricing versus arbitrage-free theory. The results are shown for an european call option with parameters St=100S_t=100, K=100K =100, Tt=1T-t = 1 in years, r=0.05r = 0.05, μ=0.1\mu = 0.1, σ=0.2\sigma = 0.2.

The first main differences is, as commented, the dependence with respect to the risk-free interest: in BSM theory, the premium is more valuable as interest rates increase, wheres in utility indifference pricing it was the other way around. In the latter, interest rates enter the formula to discount the value of future cash-flows. As interest rates grow, the opportunity cost of investing the premium for the future cash-flows of the options increases, since a risk-free deposit generates comparatively a larger yield. In other words, for an investor, the option becomes less attractive and is willing to pay less for it. However, in BSM theory used by dealers to price options, the opposite happens: bigger interest rates make funding the hedging strategy more costly, therefore requiring a larger compensation in terms of premium to execute the replication strategy.

The second big difference is of course the dependence with respect to the expected market drift, since in the BSM theory the premium is independent of it. The resulting plot is interesting because it points out to the resolution of the question of why are options traded in practice, assuming both dealers and investors have the same view of the market. As mentioned before, if both were to value the option as an investment, there would be no agreement on the premium to pay, since they hold opposite sides in the trade. However, if the dealer uses BSM theory there is room for agreement, at least for those investors whose expectations on market drifts make the premium requested by the dealer attractive. In the picture, such situation corresponds to those expected drifts that make the premium that an investor is willing to pay larger than the one commanded by the dealer.

From a classical Economics point of view, those frameworks fit well together to explain the derivatives market in terms of demand and supply. Demand for options is driven by investors looking to generate returns on investment, whereas supply comes from dealers that “fabricate” those options using replication strategies. The BSM premium is essentially the cost of “fabricating” the option, in analogy to the language used in the production of goods.

An alternative derivation: the market price of risk

An alternative derivation of the BSM equation that can be helpful to gain intuition on the theory uses the financial concept of market price of risk. The market price of risk is essentially a Sharpe ratio, commonly used in the theory of investment. The Sharpe ratio computes the excess returns of an investment, i.e. the expected returns devoted fom the risk-free interest, over their risk defined as the volatility of the returns. For the stock that is the underlying of the option, and using continuos time, this is:

λS=E[dStSt]rdtVar[dStSt]=μtrσdt\lambda_{S} = \frac{ \mathbb{E}[\frac{dS_t}{S_t}]- rdt}{\sqrt{Var[\frac{dS_t}{S_t}]}} = \frac{\mu_t - r}{\sigma}\sqrt{dt}

We can now use Ito’s formula to compute the market price of risk of the option:

λC=E[dCC]rdtVar[dCC]=Ct+μtStCSt+12σ2St22CSt2rCσStCStdt\lambda_{C} = \frac{ \mathbb{E}[\frac{dC}{C}]- rdt}{\sqrt{Var[\frac{dC}{C}]}} = \frac{\frac{\partial C}{\partial t}+ \mu_t S_t\frac{\partial C}{\partial S_t}+\frac{1}{2}\sigma^2 S_t^2 \frac{\partial^2 C}{\partial S_t^2} - rC}{\sigma S_t \frac{\partial C}{\partial S_t}} \sqrt{dt}

We can now apply a different version of the arbitrage-free theory. Since the value of the option is essentially derived from the underlying stock, which the only risk factor affecting the option price in the BSM theory, then as investment opportunities both should have the same Sharpe ratio or market price of risk, i.e. λS=λC\lambda_S = \lambda_C. Otherwise, investors would bid up the price of the one with the largest Sharpe ratio until both of them equalize. Applying this equality the terms proportional to the drift μt\mu_t cancel and we get back to the BSM differential equation.

One could of course have used the argument in reverse, reorganizing the BSM equation in terms of market prices of risk to prove that the equality of those is a consequence of the arbitrate-free argument used when building the replication portfolio.

Using the BSM framework in practice

The Black - Scholes - Merton pricing theory supposed a change of paradigm for dealers creating liquidity in option markets. The theory allows for a consistent pricing of derivatives beyond options, providing not only a way to calculate the premium but a hedging strategy that neutralizes the risk of the derivative, or from a different angle, a recipe to synthesize those derivatives from liquid tradable instruments.

However, the BSM theory is based on multiple hypothesis that are not necessarily realistic, so the dealer needs to take into account how relevant is for the pricing and hedging of derivatives that those hypothesis are not consistent with reality. For instance:

In general, these deviations from the assumptions make the premium no longer deterministic, since they introduce uncertainty in its estimation. Dealers typically will need to estimate how much do they need to increase the BSM premium to compensate for those risks. In the following plots we show the histogram of differences between the replication portfolio and the option payoff at maturity, when different assumptions of Black - Scholes - Merton theory are violated, namely:

dSt=μStdt+VtStdW1,tdVt=κ(θVt)+χVtdW2,tE[dW1,tdW2,t]=ρdtd S_t = \mu S_t dt + \sqrt{V_t} S_t dW_{1,t} \\ d V_t = \kappa (\theta - V_t) + \chi \sqrt{V_t} dW_{2,t} \\ \mathbb{E}\left[dW_{1,t}dW_{2,t}\right] = \rho dt
Histograms showing the difference between the replication portfolio using the BSM dynamic hedging strategy, and the actual payoff of an European call option. Each plot shows the impact of violating a different hypothesis of the BSM theory. We run 10000 simulations for each case. We use the parameters S_t=100, K =100, T-t = 1 in years, r = 0.05, \mu = 0.1, \sigma = 0.2 for the baseline scenario where we expect close to perfect replication when re-hedging continuously. To test the effect of different realized volatilities we use use \sigma = 0.19 and \sigma = 0.21, respectively. To test the effect of transaction cost we add a constant bid-ask half-spread of 0.005%. To test the effect of a different market dynamics we use a Heston model with parameters \kappa = 20 (mean reversion rate of the volatility squared), \theta = 0.04 (long-term volatility squared mean), \xi = 0.2 (volatility of volatility squared), \rho = -0.7 (correlation between stock and stochastic volatility risk factors).

Figure 4:Histograms showing the difference between the replication portfolio using the BSM dynamic hedging strategy, and the actual payoff of an European call option. Each plot shows the impact of violating a different hypothesis of the BSM theory. We run 10000 simulations for each case. We use the parameters St=100S_t=100, K=100K =100, Tt=1T-t = 1 in years, r=0.05r = 0.05, μ=0.1\mu = 0.1, σ=0.2\sigma = 0.2 for the baseline scenario where we expect close to perfect replication when re-hedging continuously. To test the effect of different realized volatilities we use use σ=0.19\sigma = 0.19 and σ=0.21\sigma = 0.21, respectively. To test the effect of transaction cost we add a constant bid-ask half-spread of 0.005%. To test the effect of a different market dynamics we use a Heston model with parameters κ=20\kappa = 20 (mean reversion rate of the volatility squared), θ=0.04\theta = 0.04 (long-term volatility squared mean), ξ=0.2\xi = 0.2 (volatility of volatility squared), ρ=0.7\rho = -0.7 (correlation between stock and stochastic volatility risk factors).

In the plots, we can see the different impacts that the violations produce on the distribution of the residuals. Whereas a less frequent re-hedging increases the dispersion of the residual, those are still unbiased and symmetrical. Other violations produce skewed distributions with non-zero mean. When introducing transaction costs, residuals have a negative mean, meaning that the BSM premium is insufficient to cover the actual costs of replication, as expected since the BSM derivation does not take into account such transaction costs. The effect of a different dynamics for the underlying is more nuanced: depending on the actual process, the distribution of residuals might have positive or negative mean, meaning that the BSM premium over-estimates or under-estimates, respectively, the costs of replication. For instance, if the realized volatility is lower than the one used for pricing and hedging in the BSM model, the mean of the residual is positive. This is, again, intuitive, since as we seen the premium of the option increases with volatility, so overestimating it means charging a larger premium than necessary for replication. Notice that this is not necessarily good for a dealer in competition, since if other dealers have a better estimation of market volatilities, they will be able to offer more competitive prices to clients and close more deals.

The stochastic discount factor (SDF) pricing framework

This pricing framework estipulates a fundamental pricing equation for any asset (and, in particular, financial instruments) with a similar form as the naive pricing equation for the present value pricing framework when applied to assets with uncertain future cash-flows:

pt=Et[i=1NmtiCti]p_t = {\mathbb E}_t \left[\sum_{i=1}^N m_{t_i} C_{t_i}\right]

The difference being that now mtim_{t_i} is a stochastic discount factor that does not necessarily has the form derived using the argument based on the time value of money, namely ertie^{-rt_i}. That such pricing equation is general enough to price any asset can be derived based on two hypotheses:

We can proceed now with the derivation of the pricing equation. If the law of one price holds, we can express the price of any generic instrument as:

p(X)=ip(1si)Cip(X) = \sum_{i} p(1_{s_i}) C_i

If now we multiply and divide by the probabilities of each state sis_i, denoted πi\pi_i:

p(X)=ip(1si)πiπiCiE[mX]p(X) = \sum_{i} \frac{p(1_{s_i})}{\pi_i} \pi_i C_i \equiv {\mathbb E}[m X]

where we have defined the stochastic discount factor as mip(1si)πim_i \equiv \frac{p(1_{s_i})}{\pi_i}. Notice that in order to avoid having arbitrage opportunities, this stochastic discount factor has to be strictly positive, mi>0m_i > 0. To prove it, notice that in order to avoid arbitrages, any instrument with strictly positive cash-flows Ci>0C_i > 0 has to have a positive price. Using our pricing equation:

p(X)=imiπiCi>0mi>0,ip(X) = \sum_i m_i \pi_i C_i > 0 \rightarrow m_i > 0, \forall i

since CiC_i is postive and πi\pi_i is non-negative. Notice that condition is referred as the fundamental theorem of asset pricing Cochrane, 2005.

Given the role that time has in structuring financial instruments cash-flows, it makes sense to include splicitly the time dimension into the pricing equation. Let us add a time dimension to the market states, st,iSts_{t,i} \in S_t, so now we have a complete set of possible market states for each time tt, which for the moment we consider them to belong to a discrete time grid. A generic instrument is expressed now as:

X=t,i1st,iCt,iX = \sum_{t,i} 1_{s_{t,i}} C_{t,i}

Let us focus for the moment on a single Arrow-Debreu security paying 1st,i1_{s_t, i}. If we simply extend our pricing function to add the explicit time component we would have pt0(1st,i)=Et0[mt1st,i]p_{t_0}(1_{s_t, i}) = {\mathbb E}_{t_0}[m_t 1_{s_t, i}]. However, let us consider an intermediate time t0<t<tt_0 < t' < t. According to this definition, pricing at time tt' of the same instrument would be pt(1st,i)=Et[mt1st,i]p_{t'}(1_{s_t, i}) = {\mathbb E}_{t'}[m_t 1_{s_t, i}]. If this is the case, nothing prevent us to define a financial instrument that simply pays pt(1st,i)p_{t'}(1_{s_t, i}) at time tt`. Using our pricing formula, the price of this instrument at time t0t_0 is Et0[mtpt(1st,i)]{\mathbb E}_{t_0}[m_{t'} p_{t'}(1_{s_t, i})]. But this should be consistent with simply using the Tower law in our original pricing formula:

pt0(1st,i)=Et0[Et[mt1st,i]]=Et0[pt(1st,i)]Et0[mtpt(1st,i)]p_{t_0}(1_{s_t, i}) = {\mathbb E}_{t_0}\left[{\mathbb E}_{t'}[m_t 1_{s_t, i}]\right] = {\mathbb E}_{t_0}\left[p_{t'}(1_{s_t, i})\right] \neq {\mathbb E}_{t_0}\left[m_{t'} p_{t'}(1_{s_t, i})\right]

The issue can be overcome by deflating our pricing formula by the discount factor at the pricing time, so we have:

pt0(1st,i)=Et0[mtmt01st,i]p_{t_0}(1_{s_t, i}) = {\mathbb E}_{t_0}\left[\frac{m_t}{m_{t_0}} 1_{s_t, i}\right]

Using this corrected formula:

pt0(1st,i)=Et0[Et[mtmt01st,i]]=Et0[mtmt0pt(1st,i)]p_{t_0}(1_{s_t, i}) = {\mathbb E}_{t_0}\left[{\mathbb E}_{t'}\left[\frac{m_t}{m_{t_0}} 1_{s_t, i}\right]\right] = {\mathbb E}_{t_0}\left[\frac{m_{t'}}{m_{t_0}} p_{t'}(1_{s_t, i})\right]

which is now consistent. The intuition behind this adjustment is that the stochastic discont factor implicitly defines the numeraire of the economy - that is, a reference asset used as a unit of account, that ensures that prices are consistent across time. Deflating by the discount factor at the pricing date ensures that all values are measured in the same unit of account, so that prices observed at different times can be consistently compared and aggregated.

We can now extend the pricing formula to our generic instrument XX:

pt0(X)=Et0[t>t0mtmt0Ct]p_{t_0}(X) = {\mathbb E}_{t_0}\left[\sum_{t > t_0}\frac{m_t}{m_{t_0}} C_{t}\right]

Notice that this equation implies a recursive pricing equation:

pt0(X)=Et0[mt0+1mt0(pt0+1(X)+Ct0+1)]p_{t_0}(X) = {\mathbb E}_{t_0}\left[\frac{m_{t_0+1}}{m_{t_0}}\left(p_{t_0+1}(X') + C_{t_0+1}\right)\right]

When using the temporal representation, it is mathematically convenient to derive a continuous-time representation of the pricing formula. For that we introduce a time step Δt\Delta t so that tk=t0+kΔtt_k = t_0 + k \Delta t, k=1,...,Nk = 1, ..., N and a cash-flow rate Ctk=ctkΔtC_{t_k} = c_{t_k} \Delta t. Our pricing equation becomes:

mt0pt0(X)=Et0[k=1NmtkctkΔt]m_{t_0} p_{t_0}(X) = {\mathbb E}_{t_0}\left[\sum_{k = 1}^N m_{t_k} c_{t_k} \Delta t\right]

Now we take the continuous limit Δt0\Delta t \rightarrow 0 so we get:

mt0pt0(X)=Et0[t0Tmtctdt]m_{t_0} p_{t_0}(X) = {\mathbb E}_{t_0}\left[\int_{t_0}^T m_{t} c_{t} dt\right]

where Tt0+NΔtT \equiv t_0 + N \Delta t. The one-period recursive equation now becomes:

0=mtCt+Et[d(mtpt)]0 = m_t C_t + {\mathbb E}_{t}[d(m_t p_t)]

which in the absence of cash-flows between tt and t+dtt+dt becomes simply:

0=Et[d(mtpt)]0 = {\mathbb E}_{t}[d(m_t p_t)]

or, equivalently:

0=mtEt[dpt]+ptEt[dmt]+Et[dmtdpt]0 = m_t {\mathbb E}_{t}[d p_t] + p_t {\mathbb E}_{t}[d m_t] + {\mathbb E}_{t}[d m_t d p_t]

Practical applications of the SDF pricing framework

As we have discussed extensively in this chapter, we need fundamental pricing frameworks when we don’t have access to liquid market prices of financial instruments, otherwise we can just extract the fair values using filtering techniques. If we have complete markets, the general idea is to compute the stochastic discount factor using available prices of instruments, and then use those prices to price illiquid instruments which share the same risk factors as those of tradable instruments.

Notice that if the market is not complete, we can still use this framework to find a projection of the discount factor on the subspace of instruments with available market prices. This discount factor can be used to find a consistent price of instruments that have some risk factors out of this subspace, by decomposing the general discount factor in:

This will provide us with a price that has the minimum uncertainty given the available prices.

In complete markets, we have the guarantee that a stochastic discount factor exists. Let us compute it for some representative financial instruments.

Bond pricing

We consider a standard bond paying a fixed coupon rate cc at periodic times ti=1,...,Nt_i = 1, ..., N. The day-count fraction γi\gamma_i is the annualized fraction of days between coupon payments. The bond is referred to a notional MM, so the coupon cash-flows are Ci=γicMC_i = \gamma_i c M and the principal paid at maturity T=tNT = t_N is MM.

The payoff is therefore:

X=i=1NγicM1ti+M1TX = \sum_{i=1}^{N} \gamma_i c M 1_{t_i} + M 1_T

The price at time tt is therefore given by:

Bt=Et[i=1NmtimtγicM1ti+mTmtM1T]=i=1NEt[mtimt]γicM+Et[mTmt]MB_t = {\mathbb E}_t\left[ \sum_{i=1}^{N} \frac{m_{t_i}}{m_t}\gamma_i c M 1_{t_i} + \frac{m_{T}}{m_t} M 1_T \right] = \sum_{i=1}^{N} {\mathbb E}_t\left[ \frac{m_{t_i}}{m_t}\right] \gamma_i c M + {\mathbb E}_t\left[\frac{m_{T}}{m_t}\right] M

We define the discount factor as D(t,ti)Et[mtimt]D(t, t_i) \equiv {\mathbb E}_t\left[ \frac{m_{t_i}}{m_t}\right], so we have:

Bt=i=1ND(t,ti)γicM+D(t,T)MB_t = \sum_{i=1}^{N} D(t, t_i) \gamma_i c M + D(t, T) M

How to proceed from here depends on our modelling choices regarding the risk factors that are relevant for pricing as well as the set of liquid instruments with available prices. For example, if we have a set of NN bonds from the same issuer paying coupons at the same dates tit_i but with different maturities, we could simply write the NN pricing equations and solve for the discount factors D(t,ti)D(t, t_i), without having to compute explicitely the SDF. This could be used to price non-standard bonds (e.g. with different deterministic coupons or day-count fraction conventions) as far as they pay on the same time grid. If they pay at different times, we need to make some theoretical hypothesis to be able to interporlate the value of the discount factors, or directly build a model of the SDF.

A first simple model is assuming that bonds only depend on a single risk factor, an overall macroeconomic interest rate rtr_t, for example a short-term interbank rate (e.g. one linked to collateralized contracts like overnight index swaps, see chapter {ref}`intro_financial_instruments). For the moment, we consider it deterministic and constant: rt=rr_t = r. Let us assume in this market we have access to a money-market account that accrues interest continuously. The pay-off at time TT of the money market account is βT=βter(Tt)\beta_T = \beta_t e^{r(T-t)}, for a initial investment βt\beta_t, which is also naturally the price of this instrument at time tt. Therefore, the pricing equation is given by:

mtβt=Et[mTβT]=mTβTm_t \beta_t= {\mathbb E}_t\left[ m_T \beta_T \right] = m_T \beta_T

where, in the second step, we have applied that interest rates are deterministic and also the only risk factor in our model, so the SDF becomes deterministic as well. Therefore, the SDF is given by:

mT=mter(Tt)m_T = m_t e^{-r(T-t)}

whose dynamics is: dmt=rmtdtdm_t = - r m_t dt. We can then simply compute the discount factors at any arbitrary time as D(t,ti)=er(tit)D(t, t_i) = e^{-r(t_i -t)}, and the price of a standard bond simplifies to:

Bt=i=1Ner(tit)γicM+er(Tt)MB_t = \sum_{i=1}^{N} e^{-r(t_i-t)} \gamma_i c M + e^{-r(T-t)} M

Notice that we have recovered the pricing equation derived in the present value pricing framework. As already anticipated, the SDF framework is general enough to incorporate this pricing framework, which corresponds to the case of a simple market with only one tradable instrument, the money-market account, and the hypothesis that interest rates are deterministic and constant. It is actually not difficult to generalize this result to time-dependent deterministic interest rates rtr_t. Using dmt=rtmtdtdm_t = -r_t m_t dt, we have mT=mtetTrtdtm_T = m_t e^{-\int_t^T r_t dt}, i.e. D(t,ti)=ettirtdtD(t, t_i) = e^{-\int_t^{t_i} r_t dt}.

In practice, though, it is too simplistic to consider that the price of bonds, even those issued by governments with sound finances, does not have an idiosyncratic country risk factor. This can be seen empirically, since the price of traded bonds don’t usually matches the discounting of their future cash-flows using interbank rates. The standard practice is to introduce their own interest rate risk factors, defined by the so-called yield curve y(t,Tk)y(t, T_k), which by definition is the interest rate that matches market prices of standard bonds with maturities TkT_k:

Bk,tmkt=i=1Ney(t,Tk)(tit)γickMk+ey(t,Tk)(Tkt)MkB_{k,t}^{mkt} = \sum_{i=1}^{N} e^{-y(t, T_k)(t_i-t)} \gamma_i c_k M_k + e^{-y(t, T_k)(T_k-t)} M_k

Again, in order to extend this pricing framework to other instruments with non-liquid prices, we need to be able to interporlate the yield curve to other maturities. Market practitioners might directly use interpolation schemes that ensure the yield curve is well behaved, e.g. does not produce prices that are arbitragable. There is also a large literatur on term-structure interest rate models from which consistent yield curve parametric functions can be derived, that are then fitted to market prices. We refer the reader to Brigo & Mercurio, 2006 Andersen & Piterbarg, 2010 Andersen & Piterbarg, 2010 for more details.

For the purpose of our discussion on how to build stochastic discount factor models, let us consider one of the most simple instances of such term-structure models, the Vasicek model Vasicek, 1977. This model assumes that the entire yield curve is driven by a single risk factor, represented by an instantaneous continuously compounded short rate rtr_t that drives the movements of the full yield curve y(t,T)y(t, T). The short rate rtr_t is modeled as a stochastic process following an Ornstein–Uhlenbeck mean-reverting dynamics, as discussed in Stochastic Calculus:

drt=κ(θrt)+σdWtdr_t = \kappa (\theta - r_t) + \sigma dW_t

where κ>0\kappa > 0 is the speed of mean reversion, θ\theta the long run mean level, σ>0\sigma > 0 the volatility and WtW_t a Wiener process. Notice that this short-rate is not anymore a interbank reference rate, but a funding rate linked to the issuer. As mentioned above, an alternative model could try to keep an explicit decomposition as rt=rtois+str_t = r_t^{ois} + s_t, where now rtoisr_t^{ois} is the interbank rate and sts_t the spread associated with the specific issuer, linked to specific funding, credit and liquidity characteristics of the issuer. But we will not follow this path in this section.

In order to find the SDF, we still assume there is a money-market account βt\beta_t now linked to the funding short-rate rtr_t of the issuer. Additionally, we define so-called zero-coupon bonds (ZCBs) that pay only a principal of 1atmaturity at maturity T$, whose prices are directly the discount factors, since:

Pt(1T)=Et[mTmt]=D(t,T)P_t(1_{T}) = {\mathbb E}_t \left[\frac{m_T}{m_t} \right] = D(t, T)

We propose the simplest ansatz for the SDF that preserves non-arbitrability, i.e. as we discussed in chapter Stochastic Calculus, a log-normal process whose stochastic differential equation is given by:

dmtmt=μt(rt)dt+λt(rt)dWt\frac{d m_t}{m_t} = \mu_t(r_t) dt + \lambda_t(r_t) dW_t

where μ\mu and λ\lambda are, for the moment, generic functions of time and the short-rate.

Our SDF has to price all the instruments in our market: the money-market account and the zero-coupon bonds. We first apply the continuous-time version of our pricing equation to the money-market equation, namely:

E[d(mtβt)]=0{\mathbb E}[d(m_t \beta_t)] = 0

where, recall, dβt=rtβdtd\beta_t = r_t \beta dt. Applying Ito’s Lemma:

d(mtβt)=dmtβt+mtdβt+dmtdβt=βtmt(μt(rt)+rt)dt+βtmtλt(rt)dWtd(m_t \beta_t) = dm_t \beta_t + m_t d\beta_t + dm_t d\beta_t = \beta_t m_t (\mu_t(r_t) + r_t) dt + \beta_t m_t \lambda_t(r_t) dW_t

Applying the pricing equation, we get a condition on the SDF:

E[d(mtβt)]=βtmt(μt(rt)rt)dt=0μt(rt)=rt{\mathbb E}[d(m_t \beta_t)] = \beta_t m_t (\mu_t(r_t) - r_t) dt = 0 \rightarrow \mu_t(r_t) = - r_t

Let us apply it now to the ZCBs. We make the ansatz D(t,T)=f(t,rt)D(t, T) = f(t, r_t) given that the SDF itself is Markovian on rtr_t. Applying Ito’s lemma to this expression, we get:

df(t,rt)=ftdt+frtdrt+122f2rtσ2dtd f(t, r_t) = \frac{\partial f}{\partial t} dt + \frac{\partial f}{\partial r_t} dr_t + \frac{1}{2} \frac{\partial^2 f}{\partial^2 r_t} \sigma^2 dt

where we have used the SDESDE for rtr_t given by the Vasicek model. Now we apply Ito’s on d(mtf(t,rt))d(m_t f(t, r_t)):

d(mtf(t,rt))=mt(ft+κ(θrt)frt+12σ22f2rt)+mtσfrtdWt+fmt(rdt+λtdWt)+λtmtσfrtdtd (m_t f(t, r_t)) = m_t \left(\frac{\partial f}{\partial t} + \kappa (\theta - r_t) \frac{\partial f}{\partial r_t} + \frac{1}{2} \sigma^2 \frac{\partial^2 f}{\partial^2 r_t}\right) + m_t \sigma \frac{\partial f}{\partial r_t} dW_t + f m_t (-r dt + \lambda_t dW_t) + \lambda_t m_t \sigma \frac{\partial f}{\partial r_t} dt

We use now the pricing equation:

E[d(mtD(t,T))]=0{\mathbb E}[d(m_t D(t, T))] = 0

to get the following partial differential equation for f(t,rt)f(t, r_t):

ft+κ(θrt)frt+12σ22f2rtfr+λtσfrt=0\frac{\partial f}{\partial t} + \kappa (\theta - r_t) \frac{\partial f}{\partial r_t} + \frac{1}{2} \sigma^2 \frac{\partial^2 f}{\partial^2 r_t} - f r + \lambda_t \sigma \frac{\partial f}{\partial r_t} = 0

with terminal condition f(T,rT)=1f(T, r_T) = 1. To solve this equation we make a farther simplification and consider that λt(rt)\lambda_t(r_t) is affine in rtr_t, meaning it is a linear function:

λt(rt)=λ0+λ1rt\lambda_t(r_t) = \lambda_0 + \lambda_1 r_t

In this case, the exponential ansatz f(t,rt)=A(t,T)eB(t,T)rtf(t, r_t) = A(t, T) e^{-B(t, T)r_t} transform the problem into the following two PDEs:

B˙(t,T)=1(κ+λ1σ)B(t,T)\dot{B}(t, T) = 1 - (\kappa + \lambda_1 \sigma) B(t, T)
A˙(t,T)=A(t,T)[κθB(t,T)12σ2B(t,T)2λ0σB(t,T)]\dot{A}(t, T) = A(t, T)[\kappa \theta B(t, T) - \frac{1}{2}\sigma^2 B(t, T)^2 - \lambda_0 \sigma B(t, T)]

with terminal conditions B(T,T)=0B(T, T) = 0 and A(T,T)=1A(T, T) = 1. The solution reads:

B(t,T)=1e(κ+σλ1)(Tt)κ+σλ1B(t, T) = \frac{1- e^{-(\kappa + \sigma \lambda_1)(T-t)}}{\kappa + \sigma \lambda_1}
A(t,T)=(κθσλ0κ+σλ1σ22(κ+σλ1))(B(t,T)(Tt))σ24(κ+σλ1)B(t,T)2A(t, T) = \left(\frac{\kappa \theta - \sigma \lambda_0}{\kappa + \sigma \lambda_1} - \frac{\sigma^2}{2 (\kappa + \sigma \lambda_1)}\right)(B(t, T) - (T-t)) - \frac{\sigma^2}{4(\kappa + \sigma \lambda_1)}B(t, T)^2

with this solution, now we can fit the parameters λ0\lambda_0 and λ1\lambda_1 to prices of zero coupon bonds that can be themselves be extracted from liquid bond prices. As expected, though, with two parameters we will be able to fit only approximately this term structure. In order to fit the prices of any set of liquid bonds from a given issuer, we need a model that allows for further flexibility. One such model is for example the Hull & White model Hull & White, 1990.

Stock pricing

Option pricing

Connection to previous pricing frameworks

Exercises

Footnotes
  1. We consider non-dividend paying stocks for simplicity, the extension of the theory to dividend paying stocks is relatively straightforward

  2. A well-known historical short-coming of the BSM framework is the implication that european options on the same underlying with different strikes and maturities should have the same implied volatility, equal to the expected volatility of the stock. The implied volatility is the one obtained by inverting the BSM formula given prices observed in the market, assuming for instance that there is a set of standard options that are traded in an exchange. In the first years of application of the BSM theory to price options, around the 1980s, this had the consequence of having very small premiums for options, particularly put options, with strikes very deep out-the-money (i.e. far from the underlying price at the time of quoting). This was a consequence of a Gaussian assumption on price returns, that predicted a very low probability of such options being exercised. In 1989, such prediction was contradicted when the market dramatically crashed, forcing dealers to readjust the prices with respect to the BSM formula. Multiple models such as local or stochastic volatility models, or models with jumps in the dynamics, have been proposed later to address these issues. One point to bear in mind is that the logic applied in the BSM framework can be still applied when introducing these more complex dynamics, and deterministic premiums can be derived as far as we add extra instruments in the hedging portfolio that allows the dealer to neutralize those risks (stochastic volatility, jumps, etc)

References
  1. Cochrane, J. H. (2005). Asset Pricing: Revised Edition. Princeton University Press.
  2. Gu’eant, O., & Pu, J. (2018). Mid-price estimation for European corporate bonds: a particle filtering approach (Papers 1810.05884). arXiv.org.
  3. Sinclair, E. (2010). Option Trading: Pricing and Volatility Strategies and Techniques (1st ed.). Wiley.
  4. Guo, X., Lai, T. L., Shek, H., & Wong, S. P.-S. (2017). Quantitative Trading: Algorithms, Analytics, Data, Models, Optimization. Chapman. 10.1201/9781315371580
  5. Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of Political Economy, 81(3), 637–654.
  6. Merton, R. C. (1973). Theory of rational option pricing. Bell Journal of Economics and Management Science, 4(1), 141–183.
  7. Joshi, M. S. (2003). The Concepts and Practice of Mathematical Finance. Cambridge University Press.
  8. Wilmott, P. (2007). Paul Wilmott Introduces Quantitative Finance (2nd ed.). John Wiley & Sons.
  9. Evans, L. C. (2010). Partial Differential Equations (2nd ed., Vol. 19). American Mathematical Society.
  10. Brigo, D., & Mercurio, F. (2006). Interest Rate Models: Theory and Practice: With Smile, Inflation and Credit (2nd ed.). Springer.
  11. Andersen, L. B. G., & Piterbarg, V. V. (2010). Interest Rate Modeling. Volume 1: Foundations and Vanilla Models. Atlantic Financial Press.
  12. Andersen, L. B. G., & Piterbarg, V. V. (2010). Interest Rate Modeling. Volume 2: Term Structure Models. Atlantic Financial Press.
  13. Vasicek, O. A. (1977). An equilibrium characterization of the term structure. Journal of Financial Economics, 5(2), 177–188. 10.1016/0304-405X(77)90016-2
  14. Hull, J., & White, A. (1990). Pricing interest-rate-derivative securities. The Review of Financial Studies, 3(4), 573–592.