2,289
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Mean and variance of Brownian motion with given final value, maximum and argmax

ORCID Icon
Pages 679-698 | Received 27 Aug 2020, Accepted 28 Jul 2021, Published online: 17 Aug 2021

Abstract

The conditional expectation and conditional variance of Brownian motion, B(t), is considered given B(t = 1), its maximum and its argmax, B(t|close,max,argmax), as well as those with less information: B(t|close,argmax), B(t|argmax),B(t|max,argmax) where the close is the final value: B(t=1)=c and t[0,1]. We compute the expectation and variance of a Brownian meander in time. By splicing together two Brownian meanders, the mean and variance of the constrained process are calculated. Analytic expressions are given for E[B(t|close,max,argmax)],E[B(t|max,argmax)] and E[B(t|argmax)] as well as the analogous expression for the conditional variance. Computational results displaying both the expectation and variance in time are presented. Comparison of the simulation with theoretical values are shown when the close and argmax are given. We show that using the values of max and argmax as well as the final value, c, significantly reduces the variance of the process compared to the variance of the Brownian bridge.

MATHEMATICS SUBJECT CLASSIFICATION: 2010::

1. Introduction

The study of Brownian extrema dates back to Lèvy[Citation14]. We give a brief overview of the results following Devroye[Citation7] with additional results taken from Durrett et al.[Citation8], Durrett and Iglehart[Citation9], Shepp[Citation19], and Imhof[Citation11]. Our research focuses on distribution of B(t) on [0,1] given argmax of B(t) and one or more of the maximum and the final value, B(t=1)=c. For each Brownian path on [0,1], we denote the maximal value of B(t) by h, denoting the “high.” We denote the minimal value, the “low,” by . The location of the first time when the B(t) reaches h will be denoted by θ (and occasionally by argmax). The location of the first time when the B(t) reaches will be denoted by θ. We are interested in the density of Brownian paths conditional on (c,h,,θ,θ). Unfortunately, this five dimensional parameterization is too complex for us to evaluate. Instead, we settle for evaluating the conditional density, p(x,t|c,θ,h) as well as p(x,t|θ,h) and p(x,t|θ). The distribution, P(x,t|c,θ,h), is well known[Citation6,Citation20], and has a representation in terms of two Brownian meanders[Citation3,Citation8–11]. We do give a more explicit representation of this density in Corollary 4.2. Our companion article[Citation17] examines on p(x,t|c,h,),p(x,t|c,h),p(x,t|h,) and p(x,t|h).

To better understand the conditional probability, we calculate the first and second moments of the probability conditional with respect to x. We evaluate the distribution, mean and variance of B(t|close,argmax,max),B(t|max,argmax) and B(t|argmax). Our companion article of Riedel[Citation17] examines on B(t|max),B(t|close,max), and B(t|close,max,min). For our results, we view the maximum value, the final value and the location of the maximum, θ, as statistics on the Brownian motion. In other words, given values of (c,h,θ), we can estimate the properties of a realization of Brownian motion by plugging in the values of these statistics into the density and its first and second moments. The calculation of the moments of p(x,t|c,θ,h) is new, as is the evaluation of p(x,t|θ,h) and p(x,t|θ) and their moments.

By computing the moments, E[B(t|close,max,argmax)] and Var[B(t|close,max,argmax)], we show how the values of (close, high, low, argmax) determine the behavior of B(t) over the entirety of the region, [0,1]. Both the analytic formulae of the conditional expectation and variance as well as the figures in this paper build intuition about Brownian paths.

Section 2 reviews the distribution of Brownian motion extrema. Section 3 reviews the Brownian meander and calculates its expectation and variance in Theorem 3.3. Section 4 elaborates on Williams’s construction[Citation6,Citation20,Citation21] of conditional Brownian motion as the splicing together of two Brownian meanders. Theorem 4.3 gives analytic expressions for the expectation and variance of the conditional density, E[B(t|c,θ,h)] and Var[B(t|c,θ,h)]. Section 5 calculates the density (Theorem 5.1) and moments, E[B(t|θ,h)] and Var[B(t|θ,h)], of B(t|θ=argmax,h=max). We compare the analytic moments with our simulation. Section 7 and Appendix B describe our simulation framework and compares the simulation with our analytic expressions for the conditional mean and variance. Section 6 presents our analytic B(t|argmax) including expressions for E[B(t|θ)] and Var[B(t|θ)]. It compares our analytic moments with those of the simulation. We conclude by comparing the time averaged expected variance given one or more of (close,argmax,max,min).

2. Distributions of Brownian extrema

The study of Brownian extrema dates back to the founders of the field[Citation14]. Our brief discussion follows Devroye[Citation7] with additional results taken from Borodin and Salminen[Citation4], Durrett[Citation8], Durrett and Iglehart[Citation9], Imhof[Citation11], and Shepp[Citation19]. To our knowledge, there is no known expression for p(c,h,,θ,θ). There are expressions for p(c,θ,h),p(c,h,), p(c, h), p(h,),p(h,θ),p(c,θ) as well as p(c), p(h) and p(θ).

In Karatzas and Shreve[Citation13] and Shepp[Citation19], the joint density of the close, B(t=1)c, the high, max{t1}B(t)h, the location of the high, argmax{t1}B(t)θ, is shown to be: (2.1) p(c,θ,h)=h(hc)πθ3/2(1θ)3/2exp(h22θ(hc)22(1θ)),h0,hc.(2.1)

Conditional density of θ and h given B(1) is (2.2) p(θ,h|B(1)=c)=h(hc)2πexp(c2/2)πθ3/2(1θ)3/2exp(h22θ(hc)22(1θ)),h0,hc.(2.2)

The unconditional density of θ and h is obtained by integrating Equation(2.1) on c < h[Citation1,Citation7,Citation19]: (2.3) p(θ,h)=1πhexp(h2/2θ)θ3/2(1θ)1/2,h>0,0<θ<1.(2.3)

The conditional density of h given θ[Citation19] is (2.4) p(h|θ)=hθexp(h2/2θ),h>0.(2.4)

Thus E[h|θ]=πθ/2,Var[h|θ]=(2π/2)θ. The classic result[Citation12,Citation14] derived using the reflection principle is the joint distribution of the close, c, the high, h is (2.5) P(h,c)=P(max{B(s),s[0,1]}h,B(1)=c)=ϕσ2(c)ϕσ2(2hc)(2.5) where we denote the Gaussian density by ϕσ2(x)=(2πσ2).5exp(x2/2σ2). This result has been generalized to other diffusions in Pitman and Yor[Citation16]. The marginal density of the maximum, h, and B(1): (2.6) p(h,c)=2π(2hc)exp((2hc)2/2),h0,hc.(2.6)

[Citation12,Citation14]. This can be derived by differentiating Equation(2.6) or by integrating Equation(2.1) over θ.

Let Hc be the random variable max{0<t1}B(t) conditional on B(t=1)=c. The density of Hc, p(h|c), can be computed using Devroye[Citation7] and McLeish[Citation15]: (2.7) HcmaxB(t|B(1)=c)Lc/2+c2+2E/2(2.7) where E is a standard exponentially distributed random variable and L means equality in distribution. The marginal density of h is the half normal: (2.8) p(h)=2πexp(h2/2),h>0.(2.8)

The marginal density of the location, θ, of the maximum of B(t) is the well-known arcsine distribution, p(θ)=1π1θ(1θ). Using (Citation2.Citation1), one can generate the joint distribution of (θ,M,B(1))[Citation7] as (2.9) (θ,M,B(1))L(Θ1+cos(2πU)2,2ΘE,2ΘE2(1Θ)E).(2.9)

An immediate consequence of Equation(2.9) is that the variance of B(1) given θ, the time of the maximum of B(t), is independent of θ and satisfies: (2.10) Var[B(1)|θ]=2(1π/4),E[B(1)|θ]=2(θ1θ).(2.10)

Here E, E, U are independent random variables where E and E are standard exponentially distributed and U is uniformly distributed in [0,1]. We are unaware of any previous derivation of this straightforward result. We actually discovered it during our simulation. We now evaluate p(θ,c). For c > 0: (2.11) p(θ,c)=cp(θ,h,c)dh=ch(hc)πθ3/2(1θ)3/2exp(h22θ(hc)22(1θ))dh(2.11) (2.12) =cθec22θπθ(1θ)(c21)ec22erfc(|c|(1θ)2θ)2π.(2.12)

Here erfc is the complimentary error function, erfc(z)=1erf(z). For c0, the same integral over h runs from 0 to . (2.13) p(θ,c)=|c|(1θ)ec22(1θ)πθ(1θ)(c21)ec22erfc(|c|κθ2)2π.(2.13)

This result corresponds to the conditional density p(θ|c) in Alabert and Caballero[Citation1].

A result that goes back to Lèvy[Citation5,Citation14], if not earlier, is

Theorem 2.1.

The joint distribution of the close, c, the high, h, and the low, is (2.14) P(c,h,)=P(B(1)=c,{B(s),s[0,1]}h)=k=ϕσ2(c2k(h))ϕσ2(c2h2k(h)(2.14) (2.15) =ϕσ2(c)k=0[ϕσ2(c2h2kΔ)+ϕσ2(c2+2kΔ)]+k=1[ϕσ2(c2kΔ)+ϕσ2(c+2kΔ)](2.15) where Δ(h).

Although Equation(2.14) is classic, Equation(2.15) seems to be new. By refactoring Equation(2.14) into the symmetric form, Equation(2.15), we not only treat h and symmetrically, but also show that the series is in an alternating form. Here P(c,h,) is a distribution in h, and a density in c. The density, p(h,) may be found in Riedel[Citation17].

3. Brownian meander and its moments

A Brownian meander, Bme(t), can be thought of as a Brownian motion restricted to those paths where Bme(t)0. Since this is a set of paths with measure zero, we rescale the Brownian paths to get a Brownian meander. Following Belkin[Citation2] and Durrett[Citation8], let B(t) be a Brownian motion, τ1=sup{t[0,1]:B(t)=0},Δ1=1τ1, we define the Brownian meander as Bme(t)=|B(τ1+tΔ1)|/Δ1. By Bcme(t), we denote the Brownian meander restricted to Bme(t=1)=c. Results for the distribution of meanders can be found in Biane[Citation3], Durrett[Citation8], Durrett and Iglehart[Citation9], Imhof[Citation10], and Imhof[Citation11]. Brownian meanders at a fixed time, t, are distributed as Devroye[Citation7] and Imhof[Citation10]: (3.1) Bcme(t)L(ct+t(1t)N)+2Et(1t),(3.1) where N is a standard normal variable, E is a standard exponentially distributed random variable, and N and E are independent. EquationEquation (3.1) applies for a single fixed time and not for the process, B(t|B(t=1)=c) for all t1. We are more interested in the transition density of the meander. Following Durrett et al.[Citation8], let ϕs(x)=(2πs).5exp(x2/2s),Nx(a,b)=abϕs(x)dx, then

Theorem 3.1

[Citation8,Citation10]. Bme has transition density: p(Bme(t)=y)=2yt3/2exp(y2/2t)N1t(0,y) for 0t1, y > 0. For 0<s<t1,x,y>0, (3.2) p(Bme(t)=y|Bme(s)=x)=gts(x,y)N1t(0,y)/N1s(0,x),(3.2) (3.3) gt(x,y)ϕt(yx)ϕt(y+x).(3.3)

P(Bme(1)x)=1exp(x2/2) is the Rayleigh distribution.

In Equation(3.2), we slightly abuse notation: The probability that Bme(s)=x is infinitesimally small and is a density in x: p(Bme(t)=y|Bme(s)=dx)=p(Bme(t)=y|Bme(s)=x)dx. With this understanding, we will continue to write the conditional densities without dx. The Bayesian Rule gives the conditional density given the final value of a Brownian meander:

Corollary 3.2

[Citation7]. For 0<s<t1,x,c>0, Bme has density: (3.4) p(Bme(s)=x|Bme(t)=c)=p(Bme(t)=c|Bme(s)=x)p(Bme(s)=x)/p(Bme(t)=c)=gts(x,c)xt3/2cs3/2exp(x2/2s)exp(c2/2t)=gts(x/μ,μc)xt3/2cs3/2(3.4) where μs/t. For t = 1, Equation(3.4) can be rewritten as p(Bme(s)=x|Bme(t)=c)=gs(1s)x,c=sc)xc.

Equivalent formulae to Equation(3.4) are found in Devroye[Citation7]. In Appendix A, we calculate the expectation and variance of the Brownian meander using this transition density:

Theorem 3.3.

For the Brownian meander, Bme(t), the first and second moments are (3.5) M1(s,t,c)=0xp(Bme(s)=x|Bme(t)=c)dx(3.5) (3.6) =[ts+sc2/t]erf(cs2t(ts))c+2s(ts)πtexp(sc2/2t(ts)),(3.6) (3.7) M2(s,t,c)=0x2p(Bme(s)=x|Bme(t)=c)dx=3s(ts)t+c2s2/t2. (3.7)

Proof:

These formulae are derived in Appendix A using the transition density Equation(3.4).

Theorem 3.3 is new and is the basis for the moment calculations in the remainder of the article. If t = 1, we often suppress the third argument: M1(s,c)=M1(s,c,t=1). The variance of the meander satisfies: Vr(s,c)=M2(s,c,t=1)M1(s,c,t=1)2. Thus, Vr(s, c) is the variance of a Brownian meander when the end time, t = 1 with final value c.

4. Spliced Brownian meander representation and its moments

Following Williams[Citation6,Citation7,Citation20,Citation21], we represent a Brownian motion on [0,1] given the location, θ, its maximum, h and its final value c. We represent B(t) in [0,θ] in terms of a meander Bme(t):

Theorem 4.1.

Consider a Brownian motion conditional on h=max0s1{B(s)}, θ is the smallest s such that B(θ)=h and B(1)=c. For a fixed time t, B(t|c,θ,h) is distributed as a scaled Brownian meander to the left of θ and a second scaled Brownian meander to the right of θ. (4.1) B(t|c,θ,h)LhθBrme(1t/θ)fortθwhererh/θ.(4.1)

Here Brme(t) is the meander process with Brme(t=1)=r and rh/θ. Similarly for t[θ,1], define (4.2) B(t|c,θ,h)Lh1θB˜qme(tθ1θ)fortθwhereqhc1θ(4.2) and B˜qme is an independent Brownian meander.

The formal proof may be found in Denisov[Citation6], Williams[Citation20], and Williams[Citation21]. The result is easy to understand: A Brownian meander, Bcme(t), is equivalent to a Brownian bridge restricted to non-negative paths. On either side of the first maximum, θ, the Brownian paths are restricted to not go above the maximum. Thus on either side of θ, we subtract off the appropriately scaled Brownian meander.

A rigorous notation would replace B(t|c,θ,h) with B(t|dc,dθ,dh). Theorem 4.1 is well known and sometimes called the Williams representation. We need a more explicit representation to calculate the moments. We now derive it using the meander density representation of the previous situation. We believe the following corollary is more explicit than previous analyses.

Corollary 4.2.

Under the assumptions of Theorem 4.1, the conditional density of a Brownian motion at a fixed time, t, satisfies (4.3) p(B(t)=x|c,θ,h)=gt(hx,h)θ1.5(hx)(θt)1.5hexp(h22θ)exp((hx)22(θt))fort<θ.(4.3)

For t>θ: (4.4) p(B(t)=x|c,θ,h)=g1t(hx,hc)(hx)(1θ)1.5(hc)(tθ)1.5exp((hx)22(tθ))exp((hc)22(1θ)).(4.4)

For tθ, the joint density satisfies (4.5) p(B(t)=x,θ,h,c)=(hc)(hx)π(1θ)1.5(θt)1.5gt(hx,h)exp((hx)22(θs)(hc)22(1θ))(4.5) where tθ. For t>θ, the joint density is (4.6) p(B(t)=x,θ,h,c)=h(hx)πθ1.5(tθ)1.5g1t(hx,hc)exp((hx)22(tθ)h22θ).(4.6)

Proof.

To prove Equation(4.3) and Equation(4.4), we insert the transition density of Equation(3.4) into the results of Theorem 4.1. To prove Equation(4.5), we multiply Equation(4.3) by p(θ,h,c) from Equation(2.1). To prove Equation(4.6), we multiply Equation(4.4) by p(θ,h,c).

Using Equation(4.1) and Equation(3.7), we calculate the moments of the conditional Brownian motion:

Theorem 4.3.

Under the assumptions of Theorem 4.1, the expectation of a Brownian motion satisfies E[B(t)|c,h,θ]. (4.7) E[B(t)|c,θ,h]=hθM1(1tθ,hθ)=(4.7) (4.8) hθ[thθ+h(θt)θ2]erf(hθt2tθ)+2t(θt)πθexp((θt)h22tθ)(4.8) for tθ. For tθ, the expectation satisfies: (4.9) E[B(t)|c,θ,h]=h1θM1(tθ1θ,hc1θ)fortθ.(4.9)

For tθ, E[B2(t)|c,θ,h]=h22hθM1(s,hθ)+θM2(s,hθ). Thus the variance satisfies (4.10) Var[B(t)|c,θ,h]=θVr(1tθ,hθ)fortθ,(4.10) (4.11) Var[B(t)|c,θ,h]=(1θ)Vr(tθ1θ,hc1θ)fortθ,(4.11) where Vr(s,c)M2(s,c)M1(s,c)2=3s(1s)+c2s2M1(s,c)2.

Proof.

In the representations, Equation(4.1) and Equation(4.2), the conditional density of the Brownian motion (on either side of the argmax, θ), is a linear transformation of the Brownian meander density. Thus the moments of the conditional density, Equation(4.4) and Equation(4.3), are linear transformations of the Brownian meander moments given in Theorem 3.3. The representations in Equation(4.7)Equation(4.11) follow by substituting the appropriate transformations of the Brownian meander moments into Equation(4.1)Equation(4.2).

5. Moments given high and its location

To calculate the expectation of B(t) given only its maximum and the first location of its maximum, we integrate the expressions in Equation(4.3) and Equation(4.4) over the correct conditional probability density given by Equation(2.1) and Equation(2.3): (5.1) p(x,t|θ,h)=hp(x,t|c,θ,h)p(θ,h,c)dc/p(θ,h).(5.1)

Theorem 5.1.

The conditional density, p(B(s)=x|θ,h) satisfies (5.2) p(B(s)=x|θ,h)=(hx)π(1θ)12(θs)1.5gs(hx,h)exp((hx)22(θs))(5.2) for t<θ. Here g1s(x,y)=ϕ1s(xy)ϕ1s(x+y). For t>θ, the conditional density is (5.3) p(B(s)=x|θ,h)=h(hx)πθ1.5(sθ)1.5erf(hx2(1s))exp((hx)22(sθ)h22θ).(5.3)

Proof.

We integrate Equation(4.5) and Equation(4.6) with respect to the close, c and then divide by p(θ,h) as given by Equation(2.3). We use hg1s(z,hc)dc=0ϕ1s(xz)ϕ1s(x+z)dx=erf(z2(1s)).

To calculate the expectation of B(t) given only its maximum and the first location of its maximum, we need only integrate the expressions in Equation(3.5)Equation(4.9) over the same conditional probability density: (5.4) E[B(t|θ,h)]=hE[B(t)|c,θ,h)]p(θ,h,c)dc/p(θ,h)(5.4) where p(θ,h,c) is given by Equation(2.1) and p(θ,h) is given by Equation(2.3). Thus p(θ,h,c)/p(θ,h)=(hc)(1θ)exp((hc)22(1θ)). We change variables from c to x(hc)1θ and apply (Citation9.Citation13)–(Citation9.Citation15) from Appendix A.

Theorem 5.2.

For t>θ, the expectation becomes (5.5) E[B(t|θ,h)]=h1θ0xM1(s,x)ex2/2dx=h1θG11(s),(5.5) where stθ1θ and G11(s) is defined as (5.6) G11(s)=2π[tan1(s1s)+s(1s)].(5.6)

Thus E[B(t|θ,h)]h is independent of h for tθ. For tθ,E[B(t)|c,θ,h] is independent of c and satisfies E[B(t)|θ,h]=E[B(t)|c,θ,h] as given by (Citation4.Citation7)–(Citation4.Citation8). For tθ, the variance is independent of the close and is given by (Citation4.Citation10): (5.7) Var[B(t|θ,h)]=θVr(1tθ,hθ)fortθ.(5.7)

For tθ, the variance of B(t|θ,h) (5.8) Var[B(t|θ,h)]=(1θ)[3ss2](1θ)G11(s)2.(5.8)

Proof.

The proof of (Citation5.Citation5) is in (Citation9.Citation13)–(Citation9.Citation15). The variance of B(t|θ,h) is calculated by the ensemble average of E[B2(t|c,θ,h)] and subtracting E[B(t|θ,h)]2. For tθ,E[B(t)2|c,θ,h] is also independent of c. This proves (Citation5.Citation7). For tθ, (5.9) Var[B(t|θ,h)]=hE[B2(t)|c,θ,h]p(c|θ,h)dcE[B(t|θ,h)]2(5.9) (5.10) =h22(1θ)hE[B(t|θ,h)]+(1θ)0xM2(s,x)ex22dxE[B(t|θ,h)]2.(5.10)

To evaluate (Citation5.Citation10), we insert (Citation9.Citation16) and (Citation5.Citation5) into (Citation5.Citation10). □

By ignoring the c dependence in B(t|c,θ,h), the variance increases by hE[B(t|c,θ,h)]2p(c|θ,h)dcE[B(t,θ,h)]2. One can think of this as an Analysis of Variance term. The increase occurs because the final value, c, is now a random variable.

Similar but more complicated moment calculations are possible for the case of the previous section where (θ,c) are given and one integrates over h. In Riedel[Citation17], analytic expressions for E[B(t|c,h)] and Var[B(t|c,h)] are given. It may be possible to rederive these results by integrating p(x,t|c,θ,h)p(θ,h,c)dθ/p(c,h). This integration appears to be untractable.

6. Mean and variance given only argmax

Now consider B(t|argmaxB=θ). To compute the mean and variance of B(t|θ), we integrate (Citation5.Citation4)–(Citation5.Citation5) with respect to p(h|θ)dh where p(h|θ) is given by (Citation2.Citation4):

Theorem 6.1.

For t<θ, (6.1) E[B(t|θ)]=0E[B(t)|θ,h]p(h|θ)dh=πθ20M1(1tθ,hθ)hθeh22θdh(6.1) (6.2) =πθ2θG11(s)=πθ22θπ[tan1(s1s)+s(1s)],(6.2) where s1tθ and G11 is given in (Citation5.Citation6). For t>θ, the expectation becomes (6.3) E[B(t|θ)]=πθ22(1θ)π[tan1(s1s)+s(1s)]fortθ(6.3) where stθ1θ.

For the variance, we first evaluate E[B2(t|θ)] and then subtract off E[B(t|θ)]2. (6.4) E[B2(t|θ)]=2θ4θs+θ[3ss2]fortθ(6.4)

For tθ, the variance simplifies: (6.5) Var[B(t|θ)]=Var[h|θ]+(1θ)[3ss2G11(s)2],(6.5) where Var[h|θ]=(2π/2)θ.

Proof.

We use the integrals in (Citation9.Citation13)–(Citation9.Citation15) of the Appendix A and set s=1t/θ. To prove (Citation6.Citation4), note (6.6) E[B2(t|θ)]=0E[B2(t)|θ,h]p(h|θ)dh=0[θx22θxM1(s,x)+M2(s,x)]xex22dx(6.6) and insert the expression in (Citation9.Citation17)–(Citation9.Citation18).

Our expression matches the zero dimensional results derived in the second section, E[h|θ]=πθ/2,Var[h|θ]=(2π/2)θ.

Let g(t,θ)=E[B(t|argmaxB=θ)]. plots g(t,θ) for ten quantile values of the bins in θ ranging from the second smallest bin to the second largest bin. Unsurprisingly, the g(t,θ) is maximized at t=θ for fixed θ. The curves look monotone for θ very near 0 or 1, but this is a graph and simulation resolution effect. Even at nbins = 160, there is still a small amount of artificial broadening of the peak due to averaging g(t,θ) over the bin. Using the density (Citation2.Citation6), we calculate E[h|θ]=πθ/2.

Figure 1. Expectation of B(t) given θargmax{B(t)} for various values of θ. Each color has two curves, a theoretical curve from Theorem 6.1 and the mean value of the simulation for the given parameter bin. To compute the simulation expectation, we use an ensemble of 32,000,000 realizations computed with 1500 steps and bin the results into 100 bins in θ space. The values of θ for each curve are given in the legend.

Figure 1. Expectation of B(t) given θ≡argmax{B(t)} for various values of θ. Each color has two curves, a theoretical curve from Theorem 6.1 and the mean value of the simulation for the given parameter bin. To compute the simulation expectation, we use an ensemble of 32,000,000 realizations computed with 1500 steps and bin the results into 100 bins in θ space. The values of θ for each curve are given in the legend.

plots the variance, var(t,θ)=Var[B(t)|argmaxB=θ]. For small values of θ, var(t,θ) grows linearly for most time, then it accelerates as θ approaches 1. For larger values of θ, it appears that var(t,θ)t is nearly a constant for t near zero. The growth rate decelerates in time and this deceleration occurs earlier for smaller θ. From (Citation2.Citation10), Var[B(1)|θ]=2(1π/4)0.4292.

Figure 2. Variance of B(t) given θargmax{B(t)} for various values of θ. When θ is small, the variance grows roughly linearly in time. For larger values of θ, the variance grows faster initially and then converges to the linear value.

Figure 2. Variance of B(t) given θ≡argmax{B(t)} for various values of θ. When θ is small, the variance grows roughly linearly in time. For larger values of θ, the variance grows faster initially and then converges to the linear value.

7. Overview of simulation results

We generate a large number of Brownian paths, bin the paths in (close,max,argmax) space and calculate the mean and variance for each time and bin. We order the coordinates of phase space, (q1,q2,q3), so that q1=B(1),q2=max0t1B(t) and q1=argmax0t1B(t). To see E[B(t)|givens] and Var[B(t)|givens], we simulate 10 to 36 million Brownian motions. A detailed description of the simulation method may be found in Appendix B and Riedel[Citation17]. We display simulations using 1500-1850 time steps for each realization.

For each bin, we compute the expectation and variance using (Citation4.Citation7)–(Citation4.Citation11). In Riedel[Citation18], we present a number of figures displaying and explaining the behavior of E[B(t)|c,θ,h] and Var[B(t)|c,θ,h] as t,c,θ,h vary. We omit these here for space reasons.

We evaluate the discrepancy/mean square error (MSE) between the theoretical formulae and our simulations. For each bin in parameter space, we compare the difference between (Citation4.Citation7) and the average of the curves in the bin. We then compute the mean squared error averaged over time: (7.1) MSE(c,θ,h)=01(E[B(t|c,θ,h)]Avgj[B˜j(t|c,θ,h)])2dt(7.1) where B˜j is the j curve in the bin and Avgj is the average over all of the paths in the given bin in parameter space. We then sort the bins from worst to best MSE. Not surprisingly the worst fitting bins often occur where the bins are the largest and the bias error, from having curves with slightly different parameters, is the largest.

Overall, the concatenated meander expectation fits the simulation expectation very well. To further illustrate this, we overlay all four sets of curves on the same plot. plots the curves that correspond to the worst 5%, 2%, 1%, and 0.2% difference between the simulation conditional mean and (Citation4.Citation7)–(Citation4.Citation9). There are 12031.7 million bins to compare. The curves overstruck by symbols are the simulation curves. The analytic formula curves have the same color but no symbol.

Figure 3. Comparison of Formula and Simulation: Each color has two curves, a theoretical curve from Theorem 4.3 and the mean value of the simulation for the given parameter bin. The curves overstruck by symbols are the simulation curves. The analytic formula curves have the same color but no symbol. Blue: 5% worst MSE Mean: 0.00039 at close: −0.0891, high: 0.431, argmax: 0.065. Cyan: 2% worst MSE Mean: 0.000507 at close: −0.166, high: 0.101, argmax: 0.00374. Green: 1% worst MSE Mean: 0.000608 at close: 2.246, high: 2.27, argmax: 0.643. Red: 0.2% worst MSE Mean: 0.00091 at close: 0.682, high: 1.991, argmax: 0.37.

Figure 3. Comparison of Formula and Simulation: Each color has two curves, a theoretical curve from Theorem 4.3 and the mean value of the simulation for the given parameter bin. The curves overstruck by symbols are the simulation curves. The analytic formula curves have the same color but no symbol. Blue: 5% worst MSE Mean: 0.00039 at close: −0.0891, high: 0.431, argmax: 0.065. Cyan: 2% worst MSE Mean: 0.000507 at close: −0.166, high: 0.101, argmax: 0.00374. Green: 1% worst MSE Mean: 0.000608 at close: 2.246, high: 2.27, argmax: 0.643. Red: 0.2% worst MSE Mean: 0.00091 at close: 0.682, high: 1.991, argmax: 0.37.

We conclude with a comparison of the variance of the simulation with the concatenated meander in (Citation4.Citation10)–(Citation4.Citation11). plots the curves that correspond to the worst 5%, 2%, 1%, and 0.2% difference between the simulation conditional variance and (Citation4.Citation10)–(Citation4.Citation11). In Riedel[Citation18], we display the analogous plots for E[B(t|θ,h). As expected, averaging over values of c reduces the noise in the simulations and reduces the discrepancy.

Figure 4. Comparison of Meander Variance Formula and Simulation: Each color has two curves, a theoretical variance and the variance of the simulation for the given parameter bin. The symbols are on the curve from the simulation. Blue: 5% worst MSE Var: 0.000054 at close: −1.743, high: 0.55, argmax: 0.15. Cyan: 2% worst MSE Var: 0.0000775 at close: 1.246, high: 2.177, argmax: 0.822. Green: 1% worst MSE Var: 0.0000981 at close: 2.531, high: 3.164, argmax: 0.723. Red: 0.2% worst MSE Var: 0.000159 at close: −1.437, high: 0.461, argmax: 0.0208.

Figure 4. Comparison of Meander Variance Formula and Simulation: Each color has two curves, a theoretical variance and the variance of the simulation for the given parameter bin. The symbols are on the curve from the simulation. Blue: 5% worst MSE Var: 0.000054 at close: −1.743, high: 0.55, argmax: 0.15. Cyan: 2% worst MSE Var: 0.0000775 at close: 1.246, high: 2.177, argmax: 0.822. Green: 1% worst MSE Var: 0.0000981 at close: 2.531, high: 3.164, argmax: 0.723. Red: 0.2% worst MSE Var: 0.000159 at close: −1.437, high: 0.461, argmax: 0.0208.

8. Summary

We have investigated the distribution, expectation and variance of Brownian motion conditional on the values of its maximum and its location as well as its final value. We give formulae for the mean and variance of a Brownian meander. Using the Williams construction, we evaluate the expectation and variance, E[B(t)|c,θ,h] and Var[B(t)|c,θ,h]. To analyze B(t|θ,h) and B(t|θ), we integrate the distribution and moments with respect to the close, c, and then with respect to the maximum, h. Similar but more complex formulae can be calculated for B(t|θ,c). Our simulations show good agreements with the analytic expressions. Many interesting features are displayed in the figures of the mean and variance. displays the ensemble average of the variance of B(t|givens). For example, (8.1) V(t|givens=(c,θ,h))Var[B(t|c,θ,h)]dp(c,θ,h).(8.1)

Figure 5. Comparison of Var[B(t|close)],Var[B(t|close)], versus Var[B(t|close,argmax)] versus Var[B(t|close,argmax,high)]. These curves are the ensemble average over all of the simulations.

Figure 5. Comparison of Var[B(t|close)],Var[B(t|close)], versus Var[B(t|close,argmax)] versus Var[B(t|close,argmax,high)]. These curves are the ensemble average over all of the simulations.

We compare the ensemble averages of Var[B(t|argmax)] with Var[B(t|close)],Var[B(t|close,argmax)] and Var[B(t|close,argmax,high)]. In all case, we compute the variance for each value of the givens separately and then average over the distribution of the givens. From , we see that argmax is a much less valuable statistic than either the close or max. The statistics, (close,argmax), are more valuable than the pair of statistics (max,argmax). If the final value (close) is known, the variance is symmetric in time. Otherwise, the variance increases in time.

We conclude by merging our results with those in Riedel[Citation17]. To measure how accurately we can estimate B(t|θ,h,c), we consider the time integral of the conditional variance averaged over the given variables: 01Var[B(t|c,θ,h)]dtdp(c,θ,h). If only the final value, c, is specified, the value of the integral is 1/6. presents the corresponding values for our simulations in this paper. This shows that the statistics triple (Close, High, Low) is more valuable than the triple (Close, High, ArgMax). In all cases, adding addition information such as the value of the high and/or its location significantly reduces the time averaged variance in comparison with that of the Brownian bridge (i.e. only using the final value of B(t=1)).

Table 1. Expected time average variance reduction. We multiply the variance by 6 in the third column to compare with knowing only the final value, c.

In our sister article[Citation17], we show using the high, low and close substantially improves the estimation of the log return of the SP500 over using just the close. We can reproduce this calculation/application in this article, but it will be artificial. The reason is that the location of the maximum is seldom if ever used to parameterize a stochastic process. In contrast, chartist analysis in finance heavily uses the values of the high and low. Note that shows that using the value of the maximum is better than using the value of arg maximum.

Acknowledgments

The author thanks his referees for their helpful remarks.

References

Appendix A:

Meander moments

We now compute the moments of a Brownian meander given the distribution in (Citation3.Citation4): (9.1) Mk(s,t,c)=0xkP(Bme(s)=x|Bme(t)=c)dx(9.1) (9.2) =0(ϕts(cx)ϕts(c+x))xkt3/2cs3/2exp(x2/2s)exp(c2/2t)dx(9.2) (9.3) =t3/2cs3/222πτexp(c2/2h)exp(c2/2t)0xkexp(ax2)sinh(bx)dx,(9.3) where τts,a12τ+12s=t2s(ts) and bcts. Note M0(s,t,c)=1: (9.4) M0(s,t,c)=t3/2cs3/212πτexp(c2/2h)exp(c2/2t)πb2a3eb24a=1.(9.4)

Here we use b2/4a=c2s/2t(ts) and 12h12t=s2t(ts) to show that the exponential terms cancel. The first moment of the meander satisfies (9.5) M1(s,t,c)=t3/2cs3/212πτexp(c2/2h)exp(c2/2t)[π[2a+b2]eb24aerf(b2a)4a2.5+b2a2](9.5) (9.6) =[ts+sc2/t]erf(cs2t(ts))c+2s(ts)πtexp(sc2/2t(ts)).(9.6)

Similarly, the second moment of the meander satisfies (9.7) M2(s,t,c)=t3/2cs3/212πτexp(c2/2h)exp(c2/2t)πb(6a+b2)eb24a8a7/2=(9.7) (9.8) =t3/2cs3/212(ts)b(6a+b2)8a7/2=t3/2s3/212(ts)1.5(1.5+b2/4a)2a5/2(9.8) (9.9) =(3+c2s/t(ts))(s(ts)t)=3s(ts)t+c2s2/t2.(9.9)

The integrations follow from (9.10) 0xexp(ax2)sinh(bx)dx=πb4a3eb24a(9.10) (9.11) 0x2exp(ax2)sinh(bx)dx=π[2a+b2]eb24aerf(b2a)8a2.5+b4a2(9.11) (9.12) 0x3exp(ax2)sinh(bx)dx=πb(6a+b2)eb24a16a7/2.(9.12)

For Section 5, we need the following integral: (9.13) 0xM1(s,x)ex22dx=0(1s+sx2)ex22erf(xκ2)+2πσ(s)xex22(1s)dx(9.13) (9.14) =2π[tan1(κ)+sκκ+1+s(1s)3](9.14) (9.15) =G11(s)2π[tan1(s1s)+s(1s)],(9.15) where κ=s/(1s) and σ(s)s(1s). We use the identities: 0erf(ax2)exp(x22)dx=2πtan1(a) and π20x2erf(ax2)exp(x22)dx=aa2+1+tan1(a). To evaluate the variance, we need (9.16) 0M2(s,x,1)xex22dx=0(3s(1s)+x2s2)xex22dx=[3s(1s)+2s2]=[3ss2].(9.16)

To evaluate E[B2(t|θ)] in (Citation6.Citation4), note (9.17) 0x2M1(s,x)ex22dx=0(1s+sx2)xex22erf(xκ2)+2πσ(s)x2ex22(1s)dx(9.17) (9.18) =(1s)κ1+κ+sκ2κ+3(1+κ)3/2+σ(s)*(1s)3/2=2s.(9.18)

We used the identities: 0xerf(ax2)exp(x22)dx=a1+a2 and 0x3erf(ax2)exp(x22)dx=a(2a2+3)/(a2+1)3/2.

Appendix B:

Details of Brownian simulations

In the simulation, we generate about 17 million Brownian paths. We then bin the paths in a “grid” of bins in (close,θ,high) space. We use quantile-like coordinates in the (q1,q2,q3) parameterization. In the first phase space direction, q1=c, compute bin boundaries so that the number of curves are roughly equal in each bin. For each one dimensional bin, compute bin boundaries in the second coordinate direction, h, so that the number of bins is roughly equal. Finally, for each of the two dimensional bins, compute bins in the third direction, θ. Often, we choose a grid in c to correspond to quantiles of the normal distribution. (We center the c values to be the midpoints of the c grid.) The advantage of this approach is that the number of curves in each simulations is approximately the same. The disadvantage of this approach is that a convergence analysis is hard because the coordinates vary from simulation to simulation and halving the coordinate distance cannot be done in this adaptive gridding.

Given an ensemble of Brownian paths, {Bi(t)}, we can create an equivalent ensemble of Brownian paths, {Bi(t,c)}, with right endpoint c, using the formula: Bi(t,c)Bi(t)(Bi(t=1)c)t. In our case, we transform the original 10 to 36 million Brownian paths for each value of c, calculate the maximum, h, and arg-maximum, θ, of Bi(t,c). We then bin the paths into an nbins × nbins grid in (θ,h) space based on the quantiles of the (θ,h). Typical values of nbins are from 80 to 120. For each bin in the (close,θ,h) grid, we calculate the mean and variance for each time and bin.

To describe these simulations, we call nSteps, the number of timesteps in the any one path simulation and nSim is the number of Brownian paths that we compute in the simulation before shifting the paths to have final value c.

The values of argmax occur only on the time grid and this makes the quantile buckets discrete. To regularize the discreteness of argmax, we use a quadratic interpolation of argmax, that is, for a given realization of B(t) on a discrete path, we find the argmax on the grid, {ti}. We then use quadratic intepolation of B(t) on B(ti1),B(ti) and B(ti+1) to find a better approximation to the continuous time value of argmax. A more accurate calculation is to use the exception of the argmax conditional on the three point values B(ti1),B(ti),B(ti+1) where the discrete minimum occurs at B(ti)[Citation1].