Home Mathematics Vanilla options as controls for estimating the conditional expectation of a European derivative payoff
Article
Licensed
Unlicensed Requires Authentication

Vanilla options as controls for estimating the conditional expectation of a European derivative payoff

  • Eddie W. K. Chiu ORCID logo EMAIL logo
Published/Copyright: May 22, 2025

Abstract

One way to estimate conditional expectations based on a simulation sample is to fit a parametric model to the data set and compute the conditional expectations based on the fitted model. This method may be enhanced by the control variates method. The choice of controls is critical to the effectiveness of the method. We propose the use of vanilla options as controls for estimating conditional expectations of derivative payoffs, and we provide a theoretical analysis on the uniform approximation property of vanilla options for estimating conditional expectations. Our theory suggests that vanilla options can reduce the variance of a target payoff to an arbitrarily small level subject to some assumptions. We provide examples to illustrate our proposed approach and also discuss several considerations when applying vanilla options as controls for estimating conditional expectations.

MSC 2020: 60J70; 62L20; 91G20; 91G60

A Proofs

We present the proofs of the theorems presented in Section 4.1 here. The key idea is to limit the approximation of the target payoff function within a compact region in the domain of the function, in which we can approximate the payoff function with step functions, which in turn can be approximated further with vanilla call option payoffs.

We will begin with a few technical lemmas to facilitate the proof of the theorems.

Lemma A.1

Suppose Assumption 1 and Assumption 2 are satisfied. The conditional expectation

E [ 1 { S T > M } | X = x ] 0

for all 𝑥 in 𝑆 uniformly as M .

Proof

By Assumption 1 and Assumption 2, the conditional expectations of indicators

E [ 1 { S T > M } | X = x ] , M = 1 , 2 , ,

are continuous for all 𝑥 in 𝑆 which is compact. The conditional expectations of indicators are monotonic in 𝑀 and converge to 0 as M for each 𝑥 in 𝑆. Hence, by Dini’s theorem, E [ 1 { S T > M } | X = x ] 0 for all 𝑥 in 𝑆 uniformly as M . ∎

Lemma A.2

Suppose Assumption 1 and Assumption 2 are satisfied. For any non-negative real number 𝑎, we have

E [ 1 { S T a + ϵ } | X = x ] E [ 1 { S T > a } | X = x ]

for all 𝑥 in 𝑆 uniformly as ϵ 0 .

Proof

The proof is similar to the proof of Lemma A.1. We consider a decreasing sequence of positive real numbers ϵ n such that ϵ n 0 as n and the sequence of conditional expectations of indicators

E [ 1 { S T a + ϵ n } | X = x ] .

By Assumption 1 and Assumption 2, the conditional expectations of indicators are continuous for all 𝑥 in 𝑆 which is compact. The conditional expectations of indicators are monotonic in 𝑛 and converge to E [ 1 { S T > a } | X = x ] as n for each 𝑥 in 𝑆. Again, by Assumption 2, E [ 1 { S T > a } | X = x ] is continuous for all 𝑥 in 𝑆. Hence, by Dini’s theorem, E [ 1 { S T a + ϵ n } | X = x ] E [ 1 { S T > a } | X = x ] for all 𝑥 in 𝑆 uniformly as n . ∎

Lemma A.3

Suppose Assumption 1 and Assumption 2 are satisfied. The conditional expectation

E [ V T 2 1 { S T > M } | X = x ] 0

for all 𝑥 in 𝑆 uniformly as M .

Proof

By Assumption 1 and Assumption 2, E [ V T 4 | X = x ] exists and is continuous for all 𝑥 in 𝑆 which is compact. This implies that the image of E [ V T 4 | X = x ] over 𝑆 is a compact set, which is bounded. Hence there exists a K > 0 such that E [ V T 4 | X = x ] K for all 𝑥 in 𝑆.

By Cauchy–Schwarz inequality,

E [ V T 2 1 { S T > M } | X = x ] E [ V T 4 | X = x ] E [ 1 { S T > M } | X = x ] < K E [ 1 { S T > M } | X = x ] 0

for all 𝑥 in 𝑆 uniformly as M due to Lemma A.1. ∎

Lemma A.4

Suppose a real-valued function f ( x ) is continuous on an interval [ a , b ] . There exists a sequence of functions i = 1 n p i 1 { x q i } such that

a q i b , i = 1 , 2 , , n , i = 1 n p i = f ( b ) , and i = 1 n p i 1 { x q i } f

uniformly on [ a , b ] as n .

Proof

Note that [ a , b ] is compact since it is closed and bounded in ℝ. The function 𝑓 is continuous on this compact set [ a , b ] , so it is uniformly continuous on [ a , b ] . Therefore, for any ϵ > 0 , we may choose a δ > 0 such that | f ( y ) f ( x ) | < ϵ for any x , y [ a , b ] satisfying | y x | < δ .

Now consider a partition { q i } i = 1 , 2 , , N [ a , b ] such that

a = q 1 < q 2 < < q N = b and max i = 1 , 2 , , N 1 | q i + 1 q i | < δ .

Define f N ( x ) = f ( b ) 1 { x q N } + i = 1 N 1 f ( q i ) 1 { q i x < q i + 1 } . For each 𝑥 in [ a , b ] , we have

| f ( x ) f N ( x ) | i = 1 N 1 | f ( x ) f ( q i ) | 1 { q i x < q i + 1 } < ϵ .

Hence f N f uniformly on [ a , b ] . Since 1 { q i x < q i + 1 } = 1 { x q i } 1 { x q i + 1 } , the function f N may be written as a sum of step functions i = 1 N p i 1 { x q i } . Finally, by the definition of f N , we have i = 1 N p i = f N ( b ) = f ( b ) . ∎

Lemma A.5

Suppose Assumptions 13 are satisfied. In addition, suppose the function 𝑔 stated in Assumption 3 is right-continuous everywhere. There exists a sequence of functions i = 1 n p i 1 { S T q i } such that

E [ ( V T i = 1 n p i 1 { S T q i } ) 2 | X = x ] 0

for all 𝑥 in 𝑆 uniformly as n .

Proof

For any ϵ > 0 , by Lemma A.1 and Lemma A.3, there exists an M 1 > 0 and M 2 > 0 such that

E [ 1 { S T > M 1 } | X = x ] < ϵ , E [ V T 2 1 { S T > M 2 } | X = x ] < ϵ

for all 𝑥 in 𝑆. Taking M = max ( M 1 , M 2 ) , we have

E [ 1 { S T > M } | X = x ] < ϵ , E [ V T 2 1 { S T > M } | X = x ] < ϵ

for all 𝑥 in 𝑆.

Next, by Assumption 3, we may write V T = g ( S T ) and there exists h i [ 0 , ) , i = 1 , 2 , , N J , such that

  • for each h i , both lim x h i g ( x ) and lim x h i + g ( x ) exist and lim x h i g ( x ) lim x h i + g ( x ) ,

  • 𝑔 is continuous on [ 0 , ) \ { h i } and is left-continuous or right-continuous at each h i , i = 1 , 2 , , N J .

Since we suppose 𝑔 is right-continuous everywhere in this lemma, we have that 𝑔 is continuous on [ 0 , ) \ { h i } and is right-continuous at any h i , i = 1 , 2 , , N J .

Denote

J i = lim x h i + g ( x ) lim x h i g ( x )

and define g ̃ as

g ̃ ( x ) = g ( x ) i = 1 N J J i 1 { x h i } ,

which is clearly continuous for any 𝑥 not equal to any h i . For each h i , we have

lim x h i + g ̃ ( x ) lim x h i g ̃ ( x ) = lim x h i + g ( x ) + lim x h i + J i 1 { x h i } lim x h i g ( x ) = J i J i = 0 , g ̃ ( h i ) lim x h i + g ̃ ( x ) = g ( h i ) J i lim x h i + g ( h i ) + J i = 0 .

Hence lim x h i g ̃ ( x ) = g ̃ ( h i ) and g ̃ is continuous on [ 0 , ) . Applying Lemma A.4 to g ̃ on [ 0 , M ] , we have that there exists a function i = 1 N p i 1 { x q i } such that

0 q i M , i = 1 , 2 , , N , i = 1 N p i = g ̃ ( M ) , and | g ̃ ( x ) i = 1 N p i 1 { x q i } | < ϵ

for all 𝑥 in [ 0 , M ] . The last inequality implies

| g ( x ) i = 1 N J J i 1 { x h i } i = 1 N p i 1 { x q i } | < ϵ

for all 𝑥 in [ 0 , M ] .

Now consider

(A.1) E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 | X = x ] = E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 1 { S T M } | X = x ] + E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 1 { S T > M } | X = x ] .

The first term on the right-hand side of (A.1) is small since

E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 1 { S T M } | X = x ] < E [ ( ϵ ) 2 1 { S T M } | X = x ] ϵ

for all 𝑥 in 𝑆.

Next, we break down the second term on the right-hand side of (A.1) and show that each part is small. Considering the inequality ( a + b + c ) 2 3 ( a 2 + b 2 + c 2 ) for any real numbers 𝑎, 𝑏 and 𝑐, we have

E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 1 { S T > M } | X = x ] 3 E [ V T 2 1 { S T > M } | X = x ] + 3 E [ ( i = 1 N J J i 1 { S T h i } ) 2 1 { S T > M } | X = x ] + 3 E [ ( i = 1 N p i 1 { S T q i } ) 2 1 { S T > M } | X = x ] .

Then, by Cauchy–Schwarz inequality,

( i = 1 N J J i 1 { S T h i } 1 { S T > M } ) 2 ( i = 1 N J J i 2 ) ( i = 1 N J 1 { S T h i } 1 { S T > M } ) ,

and so

E [ ( i = 1 N J J i 1 { S T h i } ) 2 1 { S T > M } | X = x ] = E [ ( i = 1 N J J i 1 { S T h i } 1 { S T > M } ) 2 | X = x ] E [ ( i = 1 N J J i 2 ) ( i = 1 N J 1 { S T h i } 1 { S T > M } ) | X = x ] = ( i = 1 N J J i 2 ) E [ i = 1 N J 1 { S T h i } 1 { S T > M } | X = x ] ( i = 1 N J J i 2 ) E [ N J 1 { S T > M } | X = x ] < N J ( i = 1 N J J i 2 ) ϵ

for all 𝑥 in 𝑆.

Then note that

E [ ( i = 1 N p i 1 { S T q i } ) 2 1 { S T > M } | X = x ] = E [ ( i = 1 N p i 1 { S T q i } 1 { S T > M } ) 2 | X = x ] = E [ ( i = 1 N p i 1 { S T > M } ) 2 | X = x ] = ( i = 1 N p i ) 2 E [ 1 { S T > M } | X = x ] < ( i = 1 N p i ) 2 ϵ = ( g ̃ ( M ) ) 2 ϵ

for all 𝑥 in 𝑆, where the second equality is due to the fact that q i M , i = 1 , 2 , , N .

Putting everything together, from (A.1),

E [ ( V T i = 1 N J J i 1 { S T h i } i = 1 N p i 1 { S T q i } ) 2 | X = x ] < ϵ + 3 ϵ + 3 N J ( i = 1 N J J i 2 ) ϵ + 3 ( g ̃ ( M ) ) 2 ϵ = ( 4 + 3 N J ( i = 1 N J J i 2 ) + 3 ( g ̃ ( M ) ) 2 ) ϵ

for all 𝑥 in 𝑆. This proves the uniform convergence.

Finally, we may rewrite i = 1 N J J i 1 { S T h i } + i = 1 N p i 1 { S T q i } as a function in the form of i = 1 N p i 1 { S T q i } by suitably defining 𝑁, p i and q i , i = 1 , 2 , , N . ∎

Theorem A.6

Suppose Assumptions 13 are satisfied. There exists a sequence of functions i = 1 n p i 1 { S T q i } such that

E [ ( V T i = 1 n p i 1 { S T q i } ) 2 | X = x ] 0

for all 𝑥 in 𝑆 uniformly as n .

Proof

By Assumption 3, we may write V T = g ( S T ) and there exists h i [ 0 , ) , i = 1 , 2 , , N J , such that

  • for each h i , both lim x h i g ( x ) and lim x h i + g ( x ) exist and lim x h i g ( x ) lim x h i + g ( x ) ,

  • 𝑔 is continuous on [ 0 , ) \ { h i } and is left-continuous or right-continuous at each h i , i = 1 , 2 , , N J .

Denote 𝑙 to be the set that indicates the points h i at which 𝑔 is left-continuous, i.e.

l = { i { 1 , 2 , , N J } : lim x h i g ( x ) = g ( h i ) } .

If 𝑙 is an empty set, 𝑔 is right-continuous at any h i , i = 1 , 2 , , N J , and we conclude the proof by invoking Lemma A.5.

Suppose 𝑙 is non-empty. We consider an auxiliary function g r defined by removing all h i where 𝑔 is left-continuous. Denote

J i = lim x h i + g ( x ) lim x h i g ( x ) .

Define g r as

g r ( x ) = g ( x ) i l J i 1 { x > h i } .

Then g r is clearly continuous for any 𝑥 not equal to any h i . For each i l , we have

lim x h i + g r ( x ) lim x h i g r ( x ) = lim x h i + g ( x ) + lim x h i + J i 1 { x > h i } lim x h i g ( x ) = J i J i = 0

and, by the definition of 𝑙,

g r ( h i ) lim x h i g r ( x ) = g ( h i ) lim x h i g ( h i ) = 0 .

Hence lim x h i g r ( x ) = g r ( h i ) . That is, g r is continuous at h i for each i l . This implies that g r is right-continuous at h i for each i = { 1 , 2 , , N J } \ l and continuous elsewhere.

Consider any ϵ > 0 . By Lemma A.5, there exists a function i = 1 N p i 1 { S T q i } such that, for all 𝑥 in 𝑆,

E [ ( g r ( S T ) i = 1 N p i 1 { S T q i } ) 2 | X = x ] < ϵ

implying

E [ ( g ( x ) i l J i 1 { S T > h i } i = 1 N p i 1 { S T q i } ) 2 | X = x ] < ϵ .

From Lemma A.2, for each i l , there exists a δ i > 0 such that

E [ ( 1 { S T > h i } 1 { S T h i + δ i } ) 2 | X = x ] < ϵ

for all 𝑥 in 𝑆.

Considering the Cauchy–Schwarz inequality, we have

E [ ( g ( x ) i l J i 1 { S T h i + δ i } i = 1 N p i 1 { S T q i } ) 2 | X = x ] = E [ ( g ( x ) i l J i 1 { S T > h i } i = 1 N p i 1 { S T q i } + i l J i ( 1 { S T > h i } 1 { S T h i + δ i } ) ) 2 | X = x ] ( N L + 1 ) E [ ( g ( x ) i l J i 1 { S T > h i } i = 1 N p i 1 { S T q i } ) 2 | X = x ] + ( N L + 1 ) i l J i 2 E [ ( 1 { S T > h i } 1 { S T h i + δ i } ) 2 | X = x ] < ( N L + 1 ) ϵ + ( N L + 1 ) ( i l J i 2 ) ϵ = ( N L + 1 ) ( i l J i 2 + 1 ) ϵ

for all 𝑥 in 𝑆, where N L is the number of elements in the set 𝑙. This concludes the proof of the uniform convergence.

Finally, we may rewrite i l J i 1 { S T h i + δ i } + i = 1 N p i 1 { S T q i } as a function in the form of i = 1 N p i 1 { S T q i } by suitably defining N , p i and q i , i = 1 , 2 , , N . ∎

Next, we extend Theorem A.6 to approximate the target payoff with call option payoffs.

Lemma A.7

Suppose Assumption 1 and Assumption 2 are satisfied. Then, for any K > 0 ,

E [ ( 1 { S T K } ( S t K + ϵ ) + ( S t K ) + ϵ ) 2 | X = x ] 0

uniformly as ϵ 0 .

Proof

Consider a decreasing sequence of positive real numbers ϵ n such that ϵ n 0 . Denote

Δ n ( x ) = ( I { x K } ( x K + ϵ n ) + ( x K ) + ϵ n ) 2 , f n ( x ) = E [ Δ n ( S T ) | X = x ] .

First, note that

Δ n ( x ) = { ( x K + ϵ n ϵ n ) 2 , x [ K ϵ n , K ) , 0 , x [ K ϵ n , K ) ,

tends to 0 as n and Δ n ( x ) is bounded by 1 for all 𝑛 and all 𝑥. Hence, by (conditional) bounded convergence theorem, for each 𝑥 in 𝑆, we have f n ( x ) 0 . Then note that Δ n ( x ) Δ n + 1 ( x ) for all 𝑛 and all 𝑥, so { Δ n } is monotonic. Applying the conditional expectation operator E [ | X = x ] to Δ n ( S T ) , we have that { f n } is monotonic. Finally, by Assumption 1, 𝑆 is compact, and by Assumption 2, { f n } are continuous. Hence, by Dini’s theorem, the convergence of f n to 0 is uniform on 𝑆. ∎

Theorem A.8

Suppose Assumptions 13 are satisfied. There exists a sequence of functions i = 1 n b i ( S T K i ) + such that

E [ ( V T i = 1 n b i ( S T K i ) + ) 2 | X = x ] 0

for all 𝑥 in 𝑆 uniformly as n .

Proof

For any ϵ > 0 , by Theorem A.6, there exists a function i = 1 N p i 1 { S T q i } such that

E [ ( V T i = 1 N p i 1 { S T q i } ) 2 | X = x ] < ϵ 4

for all 𝑥 in 𝑆.

By Lemma A.7, for each I { S T q i } , i = 1 , , N , we may choose a (small) constant r i > 0 such that

E [ ( 1 { S T q i } ( S t q i + r i ) + ( S t q i ) + r i ) 2 | X = x ] < ϵ 4 N ( i = 1 N p i 2 )

for all 𝑥 in 𝑆.

Putting everything together, we have

E [ ( V T i = 1 N p i ( ( S t q i + r i ) + ( S t q i ) + r i ) ) 2 | X = x ] = E [ ( V T i = 1 N p i 1 { S T q i } + i = 1 N p i ( 1 { S T q i } ( S t q i + r i ) + ( S t q i ) + r i ) ) 2 | X = x ] 2 E [ ( V T i = 1 N p i 1 { S T q i } ) 2 | X = x ] + 2 E [ ( i = 1 N p i ( 1 { S T q i } ( S t q i + r i ) + ( S t q i ) + r i ) ) 2 | X = x ] 2 E [ ( V T i = 1 N p i 1 { S T q i } ) 2 | X = x ] + 2 ( i = 1 N p i 2 ) ( i = 1 N E [ ( 1 { S T q i } ( S t q i + r i ) + ( S t q i ) + r i ) 2 | X = x ] ) < 2 ( ϵ 4 ) + 2 ( i = 1 N p i 2 ) ( i = 1 N ϵ 4 N ( i = 1 N p i 2 ) ) = ϵ

for all 𝑥 in 𝑆, where the first inequality is due to the inequality ( a + b ) 2 2 ( a 2 + b 2 ) for any real numbers 𝑎 and 𝑏 and the second inequality is due to Cauchy–Schwarz inequality. This proves the uniform convergence.

Finally, we may write

i = 1 N p i ( ( S t q i + r i ) + ( S t q i ) + r i )

in the form of

i = 1 N b i ( S T K i ) +

by suitably defining N and the constants b i , K i , i = 1 , 2 , , N . ∎

Proof of Theorem 4.1

For any ϵ > 0 , by Theorem A.8, there exists a function i = 1 N p i 1 { S T q i } such that

E [ ( V T i = 1 N b i ( S T K i ) + ) 2 | X = x ] < ϵ

for all 𝑥 in 𝑆.

By Cauchy–Schwarz inequality, we have

| E [ V T i = 1 N b i ( S T K i ) + | X = x ] | E [ | V T i = 1 N b i ( S T K i ) + | | X = x ] E [ ( V T i = 1 N b i ( S T K i ) + ) 2 | X = x ] < ϵ

for all 𝑥 in 𝑆.

Therefore, the conditional variance

Var ( V T i = 1 N b i ( S T K i ) + | X = x ) = E [ ( V T i = 1 N b i ( S T K i ) + ) 2 | X = x ] ( E [ V T i = 1 N b i ( S T K i ) + | X = x ] ) 2 < ϵ + ϵ = 2 ϵ

for all 𝑥 in 𝑆. This proves the uniform convergence. ∎

Proof of Theorem 4.2

Denote r ( x ) = ρ ( V T , i = 1 n b i ( S T K i ) + | X = x ) . For all 𝑥 in 𝑆, we have

Var ( V T i = 1 n b i ( S T K i ) + | X = x ) = Var ( V T | X = x ) 2 r ( x ) Var ( V T | X = x ) Var ( i = 1 n b i ( S T K i ) + | X = x ) + Var ( i = 1 n b i ( S T K i ) + | X = x ) 2 ( 1 r ( x ) ) Var ( V T | X = x ) Var ( i = 1 n b i ( S T K i ) + | X = x )

where the last inequality is due to the AM-GM inequality a + b 2 > a b for any non-negative real number 𝑎 and 𝑏.

By Theorem 4.1, we have, for all 𝑥 in 𝑆,

Var ( V T i = 1 n b i ( S T K i ) + | X = x ) 0 uniformly .

Therefore, r ( x ) 1 for all 𝑥 in 𝑆 uniformly. ∎

References

[1] C. M. Bishop, Pattern Recognition and Machine Learning, Inform. Sci. Statist., Springer, New York, 2006. Search in Google Scholar

[2] F. Black and M. Scholes, The pricing of options and corporate liabilities, J. Polit. Econ. 81 (1973), no. 3, 637–654. 10.1086/260062Search in Google Scholar

[3] P. Glasserman, Monte Carlo Methods in Financial Engineering, Appl. Math. (New York) 53, Springer, New York, 2003. 10.1007/978-0-387-21617-1Search in Google Scholar

[4] S. L. Heston, A closed-form solution for options with stochastic volatility with applications to bond and currency options, Rev. Financ. Stud. 6 (1993), no. 2, 327–343. 10.1093/rfs/6.2.327Search in Google Scholar

[5] J. Kienitz, GMM DCKE - Semi-analytic conditional expectations, preprint (2021), https://ssrn.com/abstract=3902490. 10.2139/ssrn.3902490Search in Google Scholar

[6] J. Kienitz, N. Nowaczyk and N. Geng, Dynamically controlled kernel estimation, preprint (2021), https://ssrn.com/abstract=3829701. 10.2139/ssrn.3829701Search in Google Scholar

[7] B. Lapeyre and J. Lelong, Neural network regression for Bermudan option pricing, Monte Carlo Methods Appl. 27 (2021), no. 3, 227–247. 10.1515/mcma-2021-2091Search in Google Scholar

[8] F. A. Longstaff and E. S. Schwartz, Valuing American options by simulation: A simple least-squares approach, Rev. Financ. Stud. 14 (2001), no. 1, 113–147. 10.1093/rfs/14.1.113Search in Google Scholar

[9] V. M. R. Muggeo, Estimating regression models with unknown break-points, Stat. Med. 22 (2003), no. 19, 3055–3071. 10.1002/sim.1545Search in Google Scholar PubMed

[10] S. E. Shreve, Stochastic Calculus for Finance. II, Springer Finance, Springer, New York, 2004. 10.1007/978-1-4757-4296-1Search in Google Scholar

Received: 2023-09-14
Revised: 2025-04-17
Accepted: 2025-04-22
Published Online: 2025-05-22
Published in Print: 2025-09-01

© 2025 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 13.3.2026 from https://www.degruyterbrill.com/document/doi/10.1515/mcma-2025-2013/html
Scroll to top button