Brownian motion. 8. Stochastic integrals. 9. Black-Scholes and other models. 10.
The multidimensional stochastic calculus. 11. Stochastic differential equations.
Math 635: An Introduction to Brownian Motion and Stochastic Calculus 1. Introduction and review
11. Stochastic differential equations
2. Notions of convergence and results from measure theory
12. Markov property
3. Review of Markov chains
13. SDEs and partial differential equations 14. Change of measure and asset pricing
4. Change of measure 5. Information and conditional expectations 6. Martingales
15. Martingale representation completeness
and
16. Applications and examples
7. Brownian motion 8. Stochastic integrals
17. Stationary distributions and forward equations
9. Black-Scholes and other models
18. Processes with jumps
10. The multidimensional stochastic calculus
19. Assignments 20. Problems
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
1
February 22 review • Independence • Conditional expectations: Basic properties • Jensen’s inequality • Functions of known and unknown random variables • Filtrations and martingales • Optional sampling theorem • Doob’s inequalities
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
2
1.
Introduction and review • The basic concepts of probability: Models of experiments • Sample space and events • Probability measures • Random variables • The distribution of a random variable • Definition of the expectation • Properties of expectations • Jensen’s inequality
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
3
Experiments Probability models experiments in which repeated trials typically result in different outcomes. As a means of understanding the “real world,” probability identifies surprising regularities in highly irregular phenomena. If we roll a die 100 times we anticipate that about a sixth of the time the roll is 5. If that doesn’t happen, we suspect that something is wrong with the die or the way it was rolled.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
4
Probabilities of events Events are statements about the outcome of the experiment: {the roll is 6}, {the rat died}, {the television set is defective} The anticipated regularity is that P (A) ≈
#times A occurs #of trials
This presumption is called the relative frequency interpretation of probability.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
5
“Definition” of probability The probability of an event A should be #times A occurs in first n trials n→∞ n
P (A) = lim
The mathematical problem: Make sense out of this. The real world relationship: Probabilities are predictions about the future.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
6
Random variables In performing an experiment numerical measurements or observations are made. Call these random variables since they vary randomly. Give the quantity a name: X {X = a} and {a < X < b} are statements about the outcome of the experiment, that is, are events
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
7
The distribution of a random variable If Xk is the value of X observed on the kth trial, then we should have #{k ≤ n : Xk = a} n→∞ n
P {X = a} = lim
If X has only finitely many possible values, then X P {X = a} = 1. a∈R(X)
This collection of probabilities determine the distribution of X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
8
Distribution function More generally, n
1X 1(−∞,x] (Xk ) P {X ≤ x} = lim n→∞ n k=1
FX (x) ≡ P {X ≤ x} is the distribution function for X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
9
The law of averages If R(X) = {a1 , . . . , am } is finite, then m
m
l=1
l=1
X #{k ≤ n : Xk = al } X X1 + · · · + Xn = lim al = al P {X = al } lim n→∞ n→∞ n n
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
10
More generally, if R(X) ⊂ [c, d], −∞ < c < d < ∞, then X
xl P {xl < X ≤ xl+1 } =
l
lim
m X
n→∞
l=1
xl
#{k ≤ n : xl < Xk ≤ xl+1 } n
X1 + · · · + Xn ≤ lim n→∞ n m X #{k ≤ n : xl < Xk ≤ xl+1 } ≤ lim xl+1 n→∞ n X l=1 = xl+1 P {xl < X ≤ xl+1 } =
l X
xl+1 (FX (xl+1 ) − FX (xl ))
l
Z →
d
xdFX (x) c
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
11
The expectation as a Stieltjes integral If R(X) ⊂ [c, d], define Z E[X] =
d
xdFX (x). c
If the relative frequency interpretation is valid, then X1 + · · · + Xn = E[X]. n→∞ n lim
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
12
A random variable without an expectation Example 1.1 Suppose P {X ≤ x} =
x , 1+x
x ≥ 0.
Then m X X1 + · · · + Xn ≥ lim lP {l < X ≤ l + 1} n→∞ n
= =
l=0 m X
l=0 m X l=0
l(
l l+1 − ) l+2 l+1
l → ∞ as m → ∞ (l + 2)(l + 1)
One could say E[X] = ∞, and we will. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
13
Review of basic calculus Definition 1.2 A sequence {xn } ⊂ R converges to x ∈ R (limn→∞ xn = x) if and only if for each > 0 there exists an n > 0 such that n > n implies |xn − x| ≤ . Pn P Let {ak } ⊂ R. The series ∞ a converges if lim k n→∞ k=1 ak ∈ R k=1 P∞ exists.PThe series converges absolutely if k=1 |ak | converges. (Or we write ∞ k=1 |ak | < ∞.) Examples of things you should know: lim (axn + byn ) = a lim xn + b lim yn
n→∞
n→
n→∞
if the two limits on the right exist. If limn,m→∞ |xn − xm | = 0 ({xn } is a Cauchy sequence), then limn→∞ xn exists. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
14
Examples If α > 1, then
∞ X 1 < ∞. kα k=1
For α = 1,
∞ X 1 k=1
however ∞ X (−1)k k=1
k
= lim
n→∞
n X (−1)k k=1
k
k
= lim
m→∞
= ∞;
m X l=1
∞
X 1 1 1 (− + )=− 2l − 1 2l 2l(2l − 1) l=1
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
15
The sample space The possible outcomes of the experiment form a set Ω called the sample space. Each event (statement about the outcome) can be identified with the subset of the sample space for which the statement is true.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
16
The collection of events If A = {ω ∈ Ω : statement I is true for ω} B = {ω ∈ Ω : statement II is true for ω} Then A ∩ B = {ω ∈ Ω : statement I and statement II are true for ω} A ∪ B = {ω ∈ Ω : statement I or statement II is true for ω} Ac = {ω ∈ Ω : statement I is not true for ω} Let F be the collection of events. Then A, B ∈ F should imply that A ∩ B, A ∪ B, and Ac are all in F. F is an algebra of subsets of Ω. In fact, we assume that F is a σ-algebra (closed under countable unions and complements). •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
17
The probability measure Each event A ∈ F is assigned a probability P (A) ≥ 0. From the relative frequency interpretation, we must have P (A ∪ B) = P (A) + P (B) for disjoint events A and B and by induction, if A1 , . . . , Am are disjoint m X m P (∪k=1 Ak ) = P (Ak ) finite additivity k=1
In fact, we assume countable additivity: If A1 , A2 , . . . are disjoint events, then ∞ X ∞ P (∪k=1 Ak ) = P (Ak ). k=1
P (Ω) = 1. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
18
A probability space is a measure space A measure space (M, M, µ) consists of a set M , a σ-algebra of subsets M, and a nonnegative function µ defined on M that satisfies µ(∅) = 0 and countable additivity. A probability space is a measure space (Ω, F, P ) satisfying P (Ω) = 1.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
19
Random variables If X is a random variable, then we must know the value of X if we know that outcome ω ∈ Ω of the experiment. Consequently, X is a function defined on Ω. The statement {X ≤ c} must be an event, so {X ≤ c} = {ω : X(ω) ≤ c} ∈ F. In other words, X is a measurable function on (Ω, F, P ). R(X) will denote the range of X Ω}
R(X) = {x ∈ R : x = X(ω), ω ∈
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
20
Distributions Definition 1.3 The Borel subsets B(R) is the smallest σ-algebra of subsets of R containing (−∞, c] for all c ∈ R.
Definition 1.4 The distribution of a R-valued random variable X is the Borel measure defined by µX (B) = P {X ∈ B}, B ∈ B(R).
µX is called the measure induced by the function X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
21
Discrete distributions Definition 1.5 A random variable is discrete or has a discrete distribution if and only if R(X) is countable. If X is discrete, the distribution of X is determined by the probability mass function pX (x) = P {X = x}, x ∈ R(X). Note that
X
P {X = x} = 1.
x∈R(X)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
22
Examples Binomial distribution n k P {X = k} = p (1 − p)n−k , k
k = 0, 1, . . . , n
for some postive integer n and some 0 ≤ p ≤ 1 Poisson distribution P {X = k} = e−λ
λk , k!
k = 0, 1, . . .
for some λ > 0.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
23
Absolutely continuous distributions Definition 1.6 The distribution of X is absolutely continuous if and only if there exists a nonnegative function fX such that Z b P {a < X ≤ b} = fX (x)dx, a < b ∈ R. a
Then fX is the probability density function for X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
24
Examples Normal distribution 2 1 − (x−µ) 2 2σ e fX (x) = √ 2πσ
Exponential distribution fX (x) =
λe−λx 0
x≥0 x 0 lim P {|Xn − X| ≥ } = 0.
n→∞
Convergence in distribution: The sequence {Xn } converges to X is distribution if and only if for each x such that FX is continuous at x lim FXn (x) = FX (x).
n→∞
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
34
Examples Theorem 2.2 (The strong law of large numbers) Suppose ξ1 , ξ2 , . . . are independent and identically distributed with E[|ξi |] < ∞. Then ξ1 + · · · + ξn = E[ξ] a.s. n→∞ n lim
Theorem 2.3 (The weak law of large numbers) Suppose ξ1 , ξ2 , . . . are iid with E[ξi2 ] < ∞. Then for > 0, P {|
ξ1 + · · · + ξn V ar(ξ) → 0. − E[ξ]| ≥ } ≤ n n2
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
35
Theorem 2.4 (Central limit theorem) Let ξ1 , ξ2 , . . . be iid with E[ξi ] = µ and V ar(ξi ) = σ 2 < ∞. Then Pn Z x y2 ξ − nµ 1 i √ e− 2 dy lim P { i=1√ ≤ x} = n→∞ nσ 2π −∞
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
36
Relationship among notions of convergence Lemma 2.5 Almost sure convergence implies convergence in probability. Convergence in probability implies convergence in distribution. Proof. Suppose Xn → X a.s. Then for > 0, An ≡ {sup |Xk − X| ≥ } ⊃ {|Xn − X| ≥ } k≥n
A1 ⊃ A2 ⊃ · · ·, so limn→∞ P (An ) = P (∩∞ k=1 Ak ). ∩∞ k=1 Ak = {ω : lim supn→∞ |Xn −X| ≥ } and hence limn→∞ P (An ) = 0.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
37
Suppose Xn → X in probability. Then FXn (x) ≤ FX (x+)+P {Xn ≤ x, X > x+} ≤ FX (x+)+P {|Xn −X| ≥ }. Therefore, lim sup FXn (x) ≤ lim FX (x + ) = FX (x), n→∞
→0
and similarly lim inf FXn (x) ≥ lim FX (x − ) = FX (x−). n→∞
→0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
38
Limit theorems for expectations and integrals Theorem 2.6 (Bounded convergence theorem) Let {Xn } be a sequence of random variables such that Xn → X in distribution. Suppose that there exists C > 0 such that P {|Xn | > C} = 0 for all n. Then limn→∞ E[Xn ] = E[X]. Proof. Let a0 ≤ −C < a1 < · · · < C ≤ am be points of continuity for FX . Then m−1 m−1 X X al P {al < Xn ≤ al+1 } ≤ E[Xn ] ≤ al+1 P {al < Xn ≤ al+1 }. l=0
l=0
Then the left and right sides converge to the left and right sides of m−1 X l=0
al P {al < X ≤ al+1 } ≤ E[X] ≤
m−1 X
al+1 P {al < X ≤ al+1 }.
l=0
and lim supn→∞ |E[Xn ] − E[X]| ≤ maxl (al+1 − al ).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
39
Lemma 2.7 If X ≥ 0 a.s., then limK→∞ E[X ∧ K] = E[X]. Proof. If K >
m n,
then
E[X] ≥ E[X ∧ K] ≥
m−1 X k=1
k k + 1 m→∞ bnXc k P{ ≤ X < } → E[ ]. n n n n
More generally, Theorem 2.8 (Monotone convergence theorem) If 0 ≤ X1 ≤ X2 ≤ · · · a.s. and X = limn→∞ Xn a.s., then E[X] = lim E[Xn ] n→∞
allowing ∞ = ∞.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
40
Lemma 2.9 (Fatou’s lemma) If P {Xn ≥ 0} = 1 and Xn converges in distribution to X, then lim inf E[Xn ] ≥ E[X] n→∞
Proof. lim inf E[Xn ] ≥ lim E[Xn ∧ K] = E[X ∧ K] → E[X] n→∞
n→∞
Let P {Xn = n} = 1−P {Xn = 0} = n−1 . Then Xn → 0 in distribution, and E[Xn ] = 1 for all n.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
41
Theorem 2.10 (Dominated convergence theorem) Suppose there exist Y such that |Xn | ≤ Y a.s. with E[Y ] < ∞ and limn→∞ Xn = X a.s. Then limn→∞ E[Xn ] = E[X]. Proof. E[Y ] − lim sup E[Xn ] = lim inf (E[Y ] − E[Xn ]) n→∞
n→∞
= lim inf E[Y − Xn ] n→∞
≥ E[Y − X] = E[Y ] − E[X] E[Y ] + lim inf E[Xn ] = lim inf (E[Y ] + E[Xn ]) n→∞
n→∞
= lim inf E[Y + Xn ] n→∞
≥ E[Y + X] = E[Y ] + E[X] •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
42
Computation of expectations Recall µX (B) = P {X ∈ B}. g is Borel measurable if {x : g(x) ≤ c} ∈ B(R) for all c ∈ R. Note that (R, B(R), µX ) is a probability space and a Borel measurable function g is a “random variable” on this probability space. Theorem 2.11 If X is a random variable on (Ω, F, P ) and g is Borel measurable, then g(X) is a random variable and if E[|g(X)|] < ∞, Z g(x)µX (dx) E[g(X)] = R
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
43
Lebesgue measure Lebesgue measure is the unique measure L on B(R) such that for a < b, L((a, b)) = b − a. Integration with respect to L is denoted Z Z g(x)dx ≡ g(x)L(dx) R
If f (x) =
Pm
i=1 ai 1Ai ,
R
Ai ∈ B(R) with L(Ai ) < ∞, ai ∈ R, then Z m X f (x)dx ≡ ai L(Ai ). R
i=1
Such an f is called a simple function.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
44
General definition of Lebesgue integral For g ≥ 0, Z
Z g(x)dx ≡ sup{
R
If 0,
R
R |g(x)|dx
f (x)dx : f simple, 0 ≤ f ≤ g}. R
< ∞, then setting g + (x) = g(x)∨0 and g − (x) = (−g(x))∨ Z Z Z g(x)dx ≡ g + (x)dx − g − (x)dx. R
R
R
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
45
Riemann integrals Definition 2.12 Let −∞ < a < b < ∞. g defined on [a, b] is Riemann integrable if there exists I ∈ R so that for each > 0 there exists a δ > 0 such that a = t0 < t1 < · · · < tm = b, sk ∈ [tk , tk+1 ] and max |tk+1 − tk | ≤ δ implies m−1 X | g(sk )(tk+1 − tk ) − I| ≤ . k=0
Then I is denoted by Z
b
g(x)dx a
Pm−1
g(sk )(tk+1 − tk ) is called a Riemann sum. The Riemann integral is the limit of Riemann sums. k=0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
46
Some properties of Lebesgue integrals Theorem 2.13 If g is Riemann integrable, then g is Lebesgue integrable and the integrals agree. Fatou’s lemma, the monotone convergence theorem, and the dominated convergence theorem all hold for the Lebesgue integral.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
47
Distributions with Lebesgue densities Suppose Z µX (B) =
1B (x)fX (x)dx,
B ∈ B(R).
R
Then fX is a probability density function for X. Theorem 2.14 Let X be a random variable on (Ω, F, P ) with a probability density function fX . If g is Borel measurable and E[|g(X)|] < ∞, then Z g(x)fX (x)dx E[g(X)] = R
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
48
3.
Review of Markov chains • Elementary definition of conditional probability • Independence • Markov property • Transition matrix • Simulation of a Markov chain
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
49
Elementary definition of conditional probability For two events A, B ∈ F, the definition of the conditional probability of A given B (P (A|B)) is intended to capture how we would reassess the probability of A if we knew that B occurred. The relative frequency interpretation suggests that we only consider trials of the experiment on which B occurs. Consequently, # times that A and B occur # times B occurs # times that A and B occur/# trials = # times B occurs/# trials P (A ∩ B) ≈ P (B)
P (A|B) ≈
leading to the definition P (A|B) = P P(A∩B) (B) Note that for each fixed B ∈ F, P (·|B) is a probability measure on F. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
50
Independence If knowing that B occurs doesn’t change our assessment of the probability that A occurs (P (A|B) = P (A)), then we say A is independent of B. By the defintion of P (A|B), independence is equivalent to P (A ∩ B) = P (A)P (B). {Ak , k ∈ I} are mutually independent if P (Ak1 ∩ · · · ∩ Akm ) =
m Y
P (Aki )
i=1
for all choices of {k1 , . . . , km } ⊂ I.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
51
Independence of random variables Random variables {Xk , k ∈ I} are mutually independent if P {Xk1 ∈ B1 , . . . , Xkm ∈ Bm } =
m Y
P {Xki ∈ Bi }
i=1
for all choices of {k1 , . . . , km } ⊂ I and B1 , . . . , Bm ∈ B(R).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
52
Joint distributions The collection of Borel subsets of Rm (B(Rm )) is the smallest σ-algebra containing (−∞, c1 ] × · · · × (−∞, cm ]. The joint distribution of (X1 , . . . , Xm ) is the measure on B(Rm ) defined by µX1 ,...,Xm (B) = P {(X1 , . . . , Xm ) ∈ B},
B ∈ B(Rm ).
Lemma 3.1 The joint distribution of (X1 , . . . , Xm ) is uniquely determined by the joint cdf FX1 ,...,Xm (x1 , . . . , xm ) = P {X1 ≤ x1 , . . . , Xm ≤ xm }.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
53
The Markov property Let X0 , X1 , . . . be positive, integer-valued random variables. Then P {X0 = i0 , . . . , Xm = im } = P {Xm = im |X0 = i0 , . . . , Xm−1 = im−1 } ×P {Xm−1 = im−1 |X0 = i0 , . . . , Xm−2 = im−2 } × · · · × P {X1 = i1 |X0 = i0 }P {X0 = i0 } X satisfies the Markov property if for each m and all choices of {ik }, P {Xm = im |X0 = i0 , . . . , Xm−1 = im−1 } = P {Xm = im |Xm−1 = im−1 }
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
54
The transition matrix Suppose P {Xm = j|Xm−1 = i} = pij . (The transition probabilities do not depend on m.) The matrix P = ((pij )) satisfies X pij = 1. j
P {Xm+2 = j|Xm = i} X = P {Xm+2 = j, Xm+1 = k|Xm = i} = =
k X k X
P {Xm+2 = j|Xm+1 = k, Xm = i}P {Xm+1 = k|Xm = i} (2)
pik pkj ≡ pij
k
Then
(2)
((pij )) = P 2 . •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
55
The joint distribution of a Markov chain The joint distribution of X0 , X1 , . . . is determined by the transition matrix and the initial distribution, νi = P {X0 = i}. P {X0 = i0 , . . . , Xm = im } = νi0 pi0 i1 pi1 i2 · · · pim−1 im . In general, (n)
P {Xm+n = j|Xm = i} = pij , where
(n)
((pij )) = P n .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
56
Simulation of a Markov chain Let E be the state space (the set of values the chain can assume) of the Markov chain. For simplicity, assume E = {1, . . . , N } of {1, 2, . . .}. Define H : E × [0, 1] → E by j−1 X
H(i, u) = j,
pik ≤ u
Ye } (Y − Ye )] = 0 = E[1{Y 0, E[X(t)] P {sup X(s) ≥ c} ≤ c s≤t Proof. Let τc = inf{t : X(t) ≥ x}. Then E[X(t)] ≥ E[X(t ∧ τ )] ≥ cP {τ ≤ t}. (Actually, for right continuous processes, τc may not be a stopping time unless one first modifies the filtration {Ft }, but the modification can be done without affecting the statement of the theorem.)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
101
Doob’s inequalities Theorem 6.10 Let X be a nonnegative submartingale. Then for p > 1, p p p E[sup X(s) ] ≤ E[X(t)p ] p−1 s≤t Corollary 6.11 If M is a square integrable martingale, then E[sup M (s)2 ] ≤ 4E[M (t)2 ]. s≤t
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
102
7.
Brownian motion • Moment generating functions • Gaussian distributions • Characterization by mean and covariance • Conditions for independence • Conditional expectations • Central limit theorem • Construction of Brownian motion • Functional central limit theorem • The binomial model and geometric Brownian motion • Martingale properties of Brownian motion • L´evy’s characterization • Relationship to some partial differential equations
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
103
• First passage time • Quadratic variation • Nowhere differentiability • Law of the iterated logarithim • Modulus of continuity • Markov property • Transition density and operator • Strong Markov property • Reflection principle
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
104
Moment generating functions Definition 7.1 Let X be a Rd -valued random variable. Then the moment generating function for X (it it exists) is given by ϕX (λ) = E[eλ·X ],
λ ∈ Rd .
Lemma 7.2 If ϕX (λ) < ∞ for all λ ∈ Rd , then ϕX uniquely determines the distribution of X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
105
Gaussian distributions Definition 7.3 A R-valued random variable X is Gaussian (normal) with mean µ and variance σ 2 if fX (x) = √
2 1 − (x−µ) 2 2σ e 2πσ
The moment generating function for X is λX
E[e
σ 2 λ2 ] = exp{ + µλ} 2
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
106
Multidimensional Gaussian distributions Definition 7.4 A Rd -valued random variable X is Gaussian if λ · X is Gaussian for every λ ∈ Rd . Note that if E[X] = µ and cov(X) = E[(X − µ)(X − µ)T ] = Σ = ((σij )), then E[λ · X] = λ · µ,
var(λ · X) = λT Σλ.
Note that is X = (X1 , . . . , Xd ) is Gaussian in Rd , then for any choice of i1 , . . . , im ∈ {1, . . . , d}, (Xi1 , . . . Xim ) is Gaussian in Rm . Lemma 7.5 If X is Gaussian in Rd and C is a m × d matrix, then Y = CX
(7.1)
is Gaussian in Rm . •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
107
Characterization by mean and covariance If X is Gaussian in Rd , then λT Σλ ϕX (λ) = exp{ + λ · µ} 2 Since the distribution of X is determined by its moment generating function, the distribution of a Gausian random vector is determined by its mean and covariance matrix. Note that for Y given by (7.1), E[Y ] = Cµ and cov(Y ) = CΣC T .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
108
Conditions for independence Lemma 7.6 If X = (X1 , X2 ) ∈ R2 is Gaussian, then X1 and X2 are independent if and only if cov(X1 , X2 ) = 0. In general, if X is Gaussian in Rd , then the components of X are independent if and only if Σ is diagonal. Proof. If R-valued random variables U and V are independent, then cov(U, V ) = 0, so if the components of X are independent then Σ is diagonal. If Σ is diagonal, then d Y
λ2i σi2 ϕX (λ) = exp{ + λi µ}, 2 i=1 and since the moment generating function of X determines the joint distribution, the components of X must be independent. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
109
Conditional expectations Lemma 7.7 Let X = (X1 , . . . , Xd ) be Gaussian, then E[Xd |X1 , . . . , Xd−1 ] = a +
d−1 X
bi Xi ,
i=1
where σdk =
d−1 X
bi σik ,
k = 1, . . . , d − 1
i=1
and µd = a +
d−1 X
b i µi .
i=1
Proof. The bi are selected so that Xd − (a + of X1 , . . . , Xd−1 .
Pd−1 i=1
bi Xi ) is independent
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
110
Example Suppose X1 , X2 , and X3 are independent, Gaussian, mean zero and variance one. Then X1 + X2 . E[X1 |X1 + X2 ] = 2 Let e2 = X1 + X2 − √1 X3 . e1 = X1 + X2 + √1 X3 , X X (7.2) 2 2 2 2 e1 and X e2 are independent, Gaussian, mean zero, and variance Then X one.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
111
Central limit theorem If {ξi } are iid with E[ξi ] = 0 and V ar(ξi ) = 1, then Zn = satisfies Z b 1 2 1 √ e− 2 x dx. P {a < Zn ≤ b} → 2π a If we define [nt] 1 X Zn (t) = √ ξi n i=1
√1 n
Pn
i=1 ξi
then the distribution of (Zn (t1 ), . . . , Zn (tm )) is approximately Gaussian with cov(Zn (ti ), Zn (tj )) → ti ∧ tj , Note that for t0 < t1 < · · · < tm , Zn (tk ) − Zn (tk−1 ), k = 1, . . . , m are independent.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
112
The CLT as a process
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
113
Interpolation lemma Lemma 7.8 Let Y and Z be independent, R-valued Gaussian random variables with E[Y ] = E[Z] = 0, V ar(Z) = a, and V ar(Y ) = 1. Then √ √ a a Z Z U= + Y, V = − Y 2 2 2 2 are independent Gaussian random variables satisfying Z = U + V . Proof. Clearly, Z = U +V . To complete the proof, check that cov(U, V ) = 0.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
114
Construction of Brownian motion We can construct standard Brownian motion by repeated application of the Interpolation Lemma 7.8. Let ξk,n be iid, Gaussian, E[ξk,n ] = 0 and V ar(ξk,n ) = 1. Define W (1) = ξ1,0 , and define W ( 2kn ) inductively by k−1 W ( k+1 k−1 1 k 2n ) − W ( 2n ) + (n+1)/2 ξk,n W( n) = W( n ) + 2 2 2 2 1 1 1 k−1 k+1 = W ( n ) + W ( n ) + (n+1)/2 ξk,n 2 2 2 2 2
for k odd and 0 < k < 2n .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
115
By the Interpolation Lemma, W(
k−1 1 k+1 1 k−1 1 k ) − W ( ) = W ( ) − W ( ) + ξk,n 2n 2n 2 2n 2 2n 2(n+1)/2
and k+1 k 1 k+1 1 k−1 1 ) − W ( ) = W ( ) − W ( ) − ξk,n 2n 2n 2 2n 2 2n 2(n+1)/2 are independent, Gaussian random variables. W(
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
116
Convergence to a continuous process Theorem 7.9 Let Γ be the set of diadic rationals in [0, 1]. Define Wn (t) = W (
k k n k+1 k ) + (t − )2 (W ( ) − W ( )), 2n 2n 2n 2n
k k+1 ≤ t ≤ . 2n 2n
Then sup |W (t) − Wn (t)| → 0 a.s., t∈Γ
and hence W extends to a continuous process on [0, 1].
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
117
Continuous functions on C[0, 1] Definition 7.10 Let C[0, 1] denote the continuous, real-valued functions on [0, 1]. g : C[0, 1] → R is continuous if xn , x ∈ C[0, 1] satisfying sup0≤t≤1 |xn (t) − x(t)| → 0 implies limn→∞ g(xn ) = g(x). For example, g(x) = sup x(t) 0≤t≤1
Z g(x) =
1
x(s)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
118
Functional central limit theorem Theorem 7.11 (Donsker invariance principle) Let {ξi } be iid with E[ξi ] = 0 and V ar(ξi ) = σ 2 . Define [nt]
k √ 1 X ξi + (t − ) nξk+1 , Zn (t) = √ n n i=1
k+1 k ≤t≤ . n n
Then for each continuous g : C[0, 1] → R, g(Zn ) converges in distribution to g(σW ).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
119
The binomial model and geometric Brownian motion In the binomial model for a stock price, at each time step the price either goes up a fixed percentage or down a fixed percentage. Assume the up-down probabilities are 12 − 21 and the time step is n1 . σ k+1 σ 1 k+1 k k ) = (1+ √ )Xn ( )} = P {Xn ( ) = (1− √ )Xn ( )} = . n n n n 2 n n P[nt] Then log Xn (t) = i=1 log(1+ √σn ξi ) where P {ξi = 1} = P {ξi = −1} = 1 2 . By Taylor’s formula P {Xn (
[nt]
[nt]
σ X 1 X 2 1 Xn (t) ≈ √ ξi − σ ≈ σW (t) − σ 2 t 2n i=1 2 n i=1 and
1 log Xn (t) ≈ exp{σW (t) − σ 2 t} 2 •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
120
Martingales Recall that an {Ft }-adapted process M is a {Ft }-martingale if and only if E[M (t + r)|Ft ] = M (t) for all t, r ≥ 0. Note that this requirement is equivalent to E[M (t + r) − M (t)|Ft ] = 0.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
121
Convergence of conditional expectations Lemma 7.12 Suppose Xn , X are R-valued random variables and E[|Xn − X|] → 0. Then for a sub-σ-algebra D, E[|E[Xn |D] − E[X|D]|] → 0 Proof. By Jensen’s inequality E[|E[Xn |D] − E[X|D]|] = E[|E[Xn − X|D]|] ≤ E[E[|Xn − X||D]] = E[|Xn − X|]
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
122
Martingale properties of Brownian motion Let Ft = FtW = σ(W (s) : s ≤ t). Since W has independent increments, we have E[W (t + r)|Ft ] = E[W (t + r) − W (t) + W (t)|Ft ] = E[W (t + r) − W (t)|Ft ] + W (t) = W (t), and W is a martingale. Similarly, let M (t) = W (t)2 − t. Then E[M (t + r) − M (t)|Ft ] = E[(W (t + r) − W (t))2 + 2(W (t + r) − W (t))W (t) − r|Ft ] = E[(W (t + r) − W (t))2 |Ft ] + 2W (t)E[W (t + r) − W (t)|Ft ] − r =0 and hence M is a martingale.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
123
An exponential martingale Let M (t) = exp{W (t) − 12 t}, then 1 E[M (t + r)|Ft ] = E[exp{W (t + r) − W (t) − r}M (t)|Ft ] 2 − 21 r = M (t)e E[exp{W (t + r) − W (t)}|Ft ] = M (t)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
124
A general family of martingales Let f ∈ Cb3 (R) (the bounded continuous functions with three bounded continuous derivatives), and t = t0 < · · · < tm = t + r. Then E[f (W (t + r)) − f (W (t))|Ft ] X = E[ f (W (ti+1 )) − f (W (ti ))|Ft ] X = E[ f (W (ti+1 )) − f (W (ti )) − f 0 (W (ti ))(W (ti+1 ) − W (ti )) 1 − f 00 (W (ti ))(W (ti+1 ) − W (ti ))2 |Ft ] 2 X1 +E[ f 00 (W (ti ))(ti+1 − ti )|Ft ] 2 = Z1 + Z2
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
125
Note that since Fti ⊃ Ft , E[f 0 (W (ti ))(W (ti+1 ) − W (ti ))|Ft ] = E[E[f 0 (W (ti ))(W (ti+1 ) − W (ti ))|Fti ]|Ft ] = E[f 0 (W (ti ))E[(W (ti+1 ) − W (ti ))|Fti ]|Ft ] =0 E[f 00 (W (ti ))(W (ti+1 ) − W (ti ))2 |Ft ] = E[E[f 00 (W (ti ))(W (ti+1 ) − W (ti ))2 |Fti ]|Ft ] = E[f 00 (W (ti ))E[(W (ti+1 ) − W (ti ))2 |Fti ]|Ft ] = E[f 00 (W (ti ))(ti+1 − ti )|Ft ] Z t+r X1 1 00 E[| f 00 (W (ti ))(ti+1 − ti ) − f (W (s))ds|] → 0 2 2 t X X b E[|Z1 |] ≤ CE[ |W (ti+1 ) − W (ti )|3 ] ≤ C (ti+1 − ti )3/2 → 0 Rt so f (W (t)) − f (0) − 0 21 f 00 (W (s))ds is a {Ft }-martingale. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
126
L´evy’s characterization Theorem 7.13 Suppose M and Z given by Z(t) = M (t)2 − t are continuous, {Ft }-martingales, and M (0) = 0. Then M is a standard Brownian motion. Lemma 7.14 Under the assumptions of the theorem, E[(M (t+r)−M (t))2 |Ft ] = E[M (t+r)2 −2M (t)M (t+r)+M (t)2 |Ft ] = r, if τ is a {Ft }-stopping time, then E[(M ((t + r) ∧ τ ) − M (t ∧ τ ))2 |Ft ] = E[(t + r) ∧ τ − t ∧ τ |Ft ] ≤ r, and for f ∈ Cb3 (R), Z Mf (t) = f (M (t)) − f (M (0)) − 0
t
1 00 f (M (s))ds 2
(7.3)
is a {Ft }-martingale. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
127
Proof. Taking f (x) = eiθx , iθM (t)
e
iθM (0)
−e
t
Z + 0
1 2 iθM (s) θ e ds 2
is a martingale, and hence iθM (t+r)
E[e
iθM (t)
|Ft ] = e
Z − t
t+r
1 2 θ E[eiθM (s) |Ft ]ds 2
and iθ(M (t+r)−M (t))
ϕt,θ (r) ≡ E[e
Z
r
|Ft ] = 1 − 0
Therefore
1 2 θ ϕt,θ (s)ds 2
1 2
E[eiθ(M (t+r)−M (t)) |Ft ] = e− 2 θ r . It follows that (M (t + r) − M (t)) is Gaussian and independent of Ft .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
128
Proof.[of lemma] To prove that (7.3) is a martingale, the problem is to show X E[ f (M (ti+1 )) − f (M (ti )) − f 0 (M (ti ))(M (ti+1 ) − M (ti )) 1 − f 00 (M (ti ))(M (ti+1 ) − M (ti ))2 |Ft ] 2 converges to zero. Let , δ > 0, and assume that ti+1 − ti ≤ δ. Define τ,δ = inf{t : sup |M (t) − M (s)| ≥ } t−δ≤s≤t
Then X X 3 |M (ti+1 ∧ τ,δ ) − M (ti ∧ τ,δ )| ≤ (M (ti+1 ∧ τ,δ ) − M (ti ∧ τ,δ ))2 , and the expectation of the term on the left is bounded by t. Select n → 0 and δn → 0 so that τn ,δn → ∞ a.s.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
129
Applications of the optional sampling theorem Let −a < 0 < b, and define γa,b = inf{t : W (t) ∈ / (−a, b). Then E[W (γa,b ∧ t)] = 0 and E[γa,b ∧ t] = E[W (γa,b ∧ t)2 ] ≤ a2 ∨ b2 . Let t → ∞. E[γa,b ] < ∞, so 0 = E[W (γa,b )] = −aP {W (γa,b ) = −a} + bP {W (γa,b ) = b} = −a + (a + b)P {W (γa,b ) = b} Consequently, P {W (γa,b ) = b} =
a a+b
and
E[γa,b ] = E[W (γa,b )2 ] = a2
b a + b2 a+b a+b
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
130
Relationship to some partial differential equations Let W = (W1 , W2 ) where W1 and W2 are independent, R-valued standard Brownian motions. For f ∈ Cb3 (R2 ), Z t 1 ∆f (W (s))ds Mf (t) = f (W (t)) − f (0) − 0 2 is a martingale. Let Xx (t) = x + W (t) (standard Brownian motion starting at x). Then Z t 1 Mfx (t) = f (Xx (t)) − f (x) − ∆f (Xx (s))ds 2 0 is a martingale. Let D ⊂ R2 be bounded, open, and have a smooth boundary. Suppose ∆h(y) = 0, t ∈ D, and h(y) = ϕ(y), y ∈ ∂D (the / D}. Then boundary of D). Define τDx = inf{t : Xx (t) ∈ h(x) = E[ϕ(Xx (τDx ))] •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
131
Finiteness of the exit time Lemma 7.15 For ρ > 0, let γρ = inf{t : |W (t)| ≥ ρ}. Then E[γρ ] =
ρ2 2.
Proof. Let f (x) = |x|2 for |x| ≤ ρ. then ∆f (x) = 4 for |x| ≤ ρ, and the optional sampling theorem implies E[|W (γρ ∧ t)|2 ] = E[2(γρ ∧ t)]. Letting t → ∞ gives the result.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
132
Laplace transforms for nonnegative random variables For a nonnegative random variable X, define the Laplace transform of its distribution by Z e−αx µX (dx), α ≥ 0. FX (α) = E[e−αX ] = R
Lemma 7.16 The distribution of a nonnegative random variable is uniquely determined by its Laplace transform.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
133
First passage time For c > 0, let τc = inf{t : W (t) ≥ c}. Since 1 Ma (t) = exp{aW (t) − a2 t} 2 The optional sampling theorem implies 1 1 1 = E[exp{aW (t ∧ τc ) − a2 t ∧ τc }] → E[exp{ac − a2 τc }] 2 2 since τc < ∞ a.s. Consequently, 1 2
E[e− 2 a τc ] = e−ac and
√
Fτc (α) = E[e−ατc ] = e−
2αc
.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
134
Quadratic variation For s < t, E[(W (t) − W (s))2 ] = t − s and 2 2 E[ (W (t) − W (s))2 − (t − s) ] = (t − s)2 E[ W (1)2 − 1 ]. Consequently, for 0 = t0 < · · · < tm = t, !2 m m X 2 X 2 2 (ti − ti−1 )2 (W (ti ) − W (ti−1 )) − t ] = E[ W (1) − 1 ] E[ i=1
i=1
which converges to zero as max(ti − ti−1 ) → 0. We say that [W ]t = t.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
135
Covariation For independent standard Brownian motions W1 , W2 , a similar calculation implies [W1 , W2 ]t = lim
m X
(W1 (ti ) − W1 (ti−1 ))(W2 (ti ) − W2 (ti−1 )) = 0
i=1
where the convergence is in probability.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
136
Nowhere differentiability Rt Rt Suppose that f (t) = 0 g(s)ds, where 0 |g(s)|ds < ∞ for all t > 0. Then Z t m X 2 (f (ti ) − f (ti−1 )) ≤ max |f (ti ) − f (ti−1 )| |g(s)|ds → 0 i=1
i
0
as max(ti − ti−1 ) → 0. Definition 7.17 A function f : [0, ∞) → R is nowhere differentiable if its derivative does not exist at any point t ∈ [0, ∞). Let ΓN D denote the collection of nowhere differentiable funtions. Theorem 7.18 P {W ∈ ΓN D } = 1.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
137
Law of the iterated logarithim lim sup √ t→∞
W (t) = 1 a.s. 2t log log t
c (t) = tW (1/t) is Brownian motion. V ar(W c (t)) = t2 1 = t Therefore W t W (1/t)
lim sup p t→0
2t−1 log log 1/t
c (t) W
= lim sup p t→0
2t log log 1/t
= 1 a.s.
Consequently, W (t + h) − W (t) lim sup p = 1 a.s. 2h log log 1/h h→0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
138
Modulus of continuity Definition 7.19 A function f : [a, b] → R is H¨older continuous with exponent 0 ≤ α ≤ 1, if |f (t) − f (s)| < ∞. |t − s|α a≤s τ on {τ < ∞}, and [2n t] [2n t] {τn ≤ t} = {τn ≤ n } = {τ < n } ∈ Ft . 2 2 •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
142
Conditioning up to discrete stopping times Lemma 7.24 Let E[|Z|] < ∞, and let τ be a discrete {Ft }-stopping time (R(τ ) = {t1 , t2 , . . .} ∪ {∞}). Then X E[Z|Fτ ] = E[Z|Ftk ]1{τ =tk } + Z1{τ =∞} . k
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
143
Strong Markov property Theorem 7.25 Let τ be a {FtX }-stopping time with P {τ < ∞} = 1. Then E[f (X(τ + t))|Fτ ] = T (t)f (X(τ )).
(7.5)
More generally, define Wτ by Wτ (t) = W (τ + t) − W (τ ). Then Wτ is a standard Brownian motion and Wτ is independent of Fτ .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
144
Proof. Prove first for discrete stopping times and take limits. Let τn be as above. Then X E[f (X(τn + t))|Fτn ] = E[f (X(τn + t))|Fk2−n ]1{τn =k2−n } = =
k X k X
E[f (X(k2−n + t))|Fk2−n ]1{τn =k2−n } T (t)f (X(k2−n ))1{τn =k2−n }
k
= T (t)f (X(τn )). Assume that f is continuous so that T (t)f is continuous. Then E[f (X(τn + t))|Fτ ] = E[T (t)f (X(τn ))|Fτ ] and passing to the limit gives (7.5).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
145
Lemma 7.26 If γ ≥ 0 is Fτ -measurable, then E[f (X(τ + γ))|Fτ ] = T (γ)f (X(τ )). Proof. First, assume that γ is discrete. Then X E[f (X(τ + γ))|Fτ ] = E[f (X(τ + γ))|Fτ ]1{γ=r} r∈R(γ)
=
X
E[f (X(τ + r))|Fτ ]1{γ=r}
r∈R(γ)
=
X
T (r)f (X(τ ))1{γ=r}
r∈R(γ)
= T (γ)f (X(τ )). Assuming that f is continuous, general γ can be approximated by discrete γ.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
146
Reflection principle Lemma 7.27 P {sup W (s) ≥ c} = 2P {W (t) > c} s≤t
Proof. Let τ = t ∧ inf{s : W (s) ≥ c}, and γ = (t − τ ). Then setting f = 1(c,∞) , 1 E[f (W (t))|Fτ ] = E[f (W (τ + γ))|Fτ ] = T (γ)f (W (τ )) = 1{τ c}. Since {τ < t} ∪ {W (t) = c} = {sups≤t W (s) ≥ c} and P {W (t) = c} = 0, P {sups≤t W (s) ≥ c} = P {τ < t}. Corollary 7.28 The cdf of τc = inf{t : W (t) ≥ c} is Z ∞ x2 2 e− 2t dx. Fτc (t) = 2P {W (t) > c} = √ 2πt c •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
147
Robert Brown’s original paper http://sciweb.nybg.org/science2/pdfs/dws/Brownian.pdf
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
148
8.
Stochastic integrals • Definitions • Martingale properties • Continuous integrators and cadlag integrands • Stochastic integrals as integrators • Interated integrals • Quadratic variation of a stochastic integral ˆ formula • Ito’s • Solution of a stochastic differential equation
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
149
Notation for limits through partitions Let H({ti }) denote a functional of a partition 0 = t0 < t1 < · · · satisfying limm→∞ tm = ∞. For example X f (ti )(g(T ∧ ti+1 ) − g(T ∧ ti )) H({ti }) = i
The assertion that lim H({ti }) = H0
{ti }→∗
(8.1)
will mean that for each > 0 there exists a δ > 0 such that maxi (ti+1 − ti ) < δ implies |H({ti }) − H0 | < . If H({ti }) and H0 are random variables, then the limit in (8.1) exists in probability if and only if for each > 0, there exists a δ > 0 such that maxi (ti+1 − ti ) < δ implies P {|H({ti }) − H0 | > } ≤ .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
150
Definitions Definition 8.1 A Brownian motion W is a {Ft }-Brownian motion if W is {Ft }-adapted and W (t + s) − W (t) is independent of Ft for all s, t ≥ 0. Note that FtW ⊂ Ft and that any {Ft }-Brownian motion is a {FtW }Brownian motion. In what follows, we will assume that there is a fixed filtration {Ft } and that W is a {Ft }-Brownian motion. All adapted processes are adapted to this filtration.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
151
Simple integrands Let 0 = r0 < r1 < · · · < rm and rm+1 = ∞. X(t) =
m X
ξi 1[ri ,ri+1 ) .
(8.2)
i=0
Then X is a simple process. Lemma 8.2 The simple process X given by (8.2) is {Ft }-adapted if and only if for each i, ξi is Fri -measurable. Definition 8.3 Let X be an adapted, simple process. Then Z t m X I(t) = X(s)dW (s) ≡ ξi (W (t ∧ ri+1 ) − W (t ∧ ri )). 0
i=0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
152
Martingale properties Lemma 8.4 Let X and I be as above. Suppose that for each i, E[|ξi |] < ∞. Then I is a martingale. Proof. Suppose s > t. Then, if ri ≥ t, E[ξi (W (s ∧ ri+1 ) − W (s ∧ ri ))|Ft ] = 0 = ξi (W (t ∧ ri+1 ) − W (t ∧ ri )), and if ri < t, E[ξi (W (s ∧ ri+1 ) − W (s ∧ ri ))|Ft ] = ξi E[(W (s ∧ ri+1 ) − W (s ∧ ri ))|Ft ] = ξi (W (t ∧ ri+1 ) − W (t ∧ ri )).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
153
Itoˆ isometry Theorem 8.5 Let X and I be as above. Suppose that for each i, E[ξi2 ] < ∞. Then Z t
2
X 2 (s)ds]
E[I (t)] = E[ 0
Proof. The martingale property of W implies that for i 6= j, E[ξi (W (t ∧ ri+1 ) − W (t ∧ ri ))ξj (W (t ∧ rj+1 ) − W (t ∧ rj ))] = 0. Consequently, 2
E[I (t)] = =
m X i=0 m X i=0
E[ξi2 (W (ri+1 ∧ t) − W (ri ∧ t))2 ] Z t E[ξi2 (ri+1 ∧ t − ri ∧ t))] = E[ X 2 (s)ds] 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
154
An approximation lemma Rt Lemma 8.6 Let X be adapted and E[ 0 X 2 (s)ds] < ∞ for each t > 0. Then there exists a sequence {Xn } of adapted, simple processes such that Z t lim E[ (X(s) − Xn (s))2 ds] = 0. (8.3) n→∞
0
Note that (8.3) implies Z t lim E[ (Xm (s) − Xn (s))2 ds] ≤ n,m→∞
0
Z t lim (2E[ (X(s) − Xn (s))2 ds] n,m→0 Z0 t +2E[ (X(s) − Xm (s))2 ds]) = 0. 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
155
Extension of the definition of the integral By Doob’s inequality and the martingale properties of the stochastic integral Z t Z t Z t 2 E[sup | Xn dW − Xm dW | ] = E[sup | (Xn − Xm )dW |2 ] t≤T
0
0
t≤T
Z
0 T
≤ 4E[| (Xn − Xm )dW |2 ] Z 0T = 4E[ (Xm (s) − Xn (s))2 ds] 0
Consequently, there exists stochastic process I such that R t a continuous 2 limn→∞ E[supt≤T |I(t) − 0 Xn dW | ] = 0. (Note Rthat the proof of this fact requires additional machinery.) We define XdW = I.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
156
Properties of the stochastic integral Theorem R t 2 8.7 Suppose that X1R, X2 are adapted and that for t > 0, E[ 0 Xi (s)ds] < ∞. Let Ii = Xi dW . a) Ii is adapted. Rt b) For each t ≥ 0, E[Ii2 (t)] = E[ 0 Xi2 (s)ds]. c) Ii is a martingale. R d) For a, b ∈ R, (aX1 + bX2 )dW = aI1 + bI2 .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
157
L2 convergence of stochastic integrals Rt Lemma 8.8 Let {Xn }, X be adapted processes such that E[ 0 Xn2 (s)ds] < Rt ∞ for each n and t > 0 and for each t > 0 limn→∞ E[ 0 |Xn (s)−X(s)|2 ds] = 0. Then for each T > 0, Z t Z t lim E[sup | Xn dW − XdW |2 ] = 0. n→∞
t≤T
0
0
Proof. By Doob’s inequality and the isometry, Z t Z t Z T E[sup | Xn dW − XdW |2 ] ≤ 4E[ |Xn (s) − X(s)|2 ds] t≤T
0
0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
158
Continuous integrators and cadlag integrands Let Y be a continuous, adapted process and X be a cadlag, adapted process. We define Z t X X(s−)dY (s) = lim X(ti ∧ t)(Y (ti+1 ∧ t) − Y (ti ∧ t)), (8.4) 0
{ti }→∗
if the limit on the right exists in probability for each t. Lemma 8.9 Suppose Y = W . If X is cadlag, adapted, and bounded by a constant, the two definitions of the stochastic integral are consistent.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
159
R Proof. Note that the sum on the right of (8.4) is X {ti } dW for X {ti } X (t) = X(ti )1(ti ,ti+1 ] (t) and that
Z
t
|X {ti } (s) − X(s−)|2 ds = 0.
lim
{ti }→∗
0
By the bounded convergence theorem, Z t lim E[ |X {ti } (s) − X(s−)|2 ds] = 0, {ti }→∗
0
and the lemma follows by Lemma 8.8.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
160
Elimination of the moment condition Theorem R t 2 8.10 Suppose that X is cadlag and adapted and that for each t > 0, 0 X (s)ds < ∞ a.s. Then Z t X X(s−)dW (s) ≡ lim X(ti ∧ t)(W (ti+1 ∧ t) − W (ti ∧ t)), (8.5) {ti }→∗
0
in probability. Proof. Let τc = inf{t : |X(t)|∨|X(t−)| ≥ c}. Then X τc (t) = X(t)1[0,τc ) (t) Rt is cadlag and adapted and satisfies E[ 0 |X τc (s)|2 ds] ≤ c2 t. Consequently, by Lemma 8.9, X lim X(ti ∧ t)1[0,τc ) (ti ∧ t)(W (ti+1 ∧ t) − W (ti ∧ t)) {ti }→∗
exists in probability. Since limc→∞ τc = ∞ a.s., (8.5) holds.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
161
Convergence of stochastic integrals Theorem 8.11 Suppose that Xn , X are cadlag and adapted and that for each t > 0, Z t lim |Xn (s) − X(s)|2 ds = 0 n→∞
0
in probability. Then Z lim sup |
n→∞ t≤T
t
Z Xn (s)dW (s) −
0
t
X(s)dW (s)| = 0 0
in probability.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
162
Rt Rt Proof. Let γcn = inf{t : 0 Xn2 (s)ds ≥ c} and γc = inf{t : 0 X 2 (s)ds ≥ c}. Then for each t > 0, Z t lim E[ |1[0,γcn ) (s)Xn (s) − 1[0,γc ) X(s)|2 ] = 0 n→∞
0
and hence Z lim E[sup |
n→∞
t≤T
0
t
Z 1[0,γcn ) (s)Xn (s)dW (s) −
t
1[0,γc ) X(s)dW (s)|2 ] = 0
0
and Z lim E[1{γcn ∧γc >T } sup |
n→∞
t≤T
t
Z Xn (s)dW (s) −
0
t
X(s)dW (s)|2 ] = 0.
0
For > 0 and c sufficiently large, lim supn→∞ P {γcn ∧ γc ≤ T } ≤ and the lemma follows.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
163
Stochastic integrals as integrators Theorem 8.12 Suppose that X and Z are cadlag and adapted and Y (t) = Rt 0 Z(s−)dW (s). Then Z t X X(s−)dY (s) ≡ lim X(ti )(Y (ti+1 ∧ t) − Y (ti ∧ t)) {ti }→∗ 0 Z t Z t = X(s−)Z(s−)dW (s) = X(s)Z(s)dW (s). 0
0
Rt More generally, if U is cadlag and adapted and Y (t) = 0 Z(s−)dW (s) + Rt 0 U (s)ds, then Z t Z t Z t X(s)U (s)ds. X(s−)dY (s) = X(s)Z(s)dW (s) + 0
0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
164
Proof. First check the result P for X simple. Then, for X cadlag and {ti } adapted, define X (t) = X(ti )1(ti ,ti+1 ] (t), so Z t X X(ti )(Y (ti+1 ∧ t) − Y (ti ∧ t)) = X {ti } (s−)dY (s) Z0 t = X {ti } (s−)Z(s−)dW (s). 0
Then, by Theorem 8.11, X lim X(ti )(Y (ti+1 ∧ t) − Y (ti ∧ t)) {ti }→∗ Z t X {ti } (s−)Z(s−)dW (s) = lim {ti }→∗ 0 Z t X(s−)Z(s−)dW (s). = 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
165
Interated integrals Corollary and adapted, and Rt R t 8.13 Suppose XR1 ,t X2 , U , and Z are cadlag Y (t) = 0 Z(s−)dW (s) + 0 U (s)ds and V (t) = 0 X1 (s−)dY (s). Then Z t X2 (s−)dV (s) 0 Z t = X2 (s−)X1 (s−)dY (s) Z0 t Z t = X2 (s−)X1 (s−)Z(s−)dW (s) + X2 (s−)X1 (s−)U (s)ds 0 0 Z t Z t X2 (s)X1 (s)Z(s)dW (s) + X2 (s)X1 (s)U (s)ds = 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
166
Quadratic variation as a stochastic integral Rt
X(s)dX(s) exists, and Z t 2 2 [X]t = X (t) − X (0) − 2 X(s)dX(s).
Lemma 8.14 [X]t exists if and only if
0
0
Proof. Noting that (b − a)2 = b2 − a2 − 2a(b − a), X (X(ti+1 ∧ t) − X(ti ∧ t))2 X = (X 2 (ti+1 ∧ t) − X 2 (ti ∧ t)) X −2 X(ti ∧ t)(X(ti+1 ∧ t) − X(ti ∧ t)) X 2 2 = X (t) − X (0) − 2 X(ti ∧ t)(X(ti+1 ∧ t) − X(ti ∧ t)) •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
167
Quadratic variation of a stochastic integral Rt Rt Lemma 8.15 If X and Z are adapted, 0 X 2 (s)ds < ∞ and 0 |Z(s)|ds < ∞ for each t > 0, and Z t Z t Y (t) = X(s)dW (s) + Z(s)ds, 0
then
0
Z [Y ]t =
t
X 2 (s)ds.
0
Proof. Check first for simple X and then extend by approximation.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
168
Ito’s ˆ formula Theorem 8.16 Let Z
t
Z X(s)dW (s) +
Y (t) = Y (0) + 0
t
Z(s)ds 0
and f ∈ C 2 (R). Then Z
t
Z
t
1 00 f (Y (s))d[Y ]s 0 0 2 Z t Z t 0 = f (Y (0)) + f (Y (s))X(s)dW (s) + f 0 (Y (s))Z(s)ds 0 Z0 t 1 00 + f (Y (s))X 2 (s)ds 0 2
f (Y (t)) = f (Y (0)) +
0
f (Y (s))dY (s) +
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
169
Proof. Assume that f ∈ Cb3 (R). Then X f (Y (t)) = f (Y (0)) + (f (Y (ti+1 )) − f (Y (ti ))) X = f (Y (0)) + f 0 (Y (ti ))(Y (ti+1 ) − Y (ti )) X1 + f 00 (Y (ti ))(Y (ti+1 ) − Y (ti ))2 2 X +O( |Y (ti+1 ) − Y (ti )|3 ) P P Note that if lim (Y (ti+1 ) − Y (ti )2 exists, lim |Y (ti+1 ) − Y (ti )|3 = 0. The second term on the right converges by Theorem 8.12 and the third by Problem 10.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
170
Ito’s ˆ formula take 2 Theorem 8.17 Let Z Y (t) = Y (0) +
t
Z X(s)dW (s) +
0
t
Z(s)ds 0
and f ∈ C 2,1 (R × R). Then Z
t
Z
t
f (Y (t), t) = f (Y (0), 0) + fx (Y (s), s)dY (s) + fs (Y (s), s)ds 0 0 Z t 1 + fxx (Y (s), s)d[Y ]s 0 2 Z t = f (Y (0), 0) + fx (Y (s), s)X(s)dW (s) Z t0 + (fx (Y (s), s)Z(s) + fs (Y (s), s))ds Z0 t 1 + fxx (Y (s), s)X 2 (s)ds 0 2 •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
171
Solution of a stochastic differential equation A solution of the stochastic differential equation dX = σXdW + µXdt is an adapted stochastic process satisfying Z t Z t X(t) = X(0) + σX(s)dW (s) + µX(s)ds. 0
(8.6)
0
ˆ formula suggests applying a substitution method similar to that Ito’s used to solve linear ordinary differential equaions, that is, look for a solution of the form X(t) = X(0) exp{aW (t) + bt}
(8.7)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
172
Identification of solution ˆ formula, Setting Y (t) = aW (t) + bt so that [Y ]t = a2 t, applying Ito’s we have Z t Z t 1 X(s)dY (s) + X(s)d[Y ]s X(t) = X(0) + 0 0 2 Z t Z t 1 = X(0) + aX(s)dW (s) + (b + a2 )X(s)ds 2 0 0 Consequently, setting a = σ and b = µ − 21 σ 2 , (8.7) satisfies (8.6), that is 1 X(t) = X(0) exp{σW (t) + (µ − σ 2 )t} 2
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
173
Vasicek interest rate model=Ornstein-Uhlenbeck process dR = (α − βR)dt + σdW or
Z
t
(α − βR(s))ds + σW (t)
R(t) = R(0) + 0
Employ an integrating factor setting Q(t) = eβt R(t). Then Z t Z t Q(t) = R(0) + αeβs ds + σ eβs dW (s) 0
0
Hence −βt
R(t) = e
α R(0) + (1 − e−βt ) + σe−βt β
Z
t
eβs dW (s)
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
174
Covariation Theorem 8.18 Let Z
t
Y1 (t) = Y1 (0) +
Z
t
X1 (s)dW (s) + 0
Z Y2 (t) = Y2 (0) +
U1 (s)ds 0
t
Z X2 (s)dW (s) +
0
t
U2 (s)ds. 0
Y1 and Y2 are called Itoˆ processes. Then X [Y1 , Y2 ]t ≡ lim (Y1 (ti+1 ∧ t) − Y1 (ti ∧ t))(Y2 (ti+1 ∧ t) − Y2 (ti ∧ t)) {ti }→∗ Z t Z t Y1 (s)dY2 (s) − Y2 (s)dY1 (s) = Y1 (t)Y2 (t) − Y1 (0)Y2 (0) − 0 0 Z t = X1 (s)X2 (s)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
175
Proof. X lim (Y1 (ti+1 ∧ t) − Y1 (ti ∧ t))(Y2 (ti+1 ∧ t) − Y2 (ti ∧ t)) {ti }→∗ X = (Y1 (ti+1 ∧ t)Y2 (ti+1 ∧ t) − Y1 (ti ∧ t)Y2 (ti ∧ t)) X − lim Y1 (ti ∧ t)(Y2 (ti+1 ∧ t) − Y2 (ti ∧ t)) {ti }→∗ X − lim Y2 (ti ∧ t)(Y1 (ti+1 ∧ t) − Y1 (ti ∧ t)) {ti }→∗ Z t Z t = Y1 (t)Y2 (t) − Y1 (0)Y2 (0) − Y1 (s)dY2 (s) − Y2 (s)dY1 (s)
[Y1 , Y2 ]t ≡
0
0
To verify the second equality, first assume X1 and X2 are simple. Then approximate.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
176
Integration by parts Corollary 8.19 Let Y1 and Y2 be Itˆo processes. Then Z t Z t Y1 (s)dY2 (s) = Y1 (t)Y2 (t) − Y1 (0)Y2 (0) − Y2 (s)dY1 (s) − [Y1 , Y2 ]t . 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
177
9.
Black-Scholes and other models • Derivation of the Black-Scholes model • A population model • Value of a portfolio • Hedging a contract • Pricing contracts using the Black-Scholes model
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
178
Derivation of the Black-Scholes model First consider a discrete time model of the value of a stock. At each time step, the value increases or decreases by a certain percentage, so the value becomes Yk+1 = Yk (1 + ξk+1 ) If the time steps are small, the percentage change is small suggesting 1 1 n Yk+1 = Ykn (1 + √ ξk+1 + α) n n where we assume the {ξk } are independent with E[ξk ] = 0
V ar(ξk ) = σ 2
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
179
n . Then Define Sn (t) = Y[nt] [nt]−1
log Sn (t) = log Sn (0) +
X k=0 [nt]−1
1 1 log(1 + √ ξk+1 + α) n n
[nt]−1 X 1 [nt] 1 X 1 1 √ ξk+1 + α− ( √ ξk+1 + α)2 ≈ log Sn (0) + n 2 n n n k=0
k=0
1 ≈ log S(0) + σW (t) + αt − σ 2 t 2 and
1 S(t) = S(0) exp{σW (t) + (α − σ 2 )t} 2 S is called geometric Brownian motion.
(9.1)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
180
An alternative derivation n Since Yk+1 = Ykn (1 +
√1 ξk+1 n
+ n1 α) = Ykn + Ykn ( √1n ξk+1 + n1 α), we have
[nt]−1 n Y[nt]
=
Y0n
+
X
n (Yk+1
[nt]−1
−
Ykn )
=
Y0n
+
k=0
X
1 Ykn √
n
k=0
[nt]−1
ξk+1 +
X
αYkn
k=0
1 n
P[nt]
An (t) = [nt] n , Z t Rt αSn (s−)dAn (s) Sn (t) = Sn (0) + 0 σSn (s−)dWn (s) +
n Setting Sn (t) = Y[nt] , Wn (t) =
1 √
σ n
k=1 ξk ,
0
suggesting a limit satisfying Z t Z t S(t) = S(0) + σS(s)dW (s) + αS(s)ds 0
(9.2)
0
We know that (9.1) satisfies (9.2), so this derivation looks plausible. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
181
A population model Consider the following discrete-time branching process. {ξik , i ≥ 1, k ≥ 0} iid nonnegative, integer-valued with E[ξik ] = 1 and V ar(ξik ) = σ 2 . ξik gives the number of offspring of the ith individual in the kth generation, so Zk+1 =
Zk X
ξik
= Zk +
i=1
Zk X
(ξik − 1).
i=1
Consider a sequence of branching processes with Z0n = n and define Xn (t) =
1 n Z . n [nt]
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
182
Derivation of limiting equation n
[nt]−1 Zk 1 X n 1 X X n Xn (t) = 1 + (Zk+1 − Zk ) = 1 + (ξik − 1) n n i=1 [nt]−1
k=0
k=0
Define Wn (t) = Claim: Wn and Wn (t)2 −
[nt] X
Zn
k X 1 (ξik − 1) √ p n σ n Zk i=1
k=1 [nt] n are
Z
martingales, and t
σ
Xn (t) = 1 +
p
Xn (s−)dWn (s)
0
suggesting that Xn converges to a solution of Z t p X(t) = 1 + σ X(s)dW (s) 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
183
Value of a portfolio Suppose that a market consists of a stock with price given by Z t Z t S(t) = S(0) + σS(s)dW (s) + αS(s)ds 0
0
and a money market account paying continuously compounded interest at rate r. ∆(t) denotes the number of shares of stock held by an investor at time t with the remainder of the investors wealth, X(t) − ∆(t)S(t) in the money market. The portfolio value satisfies Z t Z t X(t) = X(0) + ∆(s)dS(s) + r(X(s) − ∆(s)S(s))ds 0 0 Z t Z t (rX(s) + (α − r)∆(s)S(s))ds = X(0) + σ∆(s)S(s)dW (s) + 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
184
Hedging a contract Suppose a bank sells a contract that pays, at time T , an amount based on the the stock price at time T . That is, the bank pays h(S(T )). The contract can be hedged if there exists a portfolio whose value X at time T equals h(S(T )). We want an initial value X(0) and an investment strategy ∆ such that X(T ) = h(S(T )).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
185
Derivation of the hedging strategy Suppose X(t) = q(t, S(t)). Then q(t, S(t)) Z
t
= q(0, S(0)) + qx (s, S(s))σS(s)dW (s) 0 Z t σ2 + (qx (s, S(s))αS(s) + qxx (s, S(s))S 2 (s) + qs (s, S(s)))ds 2 0 Comparing this identity to Z t Z t X(t) = X(0) + ∆(s)dS(s) + r(X(s) − ∆(s)S(s))ds 0 0 Z t Z t σ∆(s)S(s)dW (s) + (rX(s) + (α − r)∆(s)S(s))ds = X(0) + 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
186
A PDE for q Matching coefficients gives ∆(s) = qx (s, S(s)) σ2 qx (s, S(s))αS(s) + qxx (s, S(s))S 2 (s) + qs (s, S(s)) 2 = rq(s, S(s)) + (α − r)qx (s, S(s))S(s). and hence qx (s, S(s))rS(s) +
σ2 qxx (s, S(s))S 2 (s) + qs (s, S(s)) 2 = rq(s, S(s)).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
187
Pricing contracts using the Black-Scholes model We want q to satisfy the partial differential equation 1 qs (s, x) + rxqx (s, x) + σ 2 x2 qxx (s, x) = rq(s, x) 2 with terminal condition q(T, x) = h(x). Then the price of the contract should be X(0) = q(0, S(0)). Holding ∆(s) = qx (s, S(s)) shares of stock at each time s and the remainder in the money market, the value of the portfolio at time T is q(T, S(T )) = h(S(T )).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
188
10.
Multidimensional stochastic calculus
• Multidimensional Brownian motion • Itoˆ processes • Covariation • Covariation for Itoˆ processes • Multidimensional Itoˆ formula
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
189
Multidimensional Brownian motion Definition 10.1 W = (W1 , . . . , Wd ) is a d-dimensional standard {Ft }Brownian motion if each W1 , . . . , Wd are independent (1-dimensional) standard Brownian motions adapted to {Ft } and for each t ≥ 0, W (t+·)−W (t) is independent of Ft . We will view W as a column vector.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
190
Stochastic integration with respect to W Let Mm×d denote the collection m × d matrices. Let X be Mm×d valued, cadlag and adapted. Then Z t X X(s)dW (s) = lim X(ti )(W (ti+1 ∧ t) − W (ti ∧ t)). {ti }→∗
0
Note that Z(t) =
Rt 0
X(s)dW (s) ∈ Rm and that Zi (t) =
d Z X j=1
t
Xij (s)dWj (s).
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
191
Itoˆ processes Definition 10.2 If X is Mm×d -valued, cadlag and adapted and U is Rm valued, cadlag and adapted, then Z t Z t Z(t) = X(s)dW (s) + U (s)ds 0
0
is an Itoˆ process.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
192
Covariation Definition 10.3 The covariation of Y1 and Y2 is X [Y1 , Y2 ]t = lim (Y1 (ti+1 ∧ t) − Y1 (ti ∧ t))(Y2 (ti+1 ∧ t) − Y2 (ti ∧ t)) {ti }→∗
Lemma 10.4 Z [Y1 , Y2 ]t = Y1 (t)Y2 (t) − Y1 (0)Y2 (0) −
t
Z Y1 (s)dY2 (s) −
0
t
Y2 (s)dY1 (s) 0
if the two stochastic integrals exist. Lemma 10.5 The covariation is bilinear in the sense that [aY1 + bY2 , cZ1 + dZ2 ]t = ac[Y1 , Z1 ]t + ad[Y1 , Z2 ]t + bc[Y2 , Z1 ]t + bc[Y2 , Z2 ]t
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
193
Covariation for Itoˆ processes To compute the covariations for an Itoˆ process, it is enough to know the following: Lemma 10.6 If Z
t
Y1 (t) =
Z
t
X1 (s)dW1 (s) +
U1 (s)ds
0
Z
0 t
Y2 (t) =
Z
t
X2 (s)dW2 (s) + Z0 t
Y3 (t) =
U2 (s)ds Z0 t
X2 (s)dW1 (s) +
U2 (s)ds
0
0
then
Z [Y1 , Y2 ]t = 0
[Y1 , Y3 ]t =
t
X1 (s)X2 (s)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
194
Covariation for Itoˆ processes Lemma 10.7 If Y1 and Y2 are R-valued Itˆo processes with [Y1 , Y2 ]t = X1 and X2 are cadlag and adapted, and Z t Z t Z1 (t) = X1 (s)dY1 (s) Z2 (t) = X2 (s)dY2 (s), 0
Rt 0
R(s)ds,
0
then Z [Z1 , Z2 ]t =
t
Z X1 (s)X2 (s)d[Y1 , Y2 ]s =
0
t
X1 (s)X2 (s)R(s)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
195
Matrix notation for covariation Let Z(t) =
Rt
0 X(s)dW (s) +
Rt
[Zi , Zj ]t =
0
U (s)ds. Then
d Z X k=1
t
Xik (s)Xjk (s)ds,
0
or thinking of the covariation as defining a matrix [Z]t = (([Zi , Zj ]t )), Z t [Z]t = X(s)X(s)T ds 0
where M T denotes the transpose of the matrix M .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
196
Multidimensional Itoˆ formula Theorem 10.8 Let Z be an m-dimensional Itˆo process Z t Z t Z(t) = Z(0) + X(s)dW (s) + U (s)ds 0
0
and let f ∈ C 2 (Rm ). Define the Hessian matrix Hf (x) = ((∂i ∂j f (x) )). Then Z t Z 1X t T f (Z(t)) = f (Z(0)) + ∇f (Z(s)) dZ(s) + ∂i ∂j f (Z(s))d[Zi , Zj ]s 2 i,j 0 0 Z t Z t ∇f (Z(s))T X(s)dW (s) + ∇f (Z(s))T U (s)ds = f (Z(0)) + 0 0 Z t 1 + trace(X(s)T Hf (Z(s))X(s))ds 2 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
197
Examples t
Z
1 ∇f (W (s)) dW (s) + 2 T
f (W (t)) = f (0) + 0
Z
t
∆f (W (s))ds 0
where ∆ is the Laplacian ∆f (x) =
d X
∂k2 f (x)
k=1 k X
X(t) = X(0) exp{
k=1
d
1X 2 σk )t} σk Wk (t) + (µ − 2 k=1
satisfies the stochastic differential equation Z t Z t T µX(s)ds X(t) = X(0) + X(s)σ dW (s) + 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
198
Diffusion equations If X satisfies Z
t
Z σ(X(s))dW (s) +
X(t) = X(0) + 0
t
b(X(s))ds, 0
then Z
T
Z
t
∇f (X(s)) σ(X(s))dW (s) +
f (X(t)) = f (X(0)) + 0
where
t
Lf (X(s))ds 0
1X aij (x)∂i ∂j f (x) + b(x) · ∇f (x) Lf (x) = 2 i,j
and a(x) = σ(x)σ(x)T .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
199
11.
Stochastic differential equations
• Brownian bridge • Convergence of empirical distribution functions • Kolmogorov-Smirnov test • Conditioned Brownian motion • An SDE for Brownian bridge • Stochastic differential equations • Moment estimates • Gronwall inequality • Schwarz inequality • Lipschitz conditions for existence and uniqueness •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
200
Brownian bridge Recall, that if (Y1 , . . . , Yl , X1 , . . . , Xm ) is jointly Gaussian and Cov(Yi , Xj ) = 0 of i = 1, . . . , l and j = 1, . . . , m, then Y = (Y1 , . . . , Yl ) and X = (X1 , . . . , Xm ) are independent. Let W be a scalar, standard Brownian motion. By Problem 6, W (1) is independent of B defined by B(t) = W (t) − tW (1), 0 ≤ t ≤ 1. Note that Cov(B(t), B(s)) = s ∧ t − ts. Lemma 11.1 σ(B(t) : 0 ≤ t ≤ 1) is independent of σ(W (1)). Proof. Since E[B(t)W (1)] = 0, (B(t1 ), . . . , B(tm )) is independent of W (1) for all choices of t1 , . . . , tm ∈ [0, 1]. Independence of the σalgebras follows by a standard Math 831 argument.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
201
Convergence of empirical distribution functions Let ξ1 , ξ2 , . . . be iid uniform [0, 1] random variables, and define the empirical distribution function n
1X Fn (t) = 1(−∞,t] (ξi ). n i=1 E[Fn (t)] = Fξ (t) and Cov(Fn (t), Fn (s)) =
1 (t ∧ s − ts) n
√ Define Bn (t) = n(Fn (t) − Fξ (t)), 0 ≤ t ≤ 1, then (Bn (t1 ), . . . , Bn (tm )) converges in distribution to (B(t1 ), . . . , B(tm )), t1 , . . . , tm ∈ [0, 1].
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
202
A class of continuous functions Definition 11.2 Let D[0, 1] denote the cadlag, real-valued functions on u [0, 1]. Let CD be the collection of functions g : D[0, 1] → R such that if xn ∈ D[0, 1] and x ∈ C[0, 1] satisfy sup0≤t≤1 |xn (t) − x(t)| → 0, then limn→∞ g(xn ) = g(x). For example, g(x) = sup x(t) 0≤t≤1
Z g(x) =
1
x(s)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
203
Functional limit theorem u Theorem 11.3 For each g ∈ CD , g(Bn ) converges in distribution to g(B).
In particular, lim P { sup
n→∞
0≤t≤1
√
n(Fn (t) − Fξ (t)) ≤ x} = P { sup B(t) ≤ x} 0≤t≤1
and √ lim P { sup | n(Fn (t) − Fξ (t))| ≤ x} = P { sup |B(t)| ≤ x}
n→∞
0≤t≤1
0≤t≤1
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
204
More general empirical distributions Let X1 , X2 , . . . be iid with continuous distribution function FX . DeP n 1 fine FnX (x) = n i=1 1(−∞,x] (Xi ) and √ Dn (x) = n(FnX (x) − FX (x)). Since
n
FnX (x)
1X = 1(−∞,FX (x)] (FX (Xi )) a.s. n i=1
and FX (Xi ) is uniform [0, 1], Dn has the same distribution as Bn ◦ FX . In particular, supx Dn (x) (supx |Dn (x)|) has the same distribution as sup0≤t≤1 Bn (t) (sup0≤t≤1 |Bn (t)|).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
205
Kolmogorov-Smirnov test Corollary 11.4 lim P {sup
n→∞
z
√
n(FnX (z) − FX (z)) ≤ x} = P { sup B(t) ≤ x} 0≤t≤1
and √ lim P {sup | n(FnX (z) − FX (z))| ≤ x} = P { sup |B(t)| ≤ x}
n→∞
z
0≤t≤1
√ Consequently, supz | n(FnX (z) − FX (z))| provides a test statistic to test the hypothesis that FX is the cdf for {Xi }.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
206
Conditioned Brownian motion Note that E[g( sup W (t))|W (1)] = E[g( sup (B(t) + tW (1)))|W (1)], 0≤t≤1
0≤t≤1
and since W (1) is independent of B, if we define h(y) = E[g( sup (B(t) + ty))], 0≤t≤1
we have E[g( sup W (t))|W (1)] = h(W (1)) 0≤t≤1
The same analysis works for any function of W restricted to [0, 1], and we say that the conditional distribution of W given that W (1) = b is the distribution of Bb defined by Bb (t) = B(t) + tb.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
207
Linear transformations of W Lemma 11.5 Let K : {(t, u) : 0 ≤ u ≤ t} → ∞ satisfy ∞, for all t ≥ 0, and define Z t Z(t) = K(t, u)dW (u).
Rt 0
K(t, u)2 du
0 such that |σ(x)| + |b(x)| ≤ K(1 + |x|) and that E[|X(0)|m ] < ∞. ˆ formula, Proof. Applying Ito’s Z t |X(t)|m = |X(0)|m + m|X(s)|m−2 X(s)T σ(X(s))dW (s) Z t 0 1 + m|X(s)|m−2 (X(s) · b(X(s)) + tracea(X(s)))ds 2 0 Z t 1 + m(m − 2)|X(s)|m−4 X(s)T a(X(s))X(s)ds 2 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
213
b > 0 such that Let τc = inf{t : |X(t)| ≥ c}. Then there exists K Z t∧τc m m b + |X(s)|m )ds] E[|X(t ∧ τc )| ] ≤ E[|X(0)| ] + E[ K(1 0 Z t b b + E[ K|X(s ∧ τc )|m ds] ≤ E[|X(0)|m ] + Kt 0
which by Gronwall’s inequality and Fatou’s lemma implies b Kt . E[|X(t)|m ] ≤ lim inf E[|X(t ∧ τc )|m ] ≤ (E[|X(0)|m ] + Kt)e b
c→∞
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
214
Gronwall’s inequality Lemma 11.8 Suppose that A be continuous and nondecreasing, X is cadlag, and that Z t (11.2) 0 ≤ X(t) ≤ + X(s)dA(s) . 0
Then X(t) ≤ eA(t) .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
215
Proof. Iterating (11.2), we have Z t X(t) ≤ + X(s)dA(s) 0 Z tZ s ≤ + A(t) + X(u)dA(u)dA(s) 0 0 Z t Z tZ sZ ≤ + A(t) + A(s)dA(s) + 0
0
Checking that Z t 1 A(s)dA(s) = A(t)2 , 2 0 and in general Z t Z t1
Z
0
0
0
s
X(v)dA(v)dA(u)dA(s).
0
1 A(u)dA(u)dA(s) = A(t)3 6
tn−2
A(tn−1 )dA(tn−1 ) · · · dA(t1 ) =
... 0
Z tZ
0
u
0
we see that X(t) ≤ eA(t) .
1 A(t)n , n!
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
216
Schwarz inequality Lemma 11.9 Suppose E[X 2 ], E[Y 2 ] < ∞. Then p E[XY ] ≤ E[X 2 ]E[Y 2 ]. Proof. Note that 2ab ≤ z −2 a2 + z 2 b2 , so 2E[XY ] ≤ z −2 E[X 2 ] + z 2 E[Y 2 ]. Minimizing the right side with respect to z gives the desired inequality. The same argument gives E[X · Y ] ≤
p
E[|X|2 ]E[|Y |2 ]
for random vectors. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
217
Lipschitz conditions for existence and uniqueness Theorem 11.10 Suppose that |σ(x) − σ(y)| + |b(x) − b(y)| ≤ K|x − y|. Then there exists a unique solution of Z t Z t (11.3) X(t) = X(0) + σ(X(s))dW (s) + b(X(s))ds. 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
218
e are solutions. Then by Ito’s ˆ Proof.[of uniqueness] Suppose X and X formula Z t T e 2 = e e |X(t) − X(t)| 2(X(s) − X(s)) (σ(X(s)) − σ(X(s)))dW (s) 0 Z t e e + 2(X(s) − X(s)) · (b(X(s)) − b(X(s)))ds Z0 t T e e + trace(σ(X(s)) − σ(X(s)))(σ(X(s)) − σ(X(s))) d 0
e As in the moment estimate, let τc = inf{t : |X(t) − X(t)| ≥ c}. Then b such that there exists K Z t 2 e ∧ τc )| ] ≤ b e ∧ τc )|2 ]ds, E[|X(t ∧ τc ) − X(t KE[|X(s ∧ τc ) − X(s 0
e ∧ τc )|2 ] = 0. which implies E[|X(t ∧ τc ) − X(t
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
219
Proof.[of existence] Assume, in addition, that σ and b are bounded. For h > 0, define ηh (t) = h[h−1 t], and let Xh satisfy Z ηh (t) Z ηh (t) Xh (t) = X(0) + σ(Xh (s))dW (s) + b(Xh (s))ds. 0
0
Note that Xh is constant on the interval [kh, (k + 1)h), and that Xh ((k+1)h) = Xh (kh)+σ(Xh (kh))W ((k+1)h)−W (kh))+b(Xh (kh))h. (known as the Euler-Maruyama scheme in the SDE literature). Define Z
t
Dh (t) =
Z
t
σ(Xh (s))dW (s) + ηh (t)
b(Xh (s))ds, ηh (t)
so Z Xh (t) + Dh (t) = X(0) +
t
Z σ(Xh (s))dW (s) +
0
t
b(Xh (s))ds. 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
220
b such that Then there exists K Z
2
b E[|Xh (t) + Dh (t) − Xh0 (t) − Dh0 (t)| ] ≤ K
t
E[|Xh (s) − Xh0 (s)|2 ]ds,
0
so E[|Xh (t) − Xh0 (t)|2 ] ≤ −2E[(Xh (t) − Xh0 (t)) · (Dh (t) − Dh0 (t))] Z t b +K E[|Xh (s) − Xh0 (s)|2 ]ds 0 p p ≤ 2 E[|Xh (t) − Xh0 (t)|2 ] E[|Dh (t) − Dh0 (t)|2 ] Z t b +K E[|Xh (s) − Xh0 (s)|2 ]ds 0
and hence p p b E[|Xh (t)−Xh0 (t)|2 ] ≤ 2 E[|Xh (t) − Xh0 (t)|2 ] E[|Dh (t) − Dh0 (t)|2 ]eKt . Rt But E[|Dh (t)|2 ] ≤ 2(E[ ηh (t) tracea(Xh (s))ds] + h2 supx |b(x)|2 ]) → 0 and {Xh (t), h > 0} is Cauchy in L2 as h → 0. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
221
Uniform moment bounds Suppose 2xb(x) + σ 2 (x) ≤ K1 − ε|x|2 . (For example, consider the Rt equation X(t) = X(0) − 0 αX(s)ds + W (t).) Then εt
2
e |X(t)|
Z
t
eεs 2X(s)σ(X(s))dW (s) ≤ |X(0)| + Z t 0 Z t εs 2 + e [2X(s)b(X(s)) + σ (X(s))]ds + εeεs |X(s)|2 ds 0 0 Z t Z t ≤ |X(0)|2 + eεs 2X(s)σ(X(s))dW (s) + eεs K1 ds 0 Z0 t K 1 εt ≤ |X(0)|2 + eεs 2X(s)σ(X(s))dW (s) + (e − 1), 2 0 2
and hence eεt E[|X(t)|2 ] ≤ E[|X(0)|2 ] +
K1 εt [e − 1]. ε
Therefore, we have the uniform bound E[|X(t)|2 ]] ≤ e−εt E[|X(0)|2 ] +
K1 (1 − e−εt ). ε
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
222
Vector case In the vector case Z X(t) = X(0) +
t
Z σ(X(s))dW (s) +
0
2
|X(t)|
Z
2
t
t
b(X(s))ds, 0
T
= |X(0)| + 2X(s) σ(X(s))dW (s) Z t 0 T + (2X(s) · b(X(s)) + trace(σ(X(s))σ(X(s)) ))ds . 0
As in the univariate case, if we assume, T
2x · b(x) + trace(σ(x)σ(x) ) ≤ K1 − ε|x|2 , then E[|X(s)|2 ] is uniformly bounded.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
223
12.
Markov property
• Definition • Solution of SDE as a function of the initial position • Flow property • Markov property • Strong Markov property • Differentiability with respect to initial position
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
224
Definition Recall the definition of the Markov property: Definition 12.1 If X is a stochastic process adapted to a filtration {Ft }, then X is {Ft }-Markov (or Markov with respect to {Ft }) if and only if for s, t ≥ 0, E[f (X(t + s))|Ft ] = E[f (X(t + s))|X(t)].
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
225
Solution of SDE as a function of the initial position For t ≥ r, let Z
r+t
Z σ(Xr (s − r, x))dW (s) +
Xr (t, x) = x + r
r+t
b(Xr (s − r.x))ds. r
Define Wr (t) = W (r + t) − W (r). Then Z t Z t Xr (t, x) = x + σ(Xr (s, x))dWr (s) + b(Xr (s.x))ds. 0
(12.1)
(12.2)
0
Assuming that uniqueness holds, since the distribution of Wr does not depend on r, the distribution of Xr does not depend on r. In particular, for f ∈ B(Rm ), the bounded, Borel-measurable functions on Rm , T (t)f (x) ≡ E[f (Xr (t, x))] does not depend or r. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
226
Flow property Write X(t, x) = X0 (t, x). Theorem 12.2 Let σ and b be such that solutions of the SDE are unique for all choices of the initial condition (for example, if σ and b are Lipschitz). Then X(r + t, x) = Xr (t, X(r, x)). Proof. By subtraction Z
r+t
X(r + t, x) = X(r, x) +
Z
r+t
b(X(s, x))ds(12.3)
σ(X(s, x))dW (s) + Zr t
r
Z σ(X(r + s, x))dWr (s) +
= X(r, x) + 0
t
b(X(r + s, x))ds. 0
Comparing this equation to (12.2), the proof becomes “clear”, although there are a number of technicalities. For example, we need to know that Xr (t, x) is at least a measurable function of x. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
227
Corollary 12.3 Under the assumptions of the theorem, if X(0) is independent of W and X satisfies (11.1), then X(r + t) = X(r + t, X(0)) = Xr (t, X(r)).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
228
Markov property Theorem 12.4 Let σ and b be such that solutions of the SDE are unique for all choices of the initial condition (for example, if σ and b are Lipschitz). If X is a solution of (11.1), then X is Markov. Proof. By uniqueness, X(r) is independent of Wr and hence of Xr (t, x). Consequently, for f ∈ B(Rm ), E[f (X(r + t))|Fr ] = E[f (Xr (t, X(r))|Fr ] = T (t)f (X(r)).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
229
Strong Markov property Theorem 12.5 Let σ and b be such that solutions of the SDE are unique for all choices of the initial condition (for example, if σ and b are Lipschitz). If X is a solution of (11.1), then for each finite {Ft }-stopping time τ and each f ∈ B(Rm ), E[f (X(τ + t))|Fτ ] = T (t)f (X(τ )). Proof. By the strong Markov property for Brownian motion, Wτ defined by Wτ (t) = W (τ + t) − W (τ ) is a Brownian motion independent of X(τ ), and Xτ satisfying Z t Z t Xτ (t, x) = x + σ(Xτ (s, x))dWτ (s) + b(Xτ (s.x))ds. (12.4) 0
0
has the same distribution as Xr (t, x). As with the Markov property, E[f (X(τ + t))|Fτ ] = E[f (Xτ (t, X(τ ))|Fr ] = T (t)f (X(τ )). •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
230
Continuity with respect to the initial condition Theorem 12.6 Let k ≥ 2. Under Lipschitz conditions, there exists a constant Mk such that E[|X(t, x) − X(t, y)|k ] ≤ |x − y|k eMk t . Proof. Let Zx,y (t) = X(t, x) − X(t, y). As in the proof of uniqueness k
|Zx,y (t)|
Z
k
t
= |x − y| + k|Zx,y (s)|k−2 (Zx,y (s))T (σ(X(s, x)) − X(s, y))dW (s) Z t 0 + x|Zx,y (s)|k−2 (Zx,y (s) · (b(X(s, x)) − b(X(s, y)) 0
1 + trace(σ(X(s, x)) − σ(X(s, y))(σ(X(s, x) − X(s, y))T )ds 2 1 + 2
Z
t
k(k − 2)|Zx,y (s)|k−4 |σ(X(s, x)) − σ(X(s, y)))T Zx,y (s)|2 ds
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
231
Differentiability with respect to initial position Assume d = m = 1. X(t, x) − X(t, y) x−y Z t σ(X(s, x)) − σ(X(s, y)) = 1+ ∆x,y (s)dW (s) X(s, x) − X(s, y) 0 Z t b(X(s, x)) − b(X(s, y)) + ∆x,y (s)ds X(s, x) − X(s, y) 0
∆x,y (t) ≡
and Z ∂x X(t, x) = 1 +
t
σ 0 (X(s, x))∂x X(s, x)dW (s) 0 Z t + b0 (X(s, x))∂x X(s, x)ds 0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
232
13.
SDEs and partial differential equations
• Generator for a diffusion process • One dimensional exit distributions • Exit times • Dirichlet problems • Parabolic equations • Feynman-Kac formula • Pricing derivatives
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
233
Differential operators and diffusion processes Consider Z
t
Z σ(X(s))dW (s) +
X(t) = X(0) + 0
t
b(X(s))ds, 0
where X is Rd -valued, W is an m-dimensional standard Brownian motion, σ is a d × m matrix-valued function and b is an Rd -valued function. For a C 2 function f , d Z t X f (X(t)) = f (X(0)) + ∂i f (X(s))dX(s) i=1
0
Z t 1 X ∂i ∂j f (X(s))d[Xi , Xj ]s . + 2 0 1≤i,j≤d
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
234
Computation of covariation The covariation satisfies Z tX Z t σi,k (X(s))σj,k (X(s))ds = ai,j (X(s))ds, [Xi , Xj ]t = 0
0
k
where a = ((ai,j )) = σ · σ T , that is ai,j (x) =
P
k
σik (x)σkj (x).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
235
Definition of the generator Let Lf (x) =
d X
bi (x)∂i f (x) +
i=1
1X ai,j (x)∂i ∂j f (x), 2 i,j
then Z
t T
Z
∇f (X(s))σ(X(s))dW (s) +
f (X(t)) = f (X(0)) + 0
t
Lf (X(s))ds . 0
Since a = σ · σT , X
ξi ξj ai,j = ξ T σσ T ξ = |σ T ξ|2 ≥ 0,
a is nonnegative definite, and L is an elliptic differential operator. L is called the generator for the corresponding diffusion process.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
236
The generator for Brownian motion If X(t) = X(0) + W (t), then ((ai,j (x))) = I, and Lf (x) = 21 ∆f (x).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
237
Exit distributions in one dimension If d = m = 1 and a(x) = σ 2 (x), then 1 Lf (x) = a(x)f 00 (x) + b(x)f 0 (x) 2 Suppose Lf (x) = 0 Then Z t f (X(t)) = f (X(0)) + f 0 (X(s))σ(X(s))dW (s). 0
Fix α < β, and define τ = inf{t : X(t) ∈ / (α, β)}. If supα0} ds) Xn (t) = Xn (0) + √ Ya (nλn t) − √ Yd (nµn n n 0 Z t √ √ + n(λn − µn )t + nµn 1{Xn (s)=0} ds 0
If λn → λ and µn → µ, then √ 1 Wan (t) ≡ √ Yea (nλn t) ⇒ λW1 (t) n 1 √ Wdn (t) ≡ √ Yed (nµn t) ⇒ µW2 (t)), n where W1 and W2 are standard Brownian motions.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
322
Scaled equation Defining √ n(λn − µn ) Z t Bn (t) ≡ 1{Xn (s)>0} ds cn ≡
0
Λn (t) ≡
√
nµn (t − Bn (t)) =
√
Z nµn
t
1{Xn (s)=0} ds, 0
we can rewrite Xn (t) as Xn (t) = Xn (0) + Wan (t) − Wdn (Bn (t)) + cn t + Λn (t). Λn is nondecreasing and increases only when Xn is zero.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
323
Skorohod problem Lemma 15.11 For w ∈ DR [0, ∞) with w(0) ≥ 0, there exists a unique pair (x, λ) satisfying x(t) = w(t) + λ(t) (15.1) such that λ(0) = 0, x(t) ≥ 0∀t, and λ is nondecreasing and increases only when x = 0. The solution is given by setting λ(t) = 0 ∨ sups≤t (−w(s)) and defining x by (15.1).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
324
Proof.[of uniqueness] For t < τ0 = inf{s : w(s) ≤ 0}, λ(t) = 0 and hence x(t) = w(t). For t ≥ τ0 , the nonnegativity of x implies λ(t) ≥ −w(t), and λ(t) nondecreasing implies λ(t) ≥ sup(−w(s)). s≤t
If t is a point of increase of λ, then x(t) = 0, so we must have λ(t) = −w(t) ≤ sup(−w(s)).
(15.2)
s≤t
Since the right side of (15.2) is nondecreasing, we must have λ(t) ≤ sups≤t (−w(s)) for all t > τ0 , and the result follows.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
325
Convergence to reflecting Brownian motion Λn (t) = 0 ∨ (− inf (Xn (0) + Wan (s) − Wdn (Bn (s)) + cn s)) s≤t
Consequently, if Xn (0) + Wan (t) − Wdn (Bn (t)) + cn t converges, so does Λn and Xn along with it. Assuming that cn → c, the limit will satisfy √ √ X(t) = X(0) + λW1 (t) − λW2 (t) + ct + Λ(t) √ Λ(t) = 0 ∨ sup(−(X(0) + λ(W1 (s) − W2 (s)) + ct)). s≤t
Defining
√1 (W1 2
− W2 ), √
X(t) = X(0) + X(t) ≥ 0 ∀t,
2λW (t) + ct + Λ(t)
where Λ is nondecreasing and Λ increases only when X(t) = 0. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
326
16.
Stationary distributions and forward equations
• Forward equations • Stationary distributions • Reflecting diffusions
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
327
Equations for probability distributions If
Z
t
Z
t
b(X(s))ds,
(16.1)
a = σσ T , f ∈ Cc2 (Rd ), and 1X aij (x)∂i ∂j f (s) + b(x) · ∇f (x), Lf (x) = 2 Z t f (X(t)) − Lf (X(s)) ds
(16.2)
X(t) = X(0) +
σ(X(s))dW (s) + 0
0
0
is a martingale.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
328
Forward equation Since Z
t
E [f (X(t))] = E [f (X(0)] + E Lf (X(s))ds 0 Z t = E [f (X(0))] + E [Lf (X(s))] ds, 0
defining νt (Γ) = P {X(t) ∈ Γ}, for f ∈ Cc2 (Rd ), Z Z Z tZ f dνt = f dν0 + Lf dνs ds,
(16.3)
0
which is a weak form of the equation d νt = L∗ νt . dt
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
329
Uniqueness for the forward equation Theorem 16.1 Let Lf be given by (16.2) with a and b continuous, and let {νt } be probability measures on Rd satisfying (16.3) for all f ∈ Cc2 (Rd ). If (16.1) has a unique solution for each initial condition, then P {X(0) ∈ ·} = ν0 implies P {X(t) ∈ ·} = νt .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
330
The adjoint operator In nice situations, νt (dx) = pt (x)dx. Then L∗ should be a differential operator satisfying Z Z pLf dx = f L∗ pdx. Rd
Rd
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
331
Example 16.2 Let d=1. Integrating by parts, we have Z ∞ 1 p(x) a(x)f 00 (x) + b(x)f 0 (x) dx 2 −∞ ∞ Z ∞ 1 1 d = p(x)a(x)f 0 (x) − f 0 (x) (a(x)p(x)) − b(x)p(x) dx. 2 2 dx −∞ −∞ The first term is zero, and integrating by parts again we have Z ∞ d 1 d (a(x)p(x)) − b(x)p(x) dx f (x) dx 2 dx −∞ so
d L p= dx ∗
1 d (a(x)p(x)) − b(x)p(x) . 2 dx
Example 16.3 Let Lf = 12 f 00 (Brownian motion). Then L∗ p = 12 p00 , that is, L is self adjoint.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
332
Stationary distributions Suppose
R
Lf dπ = 0 for all f in Cc2 (Rd ). Then Z Z Z tZ f dπ = f dπ + Lf dπds, 0
and hence νt ≡ π gives a solution of (16.3). Under the R t conditions of Theorem 16.1, if P {X(0) ∈ ·} = π and f (X(t)) − 0 Lf (X(s))ds is a martingale for all f ∈ Cc2 (Rd ), then P {X(t)} ∈ ·} = π, i.e. π is a stationary distribution for X.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
333
Computation of a stationary distribution if d = 1 Assuming π(dx) = π(x)dx d 1 d (a(x)π(x)) − b(x)π(x) = 0, dx 2 dx {z } | this is a constant: let the constant be 0 so 1 d (a(x)π(x)) = b(x)π(x). 2 dx Rx Applying the integrating factor exp(− 0 2b(z)/a(z)dz) Rx 1 − R0x 2b(z) dz d − 0 2b(z) a(z) a(z) dz π(x) = 0 e (a(x)π(x)) − b(x)e 2 dx −
a(x)e
Rx 0
π(x) =
2b(z) a(z) dz
π(x) = C
C R0x 2b(z) e a(z) dz . a(x)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
334
Existence of a stationary distribution Lemma 16.4 Assume a(x) > 0 for all x. If Z ∞ 1 R0x 2b(z) e a(z) dz dx < ∞, C= −∞ a(x) then
R x 2b(z) 1 π(x) = e 0 a(z) dz Ca(x) is a stationary density for (16.1).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
335
Diffusion with a boundary Suppose Z X(t) = X(0) +
t
Z σ(X(s))dW (s) +
0
t
b(X(s))ds + Λ(t) 0
with X(t) ≥ 0, and that Λ is nondecreasing and increasing only when X(t) = 0. Then Z t Z t T f (X(t)) = f (X(0)) + ∇f (X(s))σ(X(s))dW (s) + Lf (X(s))ds 0 0 Z t + f 0 (X(s))dΛ(s), 0
so
Z f (X(t)) −
t
Lf (X(s))ds 0
is a martingale, if f ∈ Cc2 and f 0 (0) = 0. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
336
Adjoint operator for reflecting process Z 0
∞
∞ Z ∞ 1 1 d 0 0 p(x)Lf (x)dx = f p(x)a(x)f (x) − (a(x)p(x)) − b(x)p(x) dx 2 2 dx 0 0 {z } | =0 ∞ Z ∞ 1 d (a(x)p(x)) − b(x)p(x) + f (x)L∗ p(x)dx = −f (x) 2 dx 0 0
and hence d L∗ p(x) = dx
1 d (a(x)p(x)) − b(x)p(x) 2 dx
for p satisfying
1 0 1 a (0) − b(0) p(0) + a(0)p0 (0) = 0 . 2 2
(16.4)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
337
Forward equation The density for the distribution of the process should satisfy d p t = L∗ p t dt and the boundary condition (16.4). The stationary density satisfies d 1 d (a(x)π(x)) − b(x)π(x) = 0 dx 2 dx and the boundary condition implies 1 d (a(x)π(x)) − b(x)π(x) = 0 2 dx giving π(x) =
c R0x 2b(z) e a(z) dz , x ≥ 0. a(x)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
338
Stationary distribution for reflecting Brownian motion Example 16.5 Let X(t) = X(0) + σW (t) − bt + Λ(t), where a = σ 2 and b > 0 are constant. Then 2b 2b π(x) = 2 e− σ2 x , σ so the stationary distribution is exponential.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
339
17.
Processes with jumps
• Poisson process • Martingale properties of the Poisson process • Strong Markov property for the Poisson process • Marked and compound Poisson processes • Stochastic equations with jumps • General counting processes • Intensities
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
340
Poisson process A Poisson process is a model for a series of random observations occurring in time. For example, the process could model the arrivals of customers in a bank, the arrivals of telephone calls at a switch, or the counts registered by radiation detection equipment. x x
x
x
x
x
x
x
t Let N (t) denote the number of observations by time t. In the figure above, N (t) = 6. Note that for t < s, N (s) − N (t) is the number of observations in the time interval (t, s].
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
341
Basic assumptions 1) Observations occur one at a time. 2) Numbers of observations in disjoint time intervals are independent random variables, i.e., if t0 < t1 < · · · < tm , then N (tk ) − N (tk−1 ), k = 1, . . . , m are independent random variables. 3) The distribution of N (t + a) − N (t) does not depend on t.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
342
Characterization of a Poisson process Theorem 17.1 Under assumptions 1), 2), and 3), there is a constant λ > 0 such that, for t < s, N (s) − N (t) is Poisson distributed with parameter λ(s − t), that is, (λ(s − t))k −λ(s−t) e . P {N (s) − N (t) = k} = k! Proof. Let Nn (t) be the number of time intervals ( nk , k+1 n ], k = 0, . . . , [nt] that contain at least one observation. Then Nn (t) is binomially distributed with parameters n and pn = P {N ( n1 ) > 0}. Then (1 − pn )n = P {Nn (1) = 0} = P {N (1) = 0} and npn → λ ≡ − log P {N (1) = 0}, and the rest follows by standard Poisson approximation of the binomial.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
343
Interarrival times Let Sk be the time of the kth observation. Then P {Sk ≤ t} = P {N (t) ≥ k} = 1 −
k−1 X (λt)i i=0
i!
e−λt ,
t ≥ 0.
Differentiating to obtain the probability density function gives 1 k−1 −λt e t≥0 (k−1)! λ(λt) fSk (t) = 0 t 1, Tk = Sk −Sk−1 . Then T1 , T2 , . . . are independent and exponentially distributed with parameter λ.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
344
Martingale properties of the Poisson process Theorem 17.3 (Watanabe) If N is a Poisson process with parameter λ, then N (t) − λt is a martingale. Conversely, if N is a counting process and N (t) − λt is a martingale, then N is a Poisson process.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
345
Proof. E[eiθ(N (t+r)−N (t)) |Ft ] n−1 X =1+ E[(eiθ(N (sk+1 )−N (sk ) − 1 − (eiθ − 1)(N (sk+1 ) − N (sk ))eiθ(N (sk )−N (t)) |Ft ] k=0 n−1 X
+
λ(sk+1 − sk )(eiθ − 1)E[eiθ(N (sk )−N (t)) |Ft ]
k=0
The first term converges to zero by the dominated convergence theorem, so we have Z r iθ(N (t+r)−N (t)) iθ E[e |Ft ] = 1 + λ(e − 1) E[eiθ(N (t+s)−N (t)) |Ft ]ds 0 iθ
and E[eiθ(N (t+r)−N (t)) |Ft ] = eλ(e
−1)t
.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
346
Strong Markov property A Poisson process N is compatible with a filtration {Ft }, if N is {Ft }adapted and N (t + ·) − N (t) is independent of Ft for every t ≥ 0. Lemma 17.4 Let N be a Poisson process with parameter λ > 0 that is compatible with {Ft }, and let τ be a {Ft }-stopping time such that τ < ∞ a.s. Define Nτ (t) = N (τ + t) − N (τ ). Then Nτ is a Poisson process that is independent of Fτ and compatible with {Fτ +t }.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
347
Proof. Let M (t) = N (t) − λt. By the optional sampling theorem, E[M ((τ + t + r) ∧ T )|Fτ +t ] = M ((τ + t) ∧ T ), so E[N ((τ +t+r)∧T )−N ((τ +t)∧T )|Fτ +t ] = λ((τ +t+r)∧T −(τ +t)∧T ). By the monotone convergence theorem E[N (τ + t + r) − N (τ + t)|Fτ +t ] = λr which gives the lemma.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
348
Marked Poisson processes Let N be a Poisson process with parameter λ and associate with each jump time Sk a mark ξk in a space E, called the mark space. Assume that the {ξk } are iid and independent of N .
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
349
An independence lemma Lemma 17.5 Let N be a Poisson process and {ξk } be independent and identically distributed and independent of N . Then Ft = σ(N (s), ξN (s) : s ≤ t) is independent of G t = σ(N (s) − N (t), ξN (t)+k : s ≥ t, k = 1, 2, . . .). Proof. The proof is left as an exercise. The lemma may look obvious, but it in general, it is false if the assumption that the ξk are identically distributed is dropped.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
350
Compound Poisson processes PN (t)
Suppose E = R and let Q(t) = Poisson process.
k=1
ξk . Then Q is a compound
Lemma 17.6 Suppose E[ξk ] = β. Then Q has independent increments, E[Q(t)] = βλt, and M (t) = Q(t) − βλt is a martingale. P Proof. Since E[ nk=1 ξk ] = nβ, E[Q(t)|N (t)] = N (t)β and E[Q(t)] = E[E[Q(t)|N (t)]] = λtβ. N (t)+N (t+s)−N (t)
E[Q(t + s) − Q(t)|Ft ] = E[
X
ξk |Ft ] = λsβ
k=N (t)+1
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
351
A stochastic equation for a jump process Theorem 17.7 Let X(0) be independent of the marked Poisson process, let H : Rd × E → Rd , and let X satisfy Z N (t) X X(t) = X(0)+ H(X(s−), ξk ) = X(0) +
H(X(s−), z)N (ds × dz)
[0,t]×E
k=1
Then X is a Markov process and Z t Z f (X(t)) − λ (f (X(s) + H(X(s), z)) − f (X(s)))µξ (dz)ds 0
E
is a martingale for each f ∈ B(Rd ).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
352
Proof. Since N (t+s)
X(t + s) = X(t) +
X
H(X(s−), ξk )
k=N (t)+1
is a function of X(t), N (t + s) − N (t), and (ξN (t)+1 , ξN (t)+2 , . . .), the Markov property follows by Lemma 17.5. Noting that E[f (X(t + h)) − f (X(t))|Ft ] = E[1{N (t+h)−N (t)=1} (f (X(t) + H(X(t), ξN (t)+1 )) − f (X(t))|Ft ] + o(h) Z −λh = λhe (f (X(t) + H(X(t), z)) − f (X(t))µξ (dz) + o(h) E
X E[f (X(t + s) − f (X(t))|Ft ] = E[f (X(ti+1 )) − f (X(ti ))|Ft ] Z t+s Z ≈ E[ λ (f (X(r) + H(X(r), z)) − f (X(r)))µξ (dz)dr t
E
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
353
Martingales associated with a marked Poisson process Lemma 17.8 If (N, {ξk }) is a marked Poisson process, then N (t+s)
E[
X
Z G(k − 1, ξk )|Ft ] = E[
t+s
Z G(N (s), z)µξ (dz)|Ft ],
λ E
t
k=N (t)+1
and hence MG (t) =
N (t) X k=1
Z G(k − 1, ξk ) −
t
Z G(N (s), z)µξ (dz)
λ 0
E
is a martingale.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
354
Lemma 17.9 Let X be cadlag and adapted and G be a bounded, continuous function. Define Z Z(t) =
t
G(X(s−), ξN (s) )dN (s) = 0
N (t) X
G(X(Sk −), ξk ).
k=1
Then Z E[Z(t + s) − Z(t)|Ft ] = E[ t
and hence
Z Z(t) −
t
Z G(X(s), z)µξ (dz)ds|Ft ]
λ E
Z G(X(s), z)µξ (dz)ds
λ 0
t+s
E
is a martingale.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
355
Proof. Note that Z(t) = lim
h→0
[t/h] X
G(X(kh), ξN (kh)+1 )(N ((k + 1)h) − N (kh))
k=0
and Z E[G(X(kh), ξN (kh)+1 )(N ((k+1)h)−N (kh))|Fkh ] =
G(X(kh), z)µξ (dz)λh E
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
356
Change of measure
Theorem 17.10 Let G be bounded and continuous, G ≥ −1, X cadlag and adapted, and Z satisfy Z t Z t Z Z(t) = 1+ Z(s−)G(X(s−), ξN (s) )dN (s)− Z(s)λ G(X(s), z)µξ (dz)ds. 0
E
0
Then Z
t
Z(t) = exp{ 0
Z t Z log(1+G(X(s−), ξN (s) ))dN (s)− λ G(X(s), z)µξ (dz)ds} 0
E
is a martingale.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
357
Transformation of martingales under a change of measure Lemma 17.11 Let Z be as in Theorem 17.10, and suppose dPe|Ft = Z(t)dP|Ft . Rt If M is a {Ft }-local martingale under P and [Z, M ]t = 0 V (s−)dN (s). Then Z t V (s−) dN (s) Y (t) = M (t) − Z(s) 0 is a {Ft }-local martingale under Pe.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
358
Proof. Z
t
Z(t)Y (t) = Z(0)Y (0) +
Z
t
Y (s−)dZ(s) + 0
Z
Z(s−)dY (s) + [Z, Y ]t 0
t
Z
t
= Z(0)Y (0) + Y (s−)dZ(s) + Z(s−)dY (s) + [Z, M ]t 0 0 Z t V (s−)G(X(s−), ξN (s) ) dN (s) − 1 + G(X(s−), ξN (s) ) 0 Z t Z t = Z(0)Y (0) + Y (s−)dZ(s) + Z(s−)dM (s) 0
0
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
359
General counting processes N is a counting process if N (0) = 0, N is right continuous, and N is constant except for jumps of +1. N is determined by its jump times 0 < σ1 < σ2 < · · ·. If N is adapted to Ft , then the σk are Ft -stopping times.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
360
Intensity for a counting process If N is a Poisson process with parameter λ and N is compatible with {Ft }, then P {N (t + ∆t) > N (t)|Ft } = 1 − e−λ∆t ≈ λ∆t. For a general counting process N , at least intuitively, a nonnegative, {Ft }-adapted stochastic process λ(·) is an {Ft }-intensity for N if Z t+∆t P {N (t + ∆t) > N (t)|Ft } ≈ E[ λ(s)ds|Ft ] ≈ λ(t)∆t. t
Definition 17.12 λ is an {Ft }-intensity for N if and only if for each n = 1, 2, . . .. Z t∧σn
N (t ∧ σn ) −
λ(s)ds 0
is a {Ft }-martingale. •First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
361
Modeling with intensities Let X be a stochastic process (cadlag, E-valued for simplicity) that models “external noise.” Let Dc [0, ∞) denote the space of counting paths (zero at time zero and constant except for jumps of +1). Condition 17.13 λ : [0, ∞) × DE [0, ∞) × Dc [0, ∞) → [0, ∞) and assume that if X is cadlag and adapted, then s → λ(s, Z, N ) is cadlag and adapted.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
362
Change of measure Theorem 17.14 On (Ω, F, P ), let N be a unit Poisson process, X be cadlag and adapted, λ satisfy Condition 17.13, and λ be bounded by a constant. Let Z satisfy Z t Z t Z(t) = 1+ (λ(s−, X, N )−1)Z(s−)dN (s)− Z(s)(λ(s, X, N )−1)ds. 0
0
Then Z is a martingale, and if dPe|Ft = Z(t)dP|Ft , then under Pe, Z t N (t) − λ(s, X, N )ds 0
is a martingale, that is, N is counting process with intensity λ(s, X, N ).
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
363
18.
Assignments
1. Due February 1: Exercises 1.5 and 1.8 2. Due February 8: Exercises 1.14 and 1.15 3. Due February 15: Exercise 2.9 and Problem 2 4. Due February 22: Problems 6, 7, 8 5. Due March 13: Problems 9, 10 6. Due April 19: Problems 11, 12 7. Due April 26: Exercises 5.8 and 5.10 8. Due May 3: Exercises 5.11 and 6.1 9. Due May 10: Exercise 5.14
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
364
19.
Problems
1. Let X, Y , and Z be random variables. X and Y are conditionally independent given Z if and only if E[f (X)g(Y )|Z] = E[f (X)|Z]E[g(Y )|Z] for all bounded measurable f and g. (a) Show that X and Y are conditionally independent given Z if and only if E[f (X)|Y, Z] = E[f (X)|Z]
(19.1)
for all bounded measurable f . (b) Let X, Y , and Z be independent, R-valued random variables, and let ψ : R2 → R and ϕ : R2 → R be Borel measurable functions. Define U = ψ(X, Z) and V = ϕ(Y, Z). Show that U and V are conditionally independent given Z. 2. Let X be uniformly distributed on [0, 1], and let D be the σ-algebra generated by {{X ≤ 21 }, { 12 < X ≤ 3 }, { 34 < X ≤ 1}}. Give an explicit representation of E[X|D] in terms of X. What is the probability 4 distribution of E[X|D]? 3. Let Q be a probability measure on (Ω, F ) that is absolutely continuous with respect to P and let L = dQ dP be the corresponding Radon-Nikodym derivative. Let PD and QD be the restrictions of P and Q to the sub-σ-algebra D. Show that QD is absolutely continuous with respect to PD and that dQD = E[L|D] . dPD
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
365
4. Suppose X and Y are independent random variables, and let µX and µY denote the distributions of X and Y . Let h be a measurable function satisfying Z ∞ Z ∞ |h(x, y)|µX (dx)µY (dy) < ∞. −∞
−∞
Define Z = h(X, Y ) and
Z
∞
g(y) =
h(x, y)µX (dx). −∞
Show that E[Z|Y ] = g(Y ). 5. Let X1 , X2 , . . . be independent random variables with E[Xk ] = 1, k = 1, 2, . . .. Let F0 = {∅, Ω} and for n = 1, 2, . . ., let Fn = σ(X1 , . . . , Xn ), that is, Fn is the smallest σ-algebra with respect to which Q X1 , . . . , Xn are measurable. Let M0 = 1 and for n = 1, 2, . . ., let Mn = n k=1 Xk . Show that E[Mn+1 |Fn ] = Mn . 6. Let W be a standard Brownian motion and define B(t) = W (t) − tW (1),
0 ≤ t ≤ 1.
Show that B is independent of W (1). (It is enough to show that B(t1 ), . . . , B(tm ) is independent of W (1) for each choice of 0 ≤ t1 < · · · < tm ≤ 1. 7. Use the result in Problem 6 to compute E[eλW (t) |W (1)].
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
366
8. Let
f (t) = tW ( 1 ). W t
f is a standard Brownian motion. Show that W 9. Let X1 and X2 be adapted, simple processes. Show that for a, b ∈ R, Z = aX1 + bX2 is an adapted, simple process and that Z Z Z ZdW = a X1 dW + b X2 dW. (Note that if U and V are D-measurable, then aU + bV is D-measurable.) n n 10. For n = 1, 2, . . ., let 0 = tn 0 < t1 < · · ·, and let {ai } ⊂ [0, ∞). Suppose that for each t > 0, X n lim ai = t. n→∞
tn i ≤t
Show that for f ∈ C[0, ∞), lim
n→∞
X tn i ≤t
n f (tn i )ai =
t
Z
f (s)ds. 0
(In fact, the result holds for all cadlag f .)
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
367
11. Consider
t
Z X(t) = 1 +
σ
p X(s)dW (s) + bt
0
and let τ = inf{t : X(t) = 0}. Give necessary and sufficient conditions on σ and b > 0 for P {τ < ∞} > 0. 12. Consider
t
Z X(t) = 1 +
t
Z σ(X(s))dW (s) +
0
b(X(s))ds 0
where d = m = 1, σ and b are continuous, and inf x∈R σ(x) > 0. Give necessary and sufficient conditions on σ and b for P {limt→∞ X(t) = ∞} = 1.
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
368
List of topics 1. Review of basic concepts of probability
16. Multivariable Itoˆ formula
2. From probability to measure theory
17. Brownian bridge
3. Review of Markov chains 4. Information and sigma algebras
18. Convergence of empirical distribution functions
5. Conditional expectations
19. Stochastic differential equations
6. Martingales
20. Examples and the Markov property
7. The central limit theorem and Brownian motion
21. SDEs and partial differential equations
8. Martingale properties of Brownian motion
23. Multidimensional SDEs
9. Sample path properties of Brownian motion
24. Examples
22. Examples
10. The Markov property
25. Change of probability measure
11. Definition of the Ito integral
26. Girsanov formula
12. Properties and examples
27. Pricing and risk neutral measures
ˆ formula 13. Ito’s
28. Martingale representation
14. Examples and applications
29. Fundamental theorems of asset pricing
15. Black-Scholes formula
30. Applications
•First •Prev •Next •Go To •Go Back •Full Screen •Close •Quit
369