Title | MTL106-1-9 - Lecture notes 1-9 |
---|---|
Author | Arpit Chauhan |
Course | Introduction to Probability Theory and Stochastic Process |
Institution | Indian Institute of Technology Delhi |
Pages | 9 |
File Size | 223.2 KB |
File Type | |
Total Downloads | 47 |
Total Views | 134 |
modes of convergence...
PROBABILITY AND STOCHASTIC PROCESS (MTL106) ANANTA KUMAR MAJEE
1. Modes of Convergence Lemma 1.1 (Markov inequality). If X is a non-negative random variable whose expected value exists, then for all a > 0, P(X > a) ≤
E(X) . a
Proof. Observe that, since X is non-negative i h E[X] = E X1{X >a} + X1{X ≤a} ≥ E[X1{X >a} ] ≥ aP(X > a). Hence the result follows.
Corollary 1.2. If X is a random variable such that E[|X |] < +∞, then for all a > 0 P(|X| > a) ≤
E(|X|) . a
Lemma 1.3 (Chebyschev’s inequality). Let Y be an integrable random variable such that Var(Y ) < +∞. Then for any ε > 0 P(|Y − E(Y )| > ε) ≤
Var(Y ) . ε2
Proof. To get the result, take X = |Y − E(Y )|2 and a = ε2 in Markov inequality.
Example 1.1. Is there any random variable X for which 1 P µ − 3σ ≤ X ≤ µ + 3σ = , 2 2 where µ = E(X) and σ = Var(X). Solution: Observe that P µ − 3σ ≤ X ≤ µ + 3σ = P |X − µ| ≤ 3σ = 1 − P |X − E(X)| > 3σ . σ2 By Chebyschev’s inequality, we get that P |X − E(X )| > 3σ ≤ 9σ 2 , and hence 1 8 P µ − 3σ ≤ X ≤ µ + 3σ ≥ 1 − = . 9 9 Since 12 < 89 , there exists NO random variable X satisfying the given condition. Theorem 1.4 (Weak Law of Large Number). Let {Xi } be a sequence of iid random variables with finite mean µ and variance σ 2 . Then for any ε > 0 Sn σ2 P | − µ| > ε ≤ 2 , n nε Pn where Sn = i=1 Xi . In particular, Sn − µ| > ε = 0 . lim P | n→∞ n Proof. First inequality follows from Chebyschev’s inequality. Sending limit as n tends to infinity in the first inequality, we arrive at the second result. An application to Real Analysis: 1
2
A. K. MAJEE
Theorem 1.5 (Weierstrass Theorem). Let f : [0, 1] → R be a continuous function. Let n X k n k f( ) Bn,f (x) = x (1 − x)n−k , x ∈ [0, 1] n k k=0
be the Bernstein polynomial of order n for the function f . Then Bn,f → f uniformly on [0, 1]. Proof.PLet x ∈ [0, 1] be fixed. Let n {Xn } be a sequence of i.i.d Bernoulli(x) random variables. Set n Sn = i=1 Xi . Then P(Sn = k) = k xk (1 − x)n−k and hence n n X X k n k k Sn f( ) f ( )P(Sn = k) = E[f ( )] = x (1 − x)n−k = Bn,f (x). n n n k k=0 k=0 Hence, for any δ > 0
Sn Sn )]| ≤ E |f (x) − f ( )| n n Sn = E |f (x) − f ( )| 1|x− Sn |>δ + 1|x− Sn |≤δ n n n Since f is continuous on [0, 1], it is uniformly continuous. Hence given ε > 0, there exists δ0 > 0 such that |f (x) − Bn,f (x)| = |f (x) − E[f (
|x − y| ≤ δ =⇒ |f (x) − f (y)| < ε .
Taking δ = δ0 in the previous inequality, we get
Sn |f (x) − Bn,f (x)| ≤ 2kf kP |x − | > δ0 + ε. n
Observe that E( Snn ) = x and Var( Snn ) =
x(1−x) n
. Hence by Chebyschev’s inequality, we get
x(1 − x) Sn 1 P |x − | > δ0 ≤ ≤ 4nδ 20 nδ02 n
∀ x ∈ [0, 1].
Combining these two inequality, we have, for all x ∈ [0, 1] |f (x) − Bn,f (x)| ≤
kf k + ε. 2nδ02
Sending n → ∞ and then letting ε → 0, we conclude the desired result, i.e., Bn,f → f uniformly on [0, 1]. We shall discuss various modes of convergence for a given sequence of random variables {Xn } defined on a given probability space (Ω, F , P). Definition 1.1 (Convergence in probability). We say that {Xn } converges to a random variable X
P defined on a the same probability space (Ω, F , P) in probability, denoted by Xn → X, if for every ε > 0,
lim P(|Xn − X| > ε) = 0.
n→∞
Example 1.2. Let {Xn } be a sequence of random variables such that P(Xn = 0) = 1 − n) =
1 n.
1 n
and P(Xn =
P
Then Xn → 0. Indeed for any ε > 0,
P(|Xn | > ε) =
(
1 n
0
if ε < n, if ε ≥ n .
Hence, limn→∞ P(|Xn | > ε) = 0.
Example 1.3. Let {Xn } be a sequence of i.i.d. random variables with P(Xn = 1) = 12 and P(Xn = Pn −1) = 12 . Then n1 i=1 Xi converges to 0 in probability. Indeed for any ε > 0, thanks to weak law of large number, we have 1 Var(X1 ) P(| Sn − µ| > ε) ≤ nε2 n
PROBABILITY AND STOCHASTIC PROCESS
3
where µ = E(X1 ). Observe that µ = 0 and Var(X1 ) = 1. Hence n 1 1X Xi | > ε) ≤ 2 → 0 as n → ∞. P(| nε n i=1
P
|Xn −X | Theorem 1.6. Xn → X if and only if limn→∞ E 1+|X = 0. n −X |
P
Proof. With out loss of generality, take X = 0. thus, we want to show that Xn → 0 if and only if |Xn | limn→∞ E 1+|X = 0. n| P Suppose Xn → 0. Then given ε > 0, we have limn→∞ P(|Xn | > ε) = 0. Now,
|Xn | |Xn | |Xn | 1|Xn |>ε + = 1|Xn |≤ε ≤ 1|Xn |>ε + ε 1 + |Xn | 1 + |Xn | 1 + |Xn | |Xn | |Xn − X| ≤ P(|Xn | > ε) + ε =⇒ lim E ≤ ε. =⇒ E n→∞ 1 + |Xn | 1 + |Xn − X| |Xn −X | Since ε > 0 is arbitrary, we have limn→∞ E 1+|X = 0. n −X | |Xn −X | x Conversely, let limn→∞ E 1+|Xn −X | = 0. Observe that the function f (x) = 1+x is strictly increasing on [0, ∞). Thus, |Xn | |Xn | ε 1|Xn |>ε ≤ 1|Xn |>ε ≤ 1 + |Xn | 1+ε 1 + |Xn | |Xn | ε P(|Xn | > ε) ≤ E =⇒ 1+ε 1 + |Xn | |Xn − X| 1+ε P lim E =⇒ lim P(|Xn | > ε) ≤ = 0 =⇒ Xn → 0. n→∞ ε n→∞ 1 + |Xn − X|
P
Exercise 1.1. Show that Xn → X if and only if limn→∞ E 1 ∧ |Xn − X| = 0.
Definition 1.2 (Convergence in r-th mean). Let X, {Xn } be random variables defined a given probability space (Ω, F , P) such that for r ∈ N, E[|X |r ] < ∞ and E[|Xn |r ] < ∞ for all n. We say that r {Xn } converges in the r-th mean to X, denoted by Xn → X, if the following holds: lim E[|Xn − X|r ] = 0.
n→∞
Example 1.4. Let {Xn } be i.i.d. random variables with E[Xn ] = µ and Var(Xn ) = σ 2 . Define Yn = Pn 2 1 n i=1 Xi . Then Yn → µ. Indeed Pn σ2 1 i=1 Xi − nµ2 ] = 1 2 E[|Sn − E(Sn )|2 ] = 2 Var(Sn ) = E[|Yn − µ| ] = E[ 2 n n n n Pn 2 where Sn = i=1 Xi . Hence Yn → µ.
Theorem 1.7. The following holds: r
P
i) Xn → X =⇒ Xn → X for any r ≥ 1. P P ii) Let f be a given continuous function. If Xn → X, then f (Xn ) → f (X).
Proof. Proof of (i) follows from Markov’s inequality. Indeed, for any given ε > 0, E[|xn − X|r ] 1 P(|Xn − X| > ε) ≤ =⇒ lim P(|Xn − X| > ε) ≤ r lim E[|xn − X|r ] = 0. n→∞ ε n→∞ εr Proof of (ii): For any k > 0, we see that {|f (Xn ) − f (X)| > ε} ⊂ {|f (Xn ) − f (X)| > ε, |X| ≤ k} ∩ {|X| > k} .
Since f is continuous, it is uniformly continuous on any bounded interval. Therefore, for any given ε > 0, there exists δ > 0 such that |f (x) − f (y)| ≤ ε if |x − y| ≤ δ for x and y in [−k, k]. This means that {|f (Xn ) − f (X)| > ε, |X| ≤ k} ⊂ {|Xn − X| > δ, |X | ≤ k} ⊂ {|Xn − X| > δ} .
4
A. K. MAJEE
Thus we have {|f (Xn ) − f (X)| > ε} ⊂ {|Xn − X| > δ} ∩ {|X| > k }
=⇒ P(|f (Xn ) − f (X)| > ε) ≤ P(|Xn − X | > δ) + P(|X| > k ).
P
Since Xn → X and limk→∞ P(|X| > k) = 0, we obtain that limn→∞ P(|f (Xn ) − f (X)| > ε) = 0. This completes the proof. In general, convergence in probability does not imply convergence in r-th mean. To see it, consider the following example. P
Example 1.5. Let Ω = [0, 1], F = B([0, 1]) and P(dx) = dx. Let Xn = n1(0,1/n) . Then Xn → 0 but r Xn 9 0 for all r ≥ 1. To show this, observe that 1 P P(|Xn | > ε) ≤ =⇒ lim P(|Xn | > ε) = 0 i.e., Xn → 0. n→∞ n On the other hand, for r ≥ 1 Z 1 n E[|Xn |r ] = nr dx = nr−1 9 0 as n → ∞. 0
Definition 1.3 (Almost sure convergence). Let X, {Xn } be random variables defined on a probability space (Ω, F , P). We say that {Xn } converges to X almost surely (or with probability 1), denoted by a.s Xn → X, if the following holds: P( lim Xn = X) = 1. n→∞
Example 1.6. Let Ω = [0, 1], F = B([0, 1]) and P(dx) = dx. Define ( 1 if ω ∈ (0, 1 − 1n ) Xn (ω) = n, otherwise. It is easy to check that if ω = 0 or ω = 1, then limn→∞ Xn (ω) = ∞. For any ω ∈ (0, 1), we can find n0 ∈ N such that ω ∈ (0, 1 − 1n ) for any n ≥ n0 . As a consequence, Xn (ω) = 1 for any n ≥ n0 . In other words, for ω ∈ (0, 1), limn→∞ Xn (ω) = 1. Define X(ω) = 1 for all ω ∈ [0, 1]. Then a.s.
P(ω ∈ [0, 1] : {Xn (ω )} does not converge to X(ω)) = P({0, 1}) = 0 =⇒ Xn → 1. Let {An } be a sequence of events in F . Define lim sup An = ∩∞ ∪m≥n Am . n=1 ∪m≥n Am = lim n→∞
n
This can be interpreted probabilistically as
lim sup An = “ An occurs infinitely often”. n
We denote this as {An i.o.} = lim supAn . n
Theorem 1.8 (Borel-Cantelli lemma). Let {An } be a sequence of events in (Ω, F , P). P∞ i) If n=1 P(An ) < +∞, then P(An i.o.) = 0. P∞ ii) If An are mutually independent events, and if n=1 P(An ) = ∞, then P(An i.o.) = 1.
P∞ Remark 1.1. For mutually independent events An , since n=1 P(An ) is either finite or infinite, the event {An i.o.} has probability either 0 or 1. This is sometimes called zero-one law. As a consequence of Borel-Cantelli lemma, we have the following proposition.
Proposition 1.9. Let {Xn } be a sequence of random variables defined on a probability space (Ω, F , P). P∞ a.s. If n=1 P(|Xn | > ε) < +∞ for any ε > 0, then Xn → 0.
PROBABILITY AND STOCHASTIC PROCESS
5
P∞ Proof. Fix ε > 0. Let An = {|Xn | > ε}. Then n=1 P(An ) < +∞, and hence by Borel-Cantelli lemma, P(An i.o.) = 0. Now c lim sup An = {ω : ∃ n0 (ω) such that |Xn (ω)| ≤ ε ∀ n ≥ n0 (ω )} := Bε . n
∞ Bc =⇒ B c = ∪r=1 Moreover, since P(Bε= r1 ) = 1, we have 1 .
Thus, P(Bε ) = 1. Let B = ∩∞ r=1 B r1 c P(B 1 ) = 0. Observe that
r
r
{ω : lim |Xn (ω )| = 0} = ∩∞ r=1 B r1 . n→∞
c
Again, P(B ) ≤
P∞
r=1
c
P(B 1 ) = 0, and hence P(B) = 1. In other words, r a.s. P {ω : lim |Xn (ω)| = 0} = 1, i.e., Xn → 0. n→∞
Example 1.7. Let {Xn } be a sequence of i.i.d. random variables such that P(Xn = 1) = 12 and P(Xn = Pn a.s. −1) = 12 . Let Sn = i=1 Xi . Then n12 Sn2 → 0. To show the result, we use Proposition 1.9. Note that ∞ X 1 1 E[|Sn2 |2 ] 1 1 a.s. ≤ P( 2 |Sn2 | > ε) < ∞ =⇒ 2 Sn2 → 0. P( 2 |Sn2 | > ε) ≤ =⇒ n n4 ε2 n n n2 ε2 n=1
Let us consider the following example.
Example 1.8. Let Ω = [0, 1], F = B ([0, 1]) and P(dx) = dx. Define Xn = 1 [
j 2k
n = 2k + j, j = 0, 1, . . . , 2k − 1,
, , j+1 k ] 2
k = 0, 2, . . . .
Note that, for each positive integer n, there exist integers j and k(uniquely determined) such that n = 2k + j, j = 0, 1, . . . , 2k − 1,
k = 0, 1, 2, . . . .
( for n = 1, k = j = 0, and for n = 5, k = 2, j = 1 and so on). Let An = {Xn > 0}. Then, clearly P
P(An ) → 0. Consequently, Xn → 0 but Xn (ω) 9 0 for all ω ∈ Ω. Theorem 1.10. The followings hold.
P i) If Xn a.s → X, then Xn → X. P ii) If Xn → X, then there exists a subsequence Xnk of Xn such that Xnk a.s. → X. a.s a.s iii) If Xn → X, then for any continuous function f , f (Xn ) → f (X).
a.s
ε ∞ Proof. Proof of i): For any ε > 0, define Aεn = {|Xn − X| > ε} and Bm = ∪n=m Aεn . Since Xn → X, ε ε P(∩m Bm ) = 0. Note that {B m } are nested and decreasing sequence of events. Hence from the continuity of probability measure P, we have ε ε ) = 0. lim P(B m ) = P(∩m Bm
m→∞
ε ) ≤ P(B εm ). This implies that limm→∞ P(Aεm ) = 0. In other words, Since Aεm ⊂ Bmε , we have P(Am P Xn → X. P
Proof of ii): To prove ii), we will use Borel-Cantelli lemma. Since Xn → X, we can choose a subsequence P Xnk such that P(|Xnk − X| > k1 ) ≤ 21k . Let Ak := {|Xnk − X| > 1k }. Then ∞ k=1 P(Ak ) < +∞. Hence, by Borel-Cantelli lemma P(Ak i.o.) = 0. This implies that 1 ∞ c P(∪∞ }) = 1 n=1 ∩m=n Am ) = 1 =⇒ P {ω ∈ Ω : ∃ n0 : ∀k ≥ n0 , |Xnk − X| ≤ k a.s. =⇒ Xnk → X. Proof of iii): Let N = {ω : limn→∞ Xn (ω) 6= X(ω )}. Then P(N ) = 0. If ω ∈ / N , then by the continuity property of f , we have lim f (Xn (ω)) = f ( lim Xn (ω)) = f (X (ω)) .
n→∞
n→∞
a.s
This is true for any ω ∈ / N and P(N ) = 0. Hence f (Xn ) → f (X).
6
A. K. MAJEE
Definition 1.4 (Convergence in distribution). Let X, X1 , X2 , . . . be real-valued random variables with distribution functions FX , FX1 , FX2 , . . . respectively. We say that (Xn ) converges to X is distribud
tion, denoted by Xn → X, if
lim FXn (x) = FX (x)
n→∞
for all continuity points x of FX .
Remark 1.2. In the above definition, the random variables X, {Xn } need not be defined on the same probability space. Example 1.9. Let Xn =
1 n
and X = 0. Then ( 1 if x ≥ n1 FXn (x) = P(Xn ≤ x) = 0, otherwise
and
( 1 x ≥ 0, FX (x) = 0, x < 0.
Observe that 0 is the only discontinuity point of FX , and limn→∞ FXn (x) = F (x) for x 6= 0. Thus, d Xn → 0.
Example 1.10. Let X be a real-valued random variable with distribution function F . Define Xn = X + 1n . Then 1 1 FXn (x) = P(X + ≤ x) = F (x − ) n n =⇒ lim FXn (x) = F (x−) = F (x) for continuity point x of F . n→∞
d
This implies that Xn → X.
d
P
Theorem 1.11. Xn → X implies that Xn → X. Proof. Let ε > 0. Since FXn (t) = P(Xn ≤ t), we have
FXn (t) = P(Xn ≤ t, |Xn − X| > ε) + P (Xn ≤ t, |Xn − X| ≤ ε)
≤ P(|Xn − X > ε) + P(Xn ≤ t, |Xn − X| ≤ ε) ≤ P(|Xn − X > ε) + P(X ≤ t + ε)
≤ P(|Xn − X > ε) + FX (t + ε),
FX (t − ε) = P(X ≤ t − ε) = P(X ≤ t − ε, |Xn − X| > ε) + P(X ≤ t − ε, |Xn − X| ≤ ε)
≤ P(|Xn − X| > ε) + P(X ≤ t − ε, |Xn − X| ≤ ε) ≤ P(|Xn − X| > ε) + P(Xn ≤ t) ≤ P(|Xn − X| > ε) + FXn (t) .
Thus, since limn→∞ P(|Xn − X| > ε) = 0, we obtain from the above inequalities
FX (t − ε) ≤ lim inf FXn (t) ≤ lim sup FXn (t) ≤ FX (t + ε). n→∞
n→∞
Thet t be the continuity point of F . Then sending ε → 0 in the above inequality, we get lim FXn (t) = FX (t),
n→∞
d
i.e., Xn → X.
Converge of this theorem is NOT true in general. Example 1.11. Let X ∼ N (0, 1). Define Xn = −X for n = 1, 2, 3, . . .. Then Xn ∼ N (0, 1) and hence d Xn → X. But ε P P(|Xn − X| > ε) = P(|2X| > ε) = P(|X| > ) 6= 0 =⇒ Xn 9 X. 2 Theorem 1.12 (Continuity theorem). Let X, {Xn } be random variables having the characteristic function φX , {φXn } respectively. Then the followings are equivalent. d
i) Xn → X. ii) E(g(Xn )) → E(g(X)) for all bounded Lipschitz continuous function. iii) limn→∞ φXn (t) = φX (t) for all t ∈ R.
PROBABILITY AND STOCHASTIC PROCESS
7
Theorem 1.13 (Strong law of large number). Let {Xi } be a sequence of i.i.d. random variables with finite mean µ and variance σ 2 . Then Sn a.s. → µ, n
where Sn =
n X
Xi .
i=1
The special case of 4-th order moment, above theorem is refered as Borel’s SLLN. To prove the theorem, we need following lemma. Lemma 1.14. Let {Xi } be a sequence of random variables defined on a given probability space (Ω, F , P). i) If {Xn } are positive, then
∞ ∞ X X E Xn = E[Xn ] .
ii) If
P∞
n=1
E[|Xn |] < ∞, then
P∞
i=1
(1.1)
n=1
n=1
Xi converges almost surely and (1.1) holds as well.
Proof of Theorem 1.13: With out loss of generality we can assume that µ = 0. Set Yn = that, thanks to independent property, n 1 X σ2 1 X E[Yn ] = 0, E[Y n2 ] = 2 E[X j2 ] = . E(Xj Xk ) = 2 n n n
Sn n
. Observe
j=1
1≤j,k≤n
Thus, limn→∞ E[Yn2 ] = 0, and hence along a subsequence, Yn converges to 0 almost surely. But we need to show that original sequence converges to 0 with probability 1. To do so, we proceed as follows. P∞ σ 2 2 P∞ P∞ 2 Since E[Y n2 ] = σn , we see that n=1 E[Yn22 ] = n=1 n=1 Y n2 n2 < +∞ and hence by Lemma 1.14, ii), converges almost surely. Thus, lim Yn2 = 0
with probability 1.
n→∞
(1.2)
Let nN. Then there exists m(n) ∈ N such that (m(n))2 ≤ n < (m(n) + 1)2 . Now 2
(m(n)) n X (m(n))2 (m(n))2 1 1 X 1 2 Xi − Yn − Y(m(n)) = Xi = 2 n n n n ( m ( n)) i=1 i=1
n X
Xi
i=1+(m(n))2
n h 2 i X n − (m(n))2 2 (m(n))2 2m(n) + 1 2 1 σ =⇒ E Yn − E[X 2i ] = σ ≤ = 2 Y(m(n))2 2 n2 n n n i=1+(m(n))2 √ √ 3σ 2 2 n+1 2 ≤ (∴ n < (m(n) + 1)2 and m(n) ≤ n) σ ≤ 3 2 n n2 ∞ ∞ h 2 i X X (m(n))2 3σ 2 E Yn − ≤ Y(m(n))2 =⇒ < +∞ . 3 n n2 n=1 n=1
Thus, again by Lemma 1.14, ii), we conclude that lim Yn −
n→∞
Obsere that limn→∞
(m(n))2 n
(m(n))2 Y(m(n))2 = 0 n
with probability 1.
(1.3)
= 1. Thus, in view of (1.2) and (1.3), we conclude that lim Yn = 0
n→∞
with probability 1.
This completes the proof. Example 1.12. 1) Let {Xn } be a sequence of i.i.d. random variables that are bounded, i.e., there a.s. exists C < ∞ such that P(|X1 | ≤ C) = 1. Then Snn → E(X1 ). 2) Let {Xn } be a sequence of i.i.d. Bernoulli(p) random variables. Then lim
n→∞
n 1X Xi = p n i=1
with probability 1.
8
A. K. MAJEE
Theorem 1.15 (Kolmogorov’s strong law of large numbers). Let {Xn } be a sequence of i.i.d. random variables and µ ∈ R. Then limn→∞ Snn = µ a.s. if and only if E[Xn ] = µ. In this case, the convergence also holds in L1 . R1 Example 1.13 (Monte Carlo approximation). Let f be a measurable function in [0, 1] such that 0 |f (x)| dx < R1 ∞. Let α = 0 f (x) dx. In general we cannot obtain a closed form expression for α and need to estimate it. Let {Uj } be a sequence of independent uniformly random variables on [0, 1]. Then by Theorem 1.15, Z 1 n 1X f (x) dx lim f (Uj ) = E[f (Uj )] = n→∞ n 0 j=1 R1 a.s. and in L2 . Thus, to get an approximation of 0 f (x) dx, we need to simulate the uniform random variables Uj (by using a random number generator). Theorem 1.16 (Central limit theorem). Let {Xn } be a sequence of i.i.d. random variables with finite −nµ mean µ and variance σ 2 with < 0 < σ 2 < +∞. Let Yn = Sσn√ . Then Yn converges in distribution to n Y , where L(Y ) = N (0, 1). Proof. With out loss of generality, we assume that µ = 0. Let Φ be the characteristic function of Xj . Since {Xj } are i.i.d., we have Pn n n hY i Y X X iu X√i u n √ i iu i=1 iu S√n iu √i σ n ]=E ΦYn (u) = E[eiuYn ] = E[e σ n ] = E[e E e σ n = Φ( √ ) . e σ n = σ n i=1 i=1 Since E[|Xj |2 ] < +∞, the function Φ has two continuous derivatives. In particular,
Φ′′ (u) = −E[Xj2 eiuXj ] =⇒ Φ′ (0) = 0, Φ′′ (0) = −σ 2 .
Φ′ (u) = iE[Xj eiuXj ],
Expanding Φ in a taylor expansion about u = 0, we have Φ(u) = 1 −
σ 2 u2 + h(u)u2 , 2
where h(u) → 0 as u → 0.
Thus, we get ΦYn (u) = e
u )) n log(φ( σ√ n
2
=e
n log(1− u 2n +
u2 nσ2
u h( σ√ )) n
=⇒ lim ΦYn (u) = e−
u2 2
n→∞
= ΦY (u).
Hence by Levy’s continuity theorem, we conclude that Yn converges in distribution to Y with L(Y ) = N (0, 1). Remark 1.3. If σ 2 = 0, then Xj = µ a.s. for all j, and hence
Sn n
= µ a.s.
One can weaken slightly the hypotheses of Theorem 1.16. Indeed, we have the following Central limit theorem. Theorem 1.17. Let {Xn } be independent but not necessarily identicallu distributed. Let E[Xn ] = 0 for all n, ane let σ n2 = Var(Xn ). Assume that sup E[|Xn |2+ε ] < +∞ for some ε > 0, n