Gaussian Random Vector Solutions
Gaussian Random Vector Solutions
1 0 1
2
= 1 − 14 = 34
σX1 |X2 ,X3 = σ22 − 1 0 4 1
0 9 0
The conditional mean does not depend on x3 since X1 and X3 are independent.
g. Let Y = 2X1 + X2 + X3 . In part (c) we found that Y ∼ N (9, 21). Thus
0 − µY −9
P{Y < 0} = Φ =Φ √ = Φ(−1.96) = Q(1.96) = 2.48 × 10−2 .
σy 21
b. The random variables X̂ and X − X̂ are jointly Gaussian since they are obtained by a linear
transformation of the GRV [ Y X ]T . By orthogonality, X̂ and X − X̂ are uncorrelated, so
they are also independent. By the same reasoning, X − X̂ and Y are independent.
c. Now write X = X̂ + (X − X̂). Then given Y = y
X = ΣXY Σ−1
Y y + (X − X̂) ,
since X − X̂ is independent of Y.
d. Thus X | {Y = y} is Gaussian with mean ΣXY Σ−1 2 −1
Y y and variance σX − ΣXY ΣY ΣYX .
4. (15 points) Noise cancellation. This is a vector MSE linear estimation problem. Since Z 1 and Z2
are zero mean, µY1 = µX + µZ1 = µ and µY2 = µZ1 + µZ2 = 0 . We first normalize the random
variables by subtracting their means to get
0 0 Y1 − µ
X = X − µ and Y = .
Y2
Now using the orthogonality principle we can find the best linear MSE estimate X̂ 0 of X 0 . To
do so we first find
P +N1 N1 P
ΣY = and ΣYX = .
N1 N1 +N2 0
Therefore
X̂ 0 = ΣTYX Σ−1
Y Y
0
1 N1 +N2 −N1
Y0
= P 0
P (N1 +N2 ) + N1 N2 −N 1 P +N 1
P Y1 − µ
= N1 +N2 −N1
P (N1 +N2 ) + N1 N2 Y2
P (N1 + N2 )(Y1 − µ) − P N1 Y2
= .
P (N1 +N2 ) + N1 N2
5. (20 points) Additive nonwhite Gaussian noise channel. The best estimate of X is of the form
n
X
X̂ = hi Y i .
i=1
7. (15 points) Convergence experiments. The Matlab code is below. The output is shown in
Figure 1.
clear all;
clf;
% Part (a)
n = 1:200;
subplot( 4, 1, 1 );
plot( n, S );
xlabel( ’n’ );
ylabel( ’Sn’ );
title( ’2(a) Sample average sequence’ );
% Part (b)
S = S./repmat( n, 5000, 1 );
% Part (c) Strong Law of Large Numbers (this loop will run for a minute)
E_m = zeros(200,1);
for m = 0:199,
end;
subplot( 4, 1, 2 );
subplot( 4, 1, 3 );
plot( n, M );
xlabel( ’n’ );
ylabel( ’Mn’ );
title( ’2(d) Mean square convergence’ );
subplot( 4, 1, 4 );
plot( n, P, ’r--’ );
hold on;
plot( n, E );
axis( [ 0 200 0 1 ] );
xlabel( ’n’ );
ylabel( ’En (solid), Pn (dashed)’ );
title( ’2(e) Convergence in probability’ );
% Produce hardcopy
orient tall
print hw6q7
0.5
Sn
0
−0.5
0 20 40 60 80 100 120 140 160 180 200
n
2(c) Strong Law of Large Numbers
1
m
0.5
E
0
0 20 40 60 80 100 120 140 160 180 200
n
2(d) Mean square convergence
1
Mn
0.5
0
0 20 40 60 80 100 120 140 160 180 200
n
2(e) Convergence in probability
En (solid), Pn (dashed)
0.5
0
0 20 40 60 80 100 120 140 160 180 200
n
8. (10 points) Convergence with probability 1. For any values of {Xn }, the sequence of Yn values
is monotonically decreasing in n. Since the random variables are ≥ 0, we know that the limit
of Yn is ≥ 0. We suspect that Yn → 0. To prove that Yn converges w.p.1 to 0, we show that
for every > 0,
lim P{|Yn − 0| < for all n ≥ m} = 1 .
m→∞
which is equivalent to limm→∞ P{|Yn − 0| ≥ for some n ≥ m} = 0. So, let m ≥ 1 and consider
P{|Yn − 0| ≥ for some n ≥ m} = P{Yn ≥ for some n ≥ m}
∞
n [ o
(a)
=P {X1 ≥ , . . . , Xn ≥ , Xn+1 < }
n=m
∞
(b) X
= P{X1 ≥ , . . . , Xn ≥ , Xn+1 < }
n=m
∞ n
(c) X Y
= P{Xn+1 < } P{Xi ≥ }
n=m i=1
∞
X
= (1 − e−λ ) e−λn = e−λm → 0 as m → ∞
n=m
3. Minimum MSE for Gaussian random vector. We are given a Gaussian random vector
0 1 2 1
X ∼ N 0 , 2 5 2 .
0 1 2 9
Means and covariances are
0 5 2 2
µ1 = 0 , Σ11 = 1 , Σ12 = 2 1 , µ2 = , Σ22 = , Σ21 = .
0 2 9 1
b. We can show that X − X̂2 and Y − EY are jointly Gaussian by representing them as a
linear function of a GRV:
" #
1 ΣXY Σ−1 E(X) − ΣXY Σ−1
X − X̂2 Y X Y E(Y)
= + .
Y − E(Y) 0 I Y − E(Y)
c. By part (a), X − X̂2 and Yi − E(Yi ) are uncorrelated for every i. By part (b), they are
jointly Gaussian. Therefore they are independent since uncorrelated jointly Gaussian ran-
dom variables are independent. Since E(X̂2 ) = E(X),
E(X − X̂2 | Y) = E(X − X̂2 ) = E(X) − E(X̂2 ) = 0 .
= 1 − hT ΣYX
αn−1
= 1 − 0 · · · 0 α ... = 1 − α2 .
α
6. Gambling By the weak law of large numbers, the sample mean n1 ni=1 Xi converges to the
P
mean E(X) in probability, so P(|Sn −µ| > ) → 0 as n → ∞. The limiting value of P(Sn < µ/2)
depends on µ.
• If µ < 0 then P(Sn < µ/2) → 1. This is because P(|Sn − µ| > ) → 0 as n → ∞ for all
positive . But this means P(|Sn − µ| < ) → 1 as n → ∞. Since Sn → µ < µ/2, we see
that P(Sn < µ/2) → 1.
• If µ > 0 then P(|Sn − µ| < ) → 1 as n → ∞. But if Sn → µ then P (Sn < µ/2) → 0.