Gaussian Random Vector Solutions
Gaussian Random Vector Solutions
                                        1 0 1
                                               
               2
                                                    = 1 − 14 = 34
                                  
              σX1 |X2 ,X3 = σ22 − 1 0 4 1
                                           0 9 0
       The conditional mean does not depend on x3 since X1 and X3 are independent.
   g. Let Y = 2X1 + X2 + X3 . In part (c) we found that Y ∼ N (9, 21). Thus
                                              
                           0 − µY           −9
            P{Y < 0} = Φ             =Φ √          = Φ(−1.96) = Q(1.96) = 2.48 × 10−2 .
                              σy              21
   b. The random variables X̂ and X − X̂ are jointly Gaussian since they are obtained by a linear
      transformation of the GRV [ Y X ]T . By orthogonality, X̂ and X − X̂ are uncorrelated, so
      they are also independent. By the same reasoning, X − X̂ and Y are independent.
   c. Now write X = X̂ + (X − X̂). Then given Y = y
                                        X = ΣXY Σ−1
                                                 Y y + (X − X̂) ,
      since X − X̂ is independent of Y.
   d. Thus X | {Y = y} is Gaussian with mean ΣXY Σ−1                2        −1
                                                  Y y and variance σX − ΣXY ΣY ΣYX .
4. (15 points) Noise cancellation. This is a vector MSE linear estimation problem. Since Z 1 and Z2
   are zero mean, µY1 = µX + µZ1 = µ and µY2 = µZ1 + µZ2 = 0 . We first normalize the random
   variables by subtracting their means to get
                                                                   
                                   0                   0     Y1 − µ
                                 X = X − µ and Y =                    .
                                                               Y2
   Now using the orthogonality principle we can find the best linear MSE estimate X̂ 0 of X 0 . To
   do so we first find                                             
                                  P +N1     N1                       P
                         ΣY =                        and ΣYX =          .
                                    N1    N1 +N2                     0
   Therefore
                        X̂ 0 = ΣTYX Σ−1
                                      Y Y
                                          0
                                                                           
                                               1           N1 +N2 −N1
                                                                         Y0
                                    
                             = P 0
                                       P (N1 +N2 ) + N1 N2  −N 1  P +N 1
                                                                           
                                      P                       Y1 − µ
                           =                      N1 +N2 −N1
                             P (N1 +N2 ) + N1 N2                  Y2
                             P (N1 + N2 )(Y1 − µ) − P N1 Y2
                           =                                .
                                  P (N1 +N2 ) + N1 N2
5. (20 points) Additive nonwhite Gaussian noise channel. The best estimate of X is of the form
                                                       n
                                                       X
                                                X̂ =         hi Y i .
                                                       i=1
7. (15 points) Convergence experiments. The Matlab code is below. The output is shown in
   Figure 1.
   clear all;
   clf;
% Part (a)
n = 1:200;
   subplot( 4, 1, 1 );
   plot( n, S );
   xlabel( ’n’ );
   ylabel( ’Sn’ );
   title( ’2(a) Sample average sequence’ );
% Part (b)
S = S./repmat( n, 5000, 1 );
% Part (c) Strong Law of Large Numbers (this loop will run for a minute)
   E_m = zeros(200,1);
   for m = 0:199,
   end;
   subplot( 4, 1, 2 );
   subplot( 4, 1, 3 );
   plot( n, M );
   xlabel( ’n’ );
   ylabel( ’Mn’ );
   title( ’2(d) Mean square convergence’ );
   subplot( 4, 1, 4 );
   plot( n, P, ’r--’ );
   hold on;
   plot( n, E );
   axis( [ 0 200 0 1 ] );
   xlabel( ’n’ );
   ylabel( ’En (solid), Pn (dashed)’ );
   title( ’2(e) Convergence in probability’ );
   % Produce hardcopy
   orient tall
   print hw6q7
                                          0.5
               Sn
                                           0
                                    −0.5
                                                0   20    40         60    80            100          120       140      160      180       200
                                                                                          n
                                                                          2(c) Strong Law of Large Numbers
                                           1
                    m
                                          0.5
               E
                                           0
                                                0   20    40         60   80           100           120        140      160      180       200
                                                                                        n
                                                                           2(d) Mean square convergence
                                           1
                Mn
0.5
                                           0
                                                0   20    40         60   80           100            120       140      160      180       200
                                                                                        n
                                                                           2(e) Convergence in probability
                En (solid), Pn (dashed)
0.5
                                           0
                                                0   20    40         60   80            100           120       140      160      180       200
                                                                                         n
8. (10 points) Convergence with probability 1. For any values of {Xn }, the sequence of Yn values
   is monotonically decreasing in n. Since the random variables are ≥ 0, we know that the limit
   of Yn is ≥ 0. We suspect that Yn → 0. To prove that Yn converges w.p.1 to 0, we show that
   for every  > 0,
                               lim P{|Yn − 0| <  for all n ≥ m} = 1 .
                                                               m→∞
   which is equivalent to limm→∞ P{|Yn − 0| ≥  for some n ≥ m} = 0. So, let m ≥ 1 and consider
               P{|Yn − 0| ≥  for some n ≥ m} = P{Yn ≥  for some n ≥ m}
                                                                                      ∞
                                                                                    n [                                                     o
                                                                            (a)
                                                                            =P                  {X1 ≥ , . . . , Xn ≥ , Xn+1 < }
                                                                                        n=m
                                                                                  ∞
                                                                            (b)   X
                                                                            =            P{X1 ≥ , . . . , Xn ≥ , Xn+1 < }
                                                                                  n=m
                                                                                   ∞                           n
                                                                            (c)   X                            Y
                                                                            =            P{Xn+1 < }                  P{Xi ≥ }
                                                                                  n=m                           i=1
                                                                                                     ∞
                                                                                                     X
                                                                            = (1 − e−λ )                    e−λn = e−λm → 0 as m → ∞
                                                                                                     n=m
3. Minimum MSE for Gaussian random vector. We are given a Gaussian random vector
                                                  
                                         0     1 2 1
                              X ∼ N  0  ,  2 5 2   .
                                         0     1 2 9
   Means and covariances are
                                                                               
                                                     0            5 2            2
                µ1 = 0 , Σ11 = 1 , Σ12   = 2 1 , µ2 =     , Σ22 =         , Σ21 =     .
                                                       0            2 9            1
   b. We can show that X − X̂2 and Y − EY are jointly Gaussian by representing them as a
      linear function of a GRV:
                    "          # 
                                   1 ΣXY Σ−1        E(X) − ΣXY Σ−1
                                                                      
                       X − X̂2            Y    X                  Y E(Y)
                                =                +                          .
                      Y − E(Y)     0   I       Y           − E(Y)
   c. By part (a), X − X̂2 and Yi − E(Yi ) are uncorrelated for every i. By part (b), they are
      jointly Gaussian. Therefore they are independent since uncorrelated jointly Gaussian ran-
      dom variables are independent. Since E(X̂2 ) = E(X),
                           E(X − X̂2 | Y) = E(X − X̂2 ) = E(X) − E(X̂2 ) = 0 .
                            = 1 − hT ΣYX             
                                                     
                                                αn−1
                            = 1 − 0 · · · 0 α  ...  = 1 − α2 .
                                                  
                                                 α
6. Gambling By the weak law of large numbers, the sample mean n1 ni=1 Xi converges to the
                                                                     P
   mean E(X) in probability, so P(|Sn −µ| > ) → 0 as n → ∞. The limiting value of P(Sn < µ/2)
   depends on µ.
    • If µ < 0 then P(Sn < µ/2) → 1. This is because P(|Sn − µ| > ) → 0 as n → ∞ for all
      positive . But this means P(|Sn − µ| < ) → 1 as n → ∞. Since Sn → µ < µ/2, we see
      that P(Sn < µ/2) → 1.
    • If µ > 0 then P(|Sn − µ| < ) → 1 as n → ∞. But if Sn → µ then P (Sn < µ/2) → 0.