Convergence Analysis of The Extended Kalman Filter Used As An Observer For Nonlinear Deterministic Discrete-Time Systems
Convergence Analysis of The Extended Kalman Filter Used As An Observer For Nonlinear Deterministic Discrete-Time Systems
 [3] J. B. Lasserre, A trace inequality for the matrix product, IEEE Trans.      The EKF consists of using the classical Kalman filter equations
     Automat. Contr., vol. 40, pp. 15001501, 1995.                             to the first-order approximation of the nonlinear model about the
 [4] M. Mrabti and M. Benseddik, Bounds for the eigenvalues of the             last estimate. The literature is vast about this subject, and we refer
     solution of the unified algebraic Riccati matrix equation, Syst. Contr.
     Lett., vol. 24, pp. 345349, 1995.                                         the reader to [7], [11], [15], [16], and [19] and the references therein.
 [5] A. W. Marshall and I. Olkin, Inequalities: Theory of Majorization and      However, few works have been performed to analyze the stability and
     Its Applications. Orlando, FL: Academic, 1979.                             convergence of the filter. The main difficulty arises from the fact that
 [6] R. T. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ.         the EKF equations are only approximate ones. So, the corresponding
     Press, 1970.
                                                                                propagation equations are available only if the estimate belongs to a
                                                                                neighborhood of the actual state.
                                                                                   An interesting study of asymptotic behavior of the EKF, used
                                                                                for the joint parameter and state estimation of linear time-invariant
                                                                                systems, was developed by Ljung in [11]. For this particular case,
      Convergence Analysis of the Extended Kalman                               some modifications, such as coupling between parameters vector
        Filter Used as an Observer for Nonlinear                                and the Kalman gain, were introduced to enhance convergence of
           Deterministic Discrete-Time Systems                                  the EKF. More recently, in a synthesis work on nonlinear discrete-
                                                                                time systems, sufficient conditions of the EKF for noisy systems
           M. Boutayeb, H. Rafaralahy, and M. Darouach                          used as a local asymptotic observer for the deterministic case have
                                                                                been established by Song et al. [15]. They have also shown that
                                                                                conditions needed to ensure the uniform boundedness of certain
    AbstractIn this paper, convergence analysis of the extended Kalman         Riccati equations are related to the observability properties of the
filter (EKF), when used as an observer for nonlinear deterministic
discrete-time systems, is presented. Based on a new formulation of the          considered nonlinear system.
first-order linearization technique, sufficient conditions to ensure local         Recently, Cicarrella et al. [18] have developed a robust observer,
asymptotic convergence are established. Furthermore, it is shown that           which uses n output values and the inverse of the observabil-
the design of the arbitrary matrix, namely Rk in the paper, plays an            ity matrix at each step, and the local convergence analysis for a
important role in enlarging the domain of attraction and then improving
                                                                                multi-input/single-output (MISO) nonlinear discrete-time system was
the convergence of the modified EKF significantly. The efficiency of this
approach, compared to the classical version of the EKF, is shown through        studied.
a nonlinear identification problem as well as a state and parameter                Motivated by the identification problem of time-invariant nonlinear
estimation of nonlinear discrete-time systems.                                  systems [3], [4], we address here the problem of stability and the
  Index TermsConvergence analysis, deterministic nonlinear discrete-           convergence of the EKF when used as a deterministic observer
time systems, extended Kalman filter, Lyapunov approach.                        for multi-input/multi-output nonlinear discrete-time systems written
                                                                                in their general form. Based on a new formulation for the exact
                                                                                linearization technique, we introduce instrumental time-varying ma-
                            I. INTRODUCTION                                     trices for the stability and convergence analysis. It is pointed out
    Since the 1960s, many research activities have been developed              that, under mild conditions, asymptotic behavior of the EKF may
to deal with the problem of state estimation for nonlinear dynamical            be improved significantly even for bad first-order approximation.
systems. The main motivation for doing so is that most physical                 To show accuracy and performances of this technique, we use the
processes are described by nonlinear mathematical models. Thus,                 modified EKF at first as a parameter estimator for the Hammerstein
several nonlinear state estimation methods have been performed to               model and secondly as a simultaneous state and parameter estimator
increase the accuracy and performances of the control system design.            of a nonlinear model.
    Generally, we distinguish two approaches for nonlinear observers
design. The first one is based on some nonlinear state transformation                                II. PROBLEM FORMULATION
using the Lie algebra. This approach brings the original system
                                                                                  The nonlinear systems considered here are of the form
into a canonical form from which the design of state observers is
performed using linear techniques in the new coordinates. Necessary                                       x  k+1 = f (xk ; uk )                     (1a)
and sufficient conditions for a standard nonlinear system to be state                                         yk = h(xk ; uk )                      (1b)
equivalent to the nonlinear canonical form, for both continuous and
discrete-time cases, have been established in [2], [8], [9], and [14].          where uk 2 Rr and yk 2 Rp are the input and output vectors at time
We notice that only a few classes of forced nonlinear systems are               instant k: We assume that f (xk ; uk ) and h(xk ; uk ) are differentiable
considered by this nonlinear transformation.                                    on Rn : The EKF for the associated noisy system that we use here
    The second approach is without transformation and is based on the           as an observer of (1a), (1b) is:
linearized model. In spite of the local convergence of this method, it             1) measurement update
is widely used in practice and generally gives good results under less
restrictive conditions than the first approach [6], [7], [12], [13], [16].               ^k
                                                                                         x +1 = x^k+1=k + Kk+1 ek+1                      (2)
One of the most popular estimation techniques largely investigated                                       T                T
                                                                                         Kk+1 = Pk+1=k Hk+1 (Hk+1 Pk+1=k Hk+1 + Rk+1 )
                                                                                                                                      01 (3)
                                                                                         Pk+1 = (I 0 Kk+1 Hk+1 )Pk+1=k
for state estimation of nonlinear systems is the extended Kalman                                                                         (4)
filter (EKF).
                                                                                  2) time update
  Manuscript received June 5, 1995; revised October 17, 1995.
  The authors are with the CRAN CNRS URA, Cosnes et Romain 54400,
France.
                                                                                                        ^k
                                                                                                        x +1=k = f (^xk ; uk )                       (5)
                                                                                                                           T
  Publisher Item Identifier S 0018-9286(97)02050-3.                                                     Pk+1=k = Fk Pk Fk + Qk                       (6)
       where                                                                       The exact equality represented here by (15) and (16) is the key
      ek+1 = yk+1 0 h(^xk+1=k ; uk+1 )
                                                                                 point of our approach since the convergence study will be performed
                                                                                 without any approximation while eik+1 and x
                                                                          (7)
                                                                                                                              ~jk+1=k are written in
        Fk = F (^xk ; uk ) = @f (xk ; uk )                                (8)    a first-order representation. Now if we introduce the signal vector,
                                 @xk       x =^
                                              x                                  we obtain
                       III. CONVERGENCE ANALYSIS                                      On the other hand, from (3) and (4), we have
  In this section we give a simple approach for setting up the                            Pk+1 HkT+1 Rk0+1
                                                                                                        1
convergence analysis of the considered nonlinear systems. It is                                            T               T
                                                                                               = Pk+1=k Hk+1 (Hk+1 Pk+1=k Hk+1 + Rk+1 )
                                                                                                                                       01                 (22)
emphasized that under the local reconstructibility condition [19], the
asymptotic convergence of the EKF is ensured when the arbitrary                  and
matrix Rk is adequately chosen.
  Let us note by x ~k+1 and x ~k+1=k the state estimation and state                               Pk0+1
                                                                                                      1 = P 01 + H T R01 Hk+1 :
                                                                                                           k+1=k  k+1 k+1                                 (23)
prediction error vectors respectively defined by
                                                                                   Substituting (22) into (21) and (21) into (12), the quadratic function
                           x~k+1 = xk+1 0 x^k+1                          (10)    becomes
                        x~k+1=k = xk+1 0 x^k+1=k
                                                                                            Vk+1 = (~xk+1=k 0 Pk+1 HkT+1 Rk0+1
                                                                                                                             1 ek+1 )T
                                                                         (11)
and the candidate Lyapunov function Vk+1 as                                                        1 Pk0+11 (~xk+1=k 0 Pk+1 HkT+1 Rk0+1
                                                                                                                                     1 ek+1 )             (24)
                         Vk+1 = x~Tk+1 Pk0+1
                                           1 x~k+1 :                     (12)
                                                                                 or
   Our aim here is to determine conditions for which fVk gk=1111 is
a decreasing sequence and to show the EKF limitations when the
                                                                                          Vk+1 = x~k+1=k Pk0+1
                                                                                                             1 x~
                                                                                                                 k+1=k 0 x
                                                                                                                          T      T    01
                                                                                                                         ~k+1=k Hk+1 Rk+1 ek+1
or equivalently                                                                                As Rk0+1
                                                                                                      1 and R01 Hk+1 Pk+1 H T R01 are symmetric matrices
                                                                                                                k+1  p 0 k+1; k+1p 0
          Vk+1 0 Vk = eTk+1 (k+1 Rk0+1
                                      1 k+1                                                 and as the interval 0 ]1  1 1 1 + 1 1 [ ]0 2[
                                                                                                                            k+1               k+1  ; ; this
                                                                                                                        2   0
                                                                                             implies that 2ik+1 0 ik+1 < ; and (40) yields to
                      0 k+1 Rk+1 0 Rk0+1
                                 0 1        1 k+1
                                                                                                max ((ik+1 0 2ik+1 )Rk+1 + Rk0+1
                                                                                                        2               0 1       1 Hk+1 Pk+1 H T R01 )
                                                                                                                                               k+1 k+1
                      + Rk0+11 Hk+1 Pk+1 HkT+1 Rk0+11 )ek+1                                                 2                     01
                                                                                                      0(ik+1 0 2ik+1 )max (Rk+1 )
                      + x~Tk (FkT k (Fk Pk FkT )01 k Fk                                              + max(Rk0+11 Hk+1 Pk+1 HkT+1 Rk0+11 )         (41)
                      0 Pk01 )~xk  0:                                                       with
     A sufficient condition to ensure that is
                                                                                               Rk0+1
                                                                                                  1 Hk+1 Pk+1 H T R01
                                                                                                                 k+1 k+1
              k+1 R01                 01         01
                    k+1 k+1 0 k+1 Rk+1 0 Rk+1 k+1                                                 = Rk0+11 Hk+1 Pk+1=k HkT+1(Hk+1 Pk+1=k HkT+1 Rk+1 )01 :
                   + Rk0+11 Hk+1 Pk+1 HkT+1 Rk0+11  0                            (31)                                                                                 (42)
and                                                                                            Therefore, under (33) and by the use of (41) and (42), it follows
                   FkT k   (
                            Fk Pk FkT   )01 k Fk 0 P 01  0:
                                                          k                       (32)
                                                                                             that (31) is satisfied.
                                                                                               Lemma 2: If we assume that
  Before we give sufficient conditions to ensure convergence of the                                            Fk is a bounded nonsingular matrix                      (43)
EKF, two lemmas are established for intermediate results.
  Lemma 1: If we assume that each ik+1 satisfies the following                              and each jk satisfies the following condition:
condition:
                                                                                                                01  jk  1       for    j = 1; 1 1 1 ; n             (44)
          1 0 1 0 1k+1 < ik+1 < 1 + 1 0 1k+1                                                then (32) is verified.
              for i = 1; 1 1 1 ; p                                                (33)           Proof: As k is a diagonal matrix, its eigenvalues are given by
with                                                                                         jk and verify
          k+1 = max (Rk+1 )max (Rk0+1
                                       1 Hk+1 P       T
                                               k+1=k Hk+1                                                                   k mj = jk mj                             (45)
                 1 (Hk+1 Pk+1=k HkT+1 + Rk+1 )01 )                                (34)       and
               ()
where max 1 represents the maximum eigenvalue and Rk+1 is                                                    mjT k = jk mTj      for    j = 1; 1 1 1 ; n            (46)
                     1            1
chosen such that k+1  ; then (31) is verified.
                                                                                             where mj is the associated eigenvector.
    Proof: As k+1 is a diagonal matrix, its eigenvalues are given
by ik+1 and verify                                                                            Under assumption (43), (32) is equivalent to
where (1) is the measure of matrix defined by where Fk and Hk are defined by (8) and (9), respectively, then the
                                      A + AT :
                                                                                                                                      ^              ^
                                                                                             system (and its associated EKF for xi=i01 and xi sufficiently close
                      (A) =         max                                                    to the true state xi ) is said to be reconstructible. Thus, we obtain the
                                                   2                                         following lemma.
584                                                                          IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 42, NO. 4, APRIL 1997
   Lemma 3: If we assume that (1a), (1b) is reconstructible (50),             Some Concluding Remarks
then we have                                                                        ik and jk are unknown factors introduced to evaluate the lin-
                              lim
                             k!1
                                    min (P 01 ) = 1
                                                k                     (53)
                                                                                     earity of the model and, consequently, to control the gain matrix
                                                                                     Kk+1   by an adequate choice of Rk in order to ensure stability
                                                                                                                                  =          ~
                                                                                     of the algorithm, particularly when ek+1 6 Hk+1=k xk+1=k and
and
                                                                                     ~          = ~
                                                                                     xk+1=k 6 Fk xk : Indeed, the sufficient conditions (33) and (44)
                                   max (Pk01 )                                      mean that fVk g is a decreasing sequence for all approximations
                          Sup klim
                               !1  (P 01 ) < 1:                      (54)
                                          min       k
                                                                                     in the form
       Proof: From (23) and under (43) it is easy to show, by induction,              ik+1 eik+1 = Hik+1 x~k+1=k        and   x~jk+1=k = jk Fjk x~k
that                                                                                 with
      P 01
       k+1   =   OeT   (1; k + 1)<(1; k + 1)Oe (1; k + 1) + 9(0; k)   (55)                    ik+1 2 ]1 0 1 0 1k+1 ; 1 + 1 0 1k+1 [
                                                                                      and        jk 2 [01; +1]:
with
                                                                                     As ik+1 and jk are unknown
                                                                                                              p factors, Rkp+1 have to be chosen
     9(0; k) = Fk0T Fk00T1 1 1 1 F00T P001 F001 1 1 1 Fk0011 Fk01: (56)              so that the interval ]1 0 1 0 1k+1 ; 1 + 1 0 1k+1 [ is large
                                                                                     enough to fulfil (33). An a priori choice of Rk+1 is to set it
Oe (1; k +1) and <(1; k +1) are defined in (51) and (52), respectively.              much higher than Hk+1 Pk+1=k HkT+1 : Notice that within the
  On a horizon time kM; if the reconstructibility Gramian
OeT (1; kM )<(1; kM )Oe (1; kM ) is decomposed into k block
                                                                                     maximum limits we have
                                    lim
                                    k!1
                                           Vk = V:                                                     IV. NUMERICAL EXAMPLES
                                                                                In order to show the efficiency of the proposed approach, we
  On the other hand, we have                                                  consider two numerical examples.
                                                01
                                          [Pk ]~xk x~k  0:
                                                                                Example 1: The first example concerns the identification prob-
                                    min
                                                            T
                           Vk                                                 lem of nonlinear time-invariant systems described by the MISO
                        tr(Pk01)      n [P 01 ]
                                            max         k
                                                                      (59)
                                                                              Hammerstein model. This kind of model is composed by a static
                                                                              nonlinearity followed by a linear dynamic system. Here we consider
According to (53) one obtains                                                 two interconnected subsystems (high order and high values of gik )
                                 lim tr(Pk01) = 1
                               k!1
                                                                      (60)    with the following pulse transfer functions:
                                                                                         V1k = g11 u1k + g12 u12k + g13 u13k + g14 u14k + g15 u15k
then
                                                                                         V2k = g21 u2k + g22 u22k + g23 u23k
                                min [Pk01 ]~xTk x~k                                           + g24 u24k + g25 u25k + g26 u26k
                            lim
                           k!1 n        [P 01] = 0                   (61)
                                          max       k                                    B1 = q01 + b11 q02
thus (54) yields to                                                                      A1 1 + a11 q01 + a12 q02
                                                                                         B2 = q01 + b21 q02 + b22 q03
                                    lim x~k = 0:
                                    k!1
                                                                      (62)
                                                                                         A2     1 + a21 q01 + a22 q02
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 42, NO. 4, APRIL 1997                                                                             585
where q 01 is the delay operator and uik is the ith input of the system
at time instant k; i = 1; 2; with
        a11 = 0:4;    a12 = :65;   a21 = 0:75; a22 = 0:9
        b11 = 0:5;    b21 = 00:6;    b22 = 0:7
        g11 = 5:2;    g12 = 02:0;     g13 = 5:2; g14 = 03:5
        g12 = 6:5;    g21 = 6:3;   g22 = 2:8; g23 = 00:02
        g24 = 3:1;    g25 = 02:3;     g26 = 5:6:
   Parameters of the static nonlinearity V1 and V2 are independent of
those of A1 ; A2 ; B1 ; and B2 :
   The invariant parameter vector x (i.e., xk+1 = xk ) to be estimated
from the input output data (yk ; u1k ; u2k )k=1;111 ; is defined as
         x = (a11 a12 a21 a22 b11 b21 b22 g11
                                                                                                         k 0 x kExample 1.
                                                                          Fig. 1. Rate of convergence of x
                                                                                                         ^k
              1 1 1 g15 1 1 1 g21 1 1 1 g26 )T 2 R18 :
                                                                                                                 k
convergence of the EKF may be improved significantly in the sense                     Dynamics and Convergence Rate of Ordinal Comparison
that the domain of attraction is enlarged. One of the main results in                          of Stochastic Discrete-Event Systems
this paper consists of introducing instrumental matrices k and k to
evaluate the linearity of the model in order to control both stability                                            Xiaolan Xie
and convergence of the EKF.
                                                                                        AbstractThis paper addresses ordinal comparison in the simulation
                            ACKNOWLEDGMENT                                           of discrete-event systems. It examines dynamic behaviors of ordinal
                                                                                     comparison in a fairly general framework. It proves that for regenerative
                                                                                     systems, the probability of obtaining a desired solution using ordinal
   The authors wish to thank the anonymous referees for their                        comparison approaches converges at exponential rate, while the variances
insightful comments.                                                                 of the performance measures converge at best at rate O(1=t2 ); where t
                                                                                     is the simulation time. Heuristic arguments are provided to explain that
                                                                                     exponential convergence holds for general systems.
                                 REFERENCES
                                                                                                                I. INTRODUCTION
 [1] J. S. Baras, A. Bensoussan, and M. R. James, Dynamic observers as
     asymptotic limits of recursive filters: Special cases, Siam J. Appl. Math.,       Optimization in discrete solution space becomes more and more
     vol. 48, no. 5, pp. 11471158, 1988.                                            important for discrete-event dynamic systems. The only general tool
 [2] D. Bestle and M. Zeitz, Canonical form observer design for non linear
                                                                                     for evaluating such systems is the simulation. A straightforward and
     time-variable systems, Int. J. Contr., vol. 38, pp. 419431, 1983.
 [3] M. Boutayeb and M. Darouach, Recursive identification method for               widely used approach consists of simulating all candidate designs to
     Hammerstein modelExtension to the nonlinear MISO case, Contr.                 obtain accurate estimation of the performance measures and selecting
     Theory Advanced Technol., vol. 10, no. 1, pp. 5772, 1994.                      the best design. Its main drawback is the requirement of a long
 [4]       , Recursive identification method for MISO Wiener-Hammerstein            simulation run to obtain accurate performance estimators. Variances
                                                                                     of performance measures converge typically at rate O(1=t) in time
     model, IEEE Trans. Automat. Contr., vol. 40, no. 2, pp. 287291, 1995.
 [5] J. J. Deyst and C. F. Price, Conditions for asymptotic stability of
     the discrete minimum-variance linear estimator, IEEE Trans. Automat.           t. and these rates are usually unsatisfactory when the number of
     Contr., pp. 702705, 1968.                                                      candidate designs is important.
 [6] A. Gelb, Applied Optimal Estimation. Cambridge, MA: MIT Press,                     Ordinal optimization approaches first proposed in [10] (see also
     1974.
                                                                                     [11] for an overview) reduce the computation burden by combining
 [7] A. H. Jaswinski, Stochastic Processes and Filtering Theory. New York:
     Academic, 1970.                                                                 some mind-set changes concerning the problem of optimization
 [8] A. J. Krener and A. Isidori, Linearization by output injection and             of discrete-event systems. The primary concern of the ordinal op-
     nonlinear observers, Syst. Contr. Lett., vol. 3, pp. 4752, 1983.              timization is the rankings of candidate solutions instead of their
 [9] A. J. Krener and W. Respondek, Nonlinear observers with linearizable           criterion values. Simulations conducted by various authors for a wide
     error dynamics, Siam J. Contr. Optim., vol. 23, pp. 197216, 1985.
[10] H. J. Kushner, Approximations to optimal nonlinear filters, IEEE              range of problems have shown that the rankings stabilize before the
     Trans. Automat. Contr., vol. AC-12, no. 5, pp. 546556, 1967.                   convergence of performance estimates. To achieve further reduction
[11] L. Ljung, Asymptotic behavior of the extended Kalman filter as a               of simulation time, ordinal optimization approaches typically relax
     parameter estimator for linear systems, IEEE Trans. Automat. Contr.,           the goal of optimization to the isolation of a set of good candi-
     AC-24, pp. 3650, 1979.
                                                                                     date solutions. Experiments indicate that it is possible to determine
[12] T. P. McGarty, Stochastic Systems and State Estimation. New York:
     Wiley, 1973.                                                                    whether a solution is good or bad very early in the simulation with
[13] R. K. Mehra, A comparison of several nonlinear filters for reentry ve-         high probability. Recent research [2], [3], [12] has demonstrated that
     hicle tracking, IEEE Trans. Automat. Contr., vol. AC-16, pp. 307319,          impressive improvement in computation efficiency can be achieved
     1971.                                                                           using ordinal optimization approaches.
[14] H. Nijmeijer, Observability of autonomous discrete-time nonlinear
     systems: A geometric approach, Int. J. Contr., vol. 36, pp. 867874,              The purpose of this paper is to provide theoretical evidence of the
     1982.                                                                           efficiency of ordinal optimization approaches. It is an extension of a
[15] Y. Song and J. W. Grizzle, The extended Kalman filter as a local               recent work [4] which considers the convergence of the probability
     asymptotic observer for nonlinear discrete-time systems, in Proc. Amer.        that the best observed design is indeed a good design. We consider
     Contr. Conf., 1992, pp. 33653369.
                                                                                     the following fundamental indicator: the probability that at least k of
[16] Special issue on applications of Kalman filtering, IEEE Trans. Automat.
     Contr., vol. AC-28, no. 3, 1983.                                                the observed top-s designs are the actual top-g designs (i.e., satis-
[17] M. Vidyasagar, Nonlinear Systems Analysis, 2nd ed. Englewood Cliffs,            factory designs). It is called alignment probability and is meaningful
     NJ: Prentice-Hall, 1993.                                                        in the perspective of simulation budget allocation. Our contributions
[18] G. Ciccarella, M. Dalla Mora, and A. Germani, A robust observer for            include:
     discrete time nonlinear systems, Syst. Contr. Lett., vol. 24, pp. 291300,
     1995.                                                                              1) monotonicity properties of the alignment probability with re-
[19] P. E. Moraal and J. W. Grizzle, Observer design for nonlinear systems                 spect to s; g; and k are established. They show that goal
     with discrete-time measurements, IEEE Trans. Automat. Contr., vol.                    relaxation improves the computation efficiency;
     40, no. 3, 1995.
                                                                                        2) the association, a kind of positive correlation, of simulated
                                                                                            systems improves the convergence rate of the probability that
                                                                                            the observed best design is the actual best one;
                                                                                        3) informal arguments of the exponential convergence rate of the
                                                                                            alignment probability in the general case;
                                                                                       Manuscript received July 14, 1995; revised June 12, 1996.
                                                                                       The author is with INRIA, Technopole Metz 2000, 57070 Metz, France.
                                                                                       Publisher Item Identifier S 0018-9286(97)02044-8.