LlNEAR AND NONLINEAR DESCRETE FiLTERiNG
FOR CONTiNUOUS SYSTEMS
Thesis for the Degree of Ph, D.
MICH§GAN STATE UNNERSWY
MARVIN A. NEEDLER
:97;
fHESl9 .. LIBRAP .
, -'Michigan State
University
This is to certify that the
thesis entitled
LINEAR AND NONLINEAR DISCRETE FILTERING
FOR CONTINUOUS SYSTEMS
presented by
Marvin A . Needler
has been accepted towards fulfillment
of the requirements for
X4244? f/gf/ I
Major professor
Date February 12, 1971
O~169
ABSTRACT
LINEAR AND NONLINEAR DISCRETE FILTERING FOR CONTINUOUS SYSTEMS
by
Marvin A. Needler
The subject of this thesis is the synthesis of discrete-time filters
for real-time estimation of the state of continuous-time dynamic systems.
The dynamic system is modeled by a set of stochastic differential
equations and the stochastic disturbance is modeled as a Brownian motion
process. The filtering process is assumed to utilize en-line digital
computation and thus a discrete-time recursiVe filter is developed to
estimate the state of the system based on sampled-data measurements.
Starting with the equations for propagation of the probability den-
sity function of the state conditioned on previous measurements, the ’
filtering equations are developed both for linear and nonlinear filtering.
In the linear filtering case, two new methods are suggested for
reducing computation and increasing accuracy. These methods are illus—
trated in the digital simulation of state estimation of a steam turbo-
generator system. In addition, the methods of divergence compensation
are reviewed.
In the nonlinear filtering case, a second—order filter is suggested,
and it is used to estimate the state of a reentry body. An original
Marvin A. Needler
adaptive filter is mechanized to reduce filter error and is shown to
improve the estimation accuracy of the linear and nonlinear filter in
reentry estimation. The adaptive filter adapts the filter gain to the
statistics of the measurement estimation errors without requiring
significant computation and.without altering the state covariance com-
putation. Based on the reentry estimation problem, the adaptive filter
produces an improved state estimate that approaches the optimum estimate
for the given filtering problem.
LINEAR.AND NONLINEAR DISCRETE FILTERING FOR CONTINUOUS SYSTEMS
By
Era:
Marvin A: Needler
A THESIS
Submitted to
‘Michigan State University
in partial fulfillment of the requirements
for the degree of
DOCTOR OF PHILOSOPHY
Department of Electrical Engineering and Systems Science
1971
ACKNOWLEDGEMENTS
I would like to express my gratitude to my major professor, Dr.
Gerald L. Park, for the guidance and inspiration that he provided
throughout my doctoral program. In addition, I am deeply grateful to
the other guidance committee members, Dr. Richard C. Dubes, Dr..J.
Sutherland Frame, Dr. William L. Kilmer, and Dr. John B. Kreer, for
their invaluable contributions to my doctoral program. .A special note
should be made of the joint effort provided by Dr. Frame in developing
Section 3-2 of this thesis.
Support was granted by Consumers' Power Company Cooperative Research
Project athichigan State University for one summer and by the National
Science Foundation for one year. This support is gratefully acknowledged.
In addition, I would like to thank Mr. Howard Bell, who introduced
me to filtering theory. Also, I am indebted to Mrs. Ann Darrenkamp for
typing the final manuscript. And finally, the original manuscript was
typed by my wife, Priscilla, whose support made this thesis possible.
ii
TABLE OF CONTENTS
Chapter I Introduction
Chapter II Analysis and Development of Filtering Theory
2-1.
2-2.
2-3.
2-4.
The System.Model
Extrapolation of the Conditional Density Function
and Its Moments
Update of the Conditional Density Function and
Its Moments
The System Disturbance Model
Chapter III Discrete Filtering of a Linear Continuous System
3.1.
3-2.
3-3.
3-4.
Linear Filtering
State and Covariance Extrapolation
An Example of State Estimation of a Steam Turbo-
Generator System
The Problem of Filter Divergence
Chapter IV Approximate Filtering of Nonlinear Systems
4-1.
4-2.
4-3.
4-4.
NOnlinear Filtering
An Example of a NOnlinear Filtering Problem
Adaptive Filtering for NOnlinear Systems
A Discussion and an Example of Adaptive anlinear
Filtering
iii
10
11
15
16
18
24
28
33
33
37
44
49
Chapter V Conclusions
5-1. Thesis Review
S-2. Results and Conclusions
5-3. Suggestions for Further Research
LIST OF REFERENCES
APPENDIX A
APPENDIX B
APPENDIX C
iv
57
'57
58
60
61
69
72
74
Figure
Figure
Figure
Figure
Figure
Figure
Figure
3-1.
3-2.
4-1.
4-2.
4-3.
4-4.
4-5.
LIST OF FIGURES
The Block Diagram of a Steam Turbo-Generator System.
The RMS Estimation Error of Frequency Deviation
for the Steam.Turbo-Generator.
.Altitude Estimation Error for the First-Order
Filter of the Reentry Estimation Problem.
.Altitude Estimation Error for the Second-Order
Filter of the Reentry Estimation Problem.
.Altitude Estimation Error for the First-Order
Adaptive Filter of the Reentry Estimation Problem.
Altitude Estimation Error for the Second-Order
Adaptive Filter of the Reentry Estimation Problem.
RMS Altitude Estimation Error for Four Similar
Filters.
27
29
42
43
51
S4
I IAHRODUCTION
The subject of this thesis is the synthesis of discrete-time
filters for real-time estimation of the state of continuous-time
dynamic systems.
Historically, the theory of estimation dates back to the least-
squares orbit estimation of Gauss [GAUl] in 1809. More recently,
Wiener [WIEl] and Kblmogorov [KDLl] derived solutions for linear least-
squares estimation. 'Wiener studied continuous-time systems using
spectral methods to obtain minimum-variance unbiased estimates of the
scalar state by observing a signal where both the state and the obser-
vation are jointly wide-sense stationary and ergodic processes.
Kblmogorov used a recursive orthonormalization.procedure to solve the
analogous discrete-time problem.
In 1960, Kalman [KALI] published a recursive solution for the
minimum-variance unbiased estimate of the state of the Markov vector
process of a discrete-time system where the system is disturbed by a
random vector and the state is observed through a non-invertible trans-
formation. Kalman and Bucy [KAL2, KALS, BUCI] and Stratonovich [S'I'Rl]
extended this work to the continuous-time filter and incorporated the
problem of observations contaminated with additive random noise. In
addition to the filtering problem (estimating concurrently in time to
the observations), solutions were developed for the prediction problem
(estimating into the future) and the smoothing problem (estimating
after the time of the observations).
During the decade following the derivation of the Kalman-Bucy
filter, much effort was involved in further refinement and develOpment
of filtering theory. MDch of the motivation comes from optimal con-
trol theory where it was feund that a simple deterministic model would
no longer suffice for a realistic model of a system that is contami-
nated by noise and whose state can not be directly observed IBELI,
SARI]. The Kalman-Bucy filter was modified to serve in aero-space
applications for guidance, navigation, and orbit and reentry estimation
[SMIl, PURl]. In addition, applications in industrial process control,
electric power systems and system identification have been developed
[SAR1, CHAl]. More recently, applications in urban and freeway traffic
flow have received attention.
In a large class of optimal control problems, namely, for linear
systems with quadratic performance index and with Gaussian noise, the
optimal control law is a feedback law, which is a computed gain times
the state [BRY1, LEEl]. In this case, it can be shown that the com-
bined optimal estimation and control scheme is the optimal estimator
and the optimal controller found independently--this principle is known
as the separation principle [JOSl, POTl, WONl]. In the nonlinear case
and/or the non-quadratic performance index case, there is no known
separation theorem; nonetheless, the separation principle provides a
reasonable, approximate design approach.
Other approaches to the optimal control problem with inaccessible
states have been developed. In particular, the methods proposed by
Pearson [PEAl] and Ferguson [FERl] obviate the need for a filter. How-
ever, it should be noted that these particular methods are restricted
to optimal linear regulators and do not consider the case of noisy
measurements or system disturbances.
Linear filtering has been rather completely exploited. Since the
appearance of Kalman's original paper using the orthogonal properties
of random vectors to minimize the error criterion, other methods such
as least squares, maximum likelihood, minimum variance, calculus of
variations, and the maximum principle [BRY1, ATHl, H01] have been used
to prove the form of the optimal filter. Other problems such as non-
zero mean, cross-correlated, and auto-correlated system and measurement
noise have been considered [BRYZ]. Filter stability and convergence
have also been given considerable attention. In addition, the Kalman-
Bucy filter has been extended to the case of nonlinear systems, called
the extended Kalman filter, first-order filter, or the quasilinear
filter [FRIl, MOWl, SCHZ, H02].
The problems remaining in linear filtering are those that involve
mechanizing a filter that uses a suitable model, i.e., does not have
detrimental modeling errors, and that is feasible in terms of computa-
tional requirements.
Nonlinear filtering theory is not as completely developed as linear
filtering theory due to the inherently more complicated structure of
nonlinear systems where superposition does not hold. By using a
probabilistic approach, it is possible to derive the equations governing
the evolution of the conditional probability density function of the
state conditioned on the sequence of observations. It has been shown
that, in general, the optimal nonlinear filter requires an infinite
number of moments to realize the conditional probability density func-
tion [KUSl, KUSZ]. However, a finite approximation can be made that
may be satisfactory [KUSS, BUCZ].
~Another problem in nonlinear filtering is finding a global filter-
ing theory that will assure filter convergence. No general convergence
theory has been found that is equivalent to that in linear filtering
theory. Indeed, the problem is so difficult that proofs of filter
convergence are restricted to the scalar case without system distur-
bances [BIRl] . A
An approach is suggested in this thesis that improves the con-
vergence properties of both the linear and nonlinear filter whenever
modeling errors are prevalent. Previous work [DEN1, PINl, SMIl] has
been concerned with the problem of filter divergence. The idea of
adaptive filtering has been suggested by Jazwinski and Dennis [JAZZ,
DENZ], whereby an artificial system disturbance is constructed to
prevent filter divergence. In this thesis a new approach is suggested
to adapt the estimates to the measured statistics of the filtering
problem; namely, the filter gain matrix is corrected to provide con-
sistent filter statistics and measurement statistics.
In many problems, it is necessary to estimate the state of a
vector-valued continuous-time system. Often, it is not possible to
measure continuously, and even if continuous-time measurements are
available, only sampled data can be used to allow for the computation
time that is necessary even in high-speed digital computers. Thus, the
structure that emerges is a discrete-time filter for the continuous-
time system. This structure has some of the aspects of both continuous
problems and discrete problems.’ For the most part, the discrete-filter
problem is a sub-problem.of the discrete-filter continuous-system prob-
lem, whereas the continuous filter problem must account for the continu-
ous observation process.
One important problem involved in a continuous-time system is that
of modeling the random.processes. .Assuming the process to be continuous-
valued and composed of stationary independent increments leads naturally
to the Brownian motion process. It shall be assumed that this process
has a Gaussian distribution function with zero mean. In order to solve.
differential equations involving the Brownian motion process, the Ito
stochastic integral is employed.
To summarize the previous discussion, the purpose of this thesis
is to study the synthesis of filters. The discrete-time filter,
continuous-time dynamic system is emphasized. In Chapter 1, the general
filtering equations are developed, and some results are derived for simu-
lation of the stochastic processes. In Chapter 2, the linear filter
equations are given, some new computational methods are presented for
reducing computation'and increasing accuracy, these methods are applied
to a simulation example and finally, the problem of modeling errors is
discussed. In Chapter 4, a nonlinear filter is derived and used in a
simulation example. Finally, a new statistical method is developed that
adapts the nonlinear filter to modeling errors to provide consistent
elements for the nonlinear filter.
II. Analysis and Development of Filtering Theory
underlying the Subject of filtering theory is the theory of
probability, random variables, and stochastic processes. The central
principles of stochastic processes that are needed here are reviewed
in.Appendix.A. (See [FEL1, RAPl, D001], for example.) In this chapter,
the equations of filtering theory are developed for discrete-time
filtering of continuous-time, discrete-parameter, dynamic systems.
2-1. The System Model
A large class of dynamic systems can be modeled by the vector-
valued stochastic differential equation,
dx = f(x,t)dt + G(x,t)dB (2-1-1)
or its formal white-noise equivalent,
i = f(x,t) + G(x,t)W’ (2-1-2)
where x is an arrandom vector, 8 is an m-vector Brownian motion process,
f is a vector function, and G is a.matrix funCtion. The prOperties of
the Brownian motion process, its relation to the white-noise process
w(t), and its physical significance are given in [D001].
The existence and.uniqueness of the solution to (2-1-1) over an
interval T, namely,
x sz f(x,t)dt + fr G(x,t)dB (2-1-3)
(where p the integrals are interpreted in Appendix A) can be proven under
thelconditions that f and G are Lipschitz in x. (See [1101 or D001] .)
The Linear System Dynamic Model
In the case that f and G are not state dependent, the system model
is linear,
dx = A(t)x dt + B(t)dB, (2-1-4)
or
x = A(t)x + B(t)w (2-1-5)
where A is an n X n matrix, B is an n x m matrix, and B is a Brownian
motion process. The linear model is of great importance for several
reasons, one of which is that the solution can be written uniquely as
t
x(t) = acntpxcto) +ft accomodem (2-1-6)
0
where 6(t,to) is the transition matrix transforming the unforced
response over the interval [to,t] and the forced response is a stochas-
tic vector integral. If x(to) is taken as a Gaussian vector indepen-
dent of B, then x(t) is a Gaussian vector, and the covariance matrix
of the state P(t,s) can be written as
P(t,s) = E {x(t)x'(s)} (2-1-7)
= 6(t,to)P(to)d'(s,to) +
min(t,s) .
ft Manama-03' (or (mm-r
o
where min(t,s)
EiKS)B'(t)} =ft Q(r)dr . (2-1-8)
0
The Observation Model
The general form of the continuous observation model is the
stochastic differential equation
dy = h(x,t)dt + dv (2-1-9)
where y is an IL-dimensional observed quantity, h is a vector function
and v is a Brownian motion vector:
min(t,s)
E {v(t)v'(s)} = ft R('r)dt.
O
In the case of discrete-time measurements, the analogous model
of (2-1-9) is
y(th) = h(x(th), th) + v(th). ‘ (2-1-10)
The additive noise v is a white-noise sequence with covariance
B{v(th)v'(t£)} = R(th) 6 kt (2-1-11)
where 5M is the Kronecker delta function: 5a: l for ~12 = fl and
6M. = 0 for I2 f L In the case that h is not state dependent, (2-1-10)
may be written-as
y(th) = M(th)x(tk) + v(th) (2-1-12)
where “M is an IL x n matrix.
For the linear observation model and system model, a necessary
condition for estimating the state of the system is that the system
be observable [KALZ]. No equivalent theory exists for the general
nonlinear problem .
2-2. Extrapolation of the Conditional Density Function and Its Moments
The Markov, Brownian motion process x(t) satisfying the Ito
stochastic differential equation (2-1-1) has a probability density
function p(x,t) that is conditioned on the previous observations,
‘Y(tk) =‘{y(t1), ..., y(tk)}. Between measurements this conditional
probability density function, called the transitional probability
density function, p(x,t|Y(th)), is given by the Kolmogorov forward
equation,
, n n
épéiiil.= - g alp(x.t)fi]. §.z z azlpcx.t)[GQG']ij] 2-2-
t i=1 TXi i=1 i=1 ( 1)
BXiBXj
where fi is the ith_component of f and [GQG]ij is the ijth_element of
GQG'.. For a derivation and discussion of this expression and diffusion
processes in general, see [FELL 1101, DYNl, 1102, M310]. The initial
condition for (2-2-1) is the value at the prior measurement,
P(X:thl YCthD-
By the use of Ito‘s Theorem.A-3, the exact equations for the
extrapolation of the moments of p(x,t) between.measurements can be
determined. First, define the expected value of o in (A-ll) as
$CXCt)lth)) = E {¢(X)| Y(th)}=;(t(X) plx.t lY(tk)] dx (2-2-2)
Then the probability density, p(x,t|Y(th)) satisfies (2-2-1), and the
extrapolation of ¢ is
d$(x(t)lth) = wai-fmthn + §tr E {GQG'¢HIY(th)}dt,
th5t
0.
The contribution of the disturbance in (2-4-5) is
V(th) = mm B(t)u(tk)u'(th)B'(t)\/At}
a B(tk)Q(tk)B'(th) At (2-4-7)
and therefore, for small enough value of At, (2-4-5) is equivalent at
each At interval to its continuous-time counterpart (2-1-5). This
approach may be considered where a sample function is desired with a
larger number of samples than in (2-4-2) and where the computation of
(2-4-4) is not necessary.
For a more general utilization of the property of discrete-time
modeling of continuous-time processes, consider (Z-l-l),
dx = f(x,t)dt + G(x,t)dB.
The discrete-time model is written as
x(t + At) = x(t) + f(x(t),t)At +\/A—t G(t)u(th) (2-4-8)
where u(th) is a white-noise sequence with covariance matrix Q(th)°
The contribution of the disturbance in (2-4-8) is the same as in
(2-4-7).
If the covariance matrix of the 8 process is given by (2-1-8),
then by Theorem.A-2,
E{ G(x,t)d3}= 0 (2-4-9)
At
14
and
E{j;tG(x,t)dB[ fAtG(x,t)dB]' =
I G(t,x)Q(t)G' (t,x)dt (2-4-10)
At
C(t.X)Q(t)G' (tut) At
for small At (i.e., the interval At must be small with respect to the
time rate of change of G and Q). Once again, the simulation generator
of this Gaussian multivariable disturbance is given in Appendix B.
III. Discrete Filtering of a Linear Continuous System
Since the conception of linear filtering theory in 1960, a great
amount of effort has been put forth, both in setting the theory straight
and in applying the theory to practice. 'Within a relatively short
period the theory was applied in space guidance and navigation programs
such as Ranger, Mariner, and Apollo. In addition, applications have
'been made of linear filtering theory to aircraft guidance, industrial
process cohtrol, and other detection problems. In most cases, it is
necessary to linearize the system differential equations about the
nominal state in order to apply a linear filter.
The explicit problems of deriving the optimal linear filter,
showing its stability, and proving asymptotic convergence are solved
[KAL1, KALZ, KALS, BUCl]. The problems remaining in linear filtering
theory are related to process modeling and filter mechanization.
The first part of this chapter is concerned.with developing the
equations for discrete-time filtering of the continuous-time system.
Next, some methods that have not appeared in the literature are
presented for mechanizing the linear filter. Then an example of an
application of linear filtering theory to a steam-turbine generator is
presented. Finally, a method for compensation for modeling errors is
presented.
15
16
3-1. Linear Filtering_
The linear set of differential equations that describe the linear
or linearized system may be written as
x = A(t)x + B(t)w I (3-1-1)
where A(t),is an integrable n x n matrix, B(t) is an integrable n X m
matrix and w is an m-vector white noise process with covariance
Q(t)6(t-r). .Although the white noise process is only a formal repre-
sentation of (2-1-4), its use is justified since in the solution of
(3-1-1), w(t)appears in the integral only and its integral, the
Brownian.motion process, is a well-defined function. The linear
?
observation process consists of sampled data of the form
V(tk) = MCtthCth) + V(tfz) . (3-1-2)
wherelM(th) is an.t X n linear non-invertifile transformation and v(th)
is an t-vector of a Gaussian white noise sequence with covariance
R(th). It is assumed that the system is observable.
The linear filtering problem is this: given initial estimates of
the state x(to) with Gaussian distribution and of P(to), the covariance
matrix, determine the mdnimum-variance recursive estimate of the system,
(3-1-1) and (3-1-2). Slightly different premises are assumed for the
discussion in Section 3-4.
With the modifications given in Section 2-4, the linear discrete-
time filter for the linear continuous-time system fits the framework
as derived by Kalman [KALl]. For the linear problem, the process x is
Gaussian and only the first and second moments of (2-2-1) and (2-3-1)
are needed, namely, the linear version of (2-2-4) and (2-2-6) for
estimation extrapolation and (2-3-4) and (2 3-5) for estimation update.
17
The linear filtering equations are given here without proof since
the derivation is commonly found in the literature and only a few steps
are required to reduce the above equations. Between measurements, the
state estimate and covariance matrix are
x(tkltk_1) a 0(th,th,1)x(th_1]th_1) (3-1-3)
P(tk|th_1) = 0(th,th_1)P(th_1Itk_1)¢'(tk,th_1)
+- fth w(th,t)'B(t)Q(t)B' (tw (thaw (3-1-4)
th_1
where o is the system transition.matrix. At the time of the measure-
ment, the state estimate and covariance matrix are
sent.) = ictklth_1) + Gummy.) - Mctkfiohltmn (3—1-5)
P(thlth) = P(th|tk-1) - G(th)M(th)P(thlth_1) (3-1-6)
where the filter gain is
ccth) = P(t,,lth_1)M'(tk)[M(tk)P(tk|th-1)M'(th) + Rctpl’l.
(3-1-7)
These five equations are the optimal linear filtering equations.
These equations are closely related to the maximum likelihood estimate
and the stochastic approximation method as shown by Ho [m1] .
When the system parameters are known exactly the solution of the
linear filtering equations is straightforward. However, the real-time
computer requirements may be exorbitant and unfeasible. The methods
discussed in the next section due to Frame and Needler [FRAZ] are
useful in decreasing computational time and increasing accuracy. In
the case that the system parameters are not known, the methods dis-
cussed in Section 3-4 can be used to correct for modeling errors.
18
3-2. State and Covariance Extrapolation
Minimizing computational time and maximizing computational
accuracy are crucial problems in.mechanizing an on-line filter for
filtering a complex process since these two criteria constrain
the filter capability in terms of (1) setting a lower limit on
sampling rate, (2) allowing local iterationsifor linearization of the
extended linear filter, and (3) allowing the system model to be suf-
ficiently complex to adequately model the physical system.
Two important aspects in both real-time system estimation and
control and in digital simulation of the same are extrapolation of the
state of the system.and extrapolation of the covariance matrix of the
system. In the case that the system may be described by (3-1-1) where
A and B are either time invariant or can be approximated as time
invariant over the time period of interest, then the computational
methods described in the sequel are useful in state and covariance
extrapolation between measurements.
The solution of (3-1-1) can be written as
t
x(t) = @(t-to)xo + Jr ¢(t-t)Bu(t)dr (3-2-1)
0
where the transition matrix ¢ is
(t) = e“ = [235 Ahth/Iz! (3-2-2)
and to is taken to be zero.
Using the transition matrix to evaluate the transient response of
a linear time-invariant unforced system was described by Liou [L101]
and others [BRA1, MMSI]. Ganapathy and Rao [GAN1]* pointed out that
*For corrections to [GANl], see [RAOl].
19
the A.matrix has to be raised to successively higher powers to allow t
to be taken over a sufficiently large interval. They in turn show that
eAt can be expanded in an infinite series of scalars instead of matrices
and that.An'1_is the largest power of.A.that is required.
An alternative method requires slightly less computation and is
possibly more straightforward than [GANl]. This new method utilizes
the conjoint algorithm to compute the coefficients d1,...,dn of the
characteristic polynomial of A,
n-l +
n
A + dll . + d A + dn = 0 (3-2-3)
n-1
and the constant coefficient matrices, B0,...,Bn of the conjoint
matrix, defined by
n-1 n-2
B(A) = adjoint (A1 -.A) = B A + B A + ... B A+-B
° 1 n-2 n-1
(3-2-4)
where Bb = I and Bn = B_1 = 0. The conjoint algorithm described by
Frame [FRAl] yields the relationships,
Bk = AB}?1 + th (3-2-5)
db = -(1/h)tr.A Bh-l (3-2-6,
Define the transition matrix as
n-l
d>(t) = z B,z,(t) (3-2-7)
j=0 J J
where
0° j+h
z. t = z c 3-2-8
1( ) h ( )
12-0 (14-12)!
20
and the ck are determined so that
m
2: dc .=0form>0;c =d =1. (3-2-9
i=0 jM'j O O )
Then s1nce Bn=B_1=0, and zj==zj._1 forj>0wehave
, n
¢(t) - 2 B z (t) (3-2-10)
' i=0 1 1
n
= 2 AB z(t)+ )3 doIzj(t)
j=1 j-lj i=0 j
n d n 00 m
=AZB. z. t+I-—2 Zd-c .t/m! 3-2-11
ill 1"]. 1." ‘1( ) dt jao im=j J m'J ( )
or
5(t) =.A ¢(t) + I - o =.A ¢(t).l (3-2-12)
Also it should be noted that
n-l
(0) = 2 BJ -jz (0) = BO 2 0(0) = B Gee; I (3-2-13)
j=0
Hence, we have the desired result:
¢(t) = eAt (3-2-14)
From equation (3-2-9) we solve for each Cl: using at most n multiplica-
tions:
m .
cm 121 djcm_, -j’ ms n
(3—2-15)
n
cm = - Z d c , m 2 n
j=1 Jm'J
By substituting these values of c into (3-2-8) and then solving
(2
(3-2-7), 9 (t) can be determined easily
21
The computational savings are considerable when compared with
using (3-2-2) truncated at some value b = N for N large compared to n.
On the other hand, when compared to the method utilizing the Hamilton-
Cayley theory [GANl], the savings in computation are less. In par-
ticular, the exponential series requires (N-l) n3 + N multiplications,
the Hamilton-Cayley method requires n4 - n3 - %n2 + %n + ZnN + N
multiplications, and the above method requires n4 - n3 - an + n
+ ZnN + N multiplications. The number of divisions is negligible and
additions are proportional to multiplications. Computer results con-
firmed the method and the previously mentioned computational savings;
for the example discussed later in this chapter, the computation was
about 30% as great as the computation of (3-2—2).
Covariance Extrapolation
In the case that u in (3-2-1) is a vector-valued, zero-mean,
white-noise process, then to determine the covariance matrix of the
error in x(t), it is necessary to evaluate*
t
V(t) =I ¢(t-t)Q'(t-t)dr (3-2-16)
0
where
Q = E{B u(t)u' (t)B'}.
*This same quadratic form also appears in optimal control problems
with a quadratic performance index over a finite time interval. See
[1.1351] and [LEVI].
22
As pointed out by Levis [LEVI], errors are present from two sources:
truncation of the series expansion of ¢(t) in (3-2-2) and approximate
numerical integration.methods. In the following, the integral is
solved exactly and then a recursive relationship is derived to solve
for V'to the desired accuracy.
Theorem The solution of (3-2-16) where Q is a symmetric matrix is
given by
w t[2+1
V(t) = 2 G}, (3-2-17)
k=0 (h+l)!
where
Gh+1 = AG!2 + Gh'.A' (3-2-18)
and the value of Go is
G =Q. (3-2-19)
Proof: By substitution of variable, (3-2-16) can be rewritten as
V(t) = th @(t)Q¢'(r)dr (3-2-20)
0
After substituting (3-2-2) into (3-2-20) and integrating, the integral
becomes
m m . j+k+1
V(t) = )3 z AjQA'h t (3-2-21)
j=0 h=0 j!h!(j+h+l)
on k - _' [2+1
= z Z (kgAJQA',2 j t (3-2—22)
h=o i=0 1 (124-1)!
23
If G!2 is defined as
h .
(3,2 = z ( k)AjQA'M (3-2-23)
j=0
then
h1_.
GkA' = z (j)AjQA'M ,(3-2-24)
i=0
Replacing j by j-l in (3-2-23) it follows that
k+1h
Gk = 2 (j -1)A11QA'
j=l
k+1’1 (3-2-25)
and multiplying by.A gives
h+1 j h+1-j .
AG = 2 ('2 )A QA' (3-2-26)
k j= —l J 1
Thus, the sum of (3-2-24) and (3-2-26) is
Iz+l
G = (
h+1 j_0
h+1)AjQA'M 1'1 (3-2-27)
and by the symmetry property of Gk?
Gh+1 = AG,2 + (AGh)'. (3-2-28)
Thus, V(t) can be given exactly by
'0. tl2+1 -
V(t) = Z Gk
h=0 (h+1)!
(3-2-29)
where Gk is given by (3-2-25) and G0 = Q.
Although the benefits from utilizing this method cannot be evaluated
analytically, the increase in accuracy and the savings in computation as
compared to numerical integration are considerable in the case of large
24
t where it is necessary to have a large number of terms in the expansion,
as was realized in the following problemu *
3-3. An Example of State Estimation of a Steam Turbo-Generator System
A steam turbo-generator system is composed of a control system, a
steam boiler plant, and a turbo-generator. Disturbances occur to the
system in terms of a changing load caused by both changing consumer
power demands and changing power supplied by alternate sources. A
model can be constructed of the steam turbo-generator system by as-
suming that the coupling to other sources is relatively loose. See for
example, Kirchmayer [KIRl], Park [PAR1, PARZ], and Stanton [STAl]. Then
the system can be linearized around its operating point to obtain a
simulation model. The problem at hand is to estimate the state of the
system using this model. The model can be written in the form of
(3-1-1) and (3-1-2):
x = Ax + Bu
y = Mx + v
where in (3-1-1)
-1/T1 o o o Kl/Tl
1/T3 -1/T3 o o o
A = o /T3 -1/T o o ,
o KZ/i‘4 (1- eel/$14 -l/T4 o
L o -D/M l/M
0 ‘1
o
B = o ,
o
L 1 .1
*In the following example, less exact numerical integration required
approximately twice as much computation time as the above method.
25
and u is a random disturbance. The reference input is neglected for
purposes of this study. An explanation of the physical constants
appearing in the A matrix and M matrix as well as their numerical
values are given in Appendix C.
From a study of the random disturbances caused by changing loads,
it was determined that a random-walk.process would be an acceptable
model of this phenomenon. Since it can be shown [PAPI] that the
random-walk process is a limiting form of the Brownian motion process,
the Brownian motion process observed at discrete-time intervals will
exhibit the stochastic characteristics of the random-walk process.
Thus, in turn, the noise source can be pre-whitened by integrating a
white noise source to yield the Brownian motion process. The addition
of one state variable is necessary to account for the disturbance. The
state can be augmented with the differential equation,
16 =‘w
where W'is a white noise source. The augmented A matrix and B matrix
now become
”-1/T1 o o o icl/Tl T
1/T3 -l/T3 o o o
A = o l/T5 -l/T5 o o ,
o
o
o
K/T (1-K2)/T -l/T o o
2o 4 o4 -D/M4 l/M l/M
o o o o o
J
HOOOOO OCO
26
In Fig. 3-1, a block diagram.is shown of the overall system where each
component is represented by its Laplace Transform system function.
An acceptable model of the observation process depends on the
assumptions concerning the measurement devices. One choice is to
observe frequency deviation, x5, and frequency rate deviation, x5.
Thus, the observation matrix is
M=[0 o o l/M D/M 1m]
0 0 0 0 l 0
The noise source V(tk) can be modeled as a two-vector white-noise
sequence since the measurements are refined to the point that bias and
other gross errors are eliminated. A
The digital simulation'was performed using parameter data from
[PARl]. In addition, representative values were taken for the measure-
ment noise and system disturbance statistics. These values are given
in Appendix C.
The digital simulation of the steam generating plant including
the random effects, and the discrete-time filter demonstrated the
effectiveness of filter applications in problems of this type. By
using the filter as mechanized in Sections (3-1) and (3-2) it was pos-
sible to observe the estimation error time-history for various noise
sample functions. An improvement on simulating sample runs is to
study the estimation error dispersion by using the filter covariance
matrix of estimation error, starting with a large initial variance.
In this way, the need for performing Mente Carlo studies is obviated.
27
.ocwnnzu one nououocom onu we a
eofiuanw new 2.onuuoen mo peoEoe one new monsoooo nooHn Hocnw one .HHoEm ma osfio> Hocaeo: one
scum cofiuon>ou one an noocna mo wouoEnxouooo on xoe nonnon Emoum eeo_oono> eooum one .oenm>
Hmcneoe on» snow newuon>op xocoooohm one on ma .nofioonuuoo on .onon: Eoumxm one we moanofinm>
oumum.onu poo .mowconu omen genome an oomnoo 3 ooconnoumno eoecon onu .3 oocouomou women
eoHHoupcou one one moanowuo> one .Eoumxm nouonocou-onhoe eoopm m we Emumofim noon one .H-m .mnd
..L
n
$5135; ”5... .5 + _
m apex 1 . .x _ a
8.2425 . mamz: - moss. , 4 239m ,
235 mzaca. usammq 03mm .6528
.—|m
28
The tradeoff between improved filter performance and the following
factors can be evaluated: . I
l.~ The required sampling frequency which limits modeling complexity,
determines the maximum.computer cycle time, and determines measure-
ment transducer required sampling rates.
2. The quality and quantity of measurements that are required.
3. The effect of system disturbances.
Many simulation runs were performed while varying model parameters,
Such as sampling rate, the measurement accuracy and the magnitude of
the disturbance variance. A representative error-study time-history is
shown in Fig. 3-2.
In summary, the application of a Kalman filter to a steam-turbo
generator plant was simulated and the results were judged to be suc-
cessful. The methods proposed in Section (3-2) were used. The study
should be extended by evaluating the effect of (l) more refined model
development, (2) more exact noise and disturbance statistics, and
(3) application of a closed-loop optimal controller for minimizing
output frequency error. An alternative approach to improved process
modeling is presented in the next section.
3-4. The Problem of Filter Divergence
Early in the study of filtering applications, it was found that
although in theory the filter estimates converged asymptotically to the
state variables, in practice the filter estimates would diverge from
the state variables, especially where modeling errors were large relative
PER UNIT FREQUENCY ERROR x 10"
-6.0"‘
8.0- 7'9
7.0-
5.0'
4.0-
3.0‘
2.0-
' l l I
0 IO 20 30
TIME (SECONDS)
Fig. 3-2. The RMS Estimation Error of Frequency Deviation f or the
Steam Turbo-Generator. Starting with a very large error, the
RMS value converges relatively quickly to a steady-state error.
30
to system disturbance statistics. This outcome may be a consequence of
modeling error in any of the parameter values or covariance specifica-
tions [SCHl]. Since the filter has no measure of the consistency of
the estimation error covariance and the actual estimation error, it 15
possible for the covariance matrix to indicate a much more accurate
estimate than is actually being performed. The filter divergence is a
result of later observations being very lightly weighted since the
covariance matrix of the estimation error has converged to a smaller
quantity than warranted. .
Several approaches have been tried on the problem of filter
divergence. One approach is to augment the state vector with the model
parameters that are only approximately known [SMIl]. By estimating
these uncertain.parameters, this cause Of filter divergence is reduced.
The accompanying disadvantage is that a larger state vector leads to
greater computational requirements and_the possibility of establishing
an unobservable system. Another approach is to artificially include
a sufficiently large system disturbance covariance such that the filter
will not converge improperly [FITl]. Naturally, a good deal of modeling
and simulation needs to be performed to insert proper disturbance
statistics. One more approach suggested in [BERl] is to prevent the
determinant of the covariance matrix from decreasing below a given
reference value. A.parameter is chosen such that after the determinant
has converged for this length of time, the covariance is scaled so
that the determinant remains constant and the covariance matrix does
not go to zero.
With these methods, adjustments must be made strictly on the basis
ofTa-priori_knowledge obtained from measurements and simulation studies.
31
For this reason, methods are desired that are adaptive to the actual
estimation environment during the filtering process. 1
».As pointed out in [JAZZ], the filter divergence problem can be
judged by comparing the statistics of the actual measurement errors
with the covariance matrix of measurement errors. The only measure
of divergence available in the filtering problem is to compare the
actual measurement, y(th),'with the measurement estimate y(th) based
on the previous state estimate; namely, the measurement estimate
error e(tk) is A
B(th) = V(th) ‘ E{Y(th)|Y(th_1)} (3-4-1)
or 1
903k) = V(th) ' M'(tk)x(thlth-l)' ’ (3'4‘2)
The measurement estimate error covariance matrix 2(th'th-1) is defined
as
Z(th]th_l) = E{e(th)e'(th)} (3-4-3)
M(th)P(thIth_1)M'(th) + R(th) (3-4-4)
Since the mean measurement error estimate E{e(th)} is zero, it is pos-
sible to judge the consistency of the filter covariance with the actual
errors in the measurement estimate by comparing (3-4-4) with the statis-
tics of e(th)e'(th).
It is further shown in [JAZZ] that in the case of no system
disturbances and scalar measurements, a disturbance covariance can be
artificially created to adapt the identical noise inputs to the
statistics of the error in the measurement estimate. By maximizing
the probability density function p[e(th)] with respect to a scalar q
32
where qI represents the system disturbance, the most probable dis-
turbance is determined at each estimate. '
Another approach suggested by Andersen, et a1.,,[AND2] is to
obtain estimates of the state x, the transition matrix o, the input
covarianCe matrix V, and the noise covariance matrix R. The assumptions
are that the system is discrete, the parameters ¢, V, and R are constant,
the random sources are zero-mean, independent, and identically distri-
buted, but probably most importantly, the obéervation matrix M is‘
invertible, that is, each state variable is directly observable through
noise. .Although this latter restriction may be severe in many cases,
it should be noted that in a class of system applications this technique
may be valuable in taking adequate experimental measurements to ascer-
tain the true values of 0, V, and R and then using these values for
real-time estimation problems--assuming the other requirements are
satisfied.
A different approach is used in this thesis to handle the problem
of adapting the estimation procedure to the measurement data. This
proposed approach has applications in filtering linear systems with
parameter uncertainties; however, it has been used more extensively in
the non-linear filtering problem. Thus, the discussion of non-linear
adaptive filtering is deferred until the next chapter.
IV. Approximate Filtering of NOnlinear Systems
The subject of this chapter is discrete-time filtering of non—
linear continuous-time systems. The case’of system disturbances and
noisy measurements is covered. A second-order nonlinear filter is
developed, that is similar to the one suggested by Athans, et al.,
[ATHZ] but is extended to the case of system disturbances. Then a new
technique is developed by the author to adapt the filter to the measure-
ment statistics such that those statistics and the filter covariance
estimates are consistent.
4-1. Nonlinear Filtering_
The original linear filtering theory was adapted to nonlinear
systems by linearizing the system about the nominal value and using the
"first-order" or "quasilinear filter” as suggested in [FRIl, MOWI, SCHl]
during the early 60's.
By the middle 60's, the problem of nonlinear filtering was being
formulated [KUSl, KUSZ, BUCZ]. The exact equations for the propagation
of the state conditional probability density function conditioned on the
observation sequence are given by (2-2-1) and (2-3-1). However, an
infinite number of moments is required to realize the exact solution of
these equations. Thus, considerable effort has been extended to obtain
an approximate filtering solution to (2-2-1) and (2-3-1). These efforts
33
‘1!—
34
include [KUS3, SORl,.ATH2, JAZl].
In this chapter, the state model is assumed to be of the form,
1 = f(x,t) + G(x,t)w (4-1-1)
where the system dynamics function f is an n-vector which is twice-
differentiable with respect to x and once differentiable with respect
to t, the input matrix G(x,t) is a continuous matrix function of x,t
and as before, w is the zero-mean'white-noise disturbance with covari-
ance Q(t)6(t-T). The observation equation is of the discrete form,
V(th) = h(X(th)) + V(th) (4-1-2)
where the r vector h is also twice differentiable with respect to x
and once differentiable with respect to t, and as before, V(tk) is a
white-noise sequence with covariance R(th)5hg° And, finally, x(to) is
assumed to be a random vector with known Gaussian distribution.
Some definitions are needed to further expound the nonlinear filter.
The Jacobian matrix of f(x) is defined as A(x) where the ijth_entry of
A is
3f.
Ali-j. =3—Xj‘ ’ 4-3.1 = 19'.'°9n‘ (4-1-3)
Define the second partials of f as the Hessian matrix, F£(x) for
Z = l,...,n, where the ijth_entry of F3 is
32f
F£(x).. - —__£;_. 2,1 = 1,...,n. (4-1-4)
L1 - axiaxj
For the output function h(x,t), define the output Jacobian matrix
as M(x) where the ijth_entry is
af- .
o. = A ' = 000 = .00 4- -5
ML] 8x, 4' 1’ 3m, .1 1’ 9n ( 1 )
J
35
Define the second partials of h as the Hessian matrix, H£(x) for
2 = l,...,m, where the thh_entry of H1 is
32hz
H£(x),, = (4‘1‘6)
Lj Axiaxj
The function f may now be expanded in a Taylor series about the condi-
tional mean ;,
f(x) é f(x) + A(x)(x-x) + %.';1¢i(x-x)'Fi(x-;) ' (4-1-7)
&=
where ¢i is the basis vector, ¢t = [0...1...0]' with the one in the
ith_row. Likewise, the function h may be expanded to give
h(x) é h(i) + M(SE)(x-§) +%_— .314) p(x-QyHLu-i). (47-1-8)
g:
Similarly, G(x,t)Q(t)G'(x,t) can be expanded about x to give an approxi-
mation for the term E{GQG'}.
Terms that are higher than second order are neglected in these
expansions, although they could be included, at least in theory. In
practice, the error distributions are assumed to be Gaussian with zero
mean; thus all odd-numbered moments vanish [WOZl], and the fourth-order
moments may be decomposed into second-order moments [ANDl]. However, it
should be noted that the addition of higher-order terms does not greatly
affect the filtering equations for small estimation errors since they
converge to zero faster than the lower-order terms as the state estimate
converges to the state [SORl]. For larger errors, another approach
is suggested in Section 4-3.
36
For the nonlinear filter, the continuous-time dynamic system can
not be transformed to an equivalent discrete-time model as was the case
for the linear system by using (2-4-2) through (2-4-4). The differen-
tial equations for the extrapolation of the system state estimate must
be solved numerically.
The extrapolation equations are found by using the expansions for
f in (4-17) and for E{GQG'} in the moment equations (2-2-4) and (2-2-6):
1:: A In A (--
x f(x,t) + §T§1¢L tr[FL(x)P(t]tk_1], tk-ls t< t,2 4 l 9)
P(t|th_1) = A(x)P(tIth_1) + P(tlth_1)A'(x)
+ E{G(§)Q(t)c'(§)}, tk_1< t < th (4-1-10)
The updated equations for the nonlinear filter are found by using
the expansion for h in (4-1-8) in the moment equations (2-3-4) and
(2-3-5):
S‘ceklth) = icthltm) + G(tptyep - hole,» -
m A
'2’Lil‘bitrmfixfiknfltkIth)]] (4-1-11)
P(th|tk) = P(th|th_1) - G(th)M(x(th))P(th|th_1) (4-1-12)
where
ccth) = Pcthltmmv(ing)[Macthnpcthltflm'(Slap) I
+ R(tk) + 0(tk)]'1 (4-1-13)
37
The term 0(th) can be approximated as
_. 1 '" : ,
0(tk) - -4 221% tr[H£(x(th))P(tkItk_l)]
m ..
£51453 tr[H,-(X(th)‘)P(thlth_ID] (4-1-14)
when only first- and second-order terms are considered, or it may be
approximated as
Oij(th)_= %-tr [Hi(;(th))P(thlth_1)Hj(;(th))P(thlth_1)I (4’1‘15)
(0 is the £jth_entry of 0) when the fourth-order central moments are
41'
considered. These updated equations are derived in [ATH2, SORl, and
JAZl], only in slightly different form. The relative performance of
linear and nonlinear filters can only be judged by experimental results.
In the following section, an estimation problem is chosen that has
nonlinearities that are widely varied by making changes in parameter.
values. In this way, the value of the nonlinear filtering equations
are tested.
4-2. An Example of a anlinear Filtering Problem
In [WAGl, GRUl, BERl, WISl, ATHZ], the problem of reentry trajectory
estimation is discussed. Typically, both the reentry system dynamics
and the observation function are severely nonlinear and accurate esti-
mates are difficult to achieve. In this section this problem is dis-
cussed and the results are given for a particular example using first-
order and second-order filtering.
38
The Reentry Prdblem
It is assumed that the force of gravity is negligible when compared
to the aerodynamic effects. Thus, the system dynamics for vertical
motion of the reentry body can be written as [ALL1, ATHZ]
Sc =-x .
1 2 (4-2-1)
- _ CDAo 2
x2- 2m X2
where x1 is altitude, x2 is velocity, CD is the drag coefficient, A is
forward area, p is atmospheric density, and m is mass. The_atmospheric
density is assumed to be an exponential function of altitude,
o = po '2'“1 (4-2-2)
_ -1 .
where y = 5 x 10 5 ft . Since the ballistic parameter CD Ap/Zm 15 not
known it can be estimated by setting
X3 = C1) ADO/2m
and augmenting the state equations such that
X1 = 'XZ
x = -x2 x e’YXl (4-2-4)
2 3
x3 = 0.
It should be noted that no system disturbances are present in this
formulation.
The observation equation consists of a scalar radar range measure-
ment taken from an altitude H and a lateral displacement M. The discrete
measurements are contaminated with additive white noise so that the
expression for the observations is
39
.— 2 2 1/2 _
V(tk) " [M + (x1(t’2) ' H) ] + V(th) (4'2‘5)
where V(th) is assumed to be zero mean with covariance
' E{V(th)V(t£)} = ROM.
For simulation purposes, numerical values were given to the system
parameters. Although a great number of cases were tried, a typical
case with rather severe nonlinear effects was to offset the radar such
that the lateral range isiM = 100,000 feet and the radar altitude is
H - 100,000 feet. This evaluation produces the maximum measurement
nonlinearity simultaneously to the maximum deceleration nonlinearity.
The measurement uncertainty was chosen as R = 10,000 feetz. Measure-
ments were assumed to be taken at the rate of one per second.
The initial state vector was chosen as
x1(0) = 300,000 feet
x2(0) = 20,000 feet/second (4-2-6)
x3 (0) = 10'3 feet.1
Initial values for the state estimates were
§1(0) = 300,000 feet
A
x2(0)
20,000 feet/second (4-2-7)
5 1
3 x 10' feet-
523(0)
In addition, the initial estimates were assumed to be uncorrelated;
thus the initial state covariance matrix is diagonal with elements
p11(0) = 106 feet2
P22(0) = 4 x 106 feetz/second2 (4—2-8)
= 10‘4 feet'2
p33(0)
40
These particular values are identical to those used by Athans et a1.
[ATHZ] and.make a rather close comparison possible for computer simula-
tion studies. In addition, by varying the values of H and M as well as
the above initial values, a wide range of nonlinearities is effected.
Reentry Simulation Results
As stated in [ATHZ], "The validity and relative advantages and
disadvantages of --- (ad hoc assumptions for nonlinear filtering) ---
can only be evaluated and analyzed based on the evidence of real or
simulation results for particular systems." 'With this viewpoint, a
large number of simlations were performed for the nonlinear problem of
reentry estimation.
Since the estimation error depends on the particular random noise
sample function, it was necessary to perform a Mente Carlo study. As
opposed to the linear filtering case, the RMS error can not be assumed
to be identical to the square root of the appropriate covariance matrix
variance term since the nonlinear filter produces only approximate
minimum-variance estimates and covariance estimates. For this reason
several different statistics were found from the Monte Carlo study,
including the average of the estimation error, the average of the
absolute value of the estimation error, and the RMS estimation error,
defined as
N A
map = [11V ,zlcxcth) - x(tpfif” (4-2-9)
L:
where RMS(tk) is a vector. The value of N, the number of trials, was
chosen to be twenty.
41
The results of the Monte Carlo study using the previously given
initial conditions are depicted for the linear filter in Fig..4-l and
for the second-order filter in Fig. 4-2. For this particular simula-
tion, the absdlute value of the average error and the RMS error can be
Compared to results obtained by.Athans et al. [ATHZ].' Overall the I
results are similar, but it is interesting to note that whereas both
the average and.RMS final values of errors are within 10% of the
respective values obtained in [ATHZ] the average values have a much
smoother shape, indicating that more cancellation is taking place in
the averaging process than in [ATHB].
Several aspects of the Mente Carlo study of the linear and non-
linear filtering problem should be noted.
1. The second-order filter was more accurate than the first-order
filter, which tended to diverge as shown in the RMS curve of Fig.
4-1.
2. The effect of the second-order terms in the state estimate update
equation (4-l-ll) and the covariance update equation (4-1-12) were
nearly negligible whereas the state estimate extrapolation non-
linear terms (4-1-9) produced a significant improvement in the
estimation accuracy.
3. Both the linear and nonlinear filter accuracy appeared to be
virtually independent of the difference between computing with
regular or extended precision.
4. Although results are not shown for velocity x2, and ballistic para-
meter X3, the contrast between linear and nonlinear filtering
parallels that for altitude in that the second-order filter shows a
marked improvement in estimation accuracy.
(FEET)
IN POSITION ESTIMATE
ERROR
500- ‘
I I]: 42
a. _
II
: '1 .,
‘ . ,I II.
400- ”
I II
I If
I ',-\
I ,\=
I “
I ‘.
:3(>Clq I ‘:
I I“
I - r.
I I /\
I I /'\/' '-
, _ I‘ .- \
. . r’ ..
200- “ : \..—
I
I _ __ \_
‘I/\. I, \/ \\\
:5: ° . --
'00"l I, \\\/°. R
,1 \_I '\._\
I K...“
I N...—
I
0 AA I I 1
0 IO 20 30
TIME (SECONDS)
Fig. 4-1. Altitude Estimation Error for the First-Order Filter of
the Reentry Estimation Problem.
Absolute Value of Average Error
....... Average of the Absolute Value of Error
-.-.-.- Square Root of the Filter Altitude Variance
-..-..- RMS Error
(FEET)
IN POSITION ESTIMATE
ERROR
500 1
43
400 -1
300 -
ZOO-
lOO-I
O l I
0 IO 20 30
TIME (SECONDS)
Fig. 4-2. Altitude Estimation Error for the Second-Order Filter of
the Reentry Estimation Problem.
Absolute Value of Average Error
_______ Average of the Absolute Value of Error
....... Square Root of the Filter Altitude Variance
-.._.._ RMS Error
44
5. Neither filter exhibited an RMS error that converged as rapidly as
the corresponding square root of the diagonal term of the covariance
matrix, although the nonlinear filter reduced the discrepancy as
compared to the linear filter.
Other comparisons may be made but they are reserved until the next
sections where other factors of filtering theory and filter divergence
are discussed.
4-3. Adaptive Filtering for NOnlinear Systems
In the customary development of linear and nonlinear filtering, it
is assumed that the system parameters--state coefficients, initial state
estimation covariance, measurement coefficients, system disturbance
covariance, and measurement noise covariance--are known precisely for
the entire filtering time interval. In addition, it is assumed that
computation routines are exact, and finally, in the case of nonlinear
filtering, that the approximate filtering equations are accurate enough
that the Gaussian assumption for the estimation error distribution is
satisfactory.
In a large class of filtering applications these assumptions are
not warranted. The initial conditions may not be available or the
measurement noise and system disturbance covariances may change unex-
pectedly. And of course, the inaccuracies of numerical solutions and
using only approximate filtering must be considered.
It was shown in Section 3-4 that most approaches to the problem of
filter divergence are not on-line solutions--they depend on prior
information to the filtering process. Although the method suggested by
45
Jazwinski [JAZZ] is on-line, it applies to linear systems with scalar
measurements and without system disturbances. In addition, all of these
filter modifications have the effect of reducing the rate of convergence
since the state covariance matrix is a priori prevented from converging.
As reported in Denham and Pines [DENl] and Wishner [WISl], another
- approach to filtering a nonlinear system is to use the first-order
filter and perform iterations at the point of measurement, called the
iterated extended Kalman filter. Simulation results [WISl] have shown
that iterations are as effective as second-order filtering in reducing
divergence. Hewever, it should be pointed out that iterations increase
the computational load and limit the measurement sampling rate. For
this reason, in many systems, the better solution is to use the non-
iterative first- or second-order filter and to use a faster measurement
sampling rate. Thus, a requirement that is placed on divergence
correction techniques is that the computation load must not be signifi-
cantly increased.
As is the case in the linear filtering problem, Section 3-4, the
only measure of filter convergence performance is to compare the
statistics of the extrapolated measurement estimation error with the
covariance of the measurement estimation error. In the nonlinear case,
an approximation must be used to obtain the extrapolated measurement
estimation error,
ecth) yep - in.)
ecth) -- yep - M(§(t,,));<(tklth_1)- (4—3-1)
46
The measurement eStimate error covariance matrix 2(thltk'1) is defined
as
zcthlt ) -E{e(tk)e' (th ‘ ‘ (4-3-2)
k-l
= M(§(tk)P(thlthf1)M'(§(tk)) + R(th)
The statistics to be used for the extrapolated measurement estimation
error are defined to be of the form
N
8(th) = .20 aie(th-L)e'(th ) (4-3-3)
"4,
where N is a constant and “L is a weighting sequence on the outer
product of the past extrapolated measurement estimation errors such that
N
2 o- = 1.
i=0 i
In the case of ideal filtering, tr 5(th) should be comparable to
tr 2(tklth-l)’ that is, for Gaussian distribution of 5(th)’ tr 8(th)
should be less than tr 2(thltk-l) on an average of 68.2% of the
measurements since the diagonal elements of Z are variances. In the
case of filter divergence, 8(th) will be on the average large compared
to Z(th|tk_1) and the above will not hold.
New that this manifestation of filter divergence is realized, it
must be decided how this quantity is to be utilized to adapt the filter
to the real-time measurements.
The author's approach is to observe that the divergence of the
filter represents the fact that 2(thlth-l) has converged to a smaller
value than warranted by actual data, and this in turn means that the
47
state covariancelmatrix P(th|th_1) has converged to a smaller value than
warranted, which has the effect of diminishing the filter gain matrix
G(th). The discrepancy in the diverging filter can thus be alleviated
by using a gain correction.matrix Gc(tk) to compute the adaptive filter
gain Gash).
Ga(th) = Gc(th)G(tk) (4-3-4)
Then the updated adaptive state estimate is
xc(thlth) = x(thltk-l) + Ga(th)e(th) (4-3-5)
The gain correction matrix is selected such that the corrected
error
ec(th) = y(tk) - M xc(tk]th) (4-3-6)
is driven toward zero.
Consider the case of a scalar system. If the state variance has
converged incorrectly to a value P(th)/C(tk) where C(th) is the diver-
gence factor, then the gain G(th) is
act-,2) = P(t,,)M/(M2P(t,,) + cctpRcthn (4-3-7)
and the gain correction should be
M2P(th) + C(th)R(th)
Gc(tI2) = 2 (4-3-8)
M p t + R t
( I2) ( k)
A good measure of the divergence C is
S t
C(tlz) = ( (2) (4-3-9)
z t
( hltk'l)
48
:Thus, for the scalar system, the gain correction can be computed exactly
based on the measurement error statistic S(tk). However, it should be
'observed that the value of S(th) is dependent upon the measurement noise
sample function and that it is only an approximation to the true measure-
ment eétimation error variance. Therefore, a reasonable approximation
to (4-3-8) is
8(th)
N|I—I
(4-3-10)
Gc(th) =
2(thlt
h-l)
The general vector filtering problem.may be handled identically to
the above except the case of multivariable measurements requires sOme
modification. One modification is to reduce the measurements to scalar
measurements processed sequentially in time. Another approach that may
be considered is to use cross-correlation to identify those components
of the state vector that are correlated with the diverging measurements
estimates and apply the gain correction in accordance with the strength
of the cross-correlation.
An important property of the diverging filter is that the diver-
gence implies that the gain matrix is too small due to the unwarranted
convergence of the state matrix. Thus, by simply testing a criterion on
divergence and increasing the gain by an appropriate factor, such as
(4-3-10), the filter adapts to the measurement errors without the need
to increase the state covariance.
In order to assure filter stability, the gain correction is tested
to assure that oscillations do not occur. The test is performed by
examdning the corrected measurement estimation error,
49
ec(th) = y(tk) -:ch(th). (4-3-11)
If this error is of the same sign as e(th), then the adaptive
estimate is unaltered. However, if the corrected error is of the
opposite sign as e(tk), then the adaptive estimate has possibly been
over-corrected and to assure the filter will be stable, the adaptive
gain is reduced such that ec(th) = 0. Therefore, the adaptive filter
gain is
Ga(th) = Gc(th)G(th), e(tk)ec(tk) 3 0
]e(t )l ’ (4-3-12)
Ga(th) = h . Gc(th)G(th), e(th)ec(th) < o
16(tt11 * lec(th)l
Using this method of adapting the filter to the measurement esti-
mation error has the advantages that (1) it is not necessary to increase
the state covariance, which reduces or obviates filter convergence,
(2) it can be used even in the case that system disturbances are present,
and (3) the increased computational load is negligible with respect to
the normal amount of filtering computation.
1-4. .A Discussion and an Example of Adaptive N0nlinear Filtering
.A criterion to establish the performance of a filter is to compare
the RMS error of the estimates obtained by a.Monte Carlo study with
the square root of the variance obtained from the filter state covari-
ance matrix. But first, it is necessary to establish the validity of
the approximate filter covariance terms.
50
Although the accuracy of the covariance matrix cannot be explicitly
demonstrated, on the other hand it can be seen that the differential
equations governing its extrapolation are smoother than those of the
state estimate. This point has been supported by experimental evidence
in several ways. I
l. The covariance matrix does not vary significantly over the runs
used in the Mente Carlo study.
2. The covariance matrix does not vary regardless of whether the first-
or second-order filter is being used.
3. 'The covariance matrix is independent of the integration accuracy for
'widely varying integration intervals.
Any of these factors would greatly affect the state estimation error,
but the constancy of the covariance matrix supports the hypothesis that
the approximate filter covariance is nearly exact.
~ In the case of the second-order filter it can be seen from Figs.
4-1 and 4-2 for the reentry problem that the RMS error is much closer
to the covariance matrix values than is the case for the first-order
filter. NOnetheless the remaining discrepancy is still large, and the
purpose of using adaptive filtering in the reentry estimation problem
is to reduce this difference.
Fig. 4-3 shows the altitude estimation error for the reentry prob-
lem and is equivalent to Fig. 4-1 except the linear filter is replaced
by the linear adaptive filter. The results are significantly improved
over the linear filter; in fact, the RMS error is even less than the
RMS error for the second-order filter is shown in Fig. 4-2. Thus, in
this particular problem, the benefit of using this adaptive filter is
(FEET)
IN POSITION ESTIMATE
ERROR
500-
51
.I
A)
400 ' I '.
I
I)
‘ I
.‘-
300- I
I:
I
‘-
200- 1 A
I \.
I . \
\AS .. A.
7‘34 \‘ ,\\/‘~./\,
Ioo - 'CI’\ I \/ VAx/IR
I, \/ . \\ \/_
. x.‘. _
I I
0 IO 20
TIME (SECONDS)
Fig. 4-3. Altitude Estimation Error for the First-Order Adaptive
Filter of the Reentry Estimation Problem.
Absolute Value of Average Error
_______ Average of the Absolute Value of Error
_._._._ Square Root of the Filter Altitude Variance
-..,_.._ RMS Error
52
greater than that of using a second-order filter. Finally, the second-
order adaptive filter was mechanized and the equivalent altitude
errors are shown in Fig. 4-4. .As can be seen from the time-history of
the error curves, the discrepancy between.RMS and covariance-originated
terms for the adaptive second-order filter is reduced by a factor of
four as compared to the second-order filter in the latter part of the
run.
The four cases are summarized in the curves shown in Fig. 4-5. The
RMS altitude estimation error is shown for the linear, nonlinear,
adaptive, and non-adaptive filters as obtained from the Monte Carlo
studies. .
Several points should be noted with regard to these simulation
results.
1. .Although only x1, the altitude estimation error, is shown in Figs.
4-1 through 4-5, similar results were found for x2, the velocity
estimation error, and x the ballistic parameter, with regard to
3.
relative filter performance.
2. The value of S(th) for these particular results was based on a
A single sample of measurement error (i.e., N = l in (4-3-3)),
although similar results have been obtained for other trials, such
as an exponential weighting function oi in (4-3-3). The difference
in results was slight.
3. The effect of using extended precision and of including a greater
number of runs in the Monte Carlo study was negligible.
(FEET)
IN POSITION ESTIMATE
ERROR
5001
400 "I
300 -
200'
1‘; ¥
. "(0 -______.o
. " ‘l "
Ioo- It; ’ \\\’/\'
-‘I v \. \x
/ \7‘?‘ ..
I73}.-- '2
I 1 ”’41
0 IO 20 30
TIME (SECONDS)
Fig. 4-4. Altitude Estimation Error for the Second-Order Adaptive
Filter of the Reentry Estimation Problem.
Absolute Value of Average Error
-- Average of the Absolute Value of Error
._ Square Root of the Filter Altitude Variance
. ._ RMS Error
ERROR IN POSITION ESTIMATE (FEET)
5001
400‘
300‘
2001
l()C)‘
.OW] \?'<:;¢x\
Fig. 4-5.
~ 54
I
I y
- I
: I.
i F:
I: ’\
.- ' ’A \
. v .
. I o’\ -\\
‘\
\
A \
. I, , ‘\._..._._..
/\\ ..
I 1
IC) 2() 3C)
TIME (SECONDS)
RMS Altitude Estimation Error for Four Similar Filters.
First-order NOn-adaptive Filter
---- Second-order NOn-adaptive Filter
.-.- First-order Adaptive Filter
...... Second-order Adaptive Filter
55
4. The simulation results did not depend on the noise generator or on
the word length; the results obtained on the Michigan State Univer-
sity IBM 1800 computer were substantiated with further runs on the
Purdue University CDC 6500 computer.
In the above study, it should be noted that the assumed values of
initial state and initial state variance were consistent; that is, the
initial estimate of each state component was within one standard
deviation of the true state, also, the sample noise function was con-
sistent with the assumed noise variance, and finally, the integration
routine was virtually exact. Thus the only source for causing diver-
gence was the approximate filtering equations since the only inexactly
known parameter x3 was included in the augmented state. The other
possible uses of adaptive filtering become apparent when other system
inconsistencies are considered. For example, it was found that a much
coarser value of the integration interval could be used for the
adaptive filter without eliciting the problem of divergence that was
evident even in the second-order non-adaptive filter. Another example
is that even for the second-order filter it was found that the initial
state error could only deviate as much as three standard deviations of
the initial dispersion estimate for convergence to be assured and this
was for a relatively small initial state covariance matrix (4-2-8)
whereas for the adaptive filter a large initial error and covariance
matrix could be assumed without filter divergence.
Several aspects of adaptive filtering are worthy of further
examination. Using the statistic S(th) (4-3-3) leads to a very diffi-
cult question about what is the optimum weighting sequence
56
“2’ L = l,...,N and what is the optimum value of N. It appears that N
should be small compared to the number of measurements that are required
for the system to transcend a sizeable nonlinear range; in addition, for
heavily weighted past data, the effect of previous gain correction on
the state estimate should be considered. The proper function for the
adaptive filter gain (4-3-12) is not analytically determined and ap-
parently must be determined empirically.
In summary, the relative performance of the nonlinear filter and
the adaptive filter as they are developed in this chapter is shown in
Fig. 4-5 for state estimation of the atmospheric reentry problem and
the increase in accuracy for either filter is demonstrated while the
best performance is obtained by both nonlinear and adaptive filtering.
V. Conclusions.
The purpose of this thesis is the investigation of problems
involved in state estimation of a class of dynamic systems.
5-1. Thesis Review
The initial problems include determining a suitable mathematical
model for the system and stochastic model for the external disturbances.~
Due to the universality of the Brownian motion proces$ for modeling
real-world phenomena, as well as the availability of useful analytic
results, the stochastic disturbance model is assumed to be a Brownian
motion process (Appendix A and Section 2-1). Due to the large class of
dynamic systems that can be modeled by a set of ordinary differential
equations, the differential model (Z-l-l) is chosen for the dynamic
system model. The output of the system is assumed to be sampled
periodically as is the case in digital computer monitors and optimal
controllers; therefore, the discrete-time observation model (2-1-10)
is used. And finally, it is assumed that real-time estimates are
necessary as in the case of an on-line controller; thus the system
state estimator is assumed to be recursive. The results and conclu-
sions in this thesis are based on these assumptions.
Using the above assumptions, the filtering equations are develcped,
based on Kolmogorov's forward equation (2-2-1), Ito's Theorem A-3 aid
Bayes' rule (2-3-1). The worth of the filter mechanization is determined
57
58
by digital computer simulation; for this reason, the discrete-time
modeling of continuous-time stochastic processes (Section 2-4) is
derived.
In the case of linear filtering problems, two methods are pro-
posed for reducing computation time and increasing accuracy. An
example of state estimation of a linearized model of a steam turbo-
generator is given. And finally, methods are discussed for handling
the problem of filter divergence due to inexactly known models.
In the case of nonlinear filtering, the nonlinear filter is
developed and used to estimate the state of a reentry body. The
adaptive filter is suggested as a method to reduce the error in the
state estimate.
5-2. Results and Conclusions
The major results and conclusions of this thesis are based on over
200 simulations of the steam turbo-generator and the reentry body as well
as the results that have been referred to in the literature.*
1. The computational methods proposed in Section 3-2 for state and
state covariance matrix extrapolation can significantly reduce
computation time and increase accuracy in real-time State estimation.
The actual improvement depends on the size of the state and is more
significant with a larger dimensional system with more than two or
three state variables.
*In addition, some preliminary results on adaptive filtering of the Van
der Pol oscillator show favorable convergence properties.
59
The results of the digital computer simulation (Fig. 3-2) suggest
that real-time state estimation of a steam.turbo-generator is
feasible either for Optimal control of a single machine or for
area-wide control.
The second-order filter developed in Section 4-l successfully
estimated the state of a reentry body as determined by the digital
computer simulation. In all cases where the nonlinearities are
significant, the second-order filter has significantly less error
than the first-order filter. Examples of this are shown in Figs.
4-1 and 4-2 for Monte Carlo studies of reentry estimation.
The nonlinear filteris superior to the linear filter in the reentry
estimation simulation due to the second-order state estimate extra-
polation term; the second-order terms in~state update and‘covarianee
update have virtually no effect on filter performance.
The adaptive filter suggested in Section 4-3 successfully reduces
the estimation error significantly as compared to the second-order
filter for both linear and nonlinear adaptive filtering (Fig. 4-3
and 4-4). The fact that the adaptive second-order filter has an
RMS error that is comparable to the filter squarevrooted variance
indicates that within the given structure, any further improvements
in filter accuracy will be slight. These results are summarized in
Fig. 4-5.
The adaptive filter as developed in this thesis is an adaptive gain
filter; i.e., only the filter gain is adapted to the divergence
which is manifested in the extrapolated measurement estimation
60
error. Other methods of divergence compensation such as adaptive
noise covariance do not appear to be nearly as satisfactory as
adaptive gain filtering according to the simulation efforts of
this research effort.
5-3. Suggestions for Further Research
In the area of linear filtering, most of the research effort lies
in the area of applications and in filtering of inexactly'modeled
processes. Naturally, the problem of state estimation and concurrent
system identification are not resolved and require further definition.
(See Mbnahan.[MDNl] for a thorough discussion of system identification.)
An example is the problem of estimation and identification.where the
observation transformation is singular and the system description is
only partially known.
In the applications area, more data are required for further
refinement of the steam.turbo-generator state estimator presented in
Section 3-3. An obvious extension is to simulate cascading an optimal
controller with the optimal filter and to test the resulting improve-
ment in the quadratic performance index as compared to presently used
feedback regulation. Furthermore, the methods of Foshe and Elgerd
[F081] for inter-area Optimal control could be mechanized by coupling
the optimal controller with the optimal filter.
The second-order filter has been reviewed and its value has been
reviewed and its value has been demonstrated in this thesis. A.natural
question arises as to the utility of mechanizing even higher-order
61
filters. .Although this topic is certainly an area for future research,
the answer appears to be in the negative for several reasons. Higher-
order filtering requires higher—order derivatives, which in practice
may be difficult to compute either analytically or numerically. .Also,
higher-order filtering requires a great deal more computation (See
[SORl], for example), especially for a larger system. And finally,
higher-order terms appear to have a limited effect in filtering
convergence.
The use of adaptive filtering appears to be more promising for
improving filter performance of severely nonlinear systems, but this
conclusion should be tested over a wide range of examples and condi-
tions. In addition, the exact form of the optimal adaptive filter is
an open question and is the subject of further research.
LIST OF REFERENCES
[ALLl]
[ANDl]
[ANDZ]
[ATHl]
[ATHZ]
[BARl]
[BELl]
[BERl]
[3151]
[BIRl]
LIST OF REFERENCES
H.J. Allen and A.J. Eggers, Jr., ”A Study,of the Motion and
Aerodynamic Heating of Missiles Entering the Earth's Atmos-
phere at High Supersonic Speeds," NACA, washington, D.C.,
Tech. NOte 4047, October, 1957.
T.Wf Anderson, "An Introduction to Multivariate Statistical
Analysis," New York, Wiley, 1958._
W.N. Anderson, Jr., G.B. Kleindorfer, P.R. Kleindorfer and
‘M.B. Wbodroofe, "Consistent Estimates of the Parameters of
a Linear System," The Annals of Mathematical Statistics,
Vol. 40, No. 6 December, 1969, pp. 2064-2075.
‘M..Athans and E. Tse, UA.Direct Derivation of the Optimal
Linear Filter Using the Maximum Principle," IEEE Trans.
Automatic Control, Vol. AC-12, pp. 690-698, December, 1967.
M; Athans, R.P.'Wishner, and A. Bertolini, "Suboptimal State
EStimation for Continuous-time NOnlinear Systems from
Discrete NOisy Measurements, 1968 Joint Automatic Control
Conf., Ann.Arbor, Michigan, pp. 364-382, June, 1968.
M.S. Bartlett, Stochastic Processes, London, Cambridge
University Press, 1961.
R.B. Bellman, Adaptive Control Processes, pp. 129-131, 152-
158, Princeton University Press, Princeton, New Jersey, 1961.
.A. Bertolini, ”A Trajectory.Ana1ysis Program," Technical Note
1967-48, Lincoln Laboratory, MIT, Lexington, Mass., September,
1967.
G.J. Bierman, "A Remark Concerning Discrete Approximation to
White Neise and Covariance Calculation," IEEE Trans. Auto-
matic Control 14, 432-433, 1969.
‘M.W. Bird, “Asympotic Convergence of NOnlinear, Continuous-time
Filters," Ph.D. Thesis, Michigan State University, 1969.
62
[BRAII
[BRYl]
[BRYZ]
(BUC1]
[BUCZ]
[CHAl]
[COCl]
[cox1]
[DENl]
[DENZ]
[D001]
[DYNl]
[FELl]
[FERl]
63
F.H. Branin, Jr., "Computer Methods of Network Analysis,"
Proc. IEEE, VOl. 55, No. 11, pp. 1787-1801, NOvember, 1967.
A.E. Bryson, Jr. and Y.C. Ho, Optimization, Estimation and.
Control, Blaisdell Publishing Co., 1969.
.A.E. Bryson and D.E. Johansen, Linear Filtering for Time-
Varying Systems Using Measurements Containing Colored
NOise, IEEE Trans. Automatic Control 10, 4-10 (1965).
R.S. Bucy, "Optimum finite-time filters for a special non-
stationary class of inputs,” Johns Hopkins University, Appl.
Phys. Lab., Baltimore, Md., Internal Memo. BBB-600, 1959.
R.S. Bucy, "NOnlinear filtering theory, IEEE Trans. Automatic
Control, (Correspondence), vol. AC-10, p. 198, April, 1965.
S.Y. Chan and K. Chuang, ”A Study of Effects on Optimal Control
Systems Due to Variations in Plant Parameters,” Int. J.
Control, 1968, V01. 7, No. 3, pp. 281-298.
W.G. Cochran, and G.M. Cox, Experimental Designs, New York:
John Wiley and Sons, 1957.
H. Cox, On the Estimation of State Variables and Parameters
for Nbisy Dynamic Systems, IEEE Trans. Automatic Control 9,
5'12, 1964 o
W.F. Denham and S. Pines, Sequential Estimation when Measurement
Function Nonlinearity is Comparable to Measurement Error,
AIAA J. 4, 1071-1076 (1966).
A.R. Dennis, Functional Updating and Adaptive Noise Variance
Determination in Recursive-type Trajectory Estimators,
Special Projects Branch.Astrodynamics Conf. NASA-GSFC,
Greenbelt, Maryland, May, 1967.
J.L. Doob, Stochastic Processes. New York: Wiley, 1953.
E.B. Dynkin,.Markov Processes, Springer, Berlin, 1965.
W. Feller, An Introduction to Probability Theory and Its
Applications, Vbl. II, Wiley, New York, 1966.
J.D. Ferguson and Z.V. Rekasius, "Optimal Linear Control
Systems with Incomplete State Measurements," Proc. 6th Ann.
Allerton Conf. on Circuit and System Theory (Urbana, 111.,
1968), p. 135, April, 1969.
[F111]
[F081]
[FRAI]
[FRAZI
[FRIl]
[GANl]
[(34111]
[(39111]
[H31]
[H32]
[1101]
[11m]
[JAZl]
[JAZZ]
64
R. J. Fitzgerald, Error Divergence in Optimal Filtering Problems,‘
Second IFAC Syrup. Automatic Control Space, Vienna, Austria,
1967.
C. E. Fosha, 'Jr. and O. I. Elgerd, "The Megawatt- Frequency Control
Problan- -A New Approach Via (hatimal Control Theory," Dept. of
Elec. Engr., University of Florida, Gainesville, Florida.
J. S. Frame, "Matrix Functions and Applications," IEEE Spectrum,
Vol. 1, No. 6, pp. 123- 131, June, 1964.
J .3. Frame and M.A. Needler, "State and Covariance Matrix
Extrapolation," Proc. IE (to be published).
B. Friedland and I. Bernstein, "Estimation of the State of :a
nonlinear process in the presence of nongauss1an noise and
disturbances," J. Franklin Inst. , vol. 281, pp. 455-480, June,
1966.
S. Ganapathy and A. Suba Rao, "Transient Response Evaluation
from the State Transition Matrix," Proc. ‘IEEE, Vol. 57, No.
3, pp. 347-349, March, 1969.
F. K. Gauss, Theory of the Motion of the Heavenly Bodies Moving
about the Sin in Game Sections. New York: Dover Publica-
tions, Inc. , 1963 (reprint).
M. Gruber, "An Approach. to Target Tracking," M.I.T. Lincoln Lab. ,
Lexington, Mass. Tech. Note 1967-8, February, 1967.
Y. C. Ho, "On the Stochastic Approximation Method and Optimal
Filtering Theory," Journal of Mathematical Anaysis and
Applications 6, pp. 152- 154, 1962.
Y.C. Ho and R.C.K. Lee, "A Bayesian approach to problems in
stochastic estimation and control," IEEE Trans. Automatic
Control, vol. AC-9, pp. 333-339, October, 1964.
K. Ito, "Lectures on Stochastic Processes," Tata Institute for
Fundamental Research, Bombay, 1961.
K. Ito and H.P. McKean, Jr., Diffusion Processes and Their
Sample Paths, Academic Press, New York, 1964.
A.H. Jazwinski, Stochastic Processes and Filtering Theory,
Academic Press, New York, 1970.
A.H. Jazwinski, Adaptive Filtering, Proc. IFAC Symp. Multi-
variable Control Systems, Dusseldorf, Germany, OctOber, 1968,
2, 1-15.
[JOSl]
[KAIl]
[W1]
[KAI-2]
[W3]
[KIRl]
[KDLl]
[KUSl]
[KUSZ]
[KUS3]
[LEEl]
[LEEZ]
[LEVl]
R.D. Joseph and J.T. Tou, "On Linear Control Theory,
T.
R.
R.
.J . Kushner ,
.J. Kushner, "Nonlinear filtering:
6S
" AIEE
Trans. Applications and Industry, 80 (1961), pp. 193-196.
Kailath and P. Frost, 'Mathematical Modeling of Stochastic
Processes," Stochastic Problems in Control, 1968 Joint Auto-
matic Control Conference, Un1versity oFMichigan, Ann Arbor,
Mich., ASHE, pp. 1- 38, June, 1968.
E. Kalinan, "A new approach to linear filtering and prediction
problems," Trans. AM, J. Basic Engrg. , vol. 82, pp. 34- 45,
March, 1960.
E. Kalman, "New methods in Wiener filtering theory," Proc. >
lst %. on Engrg. Applications of Random Function Theory
0 a 1 1ty, I... Bogdanoff and F. Kozin, Eds. New York:
Wiley, 1963. , .
.E. Kalman and R.S. 'Bucy, "New results in linear filtering and
prediction theory," Trans. ASHE, J. Basic Engrg. , set. D vol.
83, pp. 95-107, Decemher, 1961.
.K. Kirchmayer, Economic Control of Interconnected Systems,
Wiley, New York, 1959.
.N. Kolmogorov, "Interpolation and extrapolation of stationary
random sequences," Bull. Acad. Sci. USSR, Math. Ser. vol. 5,
1941. A translationjias been puElished by the RAND Corp. ,
Santa Monica, Calif ., as Memo. RM-3090-PR.
.J. Kushner, "On the differential equations satisfied by
conditional probability densities of Markov processes, with
applications," J. SIAM Control, vol. 2, pp. 106-119, 1964.
"Dynamical equations for optimum nonlinear
filtering," J. Differential Equations, vol. 3, pp. 179-190,
1967.
the exact dynamical equa-
tions satisfied by the conditional mode," IEEE Trans. Auto-
matic Control, vol. AC-12, pp. 262-267, June, 1967.
.B. Lee and L. Markus, Foundations of Optimal Control Theory,
Wiley, New York, 1967.
.C.K. Lee, Optimal Estimation, Identification, and Control,
Research Monografii No. 28, M.I.T. Press, 1964.
.H. Levis, "Some Computational Aspects of the Matrix Exponen-
tial," IEEE Trans. of Automatic Control, Vol. AC-l4, No. 4,
pp. 410-411, August, 1969.
[LIOl]
[MASl]
[MCKl]
[MONl]
[MOWl]
[NAYl]
[PAPl]
[PARl]
[PARZ]
[PEAl]
[PINl]
[POTl]
[WM]
66
M.L. Liou, "A Nbvel Method of Evaluating Transient Response,"
Proc. IEEE, Vol. 54, No. 1, pp. 20-23, January, 1966.
E.J.'Mastascusa, "A Relation Between Lioufls Method and the
Fourth-order Runge-Kutta.Method for Evaluating Transient
Response," Proc. IEEE, V01. 57, No. 5, pp. 803-804, May, 1969.
H.P. McKean, Jr., Stochastic Integrals, Academic Press, New
York, 1969.
D.D. Monahan, "System Identification by Stochastic Approxima-
tion," Ph.D. Thesis, Michigan State University, 1970.
v.0. Mowery, "Least squares recursive differential correction
estimation in nonlinear problems,P IEEE Trans. Automatic
Control, v01. AC-9, pp. 399-407, OctOber,*I965.
T.H. Naylor, J.L. Balintty, D.S. Burdick, K. Chu,
Computer
Simulation Techniques, Wiley, New York, 1966.
A. Papoulis, Probability, Random Variables, and Stochastic
Processes, McGraw-Hill, New York, 1965.
G.L. Park, J. Preminger, A.P. WUrmlinger, and W. Panyan, "A
Deterministic and Probabilistic Analysis of Response of Power
Systems to Load Changes," Prog. Rt. 2, Div. Eng. Res.,
Michigan State University, East Lansing, Michigan, October,
1967.
G.L. Park, et al, "A Deterministic and Probabilistic Analysis
of Response of Power Systems to Load Changes," Final Report,
Div. of Engr. Research, Michigan State University, East
Lansing, Michigan, October, 1969.
J.B. Pearson and C.Y. Ding, "Compensator Design for Multi-
variable Linear Systems," IEEE Trans. Automatic Control,
Vol. AC-11, p. 130, April, 1969.
S. Pines, H. WOlf, A. Bailie, and J. Mohan, "Modifications of
the Goddard.Minimum Variance Program for the Processing of
Real Data," Analytica1.Mechanics Associates, Inc., Westbury,
New York, Rept. Contract NAS 5-2535, October, 1964.
J. E. Potter, "A Guidance- -Navigation Separation Theorem," Rep.
Re- 11, Experimental Astronomy Laboratory, Mass. Institute of
Technology, Cambridge, 1964.
P. E. Purser, M. A. Faget, N. F. Smith, Eds., Manned Spacecraft:
Engineering Design and_gperation Fairchild Publications,
Inc. , New York, 1965, pp. 226—231.
[RAOl]
'[SARll
[5cm]
[scnz]
[su11]
[son1]
[STAl]
[STRl]
[STRZ]
[WAGl]
[WIEl]
[WISl]
67"-
‘V.P.C. Rao and S. Ganapathy, "Comments on Transient Evaluation
from the State Transition Matrix," Proc. IEEE, Vol. 58, No.
5, p. 814, May, 1970. A
‘VIV.S. Sarma and B.L. Deekshatula, "Optimal Control When Some of
the State variables Are th Measurable," Int. J. Control, 1968,
V01. 7, N0. 3, pp. 251-256. .
F.H. Schlee, C.J. Standish and N.F. Toda, Divergence in the
Kalman Filter, AIAA.J. , 5, 1114-1120, 1967.
F.C. Schweppe, "Algorithms for estimating a reentry body's
position, velocity, and ballistic coefficient in real time
or for post flight analysis," M.I.T. Lincoln Lab., Lexington,
IMass., Group Rept. 1964-64, DDC609524, December, 1964.
G.L. Smith, S.F. Schmidt and L.A. McGee, "Application of
Statistical Filter Theory to the Optimal Estimation of Position
and Velocity on Board a Circunlunar Vehicle," NASA TR R-135,
1962.
H.W. Sorenson and A.R. Stubberud, "Recursive Filtering for
Systems with Small but Non-negligible Non-linearities,"
Int. J. Control, Vol. 7, N0. 3, 271-280, 1968. .
K.N. Stanton, "Estimation of Turbo-Alternator Transfer Functions
using Normal Operating Data," Proc. IEE, Vol. 112, p. 173,
1965.
R.L. Stratonovich, "Application of the theory of Markov process
f0r optimum filtration of signals," Radiotekhn. 1. Electron.,
vol. 5, no. 11, pp. 1751-1763, 1960. (English transl. in
Radio Engrg. and Electronics, vol. 1, pp. 1-19.)
R.L. Stratonovich, "A New Form of Representing Stochastic
Integrals and Equations, J. SIAM Control 4, 362-371, 1966.
W.E. wagner, "Re-entry Filtering, Prediction and Smoothing,"
J. Spacecraft Rockets 3, 1321-1327, 1966.
N. Wiener, The Extrapolation, Interpolation, and Smoothing_of
Stationary Time Series with Engineering Applications. New
York: Wiley, 1949. Originally issued as a classified report
by M}I.T. Radiation Lab., Cambridge, Mass., February, 1942.
R.P. Wishner, J.A. Tabaczynski and Mt Athans, "On the Estimation
of the State of Noisy Nonlinear Multivariable Systems," Proc.
IFAC Symp. on.Multivariab1e Control Systems, Dusseldorf,
Germany, October, 1968.
[WONl]
[MONZ]
[won]
68
NM. Wonham, "On the Separation Theorem of Stochastic Control,"
SIAM Journal on Control, May, 1963, Vol. 6, N0. 2.
E. Wong, and M. Zakai, "On the Relation Between Ordinary and
Stochastic Differential Equations, Internat. J. Engrg. Sci.
3, 213-229, 1965.
J .M. W02encraft and Irwin M. Jacobs, Princi les of Communication
Engineering, Wiley, New York, pp. 205-206, 1965.
APPENDICES
APPENDIX A
Stochastic Processes and Filtering Theory
In random processes, a Guassian white noise process serves as a
model for purely random effects that appear to be pure noise with
mutually independent components. However, care must be taken since,
by definition of its correlation function, a white-noise process is
almost nowhere continuous, and thus, not Riemann integrable. For this
reason, some properties of the Ito stochastic integral are needed [1101].
Definition A-l. Let B(t,w) be a Brownian motion process defined on
the interval te[to,tN] and let f(t,w) be measureable with respect to
B(t), the smallest. 0 field induced by (t,m), te[to,tN]. Then the Ito
stochastic integral is defined as
f tN N-l
f(t,w)d8(t) = 1.i.m. Z f(t»,w)[8(t. ,w)-B(t.,w)] (A-l)
tO 1+ 0 i=0 L L+1 L
where
f(t,w) = f(ti,w), ti 3 t < ti+1’ t0 3 t < tN QA-Z)
and 1 = max {t4+1 - ti}.
The need for careful attention to solving stochastic integral and dif-
ferential equations is elaborated on in Appendix A of [BIRl]. See also
[KAIl].
69
70
The following two important properties of the Ito stochastic inte-
gral shall be stated without proofs (see [ITOl] or [DOOl]).
Theorem A-2. Let the random vector f(t,w) satisfy the previous
definition, then
E{Jf f(t,w)d8(t)} = 0 ' (A-3)
T
and
I = . 1
E{ f1 f(t,w)dB(t)[ f1 f(t,w)d8(t)] PB f1" E{f(t,w) f (t,w)}dt
(A'4)
where P = cov{B(t,w)}.
B
These relations are needed to deal with stochastic differential
equations of the form I
1 = f(x,t) + G(x,t)é (A-S)
which.must be interpreted as a formal representation of the vector Ito
stochastic differential equation
dx = f(x,t)dt + G(x,t)dB (A-6)
The solution is of the form
x = IT f(x,t)dt + IT G(x,t)dB (N7)
where the first integral can be interpreted as an ordinary Riemann
integral but the second integral has meaning only if properly inter-
preted since B(t) is of unbounded variation.
71
An alternative to the earlier defined Ito stochastic integral is
the Stratonovich integral [STRZ]:
N'l x(ti ) + x(t
)
G( ,t)d8=l. Z £+1 , _
f X “:13 4?]. G( 2 9 t4) [B(ti'i‘l) 80-1)]
08-3)
where 1.i.m. is the mean-square limit. It has been shown by W0ng and
Zakai [WONZ] that if (A-6) is defined in terms of the Stratonovich
integral, then to be interpreted in the sense of Ito, it will become
Gx ~ ’
= f(x, t)dt+ +: §-§;-El- G(x,t)dt + G(x,t)dB (A-9)
where
BG(x, ____g_) n m 36. (x(t)
Gx ,t 2 2 I2 -1, , A-10
ax.
J
and the components of dB are independent. The difference between (A-6)
and (A-9) shows that the interpretation of the stochastic integral of
the stochastic differential equation used to model the physical system
governs the solution properties. The Ito stochastic integral will be
used exclusively in this work since it is more general; however, by
using the results of W0ng and Zakai, transformation can be made from the
Ito to the Stratonovich differential equation.
The main result from Ito [ITOl] is the following theorem.
Theorem A-3. (Ito) Given LA-6), then for ¢(x,t) a scalar-valued
function of x and t,
d¢ = ¢dt + o; dx + i'tr GQG'oxxdt (A-ll)
where the subscripts denote the partials, and the necessary partials do
exist.
APPENDIX B
The Mnltivariable Normal Distribution
If a Gaussian-distributed random vector x has expected value
:1 = E{x} (13-1)
and covariance
‘V = cov {x} = E{(x-x)(x-x)'} (B-Z)
then a simple method for generating random variates with the probability
density_function
= exp[-%(x-i)'V'lcx-i)1
MK) 1 2 (B-S)
[2w] /
is given in the sequel. (See also [NAYl] or [BARl].)
Using the transformation,
x = Tu + x (3’4)
Where i is an n-vector given by (B-l), T is a lower trangular matrix,
and u is an n-vector with unit-variance, zero-mean, independent,
Gaussian-distributed random components, then the variance of x can be
expressed as
‘V = TT'. (B-S)
Due to the symmetric form of V and the fact that T is triangular, T
can be determined from the relationships
72
73
11 1'1 V11 ‘
'-1
t.. = v.. a Z t. t. t.. 1 < ' < L < n B—6
41 (41 [2:1 4'2 1h” 11 J ' ( )
4-1
2 1/ .
t = - .
ii (vii k: tth) l < L < n
Thus, the generation of the multivariable random vector x with covari-
ance V requires the solution of (B-4) where T is determined in (B-6).
The components of u are generated independently by a Gaussian random
variate generator as in [COCl], for example. -
.APPENDIX C
Parameters of the Steam Turbo-Generator Filterinngroblem
The parameters of the steam plant model are listed below.
governor time constant
steam valve motor time constant
steam system time constant
reheat steam system time constant
reciprocal of regulation coefficient
reheat coefficient
effective rotary inertia of machine
damping torque coefficient
valve control signal
valve position deviation
torque output deviation
frequency deviation
load deviation
The values of the above constants are chosen for a typical machine:
~a ~a ~a ~a
LN
II I! II II
94 25 K1 = 62.24
3-8 K2 = 0.324
113-1 M = 18,221.6
4524.0 D a 5.94
74
75
All of these quantities are on a per unit basis and must be transformed
for real values. 2
In addition to the above parameters, values for the stochastic
model are chosen. The frequency and frequency rate measurement errors
are statistically independent with zero mean and variance
r11 = 0.0001
r22 = 0.0004
The system disturbance is a scalar with zero mean and variance
q = 0.0008
The initial value for the state covariance matrix is chosen to be very
large with uncorrelated components as follows:
p11(0) = 1.0
p22(0) = 1.0
p33(0) = 1.0
p44(0) = 1.0
p55(0) = 1.0
p66(0) = 1.0
The initial state estimate is chosen to be zero, §(0) = 0, whereas the
initial true state is given the value x(0) = 0.1; that is, each component
is set at 0.1. The sampling rate was varied widely in many different
simulation runs, but a value of 2 seconds was commonly used since this
value allows a reasonable computation interval and yet is not too large
with respect to the system dynamics.
mm“
m
Vldl
mm3
UHO
Iii
I‘WIIHHIH
3
|