§ 5 Filtering

Extracting the required information from the data ( or signal ) containing errors ( meaning interference, noise ) , this data processing method is called filtering. Only a few types of least squares ( or least variance ) filtering are described below.

1. Least Squares Filtering

[ Least Squares Filtering for Growing Memory ]

Least Squares Criterion for Filtering

Given a discrete observation system

in the formula

where represents the transpose, which is the observation value at the moment , and is the n -dimensional column vector formed by , called the observation vector; yes , the matrix, where m is a constant; is the n -dimensional column vector, and its component is the error when; is the state vector ( m dimension ) . Suppose the state satisfies the equation

Among them is the m -order square matrix, its elements are all constants, and it is set to be reversible, remember .

If then

is called the optimal estimate in the least squares sense.

Assuming that it is reversible, remember , then it can be deduced

The recursive formula for filtering

The above-mentioned formula for the optimal estimate is not practical on an electronic computer, and a recursive formula is usually used on the computer. can be launched

in

is an m -dimensional row vector

satisfied .

From the above formula, all historical observations are actually used for the purpose , and as n increases, the number used increases, so this filtering recursion formula is called growth memory.

[ Weighted Least Squares Filtering ]   When n is sufficiently large, the earlier historical data is often used in an unfavorable role in estimation. Through weighting, the effect of "too old" data can gradually disappear. The following is an example of an "exponential decline" right.

Select , introduce a diagonal matrix sequence

As before, the observation system and the state equation are set as

if time

It is called the optimal valuation in the sense of least squares with the weight of "exponential decline".

can prove

in

And there is a recursive formula for calculation

in

is an m -dimensional row vector with the same definition as before.

2. Wiener filter

Wiener filtering and Kalman filtering are both least square deviation filtering, but they are based on known conditions, calculation methods and scope of application are different.

[ Wiener Filtering Criterion ]   Suppose you want to get the sequence z ( t ) ( t only takes some integer values ​​) , and actually get the sequence x ( t ) . So design a linear constant system with impulse response h ( t ) , so that its input quantity is x ( t ) , and the output quantity is

The deviation from z ( t ) is as small as possible under some criterion.

The so-called Wiener filtering problem is to take h ( t ) appropriately so that

When x ( t ) is a deterministic sequence,

When x ( t ) is a random sequence,

[ Single Least Squares Filtering ]   Suppose that the filter factor h ( t ) is a sequence consisting of s + 1 equally spaced coefficients, such as

s + 1 is called the length of the filter factor, then the least square filter factor h ( t ) satisfies

or expressed by convolution as

in

is called the autocorrelation function of the input x ( t ) ( Chapter 16 § 3 ) ,

is called the cross-correlation function of z ( t ) and x ( t ) .

How well the filtered output matches the desired output can be measured by the normalized mean squared error :

to measure, where

Obviously . At that time , the filtering effect was the best; at that time , the filtering effect was the worst.

Similar results can be obtained when the length of the filter factor is infinite. Use , , to represent the frequency spectrum * of h ( t ) , respectively , then according to the convolution of the two functions is equivalent to the product of their spectra, the frequency characteristics can be obtained

It shows that the frequency spectrum of the filter factor can be determined by the frequency spectrum of the cross-correlation function and the auto-correlation function.

[ Multi-channel Least Square Filtering ]   Multi-channel least square filtering method utilizes the repetition of multi-channel signals to provide more beneficial information.

Set as n - way input, m -way desired output, written as a matrix, respectively

is the multi-channel filter factor, written as a matrix as

-------------------------------

The frequency spectrum of a discrete time series is defined as

in

When the discrete time series has only finite items, it can be filled with many 0s to become an infinite series and the above definition can be used.

is the m -channel filtered output, written as a matrix as

If the length of each input is k +1 , and the length of all filter factors is s +1 , then the length of the filtered output of each channel is k + s +1 .

The so-called multi-channel least square filtering is to appropriately select the multi-channel filter factor ( matrix ) H ( t ) to make the total mean square error

When it is a definite sequence,

When it is a random sequence,

The multi-channel least squares filter factor satisfies the following equation

( 1 )

in

is the multi-channel least squares filter factor, is the matrix,

is the transpose matrix of the input autocorrelation function square matrix, called the multi-way autocorrelation matrix ,

A matrix of cross-correlation functions for the desired output and input .

Equation ( 1 ) is mn ( s + 1) linear simultaneous equations with the filter factor as the unknown . Its solution is the required multi-channel least squares filter factor.

The degree of agreement between the filtered output and the desired output, using the normalized mean square error

to measure, obviously, . At that time , the multi-channel least square filtering output matched the desired output best; at that time , the matching degree was the worst.

When the length of the filter factor is infinite, the least squares filter factor satisfies the following equation

Expressed in frequency spectrum, we have

=

The solution of the above equation is the desired multi-path least squares filter factor.

3. Kalman filter

[ Kalman Filtering for Linear Discrete Systems ]

The dynamic model assumes that an n -dimensional linear dynamic system and a p -dimensional linear observation system are respectively described by the following difference equations :

Or introduce the corresponding symbol simply written as

( 1 )

( 2 )

where ( k is an integer ) satisfies

x ( t )is the n -dimensional state vector,( t )is the m -dimensional dynamic noise vector, z ( t )is the p -dimensional()observation vector, and v ( t )is the p -dimensional()observation noise vector;a matrix, called dynamicnoise matrix, H ( t )is amatrix, called the observation matrix,is a non- singular matrix, called the transition matrix of the system, with the following properties:

(i ) (     for all t , where I is the identity matrix )

( ii ) (    for any )

( iii )

Linear minimum variance estimation If the estimation of the state of the system at time is determined from the dynamic model , the following conditions are satisfied :

( i ) the   estimate is a linear function of the observed value ;

( ii ) =   minimum value, where is the estimation error ; then this estimation is called the linear minimum variance estimation.

Suppose that through the p -dimensional linear observation system ( 2 ) , from the 1st time to the kth time, the state of the n -dimensional linear dynamic system ( 1 ) has been observed k times , and according to the k observation data, the state of the jth time An estimate is made , and the estimated error is , and the mean squared error of the estimate is denoted as . At the time , it was called forecasting or extrapolation, at the time , it was called interpolation. Especially when j = k is called filtering, and is abbreviated .

The Kalman filter formula is set in the above dynamic model, and the dynamic noise and observation noise are uncorrelated zero-mean white noise sequences; that is, for all k, j

mean , mean square error

,

Let the statistical characteristics of the initial state be

and are not related to

Then the optimal linear filter can be calculated recursively by the following formula

its initial value ;

Called the weighting matrix or gain matrix, it is the covariance matrix of the optimal estimation error , the I in parentheses represents the identity matrix, and the last equation is called the covariance update equation.

At this time, the optimal linear prediction ( extrapolation ) is estimated as

( j > k )

[ Kalman Filtering for Continuous Time Systems ]

The dynamic model assumes that the state equation is

( 1 )

The observation equation is

where is a random process of n -dimensional vector type, and is a random process of p -dimensional vector type. , respectively, are m -dimensional ( ) and p -dimensional vector-type, uncorrelated white noise processes with zero mean, namely

In the formula, Q ( t ) and R ( t ) are both continuously differentiable, symmetric and non-negative definite matrices for time t ; they are Dirac functions. And F ( t ) , G ( t ) and H ( t ) are respectively matrices whose elements are non-random functions or constants of t .

The linear minimum variance estimation assumes the known ( observed ) value ( ) , and is calculated by the formula

represents a linear estimate such that

= minimum value

Such an estimate is called a linear minimum variance estimate, where the filter factor matrix, each element of which is continuously differentiable for both independent variables.

The Kalman filter equation assumes that the above dynamic model satisfies the following conditions :

( i ) the   matrix R ( t ) is positive definite for all t ;

( ii ) Under the action of u ( t ) , the dynamic system ( 1 ) reaches a steady state, that is, x ( t ) is given by

Deterministic random function;

( iii ) At a certain time , the variance of the measured and the time is known;

Then the optimal filtering equation of the dynamic model is

in the formula

( weighted matrix equation )

( 2 )

( Ricati equation ( see Chapter XIII § 1 ))

The initial conditions are

In the above formula, it is called the weighting matrix, which is the covariance matrix of the optimal estimation error .

In particular, when it is a vector-type stationary random process, it can be solved in ( 2 ) .