Estimation of Mathematical Expectation |
423 |
where E(t0 + t) is the true mathematical expectation value of the investigated stochastic process at the instant t = t0. Thus, the mathematical expectation of estimate of the time-varying mathematical expectation of stochastic process in contrast to the stationary case is obtained by smoothing the estimate within the limits of time interval [t0 − 0.5T; t0 + 0.5T].
In general, as a result of considered averaging, there is the mathematical expectation bias that can be presented in the following form:
|
|
1 |
0.5T |
|
|
b[E (t0 ,T )] = |
∫ [E(t0 + t) − E(t0 )]dt. |
(12.313) |
|
T |
|
|
|
−0.5T |
|
If the magnitude E(t) is described about the point t = t0 within the limits of time interval [t0 − 0.5T; t0 + 0.5T] by the series with odd powers in the following form
|
N |
t2k −1 |
d(2k −1)E(t) |
|
|
E(t + t0 ) ≈ E(t0 ) + ∑ |
, |
|
|
|
|
|
|
|
(2k −1)! |
dt |
(2k −1) |
|
k=1 |
|
|
t0 |
|
the mathematical expectation estimate bias would be minimal. Then
b[E (t0 ,T )] ≈ 0.
The variance of the mathematical expectation estimate of the investigated stochastic process is defined in the following form:
|
1 |
0.5T |
0.5T |
|
Var[E (t0 ,T )] = |
|
∫ |
∫ R(t0 + t1, t0 + t2 )dt1dt2 , |
(12.316) |
T 2 |
|
|
−0.5T −0.5T |
|
where R(t1, t2) is the correlation function of the investigated stochastic process ξ(t).
In practice, the nonstationary stochastic processes with time-varying mathematical expectation or variance or both of them simultaneously are widely used. In doing so, the mathematical expectation and variance vary slowly in comparison with variations of the investigated stochastic process. In other words, the mathematical expectation and variance of stochastic process are constant within the limits of the correlation interval. In this case, to define the variance of the time-varying mathematical expectation estimate we can assume that the centralized stochastic process ξ0(t) = ξ(t) − E(t) is the stationary stochastic process within the limits of the interval t0 ± 0.5T with the correlation function that can be presented in the following form:
R(τ) = ξ0 (t)ξ0 (t + τ) ≈ σ2 (t0 ) (τ). |
(12.317) |
Taking into consideration the given approximation, the variance of the mathematical expectation estimate given by (12.316) after transformation of the double integral by introducing new variables τ = t2 − t1, t2 = t and changing the order of integration takes the following form:
Var[E*(t0 ,T )] ≈ σ |
2 |
(t0 ) |
2 |
T |
− |
τ |
|
|
|
1 |
|
(τ)dτ. |
(12.318) |
|
T |
|
|
|
|
∫ |
|
T |
|
|
|
|
|
0 |
|
|
|
|
424 |
Signal Processing in Radar Systems |
As we can see from (12.313) and (12.318), the dispersion of the mathematical expectation estimate is defined in the following form:
D[E*(t0 ,T )] = b2[E*(t0 ,T )] + Var[E*(t0 ,T )]. |
(12.319) |
In principle, we can define the optimal integration time T, under which the dispersion will be minimum at the instant t0 minimizing the dispersion of estimate by the parameter T. However, we can present a solution to this problem in an acceptable analytical form by giving a specific function of E(t).
Evaluate how the mathematical expectation estimate varies when the mathematical expectation E(t) deviates from the linear function. At the same time, we assume that the estimated mathematical expectation E(t) possesses the first and second continuous derivatives with respect to the time t. Then, according to the Taylor formula, we can write
E(t) = E(t0 ) + (t − t0 )E′(t0 ) + 0.5(t − t0 )2 E′′[t0 + ϑ(t − t0 )], |
(12.320) |
where 0 < ϑ < 1. Substituting (12.320) into (12.313), we obtain
|
|
1 |
0.5T |
|
|
b[E (t0 ,T )] = |
∫ t2 E′′[t0 + tϑ]dt. |
(12.321) |
|
2T |
|
|
|
−0.5T |
|
Denoting M as the maximum value of the second derivative of the mathematical expectation E(t) with respect to the time t, we obtain the top bound of the mathematical expectation estimate bias by module
| b[E (t0 ,T] | ≤ |
T 2 M . |
(12.322) |
|
24 |
|
As a rule, the maximum magnitude of the second derivative of the mathematical expectation E(t) with respect to the time t can be evaluated based on an analysis of specific physical problems.
To estimate the optimal time of integration T minimizing the dispersion of estimate given by (12.319), we assume that the correlation interval of the investigated stochastic process is much less than the integration time, that is, τcor T. Then, the following written form is true:
|
τ |
∞ |
|
∞ |
|
|
∫ |
(τ)dτ ≤ |
∫ |
| (τ) |
|
|
|
|
T |
|
|
(τ)dτ ≈ |
|
|
|
|
|
0 |
|
0 |
|
Taking into consideration (12.322) and (12.323) and based on the condition of minimization of the estimate dispersion given by (12.319), we obtain the optimal estimation of the integration time:
|
|
9σ2 (t0 )τcor |
1 |
|
|
T ≈ 2 |
5 |
|
(12.324) |
|
M |
2 |
|
. |
|
|
|
|
|
|
|
As we can see from (12.324), the larger the integration time, the larger the correlation interval and the variance of the investigated stochastic process. The lesser the integration time, the larger the maximum absolute value of the second derivative of the mathematical expectation measured. This statement agrees well with the physical interpretation of measuring the time-varying mathematical expectation.
Estimation of Mathematical Expectation |
425 |
In some applications, the time-varying mathematical expectation can be approximated by the series given by (12.256). The values minimizing the function
|
|
|
T |
|
N |
|
2 |
|
|
|
|
|
|
|
|
T |
|
∑ |
|
|
|
|
∫ |
|
|
|
ε2 |
(α1, α2 ,…, αN ) = |
1 |
|
x(t) − |
|
αiϕi (t) dt |
(12.325) |
|
|
|
|
|
|
0 |
|
i=1 |
|
|
|
can be considered as the estimates of the coefficients αi . This representation of the coefficients αi is possible only if the mathematical expectation E(t) and the functions φi(t) vary in time slowly in comparison with the variation of the first derivative of the function x0(t) with respect to the time; that is, the following condition must be satisfied:
|
|
E′(t) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
[x0′(t)]2 |
|
|
|
|
E(t) |
|
|
|
|
|
|
max |
|
|
|
|
|
|
|
|
|
|
. |
|
|
|
|
|
|
|
|
|
|
|
|
|
(12.326) |
|
|
|
|
|
|
|
|
σ |
2 |
(t) |
|
|
ϕi′(t) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
ϕi (t) |
|
|
|
|
|
|
|
|
|
|
|
|
max |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
In other words, the condition (12.325) to define the coefficients αi |
is true if the frequency band fE |
of the mathematical expectation E(t) is much less than the effective bandwidth of energy spectrum of the stochastic component x0(t). Based on the condition of minimization the function ε2, that is,
dε2 = 0, ∂αm
we obtain the system of equations to estimate the coefficients αm
N |
T |
T |
∑α*i |
∫ϕi (t)ϕm (t)dt =∫ x(t)ϕm (t)dt, m = 1, 2,…, N. |
i=1 |
0 |
0 |
Denote
∫T ϕi (t)ϕm (t)dt = cim;
0
∫T x(t)ϕm (t)dt = ym.
0
Then the estimations of the coefficients αm can be presented in the following form:
αm = AAm , m = 1, 2,…, N,
(12.327)
(12.328)
(12.329)
(12.330)
(12.331)
where the determinant A of the system of linear equations given by (12.328) and the determinant Am obtained by changing the mth column cim of the determinant A by the column yi are determined based on (12.270) and (12.271).
426 |
Signal Processing in Radar Systems |
The flowchart of measurer of the time-varying mathematical expectation estimate is similar to the block diagram shown in Figure 12.14, but the difference is that a set of functions φi(t) are assigned for the sake of convenience in generating them and the coefficients cij and values yi are formed according to (12.329) and (12.330), correspondingly. Definition of the coefficients αm is simplified essentially if the functions φi(t) are orthonormal functions; that is, the formula (12.297) is true. In this case,
αm = ∫T x(t)ϕm (t)dt. |
(12.332) |
0 |
|
The flowchart of measurer of the time-varying mathematical expectation estimate differs from the block diagram shown in Figure 12.15, and the difference is that a set of functions φi(t) are assigned for the sake of convenience in generating them; thus, there is no need to solve the integral equation (12.295).
Compute the estimate bias and mutual correlation functions between the estimates of the coefficients αm and αq . Based on investigation carried out in Section 12.7, we can conclude that the estimations of the coefficients αm of expansion in series given by (12.256) are unbiased estimates and the correlation functions and variances of estimates of the coefficients αm are defined in the following form:
* * |
|
|
1 |
|
N |
|
|
|
|
|
R(αm ,αq ) = |
|
A2 ∑ Aim Ajq ij , |
(12.333) |
|
|
|
|
|
|
i=1, j=1 |
|
Var (α*m ) = |
|
1 |
|
|
N |
|
|
|
|
∑ Aim Ajm ij , |
(12.334) |
|
A2 |
|
|
|
|
|
|
i=1, j=1 |
|
where Aim is the algebraic supplement of the determinant given by (12.270),
ij = ∫T |
∫T |
R(t1, t2 )ϕi (t1)ϕ j (t2 )dt1dt2 , |
(12.335) |
0 |
0 |
|
|
and R(t1, t2) is the correlation function of the investigated stochastic process.
With φi(t) used as the orthonormal function, the coefficients cim given by (12.329) take the following form:
In doing so, the matrix cij is transformed to the diagonal matrix and the determinant A of this matrix and algebraic supplements Aij are defined as follows:
A = 1, Aij = δij . |
(12.337) |
Based on (12.336) and (12.333), the correlation function of estimation of the coefficients αm and αq can be presented in the following form:
R(α*m, α*q ) = mq, |
(12.338) |
Estimation of Mathematical Expectation |
427 |
where ij is given by (12.335) at i = m, j = q. In doing so, the current variance of estimate given by (12.260) and the averaged variance of estimate (12.262) of the time-varying mathematical expectation take the following form:
|
N |
|
|
|
Var{E (t)} = ∑ ijϕi (t)ϕ j (t); |
(12.339) |
|
i=1, j =1 |
|
|
|
|
|
|
N |
|
|
Var{E } = |
1 |
|
∑ ii. |
(12.340) |
|
T |
|
|
i=1 |
|
|
|
|
|
|
correspondingly.
If it is possible to approximate the centralized stochastic process ξ0(t) by the “white” noise with the effective spectral density given by (12.283) in addition to the orthonormal functions φi(t), then we can write
Based on (12.341), we are able to define the current and averaged variances of the time-varying mathematical expectation estimates coinciding with the optimal estimations given by (12.336) and (12.338), which are applied to the observation of the Gaussian stochastic process with the timevarying mathematical expectation.
12.9 ESTIMATE OF MATHEMATICAL EXPECTATION BY ITERATIVE METHODS
Currently, the iterative methods or procedures of step-by-step approximation are widely used to estimate the parameters of stochastic processes. These procedures and methods are also called the recurrent procedures or methods of stochastic approximation. Essence of the iterative method applied to estimation of scalar parameter l by discrete sample with the size N is to form the recurrent relationship in the following form [18]:
l [N ] = l [N − 1] + γ[N ]{ f (x[N ]) − l [N − 1]}, |
(12.342) |
where
l*[N − 1] and l*[N] are the estimates of stochastic process parameter based on the observation of N − 1 and N samples, respectively
f(x[N]) is the function of received sample related with the transformation required to obtain the searched stochastic process parameter
γ[N] is the factor defining a value of next step to make accurate the estimate of parameter l, depending on the number of step N and satisfying the following conditions:
γ[N] > 0, |
|
|
∞ |
|
|
|
→ ∞, |
(12.343) |
∑γ[N] |
N =1 |
|
|
∞
∑γ 2[N] < ∞.N =1
428 |
|
|
|
|
|
|
|
|
|
|
|
|
|
Signal Processing in Radar Systems |
x[N ] |
|
|
|
|
|
|
|
|
|
γ[N ] |
|
|
|
|
l *[N ] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
f {x[N ]} |
|
|
+ |
|
|
× |
|
|
|
Σ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
+ – |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
FIGURE 12.16 Iterative measurer of mathematical expectation.
The relationship (12.342) allows us to image the flowchart of the iterative measurer of stochastic process parameter.
The discrete algorithm given by (12.342) can be transformed to the continuous algorithm using the limiting process for the difference equation
l[N ] − l[N − 1] = l[N ] = γ[N ]{ f (x[N ]) − l[N − 1]} |
(12.344) |
to differential equation
dl(t) |
= γ (t){ f [ x(t)] − l(t)}. |
(12.345) |
dt |
|
|
The flowchart of measurer corresponding to (12.345) is similar to the block diagram shown in Figure 12.16 where there is a need to change γ[N] on γ(t) and the summator on the integrator.
As applied to the mathematical expectation estimate of stationary stochastic process, the recurrent algorithm of measurement takes the following form:
E [N] = E [N − 1] + γ[N]{x[N] − E [N − 1]}. |
(12.346) |
The optimal magnitude of the factor γ[N] to estimate the mathematical expectation of stochastic process by uncorrelated samples can be obtained from (12.182). This optimal value must ensure the minimal variance of the mathematical expectation estimate over the class of linear estimations given by (12.183). Actually, (12.182) can be presented in the following form:
N
E [N] = N1 ∑xi = E [N − 1] + N1 {x[N] − E [N − 1]}. (12.347)
i=1
Comparing (12.346) and (12.347), we obtain the optimal magnitude of the factor γ[N]:
|
γ opt [N] = |
1 |
|
(12.348) |
|
|
|
|
N . |
|
|
The flowchart of iterative measurer of the mathematical expectation is similar to the block diagram shown in Figure 12.16, in which there is a need to exclude block f(x[N]) responsible for transformation of stochastic process.
Since the iterative algorithm (12.347) is equivalent to the algorithm (12.182), we can state that, in the considered case, the mathematical expectation estimate is unbiased and the variance of the
Estimation of Mathematical Expectation |
429 |
mathematical expectation estimate is given by (12.183). In practice, we sometimes use the constant values of the factor, γ[N] that is,
γ[N] = γ = const, 0 < γ < 1. |
(12.349) |
In this case, the estimation given by (12.346) can be presented in the following form [19,20]:
|
|
N |
|
N |
|
N −i |
|
|
N |
|
|
|
|
|
|
|
|
N |
|
|
xi |
|
E [N] = (1 |
− γ ) |
|
x[1] + γ |
|
(1 − γ ) |
|
xi = (1 − γ ) |
x[1] + γ |
|
|
|
. |
(12.350) |
|
∑ |
|
∑(1 |
i |
|
|
|
|
|
|
|
|
− γ ) |
|
|
|
|
|
i= 2 |
|
|
|
|
i= 2 |
|
|
|
As we can see from (12.350), the mathematical expectation estimate is unbiased. To define the variance of estimate, we assume that the samples xi are uncorrelated. Then, the variance of mathematical expectation estimate takes the following form:
|
Var{E } = |
2(1 − γ )2 N −1 + γ |
σ2. |
(12.351) |
|
2 − γ |
|
|
|
|
As 1 − γ < 1 and N → ∞ or N 1, (12.351) can be simplified and takes the limiting form:
Var{E } ≈ |
γ |
σ2 , |
(12.352) |
2 − γ |
that is, the considered estimate is not consistent.
The ratio between the variance of the mathematical expectation estimate defined by (12.351) and the variance of the optimal mathematical expectation estimate given by (12.183) as a function of the number of uncorrelated samples N and various magnitudes of the factor γ is shown in Figure 12.17. As we can see from Figure 12.17, for each value N there is a definite magnitude γ, at which the ratio of variances reaches the maximum.
|
|
Var(EE) |
|
|
102 |
σ2opt |
|
|
|
γ = 0.2 |
|
|
|
|
|
|
γ = 0.1 |
|
|
|
γ = 0.05 |
|
10 |
|
γ = 0.02 |
|
|
|
|
|
|
γ = 0.01 |
|
1 |
|
N |
|
102 |
103 |
|
10 |
FIGURE 12.17 Ratio between the variance of mathematical expectation estimate given by (12.351) and the variance of optimal mathematical expectation estimate given by (12.183) as a function of the number of uncorrelated samples N and various values of the factor γv.
430 |
Signal Processing in Radar Systems |
12.10 ESTIMATE OF MATHEMATICAL EXPECTATION WITH UNKNOWN PERIOD
In some applications, we can suppose that the time-varying mathematical expectation E(t) of the stochastic process ξ(t) is the periodic function
E(t) = E(t + kT0 ), k = 0,1,…, |
(12.353) |
and the value of the period T0 is unknown. At the same time, the practical case, when the period T0 is much more than the correlation interval τcor of the observed stochastic process, is of interest for us. We employ the adaptive filtering methods widely used in practice under interference cancellation to measure the time-varying mathematical expectation estimate [21]. We consider the discrete sample x(ti) = xi = x(iTs) where the sampling period is Ts. The sample can be presented in the form discussed in (12.168).
The observed samples xi come in at the inputs of the main and reference (the adaptive filter) channels (see Figure 12.18). The delay τ = kTs, where k is integer, is chosen in such a way that the samples at the main and reference channels would be uncorrelated. There is a filter with varying parameters in the reference channel. The incoming samples xi and the process yi forming at the adaptive filter output are sent at the subtractor input. At the subtractor output, the following process
εi = xi − yi = x0i + (Ei − yi ) |
(12.354) |
takes place. Taking into consideration that the samples are uncorrelated, the mathematical expectation of quadratic signal in the main and reference channels is defined as
ε |
2 |
2 |
2 |
. |
(12.355) |
|
= σ0 |
+ (Ei − yi ) |
|
The minimum of ε2 corresponds to the minimum of the second term in the right side (12.355). Thus, if the parameters of adaptive filter are changed before definition of the minimum ε2 , there is a need to use the signal yi at the adaptive filter output as the estimate Ei of time-varying mathematical expectation. As was shown in Ref. [21], under the given structure of interference and noise canceller, the value yi is the best estimate in the case of the quadratic loss function given by (11.25).
The adaptive filter with required impulse response is realized in the form of linear vector summing of signals with the weight coefficients Wj, where j = 0, 1,…, P − 1 are the numbers of
|
xi |
|
|
The main channels |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
+ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
+ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
– |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Ts |
|
|
|
|
|
|
|
Ts |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
W0 |
|
|
|
|
W1 |
|
|
WN–1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
εi = xi – yi |
|
|
|
|
|
|
yi |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Adaptive filter |
|
|
|
|
|
|
|
|
FIGURE 12.18 Adaptive filter flowcharts.
Estimation of Mathematical Expectation |
431 |
parallel channels, at that, the delay between neighboring channels is equal to the sampling period Ts. Tuning of the weight coefficients Wj is carried out in accordance with the recurrent Widrow–Hopf
algorithm [22]
Wj [N + 1] = Wj [N] + 2 x[N]x[N − d −
where
l = 0, 1,…, P − 1
P −1 |
|
|
|
|
, (12.356) |
j] − x[N − d − j]∑Wl [N]x[N − d − l] |
l= 0 |
|
|
|
|
μis the adaptation parameter characterizing the algorithmic convergence rate and the tuning accuracy of the weight coefficients
As was shown in Ref. [22], if the parameter μ satisfies the condition 0 < µ < λmax−1 , where λmax−1 is the largest eigenvalue [3] of the covariance matrix consisting of elements Cij = cicj , then the algorithm
(12.356) is converged. In the physical sense the eigenvalues of covariance matrix characterize the power of input stochastic process and, under the other equal conditions, the larger λ, the more power of input stochastic process. It was proved that
lim Wj [N] = Wj , |
(12.357) |
N→∞ |
|
where Wj are the elements of optimal vector of the weight coefficients satisfying the Wiener–Hopf equation that takes the following form:
P−1 |
|
∑CljWj = Cl+ d , l = 0,1,…, P − 1. |
(12.358) |
j= 0
in the stationary mode. The block diagram of computation algorithm for the weight coefficients of adaptive filter is shown in Figure 12.19. To stop the adaptation process we can use the procedures discussed in Ref. [22]. The most widely used procedure applied to the considered problem is based on the following inequality:
|
ε[N] = |
|
Wj [N] − Wj [N − 1] |
|
≤ ν, |
(12.359) |
|
|
|
|
Wj [N] |
|
|
|
|
|
|
where ν is the number given before.
The estimated mathematical expectation Ei is the periodical function and can be approximated by the Fourier series with finite number of terms
M |
|
Ei ≈ a0 + ∑[a cos(ωµκ) + b sin(ωµκ)], |
(12.360) |
=1
where
432 |
Signal Processing in Radar Systems |
Assignment
Wj[0], ε,
|
|
|
|
j = 0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
N = 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Reading |
|
|
|
|
x[N] |
|
|
|
|
|
|
|
|
|
|
|
N = N + 1 |
|
|
Definition of |
|
|
Wj[N] |
|
|
|
|
|
|
|
|
|
|
ε[N] < ε
Issue of
Wj[N]
j = j + 1
FIGURE 12.19 Block diagram of algorithm for determination of the weight coefficients of adaptive filters.
is the radial frequency of the first harmonic. In addition, we assume that the sampling interval Ts is chosen in such a way that the sample readings xi are uncorrelated between each other. Taking into consideration the orthogonality between components of the mathematical expectation and a definition of the covariance function (the ambiguity function) of the deterministic signal with constant component as
|
|
k |
|
CE (k) = lim |
1 |
∑E(κ)E(κ + k), |
(12.362) |
|
k→∞ 2k |
i= − k |
|
|
|
|
the covariance matrix elements given by (12.358) can be written in the following form: |
|
|
|
M |
|
C(k) = σ2δ(k) + ∑A2 cos(kωµ), |
(12.363) |
|
|
=0 |
|
where δ(k) is the discrete analog of the Dirac delta function given by (12.170); |
|
|
A02 = a02; |
(12.364) |