Добавил:
Upload Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:

Diss / 10

.pdf
Скачиваний:
143
Добавлен:
27.03.2016
Размер:
18.05 Mб
Скачать

Main Statements of Statistical Estimation Theory

363

11.5.1  Simple Loss Function

Substituting the simple loss function given by (11.23) into (11.31) and using the filtering property of the delta function

ϕ(z)δ(z z0 )dz = ϕ(z0 ).

(11.33)

−∞

 

we obtain

 

post (γ ) = 1 − ppost (γ ).

(11.34)

The a posteriori risk post(γ) and, consequently, the average risk (γ) are minimum if the a posteriori pdf ppost(γ) is the maximum for the given estimate. In other words, the a posteriori pdf ppost(γ) is maximum, if there is a single maximum, or maximum maximorum, if there are several local

maxima. This means that the probable value γm, under which the following condition

ppost m ) ≥ fpost (l), γ m , l

(11.35)

is satisfied, can be considered as the Bayes estimate. If the a posteriori pdf ppost(γ) is differentiable with respect to the parameter l, then the estimate γm can be defined based on the following equation:

dppost (l)

= 0 at

d2 ppost (l)

< 0,

(11.36)

 

 

 

 

 

 

 

dl

dl

2

 

γ m

 

 

 

γ m

 

 

and we take into consideration only the solution of this equation satisfying (11.35). Substituting the simple loss function in (11.29), we obtain the Bayes risk

m = 1− max ppost (l) = 1− ppost m ) .

(11.37)

The second term on the right-hand side of (11.37) is the average probability of correct decision making that is accurate within the constant factor. Consequently, in the case of the simple loss function, the probability of correct decision-making is maximum at the Bayes estimate. It is evident that, in the case of the simple loss function, the probability of incorrect decision-making is minimal when applying the Bayes estimate to random process parameter. At the same time, all errors have the same weight that means that all errors are undesirable independent of their values. In the case of simple loss function, the Bayes estimate is well known in literature as the estimate by maximum maximorum of the a posteriori pdf.

If the a priori pdf ppr(l) is constant within the limits of interval of possible values of the estimated random process parameter, then, according to (11.9), the a posteriori pdf is matched accurately within the constant factor with the likelihood ratio Λ(l). The estimate by maximum of the a posteriori pdf γm transforms into the estimate of the maximum likelihood lm. The maximum likelihood estimate lm is defined as a maximum maximorum position of the likelihood ratio. As a rule, the maximum likelihood ratio estimations are applied in the following practical cases:

The a priori pdf of estimated random process parameter is unknown.

It is complicated to obtain the a posteriori pdf in comparison with the likelihood ratio or function.

364

Signal Processing in Radar Systems

The method of maximum likelihood has the following advantages compared to other methods.

1.In practice, the maximum likelihood estimation is very close to the estimation by maximum a posteriori pdf for a wide class of a priori pdfs of the estimated random process parameter. In other words, the maximum likelihood estimation is the Bayes estimate of the simple loss function. Indeed, for highly accurate measurement of random process parameters, the a priori pdf about the estimate can be considered very often as the constant value and the a posteriori pdf about the estimate is matched by shape with the likelihood function or likelihood ratio. This property is very important when the a priori pdf of estimated random process parameter is unknown and a definition of the least favorable a priori pdf is associated with mathematical difficulties.

2.Estimation of random process parameter by the maximum likelihood is independent of the

mutual one-valued no inertial transformation by estimated parameter of the signal at the receiver output in the form F[Λ(t)], since the maximum likelihood point remains invariant during these transformations. This property is very important for practical realizations of optimal measurers.

3.Analytical definition of random process parameter estimate quality using the maximum likelihood ratio requires less mathematical problems and difficulties compared to other methods of estimation. As a rule, the other methods of estimation require a mathematical or physical simulation to define the characteristics of estimate quality that, in turn, complicates the design and construction of the corresponding random process parameter measurers.

4.It is defined in the mathematical statistics [5] that if there is an effective estimate, then the maximum likelihood estimate is effective too.

5.If the random process observation interval tends to approach infinity, the maximum likelihood estimate is asymptotically effective and unbiased and, in this case, it is the limiting form of Bayes estimations for a wide class of a priori pdfs and loss functions.

These and other particular advantages of the maximum likelihood estimate method are the reasons for its wide application. We need to note that the estimation of random process parameters by the maximum likelihood method has a major disadvantage if there is noise, because the noise will be analyzed jointly with the considered random signal parameters. In this case, the maximum maximorum caused by the presence of noise can appear.

11.5.2  Linear Module Loss Function

According to (11.31), the a posteriori risk for the loss function given by (11.24) can be determined based on the following equation:

post (γ ) = | γ − l | ppost (l)dl.

(11.38)

−∞

 

We can divide the integral on two integrals with the following limits of integration −∞ < l ≤ γ and γ < l ∞. Based on the definition of the extremum of post(γ), we obtain

d post (γ )

dγ γ m

γ m

 

= ppost (l)dl ppost (l)dl = 0.

(11.39)

−∞

γ m

 

It follows that the Bayes estimate is the median of the a posteriori pdf.

Main Statements of Statistical Estimation Theory

365

11.5.3  Quadratic Loss Function

 

According to (11.31), in the case of the quadratic loss function, we can write

 

post (γ ) = (γ − l)2 ppost (l)dl.

(11.40)

−∞

 

From the condition of extremum of the function post(γ), we obtain the following estimate:

 

γ m = lppost (l)dl = lpost .

(11.41)

−∞

 

Thus, the average of the a posteriori pdf of lpost, that is, the main point of the a posteriori pdf, is considered as the estimate of the random process parameter l at the quadratic loss function.

The value of post(γ) characterizes the minimum dispersion of random process parameter estimate. Since post(γ) depends on a specific form of the received realization x(t), the conditional dispersion is a random variable. In the case of the quadratic loss function, the Bayes risk coincides with the unconditional dispersion of estimate given by (11.15):

m = Var(γ m ) = Var(lpost ) = ∫∫(l lpost )2 ppost (l) p(X)dldX.

(11.42)

X

 

The estimate (11.41) is defined based on average risk minimum condition. For this reason, we can state that the Bayes estimate for the quadratic loss function makes minimum the unconditional dispersion of estimate. In other words, in the case of quadratic loss function, the Bayes estimate ensures the minimum estimate variance value with respect to the true value of estimated random process parameter among all possible estimations. We need to note one more property of the Bayes estimate when the loss function is quadratic. Substituting the a posteriori pdf (11.2) into (11.41) and averaging by the sample of observed data X, we can write

lpost = lpprior (l) p(X|l)dldX.

(11.43)

−∞ X

 

Changing the order of integration and taking into consideration the condition of normalization, we obtain

lpost = lpprior (l)dl = lprior ,

(11.44)

−∞

 

that is, in the case of the quadratic loss function, the Bayes estimate is unconditional and unbiased, forever.

The quadratic loss function is proportional to the deviation square of the estimate from the true value of estimated random process parameter; that is, the weight is assigned to deviations and this weight increases as the square function of deviation values. These losses take place very often in various applications of mathematical statistics and theory of estimations. However, although the

366

Signal Processing in Radar Systems

quadratic loss function has a set of advantages, that is, it is very convenient from the mathematical viewpoint and takes into consideration efficiently the highest value of big deviations compared to small ones, it is seldom used for the solution of optimal estimation problems compared to the simplest loss function. This is associated with the fact that the device that generates the estimate given by (11.41) has very complex construction and functioning. In the quadratic loss function case, definite difficulties can be met during computation of variance of the Bayes estimations.

11.5.4  Rectangle Loss Function

In the case of rectangle loss function (11.26), all estimate deviations from the true value of estimated random process parameter, which are less than the given modulo value η, are not equally dangerous and do not affect the quality of random process parameter estimation. Deviations, the instantaneous values of which exceed the modulo value η, are not desirable and are assigned the same weight. Substituting (11.26) into (11.31) and dividing the limits of integration on three intervals, namely, −∞ < l < γ η, γ η < l < γ + η, and γ + η < l < ∞, we obtain

γ + η

 

post (γ ) = 1 − ppost (l)dl.

(11.45)

γ − η

As we can see from (11.45), in the case of rectangular loss function, the value γ = γm is the Bayes estimate, for which the probability to satisfy the condition |γ l| ≤ η is maximum. Based on the condition of the a posteriori risk post(γ) extremum, we obtain the equation for estimation

ppost m − η) = ppost m + η).

(11.46)

This equation shows that we need to use that value of the parameter γm as the Bayes estimate, under which the values of a posteriori pdf corresponding to random process parameters deviating from the estimate on the value η both at right and left sides are the same.

If the a posteriori pdf ppost(l) is differentiable, then, at small values η, there can be a limitation by the first three terms of Taylor series expansion with respect to the point γm of the a posteriori pdfs given by (11.46). As a result, we obtain the equation coinciding with (11.36). Consequently, at small insensibility areas η, the Bayes estimations for the simplest and rectangular loss functions coincide. The Bayes risk given by (11.29) for the rectangular loss function takes the following form:

γ m+ η

m = 1 ∫ ∫

X γ m − η

ppost (l) p(X)dl dX. (11.47)

The random variable given by (11.45) is the a posteriori probability of the event that the true value of estimated random process parameter of the given realization is not included within the limits of the interval −η < γm < η. In doing so, in the case of the Bayes estimate γm, this probability is less compared to that of other types of estimations. The Bayes risk given by (11.47) defines the minimum average probability of incorrect decision making by the given criterion.

11.6  SUMMARY AND DISCUSSION

In many practical applications, the estimate criteria are selected based on assumptions about the purpose of estimate. At the same time, during the process design, choosing a definite estimate criteria is of great interest because such approach allows us to understand more clearly the essence

Main Statements of Statistical Estimation Theory

367

and characteristic property of the problem to estimate the random process parameter, which is very important, giving us a possibility to define the problem of valid selection of the estimate criterion more fairly. Owing to the finite time of observation and presence of noise and interference concurrent to observation, specific errors in the course of estimate definition arise. These errors are defined both by the quality performance and by conditions under which an estimation process is carried out. In general, the requirement of minimization of estimation error by value does not depend on just one assumption. However, if the criterion of estimation is given, the quality performance is measured based on this criterion. Then the problem of obtaining the optimal estimation is reduced to a definition of solution procedure that minimizes or maximizes the quality performance. In doing so, the parameter estimate must be close, in a certain sense, to the true value of estimated parameter and the optimal estimate must minimize this measure of closeness in accordance with the chosen criterion.

In theory, for statistical parameter estimation, two types of estimates are used: the interval estimations based on the definition of confidence interval, and the point estimation, that is, the estimate defined at the point. Employing the interval estimations, we need to indicate the interval, within the limits of which there is the true value of unknown random process parameter with the probability that is not less than the predetermined value. This predetermined probability is called the confidence factor and the indicated interval of possible values of estimated random process parameter is called the confidence interval. The upper and lower bounds of the confidence interval, which are called the confidence limits, and the confidence interval are the functions to be considered during digital signal processing (a discretization) or during analog signal processing (continuous function) of the received realization x(t). In the case of point estimation, we assign one parameter value to the unknown parameter from the interval of possible parameter values; that is, some value is obtained based on the analysis of the received realization x(t) and we use this value as the true value of the evaluated parameters.

In addition to the procedure of analysis of the estimation random process parameter based on an analysis of the received realization x(t), a sequential estimation method is used. Essentially, this method is used for the sequential statistical analysis to estimate the random process parameter. The basic idea of sequential estimation is to define the time of analysis of the received realization x(t), within the limits of which we are able to obtain the estimate of parameter with the preset criteria for reliability. In the case of the point estimate, the root-mean-square deviation of estimate or other convenience function characterizing a deviation of estimate from the true value of estimated random process parameter can be considered as the measure of reliability. From a viewpoint of interval sequential estimation, the estimate reliability can be defined using a length of the confidence interval with the given confidence coefficient.

To make the point estimation means that some number from the interval of possible values of the estimated random process parameter must correspond to each possible received realization. This number is called the point estimate. Owing to the random character of the point estimate of random process parameter, it is characterized by the conditional pdf. This is a general and total characteristic of the point estimate. A shape of this pdf defines a quality of point estimate definition and, consequently, all properties of the point estimate.

There are several approaches to define the properties of the point estimations: (a) It is natural to try to define such point estimate that the conditional pdf would be grouped very close to the estimate value; (b) it is desirable that while increasing the observation interval, the estimation would coincide with or approach stochastically the true value of the estimated random process parameter (in this case, we can say that the estimate is the consistent estimate); (c) the estimate must be unbiased or, in an extreme case, asymptotically unbiased; (d) the estimate must be the best by some criterion; for example, it must be characterized by minimal values of dispersion or variance at zero or constant bias; and (e) the estimate must be a statistically sufficient measure.

Based on the random character of the observed realization, we can expect errors in any ­decision-making rules; that is, such a decision does not coincide with the true value of a parameter.

368

Signal Processing in Radar Systems

By applying various decision-making rules, different errors will appear with different levels of probability. Since the nonzero probability of error exists forever, we need to characterize the quality of different estimations in one way or another. For this purpose, the loss function is introduced in the theory of decisions. This function defines a definite loss for each combination from the decision and parameter. The physical sense of the loss function is the following. A definite nonnegative weight is assigned to each incorrect decision. In doing so, depending on targets, for which the estimate is defined, the most undesirable decisions are assigned the greatest weights. A choice of definite loss function is made depending on a specific problem of estimation of the random process parameter. There is no general decision-making rule to select the loss function. Each decisionmaking rule is selected based on a subjective principle. Any arbitrariness with selecting losses leads to definite difficulties with applying the theory of statistical decisions. The following types of loss functions are widely used in practice: the simple loss function, the linear modulo loss function, the quadratic loss function, and the rectangle loss function.

REFERENCES

1.Kay, S.M. 2006. Intuitive Probability and Random Processes Using MATLAB. New York: Springer + Business Media, Inc.

2.Govindarajulu, Z. 1987. The Sequential Statistical Analysis of Hypothesis Testing Point and Interval Estimation, and Decision Theory (American Series in Mathematical and Management Science). New York: American Sciences Press, Inc.

3.Sieqmund, D. 2010. Sequential Analysis: Test and Confidence Intervals (Springer Series in Statistics). New York: Springer + Business Media, Inc.

4.Kay, S.M. 1993. Fundamentals of Statistical Signal Processing: Estimation Theory. Upper Saddle River, NJ: Prentice Hall, Inc.

5.Cramer, H. 1946. Mathematical Methods of Statistics. Princeton, NJ: Princeton University Press.

6.Cramer, H. and M.R. Leadbetter. 2004. Stationary and Related Stochastic Processes: Sample Function Properties and Their Applications. Mineola, NY: Dover Publications.

12 Estimation of Mathematical

Expectation

12.1  CONDITIONAL FUNCTIONAL

Let the analyzed Gaussian stochastic process ξ(t) possess the mathematical expectation defined as

E(t) = E0s(t),

(12.1)

the correlation function R(t1, t2), and be observed within the limits of the finite time interval [0, T]. We assume that the law of variation of the mathematical expectation s(t) and correlation function R(t1, t2) are known. Thus, the received realization takes the following form:

x(t) = E(t) + x0 (t) = E0s(t) + x0 (t), 0 t T,

(12.2)

where

 

x0 (t) = x(t) − E(t)

(12.3)

is the centered Gaussian stochastic process. The problem with estimating the mathematical expectation is correlated to the problem with estimating the amplitude E(t) of the deterministic signal in the additive Gaussian noise.

The pdf functional of the Gaussian process given by (12.2) takes the following form:

 

T T

 

 

 

 

 

 

 

,

(12.4)

F[x(t)| E0 ] = B0 exp 0.5∫∫ϑ(t1, t2 )[x(t1) E0s(t1)] [x(t2 ) E0s(t2 )]dt1dt2

 

 

0

0

 

 

 

 

 

 

 

where B0 is the factor independent of the estimated parameter E0; the function

 

 

 

 

 

ϑ(t1, t2 ) = ϑ(t2 , t1)

 

 

(12.5)

is defined from the integral equation

 

 

 

 

 

T

R(t1, t)ϑ(t, t2 )dt = δ(t2 t1).

 

 

(12.6)

 

0

 

 

 

 

Introduce the following function:

 

 

 

 

 

 

 

υ(t) = T s(t1)ϑ(t1, t)dt1,

 

 

(12.7)

 

 

0

 

 

 

369

370

Signal Processing in Radar Systems

which is a solution of the following integral equation:

T

R(t, τ)υ(τ)dτ = s(t).

(12.8)

0

 

 

Since the received realization of the Gaussian stochastic process does not depend on the current value of estimation E0, the pdf functional of Gaussian stochastic process can be written in the following form:

 

 

T

T

 

 

F[x(t)| E0

 

2

 

 

,

] = B1 exp E0

x(t)υ(t)dt − 0.5E0

s(t)υ(t)dt

 

 

0

0

 

 

 

 

 

 

where B1 is the factor independent of the estimated parameter E0.

As applied to analysis of the stationary stochastic process, we can write

 

s(t) = 1,

 

 

 

 

 

 

 

 

 

 

 

 

 

t2 ).

 

 

 

R(t1, t2 ) = R(t2 t1 ) = R(t1

 

 

 

 

 

 

 

 

In doing so, the function υ(t) and the pdf functional are determined as

 

 

T R(t − τ)υ(τ)dτ = 1;

 

 

 

 

0

 

 

 

 

 

 

T

 

T

 

F[x(t)| E0

 

x(t)υ(t)dt

2

 

 

] = B1 exp E0

0.5E0

υ(t)dt .

 

 

0

 

0

 

 

 

 

 

(12.9)

(12.10)

(12.11)

(12.12)

In practice, the stationary stochastic processes occur very often and their correlation functions can be written in the following form:

 

 

R(τ) = σ2 exp {−α | τ |},

 

 

 

(12.13)

 

2

 

α

 

 

 

 

R(τ) = σ

 

exp{−α | τ |} cos ω1τ +

 

sin ω1

| τ |

,

(12.14)

 

ω1

 

 

 

 

 

 

 

where σ2 is the variance of stationary stochastic process. These correlation functions correspond to the stationary stochastic processes obtained as a result of excitation of the RC-circuit, α = (RC)−1, and the RLC-circuit, ω12 = ω20 − α2, ω20 = (LC)−1 > α = R(2L)−1 inputs by the “white” noise.

Solution of (12.8) for the correlation functions given by (12.13) and (12.14) takes the following form, respectively:

υ(t) =

α

[s(t) − α −2s′′(t)] +

1

{[s(0) − α −1s′(0)]δ(t) + [s(T ) + α −1s′(T )]δ(T t)},

(12.15)

2

σ2

 

 

 

 

Estimation of Mathematical Expectation

371

υ(t) =

 

1

 

 

2

2

 

 

4

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

2

 

2 s′′′′(t) + 2 (ω0 − 2α

 

)s′′(t) + ω0s(t)

 

 

 

 

 

αω0

 

 

 

 

 

 

 

 

 

 

 

 

 

+

 

 

1

 

 

2

 

 

2

 

2

 

δ(t)

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

2

 

2 { s′′′(0) + (ω0 − 4α

 

)s′(0) + 2αω0s(0)

 

 

 

 

 

αω0

 

 

 

 

 

 

 

 

 

 

 

s′′′(T ) + (ω02

− 4α2 )s′(T )

− 2αω02s(T )

δ(t T ) + s′′(0)

− 2αs′(0) + ω02s(0)

δ′(t)

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

(12.16)

 

s′′(T ) + 2αs′(T ) + ω0s(T ) δ′(t T )}.

 

 

 

 

 

 

 

 

 

 

 

 

2

 

 

 

 

 

 

 

 

 

The notations s(t), s(t), s′″(t), s″″(t) mean the derivatives of the first, second, third, and fourth order with respect to t, respectively. If the function s(t) and its derivatives at t = 0 and t = T become zero, then (12.15) and (12.16) have a simple form. As applied to stochastic process at s(t) = 1 = const, (12.15) and (12.16) have the following form:

 

 

υ(t) =

α

+

1

 

[δ(t) + δ(T t)],

(12.17)

 

 

2

σ2

 

 

 

 

 

 

 

 

 

 

 

 

 

υ(t) =

ω02

+

1

[δ(t) + δ(t T )] +

1

[δ′(t) − δ′(t T )].

(12.18)

4ασ2

σ2

2ασ2

 

 

 

 

 

 

 

 

 

 

The following spectral densities

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

S(ω) =

 

2ασ2

 

 

 

(12.19)

 

 

 

 

 

α2 + ω2

 

 

and

4ασ2 (ω2 + α2 )

S(ω) = ( 1 ) ( )2 (12.20)

ω4 − 2ω2 ω12 − α2 + ω12 + α2

correspond to the correlation functions given by (12.13) and (12.14), respectively. It is necessary to note that there is no general procedure to solve (12.8). However, if the correlation function of stochastic process depends on the absolute value of difference of arguments |t2 t1| and the observation time T is much more than the correlation interval defined as

 

1

 

τcor =

 

| R(τ) |dτ = | (τ)| dτ,

(12.21)

σ2

 

 

0

0

 

where

 

 

 

 

 

 

 

(τ) =

R(τ)

 

(12.22)

 

 

σ2

 

 

 

 

is the normalized correlation function, and the function s(t) and its derivatives at t = 0 and t = T become zero, it is possible to define the approximate solution of the integral equation (12.8) using the Fourier transform. Applying the Fourier transform to the left and right side of the following equation

 

 

(12.23)

R(t − τ)υ(τ)dτ = s(t),

−∞

372 Signal Processing in Radar Systems

it is not difficult to use the inverse Fourier transform in order to obtain

 

 

1

S(ω)

 

υ(t) =

 

 

exp{jωt}dω,

(12.24)

S(ω)

 

 

−∞

 

 

 

where

S(ω) is the Fourier transform of the correlation function R(τ)

S(ω) is the Fourier transform of mathematical expectation of the function s(t), which can be defined as

S(ω) = R(τ) exp{− jωτ}dτ,

(12.25)

 

 

−∞

 

 

 

 

 

S(ω) = s(t) exp{− jωt}dt.

(12.26)

 

 

 

−∞

 

The inverse Fourier transform gives us the following formulae:

 

R(τ) =

1

 

S(ω)exp{jωτ}dω,

(12.27)

 

 

 

 

−∞

 

s(t) =

1

 

S(ω)exp{jωτ}dω.

(12.28)

 

 

 

 

−∞

 

If the function s(t) and its derivatives do not become zero at t = 0 and t = T and the function S(ω) is a ratio of two polynomials of pth and dth orders, respectively, with respect to ω2 and d > p, then there is a need to add the delta function δ(t) and its derivative δ′(t) taken at t = 0 and t = T. Thus, there is a need to define the solution for Equation 12.8 in the following form:

 

d −1

 

 

 

 

 

 

 

µ

(t) + cµδ

µ

 

(12.29)

υ(t) = υ(t) + bµδ

 

 

(t T ) .

 

µ = 0

Here, the coefficients bμ and cμ are defined from the solutions of equations obtained under the substitution of (12.29) in (12.8); δμ(t) is the delta function derivative of μth order with respect to the time.

In the case of stationary stochastic process, we have s(t) = 1. In this case, the spectral density takes the following form:

 

 

S(ω) = 2πδ(ω).

 

 

(12.30)

From (12.24) and (12.29), we have

 

 

 

 

 

 

 

 

 

d −1

 

 

 

 

 

υ(t) = S

−1

 

µ

(t) + cµδ

µ

 

(12.31)

 

(ω = 0) + bµδ

 

 

(t T ) .

µ = 0

Соседние файлы в папке Diss