- •Introduction to adjustment calculus
- •Introduction to adjustment calculus (Third Corrected Edition)
- •Introduction
- •2. Fundamentals of the mathematical theory of probability
- •If d'cd; then p (d1) £ lf
- •Is called the mean (average) of the actual sample. We can show that m equals also to:
- •3.1.4 Variance of a Sample
- •Is called the variance (dispersion) the actual sample. The square root 2
- •In the interval [6,10] is nine. This number
- •VVII?I 0-0878'
- •In this case, the new histogram of the sample £ is shown in Figure 3.5.
- •Is usually called the r-th moment of the pdf (random variable); more precisely; the r-th moment of the pdf about zero. On the other hand, the r-th central moment of the pdf is given by:
- •3.2.4 Basic Postulate (Hypothesis) of Statistics, Testing
- •3.3.4 Covariance and Variance-Covariance Matrix
- •X and X of a multivariate X as
- •It is not difficult to see that the variance-covariance matrix can also be written in terms of the mathematical expectation as follows:
- •3.3.6 Mean and Variance-Covariance Matrix of a Multisample The mean of a multisample (3.48) is defined as
- •4.2 Random (Accidental) Errors
- •It should be noted that the term иrandom error" is used rather freely in practice.
- •In order to be able to use the tables of the standard normal
- •X, we first have to standardize X, I.E. To transform X to t using
- •Is a normally distributed random
- •4.10 Other Measures of Dispersion
- •The average or mean error a of the sample l is defined as
- •5. Least-squares principle
- •5.2 The Sample Mean as "The Maximum Probability Estimator"
- •5.4 Least-Sqaures Principle for Random Multivariate
- •In very much the same way as we postulated
- •The relationship between e and e for a mathematical model
- •6.4.4 Variance Covariance Matrix of the Mean of a Multisample
- •Itself and can be interpreted as a measure of confidence we have in the correctness of the mean £. Evidently, our confidence increases with the number of observations.
- •6.4.6 Parametric Adjustment
- •In this section, we are going to deal with the adjustment of the linear model (6.67), I.E.
- •It can be easily linearized by Taylor's series expansion, I.E.
- •In which we neglect the higher order terms. Putting ax for X-X , al for
- •The system of normal equations (6.76) has a solution X
- •In sections 6.4.2 and 6.4.3. In this case, the observation equations will be
- •In matrix form we can write
- •In metres.
- •6.4.7 Variance-Covariance Matrix of the Parametric Adjustment Solution Vector, Variance Factor and Weight Coefficient Matrix
- •I.E. We know the relative variances and covariances of the observations only. This means that we have to work with the weight matrix к£- 1
- •If we develop the quadratic form V pv 3) considering the observations l to be influenced by random errors only, we get an estimate к for the assumed factor к given by
- •Variance factor к plays. It can be regarded as the variance of unit
- •In metres,
- •Is satisfied. This can be verified by writing
- •Into a . О
- •6.U.10 Conditional Adjustment
- •In this section we are going to deal with the adjustment of the linear model (6.68), I.E.
- •For the adjustment, the above model is reformulated as:
- •Is not as straightforward, as it is in the parametric case (section 6.4.6)
- •VeRn VeRn
- •Into the above vector we get 0.0
- •0.0 In metres .
- •In metres.
- •Areas under the standard normal curve from 0 to t
- •Van der Waerden, b.L., 1969: Mathematical Statistics, Springer-Verlag.
Is usually called the r-th moment of the pdf (random variable); more precisely; the r-th moment of the pdf about zero. On the other hand, the r-th central moment of the pdf is given by:
m' = / (x-y) ф (x) dx .
(3.25)
By inspecting the above expressions for m^ and along with equations (3.20) and (3.22), we can see that:
у = m^ (3.26a)
and a2 = = ш2 - y2 = m2 - m2- . (3.26b)
Compare the above result (3.26 a, b) with the analogy to mechanics mentioned in sections 3.1.3 and 3.1.4.
3.2.4 Basic Postulate (Hypothesis) of Statistics, Testing
The basic postulate of statistics is that "any random sample has got a parent random variable". This parent random variable xtR is usually called population and is considered to be infinite. It is common in statistics to postulate the pdf of the population for any random sample, and call it the postulated, or the underlying pdf. Such a postulate may be hence tested for statistical validity.
In order to be able to test the statistical validity we have to assume that the sample can be regarded as having been picked out, or drawn from the population, each element of the sample independently from the rest. This additional property of a sample is required by the standard definition of a random sample as used in statistical literature. However, since the present Introduction does not deal with statistical testing we shall keep using our original, more general definition.
There are infinitely many families of PDF1s• Every such family is defined by one or more independent parameters, whose values characterize the shape of its pdf. The individual members of a family vary according to the value of these parameters. It is common to use if possible, the mean and the standard deviation as the pdf1s parameters. The less parameters the family of pdf1s contains the better; the easier it is to work with.
The usual technique is that we first select the "appropriate" family of pdf1s on the basis of experience and then try to find such values
of its parameters that would fit the actual random sample the best. In other words, the shape of the postulated ф(х) is chosen first; then, its parameters are computed using some of the known techniques.
Since we shall be dealing with the samples and the random var-
iables (populations) at the same time, we shall use, throughout these notes, the latin letters for the sample characteristics, and the corresponding greek letters for the population characteristics as we have done so far.
3.2.5 Two Examples of a Random Variable
Example 3.17: As the first example, let us investigate a random variable x
with rectangular (uniform) PDF which is symmetrical
around a-value .x;= k. *-:E:it" the -probability
of x < k-q and x > k+q, be zero. Obviously, this PDF has the
following analytical form (see Figure 3.13):
0, for (x < к - q) and (x > к + q).
This can be written in an abbreviated form as:
<h, for (j.x-k|. < q). v 0, for (]jc - k| > q).
The above ф contains apparently three parameters k, q and h.
However э only two are independent, since one can be eliminated
from the condition (3*15)> i.e.:
/ ф(x) dx = 1
—00
that must be satisfied for any ф to be a PDF. Let us
eliminate for instance the parameter h. We can write: oo k-q k+q
/ ф (x) dx = / ф (x) dx + / ф (x) dx + -«о »oo k-q
00
/ ф (x) dx k+q
k+q
= 0 + / hdx + 0
k-q
= h / dx = h [x], = 2hq. = 1. . к—q
k-q u
This means that h э and therefore:
2q
<l/(2q), for (|.x - k| < q) 09 for (|x - k| > q)
T he corresponding CDF to the above ф is:
0, for (x< k-q)
X 1 x 1
¥(x) = / ф(х) dx =f~— — / dx. = — (x-k+q), for
^ k-q
—.00
(|x - fr| < q).
1, for (x > к + q) 7
and is shown in Figure 3.1*+.
From the above figure we see that the functionфis linear in the interval over which ф / 0, and is constant everywhere else. Note
dx
The mean of the given PDF is computed from equation (3.20) as follows:
\i = f хф (x)dx
J"
xc
k-q
2q
2
Jk-q
2q
= (k2 + 2kq + q2 - k2 + 2kq - q2)
This result satisfies our expectation for a symmetrical function
around k. The variance of the given PDF can be obtained from
equation (3.22), yielding
2 _ r°° p л k+q
a (х-цГ ф(х) dx = ^ / (х-кГ dx =
1 ,k+<1 2a 2k ^ A к2 ,к+1 л k~q ^ k-q 4 k-q
= T- tl-J, - 2k + k
2q 3 k-q
= ^ (k3 + 3k2q + 3kq2 + q3 - k3 + 3k2q - 3kq2 + q3) - k2
6k q ^ 2al , 2 q~
6q 6q 3
1 1
Since к = у and q = /За, then h = — - 7777-, and we can
2q 2/3a
express the given rectangular PDF, which we will denote by R, in terms of its mean у and its standard deviation о as follows:
for (-|x-ji|) < /За)
R(-y, а ; x) = ф(х) ==<^
V.O, for ( |x - ji| ) > /За) #
Similarly, we can express its corresponding CDF, which we will
denote by R^, in terms of;ц and а, as follows:
, 0, f©-apvtxT< у - /За).
R (ji,cr;x) = f(x) =^ ^7^7 (x-y,+/3a), for (|x~y| < /За).
1, for (x > у + /За) , Assume that we would like to compute the probability of x e [ y-а, y.+a], where x has the rectangular PDF . This can be done by using equation (3.l6c) and Figure 3.15 > as follows:
у+а
Р( у-а < х < у +а) = / ф(х) dx
у+а = / у-а
г7з5 ta в W [2а3
The above probability is given by the shaded area in Figure 3.15.
Similarly, for this particular uniform PDF, we find that: P(y-2a < x < у + 2a) = P( y-3a < x < p+ 3a) = 1.0
In"statistical "testing, we :6ften. need to compute the moments of the PDF (see .--secti^H:З.2.3). Let'ua, .-for instance, compute the third moment m^ about zero of the rectangular PDF,.. We will use equation Ov2U') 9 - i. e."
3 1 .
■Х
„
и
dx
°° ' у +/3с
ш - / х ф(х) dx - / * ^
-00 у-/За
1 х^ ^^За
= 27за" [Т~] /, Р -/За
= §73^ (S^ou3 + 2W3 а3р)
3 2 = у + За у
Example 3.18: As a second example let us investigate a random variable with
a triangular PDF, which is symmetrical around x = k.. Let us
Figure 3«l6
assume that the probability of x < к - q and x > к + q, be zero. We may write (see Figure 3.l6):
Ф(х)
=
(х -k + q), for (к - q < х < к).
О,
for
(х
> к + q).
— £«х + к + q) , for (к < х < к + q).
This .can be rewritten in the following abbreviated form as:
7 (q - |x - k|), for (|x - k| < q)
0 , for (|x - кj > q) .
From the above, we can see that the triangular PDF has the same parameters (k, q, h) as the uniform PDF of example 3.17• Let us again eliminate the parameter h from the condition
/ ф(х) dx = 1. This integral is nothing else but the area
.00
of the triangle, so. that we. can write.:.':—'v 2q * h = qh = 1. This: gives■. us : h = ~9 and hence,
The computations of the mean and the variance of the triangular PDF can be performed by following the same procedure as we have done for the rectangular PDF in example 3.17. We state here the results without proof, and the verification is left to the student•
The mean у of the given triangular PDF equals to к , and the
2 12
variance a comes out as q .
Since к. = у and q = /ба э we ean /agaia respites' the triangular PDF, which we will denote by T, in terms of its mean у and its standard deviation а, as follows:
1 . Ji^l э for(|x-ii| < /ба)
T (уэ cr -;x) = ф(х) =/
* 0, for ( |x-y| > /ба). The corresponding CDF is given by: '0, for (x < у - q)
x
Y(x)
y-q. q.
1, for (x > У + q), and is shown in Figure 3.17•
0,0
|
|
|
|
|
|
Д f-*^ /> 1 |
|
i ^ |
Figure 3.17
The integral in the above equation can be rewritten as:
б^
k-q'4-
i2 \
цq
+
*
-
и-
ах for
(х
< )
/
q,
+
х - у
dx
+
Х
CL^i.
for (x > у).
and we get:
x 1 1 2 x
—о / (q + x- y)dx = ~ {77 [x ] + (q-y) [x] q^ ^ n2 2 y-q ^ y-
^ y-q ч
= ~2 {|"(x2~y2+2iiq~q2) + (q-y) (x-y+q)}
12 2 2
= —— {x -2yx+y +2q(x~y) + q }
2q^
2
- (x-yi) + U-u) 1
202 4 2 ■
Similarly,
\ fX (d-x+u) dx = - +
q. U 2q q
and
л q + x- y„ 1 У-q 4
Finally, we can express the CDF , which we are going
to denote by , in terms of the mean у and the >sta1tidard deviation a, as follows:
0, for (х <_ у - /бег)
i-Mll + i*zHi + 1 for (у _ /ба }
12а /ба
Тс(у, а; х) = Ч»(х) =^
Г J2E^~ + + (u 1 х < у + /ба)
12а
1, for (х > у + /ба)
By following the same procedure as in example 3.17,- we can compute the probabilities: P (y-a <_ x у + a), P (y-2a x <_ у + 2a) and P (у-За _< x <_ у + За) as well as the third moment m^ about zero for the triangular PDF. Again, we give here the results, and the verification is left to the students
P(y-a £ x <_ у + a) = 0.66 ,
P(y-2a £ x <_ у + 2a) = 0.97 ,
P(y-3a < x < p + 3a) =1 and m^ = у3 + За2у .
3.3 Random Multivariate
3.3.1 Multivariate, its PDF and CDF
Analogically to the ideas of stochastic function and stochastic variable given in section 3.2,1, we introduce the concept of a multivalued stochastic function
Xe {U -> RS}
in the s-dimensional space.
We note that X is a vector function, i.e. , X(u) can be written
as:
/12 s s
X(u) = (x (u) , x (u) , . .. , x (u) ) e R , u e U.
The individual components x"^ (u) £ R, j = 1, 2, . .. , s are called components or constituents of X(u) • We also note that each component x^ of the stochastic function X can be regarded as a random variable (univariate) of its
i i
own. One particular value of x may be denoted by x^*) and similarly a
particular value of X may be denoted by
= ^ir ' • • •' • Note that a specific value of X is a sequence of real numbers (not a set), or a numerical vector.
The pair (X(u), ф(X)), where
ф(Х) = ф(х\ x2, . .., xS) с {RS ~* R} (3-27)
s
*The
superscripts and subscripts here are found
very useful to distinguish
between
the components x3,
j
=
1,
2, ..., s
of the multivariate X,
and
the
elements x*?,
l
~ 1,
2, • • • , n
*
of
the univariate
(random variable)
xJ
.
We can speak of a probability of X £ [Xq, X^]С R , and define it
as follows:
(3-28)
Here the integral sign stands for the s-dimensional integration; dx for
1 2
an s-dimensional differential, i.e. dX = (dx , dx ,
cbc) and
X = (x1, x2, ... xS) , X- = (x?*, x2, .xf) are assumed to satisfy о о о о 1 11 1 л
the following inequalities:
xJ > xD Xl - Xo ' j = 1, 2,
Note that in order to be able to call the function ф a PDF, the following condition has to be satisfied:
/ ф (X) dX = 1 .
s
R
(3-29)
A complete analogy to the one-dimension 1 or univariate case (section 3.2.2) is the definition of the multivariate CDF. It is defined as follows:
X
¥(X) = / ф(У) dYe {R ■> [0, 1]}
(3-30)
where Y is an s-dimensional dummy variable in the integration.
Example 3.19: Consider the univariate PDF shown in Figure 3.12. This be11-shaped PDF is known as the normal or Gaussian PDF (to be discussed later in more details), and is usually denoted by N,
in terms of its у and a we have
ф(х) = N(y, a; x) .
Then the multivariate normal PDF in two-dimensional space, i.e. 12
ф(Х) = ф(х , x ), would appear as illustrated in Figure 3~l8.
In the two-dimensional space, ф(Х) is called a bivariate PDF,
and the bivariate normal PDF illustrated above can be expressed as
ф(Х) = N(yx, У2, а±9 a2i X)
1
2
3.3.2 Statistical Dependence and Independence
The PDF, ф, of the multivariate X may have a special form,
namely
s
ф(Х) = ф.(х1).ф.(х2) ... ф (xS) = П ф. (xJ) *
s 5=1 3
In this case, the integral in equation (3-28) can he rewritten as:
X.. X ^ .
/ ф(Х) dX = /х .П_ ф. (xJ) dxJ
S X"
= П /
О
(xJ ) dx:
(3-31)
Remembering that each component xJ of the multivariate X can be regarded
as a univariate, and regarding ф. as the PDFs of the corresponding
J
univariates we can rewrite equation (3-31) as:
s j . s ...
П fXl ф . (xJ )dxJ = П P(xJ 1 xJ £ x^ ) • j=l x^ 3 j=l °
Comparing this result with equation (3-28), we get the relationship
between the probabilities
(3-32)
This relation can be read as follows: "The combined probability of all the components satisfying the condition: x^ £ xJ <_ x^, equals to the product of the probabilities for the individual components", and obviously satisfies the definition of the combined probability of
independent events (section 2.3). Hence, the components x of such a multivariate X are called statistically independent. The PDF from example 3.19 is statistically independent.
If the PDF of a multivariate cannot be written as a product of the PDF's of its constituents, then these constituents are known as statistically dependent. In this case, the probability P(X £ X £ X^) is not equal to the product of the individual probabilities.
It can be shown that for statistically independent components
we have
/ ф. (к3) dxD = 1, j = 1, 2, ..., s.
R 3
3.3.3 Mean and Variance of a Multivariate
The sequence
Г
fl = (u., u_, ... r u ) = E*(X) e R ,
12 s
(3-33)
where
у. = / xJ ф (X) dX = E* (xJ)e R, j = 1, 2,
3 Rs
(3-34)
is called the mean of the multivariate X. The argument of the operator E (i.e. the s-dimensional integral) is X. ф (X) = (х\х2, . .. ,xS) *ф (x\x2 ,. .. Similarly the variance of the multivariate X is given by
(3-35)
where
a? = / (xj-y.)2 ф(Х) dX 3 Rs 3
= E* (x]-]j.)2) £ R, j = 1, 2, . . . , s. (3-36)
Note that we can write again
E*(X-y)2 = Е*(Х~Ё*(Х))2 = E*(X2) - ]i2 , (3-37)
and
E*(xj-y.)2 = E*(xj-E*(xj))2 = E*((xj)2) - y?; .(3-38) 3 3
The variance of the multivariate does not express the statistical properties in the multi-dimensional, space as fully as the variance of the univariate does in the one-dimensional space. For this reason, we extend the statistical characteristics of the random multivariate further and introduce the so-called variance-covariance matrix (see section 3.3.4).
Let us now turn our attention to what the mean and the variance of a "statistically independent" multivariate look like. For the statistically independent components x3, j = 1, 2, ..., s of a multivariate X, we obtain
-i s о 0
v. = f xJ П ф.(х") dx* 3 rs n=i *
= / [x3 $.(x3) ( П ф0 (x£)dxX)dxj] (3-39)
= / x^.(xj)dxj • П / ф0(x£)dxЛ . R 3 £=1 R 1
Here, according to section 3.3.2, all the integrals in equation (3-39) after the П-sign are equal to one, and thus we have
Уц = /DxV (xj) dxj, j = 1, 2, s . (3-40)
J K D
Similarly,
— 1, 2,
s.
(3-41)
Thus for the statistically independent X, we can compute the mean and
the variance of each component x3 separately , as we have computed
a0 of the PDF from example 3.19.