- •Introduction to adjustment calculus
- •Introduction to adjustment calculus (Third Corrected Edition)
- •Introduction
- •2. Fundamentals of the mathematical theory of probability
- •If d'cd; then p (d1) £ lf
- •Is called the mean (average) of the actual sample. We can show that m equals also to:
- •3.1.4 Variance of a Sample
- •Is called the variance (dispersion) the actual sample. The square root 2
- •In the interval [6,10] is nine. This number
- •VVII?I 0-0878'
- •In this case, the new histogram of the sample £ is shown in Figure 3.5.
- •Is usually called the r-th moment of the pdf (random variable); more precisely; the r-th moment of the pdf about zero. On the other hand, the r-th central moment of the pdf is given by:
- •3.2.4 Basic Postulate (Hypothesis) of Statistics, Testing
- •3.3.4 Covariance and Variance-Covariance Matrix
- •X and X of a multivariate X as
- •It is not difficult to see that the variance-covariance matrix can also be written in terms of the mathematical expectation as follows:
- •3.3.6 Mean and Variance-Covariance Matrix of a Multisample The mean of a multisample (3.48) is defined as
- •4.2 Random (Accidental) Errors
- •It should be noted that the term иrandom error" is used rather freely in practice.
- •In order to be able to use the tables of the standard normal
- •X, we first have to standardize X, I.E. To transform X to t using
- •Is a normally distributed random
- •4.10 Other Measures of Dispersion
- •The average or mean error a of the sample l is defined as
- •5. Least-squares principle
- •5.2 The Sample Mean as "The Maximum Probability Estimator"
- •5.4 Least-Sqaures Principle for Random Multivariate
- •In very much the same way as we postulated
- •The relationship between e and e for a mathematical model
- •6.4.4 Variance Covariance Matrix of the Mean of a Multisample
- •Itself and can be interpreted as a measure of confidence we have in the correctness of the mean £. Evidently, our confidence increases with the number of observations.
- •6.4.6 Parametric Adjustment
- •In this section, we are going to deal with the adjustment of the linear model (6.67), I.E.
- •It can be easily linearized by Taylor's series expansion, I.E.
- •In which we neglect the higher order terms. Putting ax for X-X , al for
- •The system of normal equations (6.76) has a solution X
- •In sections 6.4.2 and 6.4.3. In this case, the observation equations will be
- •In matrix form we can write
- •In metres.
- •6.4.7 Variance-Covariance Matrix of the Parametric Adjustment Solution Vector, Variance Factor and Weight Coefficient Matrix
- •I.E. We know the relative variances and covariances of the observations only. This means that we have to work with the weight matrix к£- 1
- •If we develop the quadratic form V pv 3) considering the observations l to be influenced by random errors only, we get an estimate к for the assumed factor к given by
- •Variance factor к plays. It can be regarded as the variance of unit
- •In metres,
- •Is satisfied. This can be verified by writing
- •Into a . О
- •6.U.10 Conditional Adjustment
- •In this section we are going to deal with the adjustment of the linear model (6.68), I.E.
- •For the adjustment, the above model is reformulated as:
- •Is not as straightforward, as it is in the parametric case (section 6.4.6)
- •VeRn VeRn
- •Into the above vector we get 0.0
- •0.0 In metres .
- •In metres.
- •Areas under the standard normal curve from 0 to t
- •Van der Waerden, b.L., 1969: Mathematical Statistics, Springer-Verlag.
5.4 Least-Sqaures Principle for Random Multivariate
So far, we have shown that the least-squares principle spells out the equivalence between the sample mean £ and the estimate for the parent population mean у determined from the condition that the sum of discrepancies be minimum. We have also shown that £ is the most probable estimate for providing the parent population is postulated to have normal or any other symmetrical PDF. We shall show now that the same principle is valid even for random multisample if we postulate the underlying PDF to be statistically independent (see Section 3.3.2).
Denoting the multisample by L and its components by I?, j = 1, 2, —, s, and remembering that each I? is a sample on its own, we can write:
(5-19)
L =: (L , L , • • • , L )
L = (£_, £ _ , • . . , £ ) с R
12 n.
3
Assuming a particular value Lq for the multisample L, where
L = (£1, £2, ..., £S) eRS (5-20)
о о о о
is a numerical vector (sequence of real numbers), the associated discrepancies V, which can be regarded as a multisample as well, are:
V = (L-Lq) = (V1, V2, VS) . (5-21)
Here, each V~*, j = 1, 2, .. ., s is a sample of discrepancies on its own, i.e.
VD = (V?, vj, ..., VD ) eRS . (5-22)
12 nj
Making use of formula (.3-52), we can write analogically to (5.2) s*2 = ^- ZJ (l* - £J)2 = E[(VJ)2]. (5-23)
i 2
The minimization of the variances, i.e. minimization of each E[(VJ) ],
2
is equivalent to the minimization of each S* , or as we usually write:
J
(min [E(VJ)2]5 j = 1, 2, .., s)s min [trace Zf] *\ (5-2*0
L eRS L eRS L
о о
where Tff is the variance-covariance matrix of the multisample L (see section 3.3.6). By carrying out this operation^ similar to section 5»15 we will find that the vector
Lq = (J1, I2, . .., Is) eRS (5-25)
satisfies the condition (5-2*0. On the other hand, the result (5-25)
is nothing else but the mean L of the multisample, i.e.:
L = L eRS . (5-26)
о
Postulating a normal PDF, N(£J , S*; £J), for each component LJ of the multisample L, the multivariate PDF of the parent population can be written as:
^ s
ф(А) = ni(HJ, S*; tJ) j=l ° J
=
П
—±
exp
[-
p°
], (5-27)
j=l S*/(2tt) 2S*
J J
where I3 is the random variable having mean and standard deviation S*.
о J
Following a similar procedure as in section 5.2/ we end up again with with the discovery that the vector
*) Trace of a matrix is the sum of its diagonal elements.
LQ S L cs ( I1, £2, £S) eRS (5-28)
maximizes the probability that the members of the parent population will occur at the same places as the members of the multisample L.
*"** s
Hence LeR is, under the above conditions,*) the maximum probable estimator for the mean у of the postulated parent multivariate PDF, where
T= (Pr U2* • , Уд) £RS . (5-29)
5* 5 Exercise 5
1. Prove that the mean у of a continuous PDF, ф(х) , defined as:
00
у » / x ф (x) dx
— со
2
minimizes the PDF variance a , defined as : a2 = / (x-y)2 ф(х) dx.
— 00
as*
2. Prove that = 0, is the necessary and sufficient condition for the
rectangular (uniform) PDF, R( I S*; l) , to be the most probable
2
underlying PDF for a sample L with mean % and variance S . Note that the analytic expression for the uniform PDF is given in example 3.175 section 3.2.5-
*) It can be shown that L is the maximum probability estimator of у even . when we postulate a statistically dependent multi dimensional PDF from a certain family of PDFT s.
3. Prove that the same holds for "the triangular- PDF, T(£°, S*; £), using its analytic expression given in example 3.18, section 3.2.5.
бв ^ШРА^№\ХЗ OF ADJUSTMENT CALCULUS 6.1 Primary and Derived Random Samples
So far, we have been dealing with random samples (multisamples) that had been obtained through some measurement or through any other data collecting process. These samples may all be regarded as primary or original random samples (multisamples).
In practice, we are often interested in other samples that would be derived from the primary samples by means of a computation of some kind. Such samples may be called derived random samples (multisamples).
From the philosophical point of view, there is not much difference between these two, since even the "primary" samples may be regarded as derived from the samples of physical influences or physical happenings. However, it is necessary to distinguish between them to be able to speak about the transition from one to the other.
6.2 Statistical Transformation, Mathematical Model The transition from a primary to a derived sample (multisample) along with the associated variances and covariances may be called statistical t r an s format i on. We have already met two examples of such t r an s f or ma t ion although applied to random variable rather than sample (see sections h.5 and 4.6), namely the transformation of the Gaussian PDF to the normal and to the standard normal PDFfs, respectively.
Such statistical transfоrmation may not always be as simple as in the above two cases. As a matter of fact, it may not be even possible to derive the sample at all from the primary sample which is usually
the case with multisamples. In other words, it might not be possible to express the derived sample explicitly in terms of the primary sample.
Let us consider a primary multisample L = (L1), i = 1, 2j( . .. s,
that has s constituents. Each constituent L1 = (£ 1), к = 1, 2, . . .^n.,
к. I
is a random sample on its own and represents a distinct physical quantity
(i.e. the observations £^.1, к = 1, 2, . . .^n^ are all representing the same physical quantity £_^). Now, we may be interested in deriving a multi-sample X having n constituents, ie.
X = (XJ) , j=l, 2 y . . ,n,
i
from the original multisample L; noting again that each constituent X represents a distinct physical quantity x., J - 1, 2, . . .,n. The formulae (relationships) relating the physical quantities £ and x, where
£ — (£-^ 5 ^2* * * " } ) 9
and \ (6.D»
x = (^~2_5 ^25 * • • > ^"jj ) are called the mathematical model for the statistical transformation; and is usually expressed as:
F (£ , x) = 0
where F denotes the vector of functions f , i = 1, 2, . . . , r(having r components)that can be established between £ and x.
To be able to derive x from £, the mathematical model (6.2) should be formulated as:
* Note that £ and x are nothing else but the multivariates corresponding to the multisamples L and X respectively.
x = F (£), (6.3)
which gives x as explicit function of I.
Example 6.1: After having measured the two perpendicular edges a and Ъ
of a rectangular desk (see Figure 6.1), suppose that we are interested in learing something about the length of the diagonal d, and about the surface area a of this disk. In this case, the mathematical model will be written as: x = F(£), where
x = (x1? x2) = (d, a ), and
1 = ^1' l2^ = ^5 Ъ^ *
To derive the components of x from I we write: d = f (a, b) = /(a2 + b2) , and
x-, |
|
м. - d |
1 |
— |
|
x0 |
|
a |
. 2 |
|
a = f (a, b) = ab . In vector notation, we can write:
V(a2 + Ъ*)
I JL 1
x =
ab
The possibility .of carrying out the statistical transformation depends basically on three factors:
(i) complexity of the mathematical model, i.e., the possibility of expressing x explicitly in terms of i (x - F(i));
(ii) "completeness" of the primary multisample L, i.e. whether all its con- stituents have the same number of elements in order to deduce the variance- covariance matrix ST;
(iii) our willingness to match the individual s-tuples of elements from the primary multisample L with the n-tuples of elements from the derived multisample X, which creates much of a problem.
Particularly the last two factors are so troublesome that we usually do not even try to carry out the transformation and put up with some statistical estimates, i.e. representative values E(X) and I for the derived multisample instead. To do so, we first evaluate E(L) and ZL for the primary multisample L, from which we then compute the statistical estimates E(x) and £ for X.
-A.
According to the basic postulate of the error theory and to make the subsequent development easier, we generally postulate at this stage the PDF of the parent multivariate to the multisample L and assume
E(L) = E*(£), EL.= E*r E(X) = E*(x), and Zx = E£ (6Л)