
Словари и журналы / Психологические журналы / p177British Journal of Mathematical and Statistical Psycholo
.pdf
177
British Journal of Mathematical and Statistical Psychology (2002), 55, 177–192
© 2002 The British Psychological Society
www.bps.org.uk
In• uential observations in the estimation of mean vector and covariance matrix
Wai-Yin Poon1 * and Yat Sun Poon2
1The Chinese University of Hong Kong
2University of California at Riverside, USA
Statistical procedures designed for analysing multivariate data sets often emphasize different sample statistics. While some procedures emphasize the estimates of both the mean vector m and the covariance matrix S, others may emphasize only one of these two sample quantities. In effect, while an unusual observation in a data set has a deleterious impact on the results from an analysis that depends heavily on the covariance matrix, its effect when dependence is on the mean vector may be minimal. The aim of this paper is to develop diagnostic measures for identifying in•uential observations of different kinds. Three diagnostic measures, based on the local in• uence approach, are constructed to identify observations that exercise undue in• uence on the estimate of m, of S, and of both together. Real data sets are analysed and results are presented to illustrate the effectiveness of the proposed measures.
1. Introduction
The multivariate normal distribution is a common distribution in analysing multivariate data. The distribution is speciŽed by a mean vector m together with a symmetric and positive deŽnite matrix S. The estimates of these parameters are usually obtained by the method of maximum likelihood, and various multivariate statistical procedures have been developed based on one or both of these sample quantities. However, it is well known that the accuracy of the estimates is affected by unusual observations in the data set, so-called outliers.
Outlier identiŽcation is an important task in data analysis because outlying observations can have a disproportionate in•uence on statistical analysis. When such distortion occurs, the outliers are regarded as in•uential observations. However, an observation
* Requests for reprints should be addressed to Wai-Yin Poon, Department of Statistics, The Chinese University of Hong Kong, Shatin, Hong Kong (e-mail: wypoon@cuhk.edu.hk).
178 Wai-Yin Poon and Yat Sun Poon
that substantially affects the results of one analysis may have little in•uence on another because the statistical procedures may emphasize different sample quantities. Therefore, in•uential observations must be identiŽed according to context.
The identiŽcation of in•uential observations or outliers of different kinds is well documented in the literature on the regression model; however, less has been achieved for multivariate analysis. Although the detection of multivariate outliers has received considerable attention (see Rousseeuw & van Zomeren, 1990; Hadi, 1992, 1994; Fung, 1993; Atkinson & Mulira, 1993; Atkinson, 1994; Rocke & Woodruff, 1996; Poon, Lew, & Poon, 2000), the emphasis has been on the identiŽcation of location outliers that in•uence the estimate of m. Observations that in•uence the estimate of S are addressed by studying their effect on the measures used to identify the location outliers. In view of the fact that statistical procedures may depend heavily on different sample quantities, methods for identifying in•uential observations of different kinds in multivariate data sets are necessary, and the aim of this paper is to develop diagnostic measures for this purpose.
There are two major paradigms in the in•uence analysis literature: the deletion approach and the local in•uence approach. The deletion approach assesses the effect of dropping a case on a chosen statistical quantity, and a typical diagnostic measure is Cook’s distance (Cook, 1977). The concept of Cook’s distance was Žrst introduced in the context of linear regression and was subsequently generalized to other statistical models (McCullagh & Nelder, 1983; Bruce &Martin, 1989). While intuitivelyconvincing measures of this kind have become very popular in in•uence analysis, it is also well known that they are vulnerable to masking effects that arise in the presence of several unusual observations. Diagnostic measures derived from deleting a group of cases are also well documented in the literature, but their practicality is in doubt because of combinatorial and computational problems.
On the other hand, the local in•uence approach is well known for its ability to detect joint in•uence. In this approach diagnostic measures are derived by examining the consequence of an inŽnitesimal perturbation on the relevant quantity (Belsley, Kuh, & Welsch, 1980; Pregibon, 1981); a general method for assessing the in•uence of local perturbation was proposed by Cook (1986). The approach starts with a carefully chosen perturbation on the model under consideration and then uses differential geometry techniques to assess the behaviour of the in•uence graph of the induced likelihood displacement function. In particular, the normal curvature Cl along a direction l, where lTl = 1, at the critical point of the in•uence graph of the displacement function is used as the diagnostic quantity. Alarge value of Cl is an indication of strong local in•uence and the corresponding direction l indicates how to perturb the postulated model to obtain the greatest change in the likelihood displacement. Moreover, the conformal normal curvature Bl transforms the normal curvature onto the unit interval and has been demonstrated to be another effective in•uence measure (Poon & Poon, 1999).
The current study uses the local in•uence approach to develop diagnostic measures for identifying in•uential observations that affect the estimate of the mean, of the covariance matrix, and of both. These measures inherit the nice property of the local in•uence approach in its abilityin detecting joint in•uence, hence multiple outliers that share joint in•uence can be identiŽed simultaneously. It is worth noting that while a ‘masking’ effect arises in the presence of multiple outliers, there are two distinct notions of masking effect, namely the joint in•uence and the conditional in•uence (Lawrance, 1995; Poon & Poon, 2001). The proposed diagnostic measures, like other measures

In• uential observations 179
developed by the local in•uence approach, are capable of addressing masking effects under the category of joint in•uence but not of conditional in•uence.
Particulars about the local in•uence approach will be given in the next section where we will also use the approach to develop measures for identifying in•uential observations of different kinds The results of analysing real data sets using the proposed measures are then presented as illustrations.
2. In• uential observations in the estimation of normal distribution parameters
2.1. Case-weights perturbation
Let X be a p ´ 1 random vector distributed as N(m, S), where S = f jabg is a symmetric and positive deŽnite matrix, and let f xi , i = 1, . . . , ng be a random sample. The maximum likelihood estimates mˆ of m and Sˆ of S are obtained by maximizing the loglikelihood given by
L(u) = |
i= 1 |
³± |
2 |
p log(2p) ± |
2 log | S| ± |
2 (xi ± m)TS± 1(xi ± m)´, |
(1) |
|
n |
|
|
|
|
|
|
|
X |
|
1 |
|
1 |
1 |
|
|
|
|
|
|
where u = (mT, sT)T is a q = p + p vector storing the elements in m and the lower triangular elements of S, and p
= p( p + 1)/2. It can be shown that the maximum likelihod estimate uˆ = (mˆ T, sˆ T)T of u is given by (Anderson, 1958)
mˆ = x¯ = |
Sin= 1xi |
|
ˆ |
S = |
Sin= 1 |
(xi ± x¯ )(xi ± x¯ )T |
|
(2) |
n |
, |
S = |
|
n |
. |
|||
|
|
|
|
|
|
|
In order to assess the in•uence of individual observations on the estimate of u, we follow Cook (1986) and introduce the case-weights perturbation to the log-likelihood. The resulting perturbed likelihood is given by
|
n |
vi ³± |
1 |
1 |
1 |
m)TS± 1(xi ± m)´, |
|
||
L(u, | v) = |
i= 1 |
(3) |
|||||||
2 p log (2p) ± |
2 log | S| ± |
2 (xi ± |
|||||||
|
X |
|
|
|
|
|
, . . . , q )T is deŽned on |
||
where q , i = 1, . . . , n, are perturbation parameters and v = |
(q |
||||||||
i |
|
|
|
|
|
1 |
n |
|
|
a relevant perturbation space Q of fn . For example, Q may be the space such that |
|||||||||
|
|
ˆ |
ˆ |
|
|
|
|
|
0 #qi #1 for all i. Let u and uvbe the quantities that maximize (1) and (3), respectively;
then the discrepancy between them can be measured by the likelihood displacement function
ˆ |
ˆ |
(4) |
f (v) = 2(L(u| v0) ± L(uv| v0)). |
This function has its minimum value at v = v0 and we have L(u| v0) = L(u) if v0 = 1, which is an n ´ 1 vector of 1s. When the perturbation speciŽed by v causes a substantial deviation of uˆ v from uˆ , a substantial deviation of the function f (v) from f (v0) is induced. Therefore, examining the changes in f (v) as a function of v will enable the identiŽcation of in•uential perturbations that in turn will disclose in•uential observations. Cook (1986) proposed quantifying such changes bythe normal curvature Cl of the in•uence graph g of f (v) along a direction l at the optimal point v0, where l with lTl = 1 deŽnes a direction for a straight line in Q passing through v0. A large value of Cl is an indication that the perturbation along the corresponding direction l induces a considerable change in the likelihood displacement.

180 Wai-Yin Poon and Yat Sun Poon |
|
|
|
|
|
||
2.2. Normal curvature as an in• uential measure |
|
|
|
||||
Let |
È |
|
|
|
|
|
|
L and D be q ´ q and q ´ n matrices with elements respectively given by |
|
||||||
|
¨ |
¶2L(u| v0) |
|
|
¶2L(u| v) |
|
(5) |
|
Lij = |
¶vi ¶vj |
, |
Dij = |
¶vi ¶qj |
. |
|
|
|
ˆ |
|
ˆ |
|
||
|
|
|
| u= u |
|
|
| u= u,v= v0 |
|
Cook (1986) noted that the normal curvature of g in a direction l at the point v0 could be computed by
C = ± 2l |
T T È ± 1 |
D)l |
ˆ |
T È |
ˆ |
. |
(6) |
(D L |
= ± 2l Fl |
||||||
l |
|
| u= u,v= v0 |
|
| u= u,v= v0 |
|
|
The direction lmax along which the greatest change in the likelihood displacement is observed identiŽes the most in•uential observations. The direction lmax is that which
gives Cmax = maxl Cl, and Cmax together with |
lmax are the largest eigenvalue and |
associated eigenvector of the symmetric matrix |
È |
± 2F in (6). |
When Cl is deŽned on an unbounded interval and it may be difŽcult to judge its magnitude, the conformal normal curvature is a one-to-one transformation of the normal curvature onto the unit interval. Along the direction l at the critical point v0, the conformal normal curvature is given by
TDT È ± 1D
l L l
Bl = ± q •
tr DT È ± 1D 2
( L ) | u=
|
T È |
|
|
l Fl |
(7) |
uˆ ,v= v0 |
= ± ptr FÈ 2•| u= uˆ ,v= v0 . |
Let Ej, j = 1, . . . , n, be vectors of the n-dimensional standard basis. Poon and Poon (1999) demonstrated that BEj = Bj, j = 1, . . . , n, are effective measures for identifying the in•uential perturbation parameters when Cmax is sufŽciently large. Moreoever, the computation of Bj, j = 1, . . . , n, is easy because Bj is the jth diagonal element of the matrix
È |
|
|
T È |
± 1 |
D |
|
|
|
F |
|
D L |
|
(8) |
||
± |
|
|
= ± |
qtr (DTLÈ ± 1D)2•| u= uˆ ,v= v0 . |
|||
ptr FÈ 2•| u= uˆ ,v= v0 |
Clearly, to develop diagnostic measures for uncovering observations that in•uence
|
È |
the estimates of both m and S, it is necessary to compute the matrix F. We discuss such |
|
computation in the next subsection. |
|
2.3. Observations in• uencing the estimates of m and S |
|
È |
È |
There are two components in the matrix F: the q ´ q matrix L and the q ´ n matrix D. |
È |
|
|
|
|
|
|
|
|
|
|
|
|
Because ± L is the observed information for the postulated model and the maximum |
||||||||||||
likelihood estimates mˆ of m and sˆ |
of s are statistically independent, |
È ± |
1 |
is a diagonal |
||||||||
L |
|
|||||||||||
block matrix given by |
|
± Cov(sˆ ) |
|
Á f |
|
|
g |
LÈ ± 1 |
!, |
|
(9) |
|
LÈ 1 = |
0 |
= |
0 |
|
|
|||||||
± |
³ |
± Cov(mˆ ) |
0 |
´ |
|
È ± |
1 |
|
0 |
|
|
|
|
Lab |
|
|
|
|
|||||||
|
|
|
|
|
|
|
f (ab)(gr)g |
|
|
|
where Cov(mˆ ) is a p ´ p matrix storing the covariance matrix of mˆ and Cov(sˆ ) is a p ´ p
matrix storing the covariance matrix of sˆ . For the sake of clarity, we denote an element
in ± Cov( ˆ ) by È ± 1 when it relates to the covariance between the ath and bth elements
m Lab
mˆ a and mˆ b of ˆ , and an element in Cov( ˆ ) by È ± 1 when it relates to the covariance
m ± s L(ab)(gr)

|
|
|
In• uential observations |
181 |
between jˆ ab and jˆ gr. Using such notation, we have (Anderson, 1958) |
|
|||
È ± 1 |
= ± (Cov(mˆ ))ab = ± Cov(mˆ a , mˆ b) = |
|
1 |
(10) |
Lab |
± |
n jab, |
||
and |
|
|
|
|
È ± 1 |
|
1 |
|
|
L(ab)(gr) = ± (Cov(sˆ ))(ab)(gr) = ± Cov(jˆ ab, jˆ gr) = ± |
n |
(jagjbr + jarjbg). |
(11) |
Moreover, we denote the elements in the ith column of D by |
|
|||||
Dai = |
¶2L |
or |
D(ab)i = |
¶2L |
, |
(12) |
¶ma ¶qi |
¶jab¶qi |
depending on whether they correspond to ma in m or jab in S, respectively. Using (6) |
||||||||
|
|
|
|
È |
|
|
|
|
and (9), the (i, j)th element of the matrix F becomes |
|
|
|
|||||
|
|
¨ |
È ± 1 |
|
|
È ± 1 |
|
(13) |
|
|
Dai Lab |
Dbj + |
D(ab)i L(ab)(gr)D(gr) j . |
|
|||
|
|
Fij = |
|
|||||
|
|
|
1#a,b#p |
1#b#a#p |
|
|
|
|
|
|
|
X |
# # |
|
|
|
|
|
|
|
1#Xr g |
p |
|
|
|
|
By (3), |
|
|
|
|
|
|
|
|
|
¶L |
|
1 |
1 |
1 |
m)TS± 1(xi ± |
|
|
|
|
= ± |
2 p log (2p) ± |
2 log | S| ± |
2 (xi ± |
m). |
(14) |
|
|
¶qi |
Let ea be a p ´ 1 vector with 1 as its ath element and zeros elsewhere, let yi = xi ± m be |
||||||||||
a p ´ 1 vector with ybi |
as its bth coordinate, and denote the (a, b)th element of S± 1 by |
|||||||||
jab± 1. From (14), |
|
|
|
|
|
|
|
|
|
|
|
|
¶2L |
Xb |
|
|
Xb |
|
|
||
Dai = |
|
|
= yiTS± 1ea = |
ybijba± 1 = |
jab± 1 ybi . |
(15) |
||||
|
¶ma ¶qi |
|||||||||
Therefore, the Žrst summand on the right-hand side of (13) becomes |
|
|||||||||
X |
1 |
|
1 |
X |
|
|
1 |
|
||
a,b,c,d jac± 1 yci³± n jab´jbd± 1 ydj = ± |
n c,d |
yci jcd± 1 ydj = ± n yiTS± 1 yj . |
(16) |
The second summand on the right-hand side of (13) can be obtained numerically by
the expression in (11) and the following expression derived from (14): |
|
||||||||||
|
¶2L |
|
1 ¶ log | S| |
|
1 |
¶S± 1 |
yi = ± Ájab± 1 ± |
Xa,b |
|
!. (17) |
|
D(ab)i = |
|
= ± |
2 |
|
± |
2 yiT |
|
jaa yai jb±b1 ybi |
|||
¶jab¶qi |
¶jab |
¶jab |
It is also possible to compute the second summand as follows:
b X |
È ± 1 |
X |
È ± 1 |
X |
|
D(ab)i L(ab)(gr)D(gr) j = |
|
D(aa)iL(aa)(gg)D(gg) j + |
D(ab)i |
#a,r#g |
|
a,g |
|
b<a,r<g |
È ± 1 D
L(ab)(gr) (gr) j
|
|
X |
|
|
|
bX |
|
|
+ |
È |
± 1 |
+ |
È ± 1 |
||
|
D(aa)iL(aa)(gr)D(gr) j |
D(ab)i L(ab)(gg)D(gg) j |
|||||
|
|
a,r<g |
|
|
|
<a,g |
|
|
1 |
X |
|
± 1 |
|
X |
± 1 |
= |
4 |
( a,g |
D(aa)i LÈ |
(aa)(gg)D(gg) j |
+ |
a,g,r D(aa)iLÈ |
(aa)(gr)D(gr) j |
|
+ aX,b,g D(ab)i LÈ (±ab1 )(gg)D(gg) j + |
aX,b,g,r D(ab)i LÈ (±ab1 )(gr)D(gr) j). (18) |

182 Wai-Yin Poon and Yat Sun Poon
Using (11) as well as (17), we obtain the following expressions for the summands of
(18): |
|
|
a g |
(jaa± 1jag2 jgg± 1 ± jaa± 1jag2 |
"Á a |
jg±a1 yaj |
! |
a,g D(aa)i LÈ (aa)(gg)D(gg) j = ± n |
|||||||
X |
|
2 |
XX |
|
X |
|
2 |
|
± 1 |
|
|
|
|
|
|
|
+ Á |
jg±a1 yai! |
# + jag2 |
Á |
ja± a1 yai ! Á |
jg±a1 yaj! |
); |
|||||
|
|
Xa |
|
|
2 |
|
|
Xa |
2 |
Xa |
|
2 |
|
|
|
|
|
|
|
|
Á a |
! |
#); |
||||
a,g,r D(aa)iLÈ (aa)(gr)D(gr) j = ± n a |
([jaa± 1 ± ya2j]"jaa± 1 ± |
ja± a1 yai |
|||||||||||
X |
|
|
2 |
X |
|
|
|
|
X |
|
2 |
|
|
|
± 1 |
|
|
([jgg± 1 ± yg2i ]"jgg± 1 ± |
Á a |
|
! #); |
||||||
|
D(ab)iLÈ (ab)(gg)D(gg) j = ± n g |
jg±a1 yaj |
|||||||||||
aX,b,g |
|
|
2 |
X |
|
|
|
X |
|
2 |
|
|
|
|
± 1 |
|
|
|
|
|
|
|
|
|
|
|
|
aX,b,g,r D(ab)iLÈ |
± 1 |
|
2 |
p ± yjTS± 1yj ± yiTS± 1yi + (yiTS± 1yj)2g . |
|
||||||||
(ab)(gr)D(gr) j |
= ± |
n f |
|
(19)
(20)
(21)
(22)
Using (13), (16), and (19) to (22), it is possible to compute |
È |
and hence the |
F |
diagnostic measures. SpeciŽcally, observations that unduly in•uence the estimates of
the mean and the covariance matrix can be isolated byexamining lmax or Bj, j = 1, . . . , n. That is, the elements with large Bj values or large magnitudes in lmax are the group of
in•uential observations.
Note that observations so identiŽed are in•uential on the estimates of both m and S. However, as there may be observations in the data set that in•uence the estimate of m but not the estimate of S, or vice versa, it is also of interest to develop diagnostic measures for identifying these.
2.4. Observations in• uencing the estimate of m or S
The diagnostic measures developed in Section 2.3 are developed based on the in•uence graph given in (4) and hence the effects of the perturbation on estimates of all parameters in u are taken into account. When the effects on only a subset of the parameters is of interest, Cook (1986) demonstrated that the effects could be assessed by examining the normal curvature of the in•uence graph of an objective function deduced from (4). SpeciŽcally, if one is only interested in the effects on the estimate of m, one can examine the normal curvature given by
m |
|
T |
T È m ± 1 |
D)l |
ˆ |
|
|
T |
È m |
l |
ˆ |
, |
(23) |
||
C |
= ± 2l (D L |
) |
= ± 2l F |
|
|||||||||||
l |
|
|
|
|
|
| u= u,v= v0 |
|
|
|
|
| u= u,v= v0 |
|
|
||
where |
(LÈ ) 1 = |
³ |
|
0 |
0 |
´ = |
Áf |
0 |
|
g |
|
0 ! |
|
(24) |
|
|
|
1 |
|
|
|||||||||||
|
m ± |
|
|
± Cov(mˆ ) 0 |
|
|
È ± |
|
|
0 |
|
|
|||
|
|
|
|
|
Lab |
|
|
|
|
and D is as in Section 2.3 (see (12)). Using (10), (12), (15) and (16), the (i, j)th element of
È m is found to be
F
m |
± 1 |
|
1 |
|
FÈ i j = |
1#Xa,b#p DaiLÈ ab |
Dbj = ± |
n yiTS± 1yj. |
(25) |
In• uential observations 183
By similar arguments to those given in Section 2.2, observations that exert a dispropor-
tionate in•uence to the estimate of m can be located by examining the eigenvector lm |
||||||||||
|
|
|
|
|
È m |
|
|
max |
||
|
|
|
|
|
or by examining the diagonal elements |
|||||
associated with the largest eigenvalue of ± 2F |
|
|||||||||
atj |
u = uˆ and v = v0. |
|
|
p |
• |
|
|
|
|
|
m |
|
È m |
È m |
2 |
|
|
|
È m |
are evaluated |
|
B |
, j = 1, . . . , n, of the matrix ± F |
/ tr (F ) |
|
, where the unknowns in F |
||||||
|
Similarly, let |
|
|
|
|
|
|
|
|
|
|
È s± 1 |
|
0 |
0 |
´ |
|
0 |
0 |
|
|
|
(L ) |
= |
³0 |
± Cov(sˆ ) |
= |
Á 0 |
f LÈ (±ab1 )(gr)g !. |
(26) |
The normal curvature that re•ects the in•uences of the perturbation on the estimates of S is given by
|
|
|
|
|
s |
|
T |
T |
È s |
± |
1 |
D)l |
ˆ |
|
T È s |
l |
ˆ |
, |
(27) |
|
|
|
|
|
|
C = ± 2l |
|
(D |
(L |
) |
|
|
, = ± 2l F |
||||||||
|
|
|
|
|
l |
|
|
|
|
|
|
|
|
| u= u,v= v0 |
|
|
| u= u,v= v0 |
|
||
where |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
È s |
|
|
|
|
È ± 1 |
|
|
|
|
|
||
|
|
|
|
|
|
|
|
Fi j = |
|
|
|
D(ab)i L(ab)(gr)D(gr) j , |
|
|
|
|
||||
|
|
|
|
|
|
|
|
|
|
1#b#a#p |
|
|
|
|
|
|
|
|||
|
|
|
|
|
|
|
|
|
|
|
|
# # |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1#Xr g p |
|
|
|
|
|
|
|
|||
which can be computed using (18) to (22). Let lmaxs |
be the eigenvector associated with |
|||||||||||||||||||
the largest |
eigenvalue |
of |
|
È s |
|
|
ˆ |
|
and |
s |
be the |
|
jth |
diagonal |
element of |
|||||
± 2F |
|
|
|
B |
|
|||||||||||||||
È s |
p |
È s |
• |
|
|
|
|
| u= u,v= v0 |
j |
|
|
|
|
|
||||||
|
2 |
ˆ |
; observations that in•uence the estimates of S can be detected |
|||||||||||||||||
± F / |
|
tr(F |
) |
|
||||||||||||||||
by ls |
|
|
|
|
| u= u,v= v0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
or Bs, j = 1, . . . , n. |
|
|
|
|
|
|
|
|
|
|
|
|
|
||||||
max |
|
j |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3. Examples
Example 1: Brain and body weight data set
As an illustration, we Žrst consider the brain and body weight data set (in logarithms to base 10) which is available in Rousseeuw and Leroy (1987, p. 58). The data set consists of observations for 28 species on two variables: body weight and brain weight. From a scatter plot of the data (Fig. 1) it can be seen that the two variables exhibit a positively correlated pattern. Many analyses of this data set have been carried out in the context of outlier identiŽcation. For example, Rousseeuw and van Zomeren (1990) used a robust version of the Mahalanobis distance to conclude that cases 25, 6, 16, 14 and 17, in this order, are outlying observations and that the effect of case 17 is marginal. Atkinson and Mulira (1993), on the other hand, reached a similar conclusion using the Mahalanobis distance in a forward identiŽcation technique with results summarized visually in a stalactite plot. Poon et al. (2000) used the local in•uence approach to identify location outliers under various metrics. Different cases were identiŽed as outliers under different metrics, but cases 25, 6 and 16 were always •agged as outliers when the metric used had taken into account the shape of the data set.
We reanalysed the data set using the proposed procedure and plotted in Fig. 2 the |
|||||
m |
s |
, j = |
È È m |
È ss |
|
values of Bj, Bj |
and Bj |
1, . . . , n, computed from F, F |
and F |
respectively. The |
results show that cases 25, 6 and 16, in this order, are most in•uential on the estimates of m and S. From Figs. 2b and 2c, we conclude that case 20 affects the estimate of m substantially but its effect on the estimate of S is less pronounced. Note from Fig. 1 that as case 20 lies at the lower left-hand corner with smallest values in both variables, it therefore affects the location of the data set substantiallybut its effect on the dispersions or correlation of the variables is not conspicuous. A similar phenomenon can also be

184 Wai-Yin Poon and Yat Sun Poon
Figure 1. Scatter plot of the brain and body weight data set.
observed for cases 15, 19 and 7. On the other hand, the three extreme cases, 25, 6 and 16, are located outside the ellipse formed by the majority of the data; while they are in•uential on the estimates of both m and S, their effects on S are more pronounced than those on m. These results, therefore, indicate that the proposed measures have identiŽed what they are supposed to identify.
Example 2: Head data set
The second data set, originally from Frets (1921), is also contained in Seber (1984, p. 263). It consists of measurements of head lengths and breadths of the Žrst and second adult sons in 25 families. Pairwise scatter plots among the four variables are given in Fig.
3, and several cases are marked for further discussion. The computed values of Bj, Bjm |
||
ss |
È È m |
È ss |
and Bj , j = |
1, . . . , n, from F, F |
and F are presented in the form of index plots in Fig. 4. |
Three cases, 2, 17 and 6, are classiŽed as in•uential on the estimates of both m and S. From Fig. 3, we see that these three observations are usually located at the boundaries of

In• uential observations 185
Figure 2. Index plots of the in• uential measures for the brain and body weight data set: (a) Bj; (b) Bmj ;
(c) Bsj .
the data point clouds formed by different pairs of variables, hence they are in•uential on both the estimates of m and S. On the other hand, Figs. 4b and 4c show that Bm16 is relatively large but Bs16 is not, which leads to the conclusion that the in•uence of case 16 on the estimate of m is substantial but on S is not noticeable. This Žnding makes sense

186 Wai-Yin Poon and Yat Sun Poon
Figure 3. Scatter plots for the head data set.
because it can be seen from Fig. 3 that case 16, which has the smallest observed values on all four variables, always lies in the lower left-hand corner of the scatter plots; its in•uence on the location of the data set is therefore considerable.
Following the suggestion of a reviewer, we reduced the dimensionality to p = 2 by