Добавил:
Upload Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:
FirstCourseGR.pdf
Скачиваний:
38
Добавлен:
07.03.2016
Размер:
7.41 Mб
Скачать

37

 

2.2 Vector algebra

 

 

 

 

 

 

In the last line we use the summation convention (remember always to write the index on e as a subscript in order to employ the convention in this manner). The meaning of Eq. (2.11)

 

 

 

 

A is the linear sum of four vectors A0e , A1e , etc.

is that

 

0

1

Transformation of basis vectors

The discussion leading up to Eq. (2.11) could have been applied to any frame, so it is equally true in O¯ :

 

 

 

 

α

 

 

 

 

 

 

A

=

Aα¯ e

.

 

 

 

 

 

 

¯

 

¯

 

¯

 

A is also the sum of the four vectors A0¯

1¯

 

e0

e1, etc. These are not the same

This says that

 

 

 

 

, A

 

 

 

 

 

 

 

 

O

four vectors as in Eq. (2.11), since they are parallel to the basis vectors of ¯ and not of

 

 

A. It is important to understand that the expressions

O, but they add up to the same vector

 

 

 

 

 

 

 

¯

are not obtained from one another merely by relabeling dummy indices.

Aα eα

and Aα¯ eα

Barred and unbarred indices cannot be interchanged, since they have different meanings.

Thus,

{

Aα¯

}

is a different set of numbers from

{

Aα

}

, just as the set of vectors

{ ¯

}

is different

 

 

 

 

eα

 

from {eα }. But, by definition, the two sums are the same:

 

 

 

 

 

 

 

Aα e

 

 

Aα¯ e

,

 

 

(2.12)

 

 

 

 

α =

 

α¯

 

 

 

 

and this has an important consequence: from it we deduce the transformation law for the basis vectors, i.e. the relation between {eα } and {eα¯ }. Using Eq. (2.7) for Aα¯ , we write Eq. (2.12) as

α¯ β Aβ eα¯ = Aα eα .

On the left we have two sums. Since they are finite sums their order doesn’t matter. Since the numbers α¯ β and Aβ are just numbers, their order doesn’t matter, and we can write

 

 

 

Aβ α¯

 

e

=

Aα e .

 

 

 

 

¯

 

 

 

β α¯

 

 

 

 

α

¯

 

Now we use the fact that β and

 

 

 

 

 

 

 

 

 

 

 

 

 

¯,

α are dummy indices: we change β to α and α to

 

 

 

 

¯

 

 

 

 

 

 

 

 

 

 

 

β

 

 

 

A

α e

 

 

 

A

α

eα .

 

 

 

 

 

α

 

β

 

¯

=

 

 

 

 

 

 

 

 

 

 

 

 

 

β

 

 

 

 

 

 

 

 

 

This equation must be true for all sets

{

Aα

}, since

A is an arbitrary vector. Writing it as

 

 

 

 

 

 

 

 

 

 

 

 

 

 

A ( ¯ α e

 

 

 

eα ) 0

 

 

 

 

α

 

β

 

 

¯

 

 

 

=

 

 

 

 

 

 

 

 

 

β

 

 

 

 

 

 

 

 

 

we deduce

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

β

 

¯

=

 

 

 

 

 

 

 

 

 

 

 

 

¯

0 for every value of α,

 

α e

β

eα

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

or

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

=

 

β

 

 

 

¯

 

 

 

 

 

 

 

 

 

¯

 

 

.

 

 

(2.13)

 

 

 

 

eα

 

 

α e

β

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

38

Vector analysis in special relativity

This gives the law by which basis vectors change. It is not a component transformation: it gives the basis {eα } of O as a linear sum over the basis {eα¯ } of O¯ . Comparing this to the law for components, Eq. (2.7),

A

¯

=

¯

α A ,

 

β

 

β

α

we see that it is different indeed.

The above discussion introduced many new techniques, so study it carefully. Notice that the omission of the summation signs keeps things neat. Notice also that a step of key importance was relabeling the dummy indices: this allowed us to isolated the arbitrary Aα from the rest of the things in the equation.

An example

O

 

 

 

 

 

 

 

 

O

 

β

 

 

 

 

 

 

 

 

. Then the matrix [ ¯ α ] is

Let ¯ move with velocity v in the x direction relative to

 

 

 

( ¯ α )

 

 

0

0

,

 

 

 

 

γ

 

 

 

 

β

=

 

γ

 

0

0

 

 

 

0

0

 

1

0

 

 

 

 

0

0

 

0

1

 

 

 

 

 

 

 

 

 

 

 

 

 

where we use the standard notation

 

 

 

 

 

 

 

 

 

 

γ := 1/(1 v2).

 

 

 

A

(5, 0, 0, 2), we find its components in

O

by

 

 

Then, if

 

 

 

 

 

 

¯

 

 

 

 

O

¯ ¯ ¯

A0 = 00A0 + 01A1 + · · ·

=γ · 5 + () · 0 + 0 · 0 + 0 · 2

=5γ .

Similarly,

 

 

 

 

A1¯ = −5,

 

 

 

 

A2¯ = 0,

 

 

 

 

 

 

A3¯ = 2.

 

 

A

(5γ ,

5, 0, 2).

 

 

 

 

 

Therefore,

 

 

 

 

 

 

 

 

¯

 

 

 

 

 

 

 

 

 

O

 

 

 

 

 

 

 

 

The basis vectors are expressible as

 

 

 

 

 

 

 

 

 

eα

 

 

 

¯

α e

 

 

 

 

 

=

 

β

¯

β

or

¯ ¯

e0 = 00e¯ + 10e¯ + · · ·

0 1

= γ e¯ e¯ .

0 1

39

 

2.2 Vector algebra

 

 

 

 

 

Similarly,

 

 

 

 

 

e1 = −e0¯ + γ e1¯ ,

 

 

e2

= e2¯ ,

 

 

e3

= e3¯ .

This gives O’s basis in terms of O¯ ’s, so let us draw the picture (Fig. 2.1) in O¯ ’s frame: This transformation is of course exactly what is needed to keep the basis vectors pointing along the axes of their respective frames. Compare this with Fig. 1.5(b).

Inverse transformations

¯

The only thing the Lorentz transformation β α depends on is the relative velocity of the two frames. Let us for the moment show this explicitly by writing

 

 

¯

α

=

¯

α (v).

 

 

β

 

 

β

 

 

 

Then

 

 

 

 

 

 

 

 

 

 

=

β

 

¯

.

(2.14)

 

 

 

eα

 

 

¯ α (v)e

 

 

 

 

 

 

β

 

 

If the basis of O is obtained from that of O¯ by the transformation with velocity v, then the reverse must be true if we use v. Thus we must have

eμ¯ = ν μ¯ (v)eν .

(2.15)

In this equation I have used μ¯ and ν as indices to avoid confusion with the previous formula. The bars still refer, of course, to the frame O¯ . The matrix [ ν μ¯ ] is exactly the

matrix [ ¯ α ] except with changed to . The bars on the indices only serve to indicate

β v

v

the names of the observers involved: they affect the entries in the matrix [ ] only in that the matrix is always constructed using the velocity of the upper-index frame relative to the

t

t

 

 

 

e0

e0

 

 

 

 

 

e1

 

 

x

 

 

 

 

e1

 

 

x

 

 

 

 

Figure 2.1

Basis vectors of

O

and

O¯

O¯

 

 

 

as drawn by .

40

Vector analysis in special relativity

lower-index frame. This is made explicit in Eqs. (2.14) and (2.15). Since v is the velocity of O¯ (the upper-index frame in Eq. (2.14)) relative to O, then v is the velocity of O (the upper-index frame in Eq. (2.15)) relative to O¯ . Exer. 11, § 2.9, will help you understand this point.

We can rewrite the last expression as

 

e

 

=

ν

¯

(

v)eν .

 

 

¯

 

 

 

 

¯

 

β

 

 

β

 

 

 

 

β

 

 

 

 

 

 

 

,

Here we have just changed μ to

¯. This doesn’t change anything: it is still the same four

 

β

 

 

 

 

 

 

 

 

equations, one for each value of

 

 

 

 

 

 

 

 

¯. In this form we can put it into the expression for eα

 

Eq. (2.14):

 

 

 

 

 

 

 

 

 

¯

eα = β α (v)e ¯

β

¯

= β α (v) ν ¯ (v)eν . (2.16)

β

In this equation only the basis of O appears. It must therefore be an identity for all v. On

the right there are two sums, one on

¯ and one on

ν

. If we imagine performing the

¯ sum

 

β

 

β

first, then the right is a sum over the basis {eν } in which each basis vector eν has coefficient

¯

 

ν

 

 

 

 

 

β

 

¯

 

 

¯

α (v)

 

β

(

 

v).

(2.17)

 

 

 

 

 

 

 

β

Imagine evaluating Eq. (2.16) for some fixed value of the index α. If the right side of Eq. (2.16) is equal to the left, the coefficient of eα on the right must be 1 and all other coefficients must vanish. The mathematical way of saying this is

¯

β α (v) ν ¯ (v) = δν α ,

β

where δν α is the Kronecker delta again. This would imply

eα = δν α eν ,

which is an identity.

Let us change the order of multiplication above and write down the key formula

ν

 

(

 

v)

β

α (v)

 

δν

α .

(2.18)

¯

¯

=

 

β

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

β

 

 

 

 

¯

 

 

 

 

 

β

This expresses the fact that the matrix [ ν

β

(

 

v)] is the inverse of [ ¯ α (v)], because the

sum on ¯ is exactly the operation we perform when we multiply two matrices. The matrix (δν α ) is, of course, the identity matrix.

The expression for the change of a vector’s components,

 

 

 

 

 

 

 

A

 

¯

=

 

¯

α

(v)A

α

,

 

 

 

 

 

 

 

 

 

 

 

 

β

 

 

β

 

 

 

 

 

 

 

 

 

also has its inverse. Let us multiply both sides by ν

¯

 

v

 

 

β

 

 

 

 

 

 

 

¯

 

 

 

 

 

 

 

β (

 

 

) and sum on

¯. We get

 

ν

 

( v)A

 

 

ν

 

( v)

¯ α

(v)A

α

 

 

¯

 

 

β

=

 

 

¯

 

 

β

 

 

 

 

 

β

 

 

 

 

 

 

 

 

 

β

 

 

 

 

 

 

 

 

 

= δν α Aα

= Aν .

Соседние файлы в предмете [НЕСОРТИРОВАННОЕ]