<<

1 Space Vectors and 1.1 Representations

At each point P of a M, there is a TP of vectors. Choos- ing a set of vectors e TP provides a representation of each vector u TP in terms of components u .2 2

u = u e 0 1 2 = u e0 + u e1 + u e2 + ::: = [u][e] where the last expression treats the basis vectors as a column [e] and the vector components as a row matrix [u]. There is also a T^P of one-forms at each point. The basis one-forms which are dual to the basis vectors e are just the component-taking operators: ! (u) = ! u = u ! = u   or, in terms of a row matrix [!] of basis forms, u [!] = [u] :  When one of these operators acts on a basis vector, it produces one or zero according to:

! (e ) =  or, in matrix notation, [e] [!] = [1]  An arbitrary one-form ' can be represented by its components ' in the form:

' = ' ! = [!]['] where the components are now arranged as a column matrix. Here are various ways to represent a form acting on a vector to give a number:

' (u) = u (') = ' u = u ' = u ' 0 1  2  3 = u '0 + u '1 + u '2 + u '3 + ::: = ['][u]

1.2 Holonomic Basis Given a coordinate patch — a set of independent functions x which can be used to label points in some part of M — a special set of basis one-forms can be de…ned: ! = dx dx (u) = u (x ) Recall that each u is really a directional . d u = d C

1 where C is a whose tangent vector at P is u. Rewrite the de…nition of the dx in various ways:

dx (C ()) dx u = u dx = dx (u) =   d The basis vectors dual to the one-forms dx are just our old friends the partial : @ @ = @x This type of basis is called a holonomic basis. For a vector u, here are various expressions in such a basis:

u = u @ ; u = dx u dx  u = u (x ) = d =x _ dx @ u =x _ @ = d @x For a one-form ' here are various expressions in a holonomic basis:

' = ' dx ;' = ' @ 

@f @f df @ = ; df = dx  @x @x Some more expressions in a holonomic basis:

u ' = (u @ ) ' dx = u '   dx @f u df = u (df) =  d @x df u df = u (f) =  d 1.3 n-beins, Vierbeins, Tetrads, etc. Holonomic bases are not always appropriate. For example, local observers al- ways want to use orthonormal basis vectors but it is not usually possible for a holonomic basis system to be orthonormal everywhere. However, if one has a curve C which is de…ned in terms of coordinates and wants its tangent vector, then the holonomic basis is needed. Thus, non-holonomic basis systems are usually related to holonomic ones like this:

a a ea = ba @ ;! = dx b with the inverse relations:

b b @ = b eb; dx = ! bb

2 Here, I am using Latin indexes for the non-holonomic basis and Greek indexes for the holonomic basis. Sometimes the two kinds of indexes are distinguished in other ways: ( ) ( ) e( ) = b( ) @ ;! = dx b Underlines, primes, carets, and other doodads are used to mark one set of indexes as distinct from the other. Notice that the index positions are absolutely critical. If you are using Scienti…c Notebook or Scienti…c Workplace, you need to prevent superscripts and subscripts from lining up one over the other by inserting a zero space between them. In terms of matrices, the relation between holonomic and non-holonomic representations would look like this:

1 [e] = [b][@] ; [!] = [dx][b] 1 [@] = [b] [e] ; [dx] = [!][b] Here are the non-holonomic components of the velocity of a particle which is located at coordinates xi (t) at time t:

a a (a) (a) v = v @a = v ba e(a) = v e(a)

(a) a (a) v = v ba dxa va = =x _ a dt (a) a (a) v =x _ ba In a four-dimensional , a non-holonomic basis system is often called a tetrad or a vierbein or a . The two types of indexes are not (a) always integrated into a single formalism connected by tetrad coe¢ cients ba . Sometimes we just perform a when it is convenient but use only one type of basis at a time.

1.4 Higher rank tensors Vectors and one-forms are special cases of tensors. Higher rank tensors are multilinear functions of vectors and one-forms. They can be built up from products de…ned like this: u v ( ; ) = u ( ) v ( ) ; (u; v) = (u) (v)

u ' (; u; v) = u () (u) ' (v)

The space of tensors that act on pairs of vectors can be built from the tensor products e = e e and is designated TP TP . Other tensors are built up similarly. An example of such a tensor would be

T = T e e

3 2 Connections 2.1 Derivatives A is a way of di¤erentiating vector and tensor …elds. The essence of any type of “derivative”is that it obeys Leibniz’sproduct rule:

D (AB) = D (A) B + AD (B) as well as simple linearity

D(B + C) = D (B) + D (C)

Evidently derivatives can be de…ned on any algebra which permits multiplica- tions and additions. In more exotic cases, such as the algebra of square matrices for example, the linear Leibnizian operators are referred to as “derivations”. Consider the operator @D de…ned on square matrices by

@D (A) = [A; D] = AD DA and show that it is a derivation.

2.2 Directional Derivatives The simplest kind of derivative takes a tensor …eld into another tensor …eld of the same rank. Suppose that Kop is this type of derivative and work out how it must act on a vector …eld that has been expressed in terms of a particular moving frame e : f g

Kop (w) = Kop (w e )

= Kop (w ) e + w Kop (e )

The problem is reduced to two parts (1) knowing how the operator acts on functions such as the components w (2) knowing the vector that the operator assigns to each basis vector. The …rst part tells us a lot about the nature of this operator because there is only one derivative operator on functions, the , and each such operator corresponds to a vector …eld. Thus, our operator should be labeled by the vector …eld which it turns into when it acts on functions. If Kop (f) = vf then write Kop as Dv. Now we have

Dvw = v (w ) e + w Dve

That leaves part (2). We need to specify a vector-valued function (v) of a vector which maps v into Dve for each value of . Then, (changing the last dummy index with malice aforethought)

Dvw = v (w ) e + w (v)

4 Expanding the vector-valued function in terms of its components,

(v) = e (v)

Dvw = v (w ) e + w e (v) Collecting the results, we have

Dvw = v (w ) + w (v) e

where the connection functions : TP represent the action of the derivatives on basis vector …elds ! <

Dve = e (v) or equivalently,

(v) = (Dve ) ! :  If you wish to push the matrix notation a bit farther, you can write this de…nition as T (v) = (Dv [e]) [!]  and express the row of components  of Dvw as

T [Dvw] = v [w] + [w] (v) :  2.3 A¢ ne Connection

The directional derivative operator Dv has a very nice property when it acts on functions: For any functions h and f

Dhv f = hvf = hDv f and, for any vectors u and v,

Du+v f = (u + v) f = uf + vf = Du f + Dv f

Evidently we have the option of extending this nice property to the action of Dv on vector …elds so that, in general, for any function h and vector …eld v,

Dhv = hDv and, for any vectors u and v,

Du+v = Du + Dv

When this choice is made, the result is called an a¢ ne directional derivative. For an a¢ ne directional derivative, the connection functions obey

(hv) = ! Dhve = h! Dve   = h (v)

5 as well as

(u + v) = (u) + (v) : For a¢ ne directional derivatives, the connection functions are locally linear functions on the TP — they are one-forms called the connection forms.

(v) = ! v  In terms of components,  (v) = v or  ! = ! and some various ways of expressing the directional derivative of an arbitrary vector …eld w are:

Dvw = v (w ) + w (! v) e   = v (w ) + w v e     = v e (w ) + w v e   = v e (w ) + w  e  

By specifying the connection coe¢ cients  in each of the overlapping coordinate charts of a manifold so that the resulting directional derivatives agree in the overlap regions, one de…nes an a¢ ne connection on the manifold.

2.4 Extending the de…nition using Leibniz’srule Given an a¢ ne connection de…ned on vector …elds, one can extend it by insisting that it obey the for every conceivable type of product: For a form- …eld ' and a vector-…eld w; insist that

Dv (' w) = (Dv') w + ' Dvw    and …nd that

(Dv! ) e = ! v   and  Dv' = v e ' '  ! In terms of matrices, the column of components  of D v' would be

T [Dv'] = v ['] (v)['] For any , u w, just do  Dv (u w) = (Dvu) w + u Dvw:

Since any tensor in TP TP can be expanded in terms of such tensor products, we can now …nd the directional derivative of any such tensor. Try it on the tensor S = S e e :

DvS = Dv S e e   = DvS e e + S Dv (e) e + S e Dve   6 where I have changed two dummy indexes to  to make my life easier later. Continue:    DvS = v e S e e + v S e e   +v S e e     DvS = v e S + S  + S  e e

Try something more involved,  Q = Q e e !. The result is

   DvQ = v e Q + Q  + Q   Q   e e !   The pattern should be fairly clear by now: Each tensor index is contracted with one of the …rst two indexes on a connection coe¢ cient with plus signs for the terms in which a superscript tensor index in contracted and minus signs for the terms in which a subscript tensor index is contracted.

2.5 Covariant Derivatives

When an a¢ ne directional derivative Dv acts on a vector-…eld w, it produces another vector …eld Dvw. Thus,

Dvw TP jP 2

In terms of components, the vector Dvw has components

 (Dvw) = v ew + w  :

But, looking at this expression makes it obvious that its dependence on the vector-…eld v is completely local (the vector …eld is never di¤erentiated) and linear. Thus, we can de…ne a second-rank tensor …eld Dw according to

Dw ('; v) = ' Dvw: 

The components of this tensor …eld are traditionally denoted by w ; where

 Dw = w ;e !

 Dvw = v w ;e and evidently

w ; = ew + w  This tensor-…eld is called the covariant derivative of the vector-…eld w. For any tensor-…eld, one has a covariant derivative tensor that is one rank higher. The tensor …eld with components Q considered earlier has a covariant derivative with components

   Q ; = e Q + Q  + Q  Q   

7 The index-free notation for the covariant derivative DQ has the disadvantage of hiding the slot for the di¤erentiating vector …eld. The directional derivative notation displays the di¤erentiating vector …eld but at the expense of not show- ing o¤ the full tensor nature of the result. The standard in the previous equation shows the dependence on the di¤erentiating vector …eld by showing the corresponding index — call it the di¤erentiating index. For example, in  Dvw = Dw ( ; v) = v w ;e () the di¤erentiating index is . However the standard index notation obscures the operator nature of the covariant derivative. A modi…cation of the index notation is sometimes used to …x this de…ciency.

Instead of denoting the components of Dw by w ; just denote them by Dw . In this spirit,

   DQ = e Q + Q  + Q  Q  : I will not be using this notation. If you encounter it, be very careful to distin- guish between the directional derivative Dv and the operator notation D for the components of the covariant derivative.

2.6 The Perils of second derivatives

In the notation being used here, the derivative Dvw of a vector …eld w in the direction of a vector …eld v is interpreted as a second rank tensor …eld with v inserted as one of its arguments:

Dvw = Dw (_; v) :

Thus, the second derivative DuDvw is evaluated as the derivative of this second rank tensor …eld. DuDvw = DDw (_; v; u)

An anthropomorphic way of saying this is that the operator Du “sees” the di¤erentiating vector …eld v as a tensor argument. The expression for DuDvw must then be locally linear in that argument so that, for any function f on the manifold, DuDfvw = fDuDvw: Often, however, we do not want to think of a di¤erentiating vector …eld as a tensor argument. Here, we handle this situation by using a di¤erent symbol for the derivative. The derivative vw is de…ned as identical to Dvw except that r v is not regarded as a tensor argument. Thus, the expression u vw is simply r r the derivative of the vector …eld vw. One evaluates it just like the derivative of any other vector …eld. For a functionr f, that means

u fvw = u ( fvw) r r r r = u (f vw) r r = ( uf) w + f u ( vw) r r r

8 so this expression is not tensorial in v. It is easy to check that the tensorial derivative DuDvw can be expressed as

DuDvw = u vw uvw: r r rr 2.7 Overloaded Operators In order to avoid total confusion, it is very important to notice that the operators we are using act in di¤erent ways on di¤erent kinds of objects. Computer scientists refer to this type of operator as ‘overloaded’. Thus, if f is a scalar function and v a vector …eld, then one can see v acting in two very di¤erent ways in the equation vf = v (f) = v (df) = v df:  On a scalar …eld, it acts as a derivative. On a one-form …eld it is a local linear operator. Similarly, the between two objects depends on the nature of the objects. For vectors u; v and a one-form ,

u v = g (u; v)  u = (v) :  Suppose that k and f are functions on a manifold and v is a vector …eld. Com- pare the following two relations:

v (kf) = v (k) f + kv (f) v (kdf) = kv (df) :

This type of notation places a heavy burden on the reader to remember exactly what type of object each symbol stands for. Physicists sometimes su¤er a certain amount of culture shock because they are used to typographical clues to the nature of each symbol: vector signs or bold face type for vectors, gothic letters for tensors, and so on. Although it is customary to adopt some sort of typographical conventions such as using greek letters for forms and latin letters for vectors, these conventions are not absolute. Thus, you might be told that f is a one-form …eld while k is a scalar and v is a vector. In that case, you would then know that v (kf) = kv (f) rather than the relation given above.

3 3.1 De…nition A is the natural generalization of a straight line in ordinary space. The basic de…nition is that such a path does not turn. Given the path C : M < ! and the corresponding tangent vector …eld u () TC() de…ned along it we can calculate the directional derivative of this vector2 …eld along the curve.

  uu = u (eu + u ) e r   = uu + u u  e  9 A geodesic is a curve for which the tangent vector u does not turn. In other words, its rate of change must be in the same direction as u so that the tangent vector simply gets longer or shorter. Thus, we require

uu = ku r for some scalar function k along the curve.

3.2 A¢ ne parameters This geodesic de…nition is awkward. If we change the curve parameter from  to @ some parameter ~ then the quantity uu changes accordingly since u u~ = r ! @~ d d~ u~ = u; u = u~ d~ d

d~ De…ne  = d so that u = u~ and the geodesic condition becomes

u~ (u~) = ku~ r or  u~ (u~) = ku~ r which becomes u~ (u~) = ku~ r or u~ ()u ~ +  u~u~ = ku~ r r Cast this expression in the form of the de…nition of a geodesic

k u~ () u~u~ = r u~ r  or k u~ () u~u~ = k~u;~ k~ = r r  Thus, the arbitrary function k or k~ is linked to the arbitrary choice of curve parameter along the geodesic. Now eliminate the arbitrariness by choosing the new curve parameter so that

u~ () = k: r so that the equation for a geodesic with this particular curve parameter is just

u~u~ = 0: r We need to show that this new curve parameter can always be found. Rewrite its de…ning condition as d d d = k = k d~ ! d d~

10 but d~  = d so that  d = k  =  e k(0)d0 d ! 0 or R ~   0 d k(0)d0 k(00)d00 =  e 0 ~ = ~0 +  e 0 d 0 ! 0 R Z 0 R This expression exhibits the new curve parameter explicitly in terms of the old one. Notice that choosing the new parameter ~ to be an increasing function of d~ d~ the old parameter at just one point — d = 0 > 0 — guarantees that d is 0 always positive so that the new parameter is an increasing function of the old one everywhere and is guaranteed to be a good curve parameter. In terms of this new parameter, u~u~ = 0: r The resulting curve is called a geodesic with an a¢ ne parameter along it. Suppose that the parameters  and ~ are both a¢ ne parameters. Just set k = 0 in the above discussion and notice that the solution degenerates to ~ ~  = 0 + 0 ~ where 0 and 0 are both constants. Thus, di¤erent a¢ ne parameters along a given curve are all linearly related to one another.

3.3 Riemannian Coordinates A geodesic curve with an a¢ ne parameter is completely determined once its tangent vector at one point is given. Use this fact to produce a special coordinate patch around a chosen point P0. The simplest way to start is by working backwards — given a set of real number coordinates x specify the point P which is labeled by those points. In other words, de…nef ag map

exp : n U M < !  by specifying the point exp x0; x1; x2;::: . To specify this point, choose a set of basis vectors e for the tangent space TP and start an a¢ nely parameterized f g  0 geodesic curve Cx from P0 with the initial value of the parameter set to zero  and the initial tangent vector set to x e. @ C (0) = P ; = xe x 0 @  Cx;=0

The image point is then de…ned to be the point with parameter value equal to 1. exp (x) = Cx (1) :

11 The map exp will produce a well-behaved near the refer- ence point P0. It can break down if the geodesics start crossing each other — as they would on a for example. These coordinates are usually called Riemannian coordinates and the map is often called the exponential map of the tangent space into the manifold. It has some special properties that we need to work out. For each set of coordinates x, there is a geodesic Cx such that Cx (0) = P0 is the reference point and Cx (1) = P is a point which corresponds to the coordinate set x. So long as we stay in a subset U which is close enough to the reference point, the map from coordinate space n to U is one-to-one and can be inverted to provide coordinate functions X

What are the coordinates assigned to the point Cx ( ) for some number < 1? This point will be reached by a geodesic Cy which di¤ers from Cx only by having a di¤erent a¢ ne curve parameter so that

Cx ( ) = Cy (1)

The coordinates y assigned to this point correspond to the initial tangent vector to the curve Cy, so that tangent vector is what we must …nd. The tangent vector d is just where 0 is the a¢ ne parameter along Cy. Di¤erent a¢ ne parameters d0 for geodesics which go through the same set of points must be linearly related. In this case, all of the a¢ ne parameters start at zero, so they must be related by constant factors. The a¢ ne parameter 0 on Cy must be some constant k times the a¢ ne parameter  on Cx.

0 = k

But the parameter values 0 = 1 and  = correspond to the same point so that 1 = k or 1 k =

12 1 and the a¢ ne parameter on Cy is just 0 = : Now …nd the tangent vector to Cy. d d = d0 d or, in components

y e = x e which yields y = x

In terms of the coordinate functions and the curve Cx

   X (Cx ( )) = x = X (Cx (1))

For any geodesic through the reference point P0, the coordinate functions are given by X (C ( )) = X (C (1)) In terms of the simpler notation x (t) = X (C (t)), this result implies that the coordinates along any geodesic through P0 obey the relation

x (t) = tx (1) = tu where u is the initial tangent vector. Now consider what the geodesic equation will look like in this coordinate system. Let the functions x (t) describe the coordinate image of a curve passing through the reference point. The tangent vector to this curve is given by d dx u = = @ =x _ e dt dt   Now calculate for a bit:

   uu = u (u @) = ( uu ) @ + u u@ r r r r    uu = u (u ) @ + u u @ @ r d    r = dt u @ + u u @      =u _ @ + u u @      = (•x + u u  ) @ and …nd that the geodesic equation becomes

    x• + u u  = 0

But, in these coordinates, x (t) = tx (1) so that x• = 0

13 Putting this result back into the geodesic equation, we get the constraints

   u u  (P0) = 0 for any values of the components u. Split the connection into symmetric and antisymmetric parts:

    = () + [] and notice that, by antisymmetry

uu  = uu  [] [] and, by renaming dummy indexes,

uu  = u u = uu  [] [] [] so that    u u [] = 0: The antisymmetric part of the connection drops out of the constraints, leaving

   u u () (P0) = 0 which can be satis…ed for all values of u only if

 () (P0) = 0: Thus, the symmetric part of the connection coe¢ cient array vanishes at the reference point of a Riemannian coordinate system.

4 Connection Tensors 4.1 Torsion Consider how covariant derivatives commute when they act on a function.

[Du;Dv] f = DuDvf DvDuf But

DuDvf = u vf uvf r r rr and these derivatives of functions are just the vectors themselves acting as op- erators

DuDvf = uvf ( uv) f r = uv uv f f r g The commutator is then

[Du;Dv] f = uv uv f vu vu f f r g f r g = [u; v] uv + vu f f r r g

14 or, since f is an arbitrary function, the commutator is just a vector …eld.

[Du;Dv] = [u; v] uv + vu: r r The torsion is de…ned to be a third rank tensor …eld that yields the vector …eld

T (u; v) = [Du;Dv] = [u; v] + uv vu r r for any vector …elds u; v. For future reference, note that the de…nition implies that the arguments u; v of the are both di¤erentiating vector …elds. Thus, a derivative operation such as wT (u; v) will treat T (u; v) as a simple vector …eld and will not include counterr terms for its arguments. In a holonomic frame, the components of the torsion tensor are

   T = dx T (e ; e ) = dx e e e e     r r =  so the torsion is essentially the antisymmetric part of the connection coe¢ cient array. The torsion is a tensor — it has to be because we constructed it from covari- ant derivatives — and no coordinate transformation can ever make it vanish. As we have seen, the rest of the connection, the symmetric part, can be made to vanish at a point by choosing Riemannian coordinates there.

4.2 De…nition of Now consider how covariant derivatives commute when they act on a vector …eld. As we saw earlier, the object with the simplest properties is

(u; v) w = [ u; v] w w: R r r r[u;v] Here, is regarded as a linear operator on the vector …eld w. This de…nition uses derivativesR that do not treat the di¤erentiating vector …elds u; v as tensor arguments. Nevertheless, it produces a fourth-rank tensor. With all of its arguments made explicit, this Riemannian Curvature tensor is

Riem ( ; w; u; v) = (u; v) w  R = [ u; v] w w  r r  r[u;v] where is a form and w; u; v are vectors. The components of this tensor can be expressed in terms of the connection and commutation coe¢ cients as

   R  = e  e  +     c  : r  r  where the derivatives are regarded as acting on ordinary scalar …elds.

15 In an environment such as Scienti…c Notebook, where matrix operations are carried out automatically, it is useful to organize the curvature and connection in terms of matrices:

0 0 0 0 0 1 2 3 1 1 1 1 0 1 2 3 [] = 2 2 2 2 2 3 0 1 2 3 3 3 3 3 6 0 1 2 3 7 6 7 4 5 0 0 0 0 R 0 R 1 R 2 R 3 1 1 1 1 R 0 R 1 R 2 R 3 [R] = 2 2 2 2 2 3 R 0 R 1 R 2 R 3 3 3 3 3 6 R 0 R 1 R 2 R 3 7 6 7 so that the curvature expression4 becomes 5

 [R] = e [] e [] + [] [] [] [] c  []  r   r        Because this expression is manifestly antisymmetric in the indexes ;  there are just six independent matrix expressions to evaluate.

4.3 Ricci Identities Calculate the commutator of covariant derivatives acting on a vector …eld as follows:

[Du;Dv] w = DuDvw DvDuw = u vw uvw v uw v uw fr r rr g fr r rr g = [ u; v] w + v uw uvw r r rr rr = [ u; v] w + v u uvw r r rr r = (u; v) w + [u;v]w + v u uvw R r rr r = (u; v) w [u;v]+ uv v uw R r r r = (u; v) w w R rT (u;v)

[Du;Dv] + D w = (u; v) w T (u;v) R In terms of components, this identity takes the form

  w ; w ; + T  w ; = w R 

The operator [Du;Dv] + DT (u;v) obeys Leibniz’s rule so that its action on tensors of various rank follows a pattern similar to that of covariant derivatives, with each index in turn getting contracted with the curvature tensor.

  ' ' + T  ' = ' R  ; ; ;     K ; K ; + T  K ; = K R  K R     g ; g ; + T  g ; = g R  g R 

16 4.4 Bianchi Identities 4.4.1 Jacobi Identity The fundamental consistency relation for any system of commutors is the Jacobi identity [[A; B] ;C] + [[C;A] ;B] + [[B;C] ;A] = 0 which can be veri…ed just by writing out the expression in terms of products — everything cancels. It does not matter what sort of objects are being commuted. So long as they obey an (where factors can be regrouped), then the Jacobi identity must hold. Any time that a set of commutation relations is given, the Jacobi identity needs to be true.

4.4.2 Torsion Take the operators A; B; C in the Jacobi identity to be covariant derivatives and apply the identity to a scalar …eld — a function on the manifold.

0 = [Dw; [Du;Dv]] f + [Dv; [Dw;Du]] f + [Du; [Dv;Dw]] f

From here, just write out the commutators, working from the outside in and recognize the torsion and curvature terms as they occur. The result is the Torsion Bianchi Identity:

0 = DwT (v; u) + DvT (u; w) + DuT (w; v) + (u; v) w + (w; u) v + (v; w) u R R R In the case of zero torsion, this identity takes the form of a symmetry of the curvature tensor

(u; v) w + (w; u) v + (v; w) u = 0 R R R or, in terms of components

   R + R + R = 0:

4.4.3 Curvature Now apply the identity to a vector …eld on the manifold

0 = [ t; [ u; v]] w + [ v; [ t; u]] w + [ u; [ v; t]] w r r r r r r r r r and, after applying the de…nitions of the curvature and torsion tensors, obtain the curvature Bianchi Identity in the form

0 = Dt (u; v) + Dv (t; u) + Du (v; t) R R R (u; T (v; t)) (v; T (t; u)) (t; T (u; v)) : R R R

17 In the case of zero torsion, this identity is a symmetry of the covariant derivative of the curvature tensor

0 = Dt (u; v) + Dv (t; u) + Du (v; t) R R R or, in terms of components

   R  ; + R  ; + R  ; = 0 where the last three indexes are permuted cyclically from one term to the next.

4.5 In order to specify the lengths of vectors, and the angles between vectors, a tensor g which eats pairs of vectors is needed — a second rank covariant tensor. Thus, the length of the vector v is given by g (v; v) while the angle  between spacelike vectors u and v is given by cos  = g(u;v) . In the spacetime of ppg(u;u)g(v;v) relativity, we insist that there should be a set of basis vectors such that these expressions take their forms. For such basis vectors e , the metric has components

g00 = g (e0; e0) = 1 g11 = g (e1; e1) = +1 g22 = g (e2; e2) = +1 g33 = g (e3; e3) = +1 g = g (e; e ) = 0 for  =  6 or, more brie‡y, [g] =diag( 1; +1; +1; +1). Mathematically, we can guarantee the existence of such a basis set for either g or g by requiring three things of the metric of components [g]:

(1) det [g] < 0  (2) [g] is symmetric — [g]T = [g]  1 (3)[g] is invertible — [g] exists. 

The tensor g can be regarded as a map g : TP T^P where the form g (u) which is assigned to a vector u acts on the vector v!according to the de…nition

g(u)(v) = g (u; v)

In terms of components,

g(u) = g v : The inverse of this map, has the component representation:

1 g (') = g ' :

18 Putting these last two expressions together, we …nd

g g =  or, in terms of matrix arrays of components,

1 1 1 g [g] = I; g = [g] : ! Now de…ne a second rank  tensor which acts on pairs of forms according to

1 1 1 g ('; ) = g (') = ' g ( ) :   This contravariant second rank tensor has the components

1 1 g ! ;! = g ! ! = g  Historically, the metric tensor …eld has been regarded as the fundamental …eld variable of with everything else derived from it. That point of view has changed radically during the last few years. It now looks as if the connection is the fundamental gravitational variable while the is a derived quantity which is sometimes needed in order to describe interactions between gravity and matter.

4.6 Diagonalization of the Metric For a basis transformation  e 0 = U 0 e to yield an orthonormal basis, the coe¢ cients U 0 must obey the conditions 1 for = = 0   U 0 e U 0 e =  = 1 for = = 0  8 0 for = 6 < 6 or   : U 0 U 0 g =  In matrix notation, these conditions take the form [U][g][U]T = [] Notice that these are not the same as the conditions on a similarity transforma- tion that diagonalizes the matrix [g]. A similarity transformation would take 1 the form A [g] A . If the original basis is already orthonormal, the condition on a basis change to a new orthonormal set is [U][][U]T = [] In this case, the transformation

[e0] = [U][e] is a Lorentz Transformation.

19 4.7 Metricity The link between the metric tensor and the connection is the covariant derivative of the metric tensor — the metricity tensor. For an arbitrary metric in an arbitrary connection, the metricity tensor is just

1 Q ( ; ; v) = Dvg ( ; ) or   Q = De g = e g g  g  : I am de…ning the metricity in terms of the inverse metric tensor because that turns out to be more useful in calculations of gravitational dynamics than the regular metric tensor. From matrix algebra, any derivative operator D acting on the inverse of a matrix yields

1 1 1 D m = m (Dm) m so that an equivalent de…nition of metricity is

Q (a; b; v) = Dvg (a; b) or  Q = g Q g = De g = g ;   = e g g g  Ordinarily, one requires the metricity tensor to vanish so that the covariant derivatives of inner products obey Leibniz’srule:

De (u v) = De u g v  = De u g v + u Q v + u g De v  The metricity amounts to the covariant derivative of the “dot”. A connection for which the metricity tensor vanishes is called a “metric compatible connec- tion”. This condition is usually interpreted as a constraint on the connection coe¢ cients:   e g = g + g  I will follow the usual interpretation for a while and get expressions for the connection coe¢ cients in terms of the metric. However, before we mess up this expression any further, notice that it is a beautifully simple system of linear …rst order di¤erential equations in the metric components. Given the connection coe¢ cients, we can recover the metric tensor by solving this system of linear equations. The standard approach de…nes the object

 = g  and notices that we already know that the torsion tensor is given by

 T = g c 2 [ ]

20 so that 1  = (g c T ) [ ] 2 The metric compatibility condition takes the form

e g = + or 1 = e g : ( ) 2 This system of linear equations can be solved for the connection coe¢ cients. The system to be solved is 1 1 = b ; = a ( ) 2 [ ] 2 or, equivalently + = b = a Cyclically permute the argments of this last equation and write it as

= + a which can be substituted into the …rst equation of the pair

+ + a = b and gives the key result:

+ = b a : Now denote cyclic index permutations by

(C) = and B = b a so that the equation that we have to solve becomes just

+ C = B:

Operate on it twice with the cyclic permutation operator

C + C2 = CB

C2 + = C2B: Now we have three equations in the three unknowns ;C;C2.

21 Eliminate C2 using the last equation C2 = C2B : Next, eliminate C by using middle equation C = CB C2 = CB C2B + : The remaining equation them becomes + CB C2B + = B or 2 = C2B CB + B 1 = C2B CB + B 2 1  = (B B + B ) 2 Now unravel the de…nitions

B = b a  = e g g c + T and get the …nal expression for the connection coe¢ cients in terms of the metric and torsion tensors. 1   = 2 g (e g g c  + T    +e g  gc + T eg + g c  T ) In the case of a holonomic frame and zero torsion, this expression reduces to just a combination of derivatives of the metric tensor components.

1  = g (e g + e g  eg ) 2 4.8 Curvature-Metricity Identity Let the curvature operator act on the metric tensor

[Du;Dv] + D g (a; b) = g ( (u; v) a; b) g (a; (u; v) b) T (u;v) R R or 

DuDvg (a; b) DvDug (a; b)+D g (a; b) = g ( (u; v) a; b) g ( (u; v) b; a) T (u;v) R R which, from the de…nition of metricity yields

DuQ (a; b; v) DvQ (a; b; u)+Q (a; b; T (u; v)) = g ( (u; v) a; b) g ( (u; v) b; a) R R or, in terms of indexes,  Q ; Q ; + Q T  = R  R  which connects the metricity to the curvature tensor. When the metricity is zero, this identity becomes an additional index symmetry of the curvature tensor

R  = R 

22 4.9 Index Symmetries of the Curvature Tensor From its de…nition, the curvature tensor obeys

R  = R  : In the most general case, this is the only symmetry which the tensor has. Count its components in a four dimensional space: Array its independent components in rows according to the last two indexes and columns according to the …rst two. That makes six independent rows and sixteen independent columns for a total of 96 independent components. In the most common case, the connection is metric compatible, which, from the curvature-metricity identity, guarantees antisymmetry in the …rst two in- dexes. The array now has just six columns and six rows and the curvature has 36 independent components. If, in addition to being metric compatible, the connection is torsion-free, then the torsion Bianchi identities put additional constraints on the curvature components:    R + R + R = 0 These additional 16 equations reduce the number of components to just 20. Lower an index in the torsion Bianchi identities and start manipulating them, along with the antisymmetries in the …rst two and second two indexes:

R + R + R = 0 (1) symmetrize on ;  to so that the …rst term vanishes and the remaing two yield

R + R  + R + R  = 0 or, re-grouping the …rst term with the last and the second with the third,

(R + R ) + (R  + R ) = 0 and using the antisymmetry of the …rst and second pairs of curvature indexes

(R  R  ) + (R R  ) = 0 (2) Now symmetrize equation(1) on ;  so that the last term vanishes

R + R  + R + R  = 0 or (R R  ) + (R  R ) = 0 (3) Then symmetrize equation (1) on ;  so that the middle term vanishes

R + R  + R + R  = 0 or (R R  ) + (R  R  ) = 0 (4)

23 The three equations (2,3,4) have the form

z + x = 0 y + z = 0 x + y = 0 where x; y; z are the tensor combinations in parentheses. The solution to this system is easily seen to be x = y = z = 0 so that we obtain the index symmetry

R  = R  :

Notice that this last index symmetry is derived from the torsion Bianchi identities for zero torsion. It does not contain quite all of the information which is in those identities, however. One way to see this is to count the independent components of the Riemann tensor again, using this index symmetry. Think of the array of components as having rows and columns labeled by antisymmetrized index pairs. That makes it a six by six array. The last index symmetry makes 6 7 that array symmetric which means that it has 2 independent components — 21. Since we know from counting the torsion Bianchi identities that there are only 20 independent components, there is evidently one constraint that is not captured by any of the index symmetries. To see what the remaining constraint is, notice that we symmetrized various indexes with the one …xed index in the torsion identity. That procedure would miss whatever constraint is gotten by antisymmetrizing those indexes. Since the torsion bianchi identities are already antisymmetric in three indexes and have the form R[ ] = 0 the only possible result of antisymmetrizing the remaining …xed index  with any other index is the totally antisymmetrized expression

R[  ] = 0:

The one component of the curvature tensor that is permitted by the index symmetries but forbidden by this remaining constraint is

R  = "  where  is an arbitrary function and "  is the totally antisymmetric Levi- civita symbol. The missing constraint is that this totally antisymmetric com- ponent of the Riemann tensor must be zero.

 " R  = 0:

24