This Chapter connects some of the essential ideas in Linear Algebra with concepts in Tensor
Calculus and, in turn, discusses how to express indicial equations involving first- and
second-order systems in matrix terms. This Chapter may present a challenging read as it attempts to
answer questions that the reader may not have asked. After all, as we will state early in this
Chapter, Tensor Calculus does not need matrices. As a matter of fact, Linear Algebra and matrices
benefit from the ideas of Tensor Calculus more than the other way around. Nevertheless, I hope that
the reader gives this Chapter a thorough read at this time and returns to it later when the
questions discussed here arise naturally at some point in the future.
19.1Preliminary remarks
19.1.1Sidestepping matrices
Throughout our narrative, we have tried, as much as possible, to avoid the language of matrices.
This was, in part, to show that the structures that present themselves in the course of a tensorial
analysis can be handled without the use of matrices. Scalar variants of any order can be thought of
simply as indexed lists of numbers that need not be organized into tables. For example, it has been
sufficient to summarize the elements of the Christoffel symbol
in cylindrical coordinates by listing its three nonzero elements
This economic way of capturing
proved quite effective in helping us interpret the formula for the Laplacian
in cylindrical coordinates, where
That calculation was a microcosm of
the general realization that Tensor Calculus does not need Matrix Algebra.
19.1.2The advantages of matrices
Yet, a few of the distinct advantages of Matrix Algebra are impossible to ignore. First, the Matrix
Algebra treatment of systems as whole indivisible units helps stimulate our algebraic intuition
which is deeply ingrained in our mathematical culture. Consider, for example, the concept of the
matrix inverse and the identity for the inverse of the product of two matrices:
This identity is a concise
expression of the fundamental idea that to reverse a combination of two actions, we must combine
the individual reverse actions in the opposite order. For example, if our trip to work consists of
traveling by train followed by traveling by bus then the trip home will require traveling back by
bus followed by traveling back by train. The indicial notation does not have an effective way of
capturing this elementary idea.
To prove the formula
simply multiply by
. We have
as we set out to show. In this
concise derivation, the great utility of treating matrices as indivisible whole units is on full
display. Furthermore, the formula
can be used to derive the analogous equation for the inverse by
the following chain of identities
where the justification of each step
is left as an exercise. In summary,
From this derivation, it is clear
that the inverse of the product of any number of matrices equals the product of the individual
inverses in the opposite order. This derivation is not only straightforward, but also highly
insightful from the algebraic point of view. Meanwhile, the same set of ideas cannot be effectively
expressed in the indicial notation.
We should also call attention to another purely algebraic fact which we have relied upon in our
narrative to a significant degree. Namely, it is the fact that
It is because of this fact that we
have been able to contract the product of the covariant metric tensor
and the contravariant metric tensor
on any valid combination of indices to produce the Kronecker delta.
Let us give the classical matrix-based proof of this fact which will once again demonstrate the
utility of the algebraic way of thinking afforded to us by matrices. It follows from Linear Algebra
considerations that if is a square matrix with linearly independent columns then
there exists a unique matrix such that
We will refer to as the "right inverse" of . A square matrix with linearly independent columns has
linearly independent rows and therefore there exists a unique matrix , referred to as the left inverse of , such that
Our goal is to show that the left
inverse and the right inverse are the same matrix, i.e.
Subsequently, if we denote the
common value of and by , then the equivalence of and (both being the identity matrix) becomes the
equivalence of and .
To show that
multiply both sides of the identity
by on the right, i.e.
Since , , and , we find
as we set out to show. Note that we
glossed over an application of the associative property of matrix multiplication. Indeed,
multiplying by yields . By the associative property, we find
that and the rest of the argument can proceed as
described above. Thus, the equivalence of the left and right inverses is a direct consequence of
the associative property of multiplication.
Much like the preceding proofs, this proof, too, cannot be effectively carried out in indicial
notation. In fact, one could argue that this fundamental property of matrices cannot be discovered
in indicial notation. This proof completes our discussion of the significant logical advantages of
the matrix notation.
Another advantage of the matrix notation is that there are numerous software packages that
implement various matrix operations. Consequently, any calculation formulated in terms of matrices,
matrix products, and even more advanced matrix operations, can then be very effectively carried out
by computers. Most readers can probably hardly remember the last time they multiplied two matrices
by hand.
Finally, we must acknowledge the ubiquity of the language of matrices. Matrices represent
our go-to way of visualizing data, economizing notation, and injecting algebraic structure into
otherwise unstructured frameworks in a way that enables the leveraging of ideas from Linear
Algebra.
Tensor Calculus which, as we stated, does not need matrices, is a case in point. After all, it is
practically impossible not to think of first- and second-order systems as matrices. We have
consistently associated metric tensors with matrices, even though we preferred the phrase
corresponds to to the word equal in order to preserve the logical separation between
systems and matrices. For example, we would say that, in spherical coordinates,
Consequently, in spherical
coordinates, the dot product of vectors and with components and
,
given by the tensor identity
can be expressed in the language of
matrices as follows:
This form was frequently mentioned
in Chapter 10. To many readers it likely helped
clarify the interpretation of the then-novel expression .
19.1.3The advantages of the tensor notation
Having just described the indisputable benefits of the language of matrices, we ought to reiterate
some of the advantages of the tensor notation.
First, by virtue of treating systems not as whole indivisible units but as collections of
individual elements, the tensor notation allows direct access to those elements. We have taken
great advantage of this feature on numerous occasions, including the analysis of quadratic form
minimization in Section 8.7 and in establishing the
derivative of the volume element
with respect to the coordinate in
Section 16.12.
Second, the tensor notation continues to work for systems of order greater than two. This is the
case for a number of objects that we have already encountered, including the Christoffel symbol
,
the Levi-Civita symbols
and ,
and the Riemann-Christoffel tensor .
In order for an identity to be beyond the scope of the matrix notation, it need not be as
complicated as the celebrated tensor equation
Even an identity involving
second-order systems may require the tensor notation. For example, the identity
requires the tensor notation since
it has four live indices.
Last but not least, the tensor notation has its own unique way of inspiring algebraic intuition. In
particular, the natural placement of indices always accurately predicts the way an object
transforms under a change of coordinates and suggests the ways in which it can be meaningfully
combined with other objects. This feature of the tensor notation will be showcased throughout this
Chapter.
19.2Vectors, inner products, bases
In Section 2.7, we contrasted our approach to vectors to
that often taken by Linear Algebra. According to our approach, vectors are directed segments,
subject to addition according to the tip-to-tail rule and multiplication by numbers. It can be
demonstrated that these operations satisfy a number of desirable properties, such as
associativity and distributivity. In Linear Algebra, vectors are defined as generic objects subject
to abstract operations of addition and multiplication by numbers that, by definition,
satisfy the same set of desirable properties. That makes geometric vectors, i.e. directed
segments, a special case of the abstract vector as defined by Linear Algebra. To distinguish
between the two categories of vectors, we have used bold capital letters, such as , , and , for geometric vectors, and bold
lowercase letters, such as , , and , for vectors in the sense of Linear Algebra.
Tensor Calculus and Linear Algebra take similarly different approaches to the related concepts of
the dot product and inner product. The dot product of two geometric vectors and is defined by the equation
It can be demonstrated that
the dot product satisfies commutativity,
and distributivity
The distributive property can also
be referred to as linearity and is often described by stating that the dot product is
linear in each argument. It is also obvious that the dot product of a nonzero vector with itself is positive, i.e.
This property, known as positive
definiteness, is hardly worthy of mention for geometric vectors but becomes part of the
definition for generic vectors.
The Linear Algebra concept of an inner product is an axiomatic adaptation of the classical
dot product. An inner product is an operation that takes two vectors and produces a number.
The inner product of vectors and is denoted by and is defined by commutativity
distributivity, also known as
linearity,
and positive definiteness
A vector space endowed with an inner
product is called a Euclidean space -- so close is the analogy with Euclidean spaces as we
introduced them in Chapter 2.
The final difference between our approach to vectors and that of Linear Algebra is the way in which
a basis emerges. In Tensor Calculus, a coordinate system is chosen arbitrarily and the covariant
basis is
constructed by differentiating the position vector function with respect to the coordinate , i.e.
Thus, the covariant basis (at a
given point) is specific to the chosen coordinate system. The dependence of on the
choice of coordinates is underscored by the term variant. In Linear Algebra, a basis is
arbitrarily selected. Any complete linearly independent set of vectors represents a legitimate
basis. Thus, in both approaches we can talk about a change of basis, although in Tensor Calculus, a
change of basis is induced by a change of coordinates.
Having contrasted the origins of the key concepts, we must now point out that their
uses in the two subjects are almost identical. Tensor Calculus and Linear Algebra share the
concept of the component space in which analysis is performed in terms of the components of
vectors rather than vectors themselves. For example, with the help of the covariant metric tensor
defined by
the component space expression for
the dot product of vectors and with components and
reads
Similarly, the inner product matrix
, whose entries
are
can be used to express the inner
product of vectors and with components and
as
follows:
In the language of matrices, which
is one of the topics discussed in this Chapter, the same operation is captured by the equation
where and are the column matrices consisting of
the elements and
.
The covariant metric tensor
and matrix are essentially the same object,
differing only in notation and the terminology used to describe them. We will therefore use in many of the same ways that we have
used ,
including utilizing the symbol
to denote the entries of , and
performing index juggling by contracting with
and .
19.3Correspondence between systems and matrices
As we have already mentioned, only first- and second-order systems can be effectively represented
by matrices. As a matter of convention, let us agree to represent first-order systems by matrices. Thus, a system in a
three-dimensional space will be represented by the column matrix
We could have also represented by a
row matrix . However, limiting ourselves to column matrices
only will help reduce the number of possible matrix representations of contractions.
The flavor of the index, whether it indicates a tensor property of the corresponding variant or is
used for convenience, has no bearing on the matrix representation of a system. Thus, a system is
also represented by a column matrix, i.e.
Second-order systems correspond to square matrices. Note that in Chapter 8, we encountered systems that correspond to rectangular matrices. As a
matter of fact, systems of that sort will begin to arise naturally when we study embedded surfaces
in the next volume. Nevertheless, in this Chapter, we will limit our focus to square matrices.
However, all points made in this Chapter will remain valid for systems corresponding to rectangular
matrices.
As we have done throughout the book, and as the most commonly accepted convention dictates, the
first index of the system corresponds to the row the element is in while the
second index corresponds to the column. Thus, it is essential to have complete
clarity as to the order of indices. For systems with two superscripts or two subscripts, the order
is obvious. For mixed systems, i.e. systems with one superscript and one subscript, the
order of the indices is usually indicated by the dot placeholder technique introduced in Section 7.2. For example, in the symbols
and
the dot makes it clear that is the first index and is second.
Interestingly, we have not seen the dot placeholder that much in our narrative so far. The reason
for this is that the only mixed second-order system that we have consistently encountered is the
Kronecker delta
which corresponds to the symmetric matrix
and, therefore, the order of the
indices does not matter. However, as we will describe below in Section 19.5, this exception can be made only for symmetric
systems, in the sense of the term symmetric that differs from that for matrices.
Another category of mixed second-order systems for which the use of the placeholder is not required
is the Jacobians and
,
for which we can simply agree that the superscript is first and the subscript is second. What makes
this convention reliable in this case is the fact that the indices of a Jacobian are never juggled.
In other words, the symbol is
never used to represent the combination .
Had index juggling been allowed, the symbol
would be ambiguous and the use of the dot placeholder would be in order. More generally, the
possibility of index juggling is the very reason for not being able to use the flavors of indices
as a mechanism for determining their order.
19.4Matrix multiplication
Any combination of contractions involving first- and second-order systems can be represented by
matrix multiplication. In this Section, we will review the basic mechanics of matrix multiplication
and subsequently show how some of the most common contractions can be expressed by matrix products.
19.4.1The mechanics of matrix multiplication
Consider three matrices , , and with entries ,
and .
Suppose that is an matrix, is , and is . Then, by definition, is the product of and , i.e.
if
Note that we are following the
Linear Algebra tradition of using only subscripts to reference the individual entries of a matrix.
An essential element of matrix multiplication is that the summation takes place over the
second index of and the first index of . Since the first index indicates the
row of the entry and the second indicates the column, matrix multiplication combines the -th row of with the -th column of to produce ,
i.e. the entry in the -th row and -th column of . These mechanics are illustrated in
the following figure:
Thus, the mechanics of matrix
multiplication are very rigid and it is up to us to use the two available "levers" -- the order of
the operands and the transpose -- to make sure the products reflect the contractions in a given
tensor expression. We will now go through a series of examples converting contractions to matrix
products in increasing order of complexity.
19.4.2Contraction of first-order systems
Let us start with the contraction
that represents the inner product of
vectors and . Since
corresponds to
and
corresponds to
the combination
corresponds either to
In either case, the matrix product
involves the transpose of one of the matrices. We will find this to be the case in the more
complicated situations as well.
Note that the tensor product
corresponds to the matrix product
Thus, can
also be captured evaluating the trace of the resulting matrix, i.e.
19.4.3Contraction of a second-order system with a first-order system
Let us now turn our attention to the contraction
involving a second-order system
and a first-order system .
If is the matrix corresponding to
and is the matrix corresponding to , then
Note that the convention that the
first index indicates the row while the second indicates the column is essential for reaching this
conclusion. Also note that the alternative product
results in the same values, although arranged into a row matrix. Therefore, by our convention, where
first-order systems are represented by column matrices,
does not properly represent the resulting system.
As we already mentioned, the flavor of the indices has no bearing on the matrix representation.
Therefore, the combinations
which represent variations in the
flavors of indices but not their order, are also represented by the product .
The contraction
on the other hand, is principally
different in that it is the first index of
that is engaged in the contraction. As a result, the only way to represent by a
matrix product is
As before, the product produces the same values as a row
matrix and therefore does not properly represent the result.
Let us now express the transformation rules for first-order tensors in matrix form. Recall from
Chapter 17, that we let
Now suppose that is a
contravariant tensor, i.e.
Since first-order tensors correspond
to column matrices, we will use lowercase letters to denote those matrices. If the matrix represents and
represents ,
then
Similarly, if is a
covariant tensor, i.e.
then its matrix representations
and are
related by the equation
19.4.4Contraction of second-order systems
A contraction of two second-order systems produces another second-order system. Thus we expect that
a tensor identity featuring such a contraction corresponds to one of the many variations of the
matrix equation
that differ by the order of and and their transposition. In fact, we
can construct different variations, such as
and so on. The fact that
makes half of the variations equivalent. For example,
are equivalent as can be seen by
taking the transpose of both sides of the first equation. This reduces the number of truly distinct
equations to and also means that every indicial relationship can be
captured in two different, albeit equivalent, ways.
Recall, once again, that the flavors of indices have no bearing on the corresponding matrix
representations. Thus, we will choose the placements of indices completely at will.
Consider three second-order systems represented by the matrices , , and . For our first example, consider the
relationship
This relationship very clearly
corresponds to the matrix identity
However, if the order of the indices
on is
switched,i.e.
then the corresponding matrix
identity is
or, equivalently,
As the above two examples illustrate, the question of representing the contraction by
a matrix product has two legitimate answers: and ,
even though these expressions result in two different matrices. The two different interpretations
are possible because the combination ,
on its own, does not give us an indication of the order of the indices in the resulting system. If,
in the result, is the first index and is the second, then
corresponds to . Otherwise,
corresponds to .
For another example, consider the equation
This contraction takes place on the
second index of
and the second index of .
The entries in the matrices and involved in the contraction for and are illustrated in the following figure:
Clearly, this arrangement does not
correspond to a valid matrix product. In order to invoke matrix multiplication, the matrix needs to be transposed. Thus,
corresponds to
Finally, consider the identity
The summation on the right takes
place on the first index of
and the first index ,
which suggests the combination .
Additionally, as indicated by the symbol ,
the result of the contraction must be arranged in such a way that is the first index and is the second. Thus, the matrix form of the identity must
use
instead of . In summary, the above relationship
corresponds to the matrix identity
or, equivalently,
Let us now express the transformation rules for second-order tensors in matrix form. Suppose that
,
,
and
are tensors of the type indicated by their indicial signatures, i.e.
If the symbols and
represent the matrices corresponding to the alternative manifestations of each of the three
tensors, then the transformation rules in the matrix notation read
19.5Symmetric systems
Two systems
and ,
related by the identity
correspond to matrices that are the
transposes of each other. For example, if
then
The best way to convince yourself of
this relationship is to recall the technique of unpacking described in Section 7.4. For example, equals
which
equals . In other words, the entry in the
row and
column of the matrix corresponding to is
-- the same as the entry in the
row and
column of the matrix corresponding to .
Thus, the two matrices are, indeed, the transposes of each other.
The same statement holds for systems
and
with superscripts. If
then the two systems correspond to
matrices that are the transposes of each other.
Similarly, two mixed systems
and
related by the identity
correspond to matrices that are the
transposes of each other. It is noteworthy, however, that in order for this relationship to hold,
the order in which the superscript and the subscript appear in the two systems must be reversed.
A system is
called symmetric if it satisfies the identity
Note that a symmetric system
corresponds to a symmetric matrix. Recall that a matrix is symmetric if it equals its transpose, i.e.
The term symmetric makes
sense since the values of the entries exhibits a mirror symmetry with respect to the main diagonal,
e.g.
Similarly, a system
with two superscripts is called symmetric if
Such a system, too, corresponds to a
symmetric matrix.
Interestingly, the concept of symmetry appears to be problematic for a mixed system .
The would-be definition of symmetry
is invalid on the notational level.
This should give us serious pause. Over the course of our narrative, we have learned to trust the
intimation given to us by the tensor notation. We must, therefore, accept the fact that the concept
of symmetry cannot be applied to a mixed system in the above form. If we ignore what the notation
is telling us, and call a mixed system symmetric if it corresponds to a symmetric matrix, we will
run into contradictions later on. Specifically, suppose that a "symmetric" is
the manifestation of a tensor in some coordinate system or, to use the language of Linear Algebra,
the manifestation of a tensor with respect to a particular basis. Then transforms under a change of coordinates (or a change of
basis) according to the rule
Thus, if
corresponds to a symmetric matrix, then
will most likely not. For a simple illustration, note that for a symmetric matrix
the combination
equals
which is no longer symmetric. Thus,
when a mixed system
corresponds to a symmetric matrix, it is as much a characterization of the particular coordinate
system (or basis) as it is of
itself. As we mentioned earlier, the Kronecker delta is
the sole exception to this rule because it corresponds to the identity matrix in all coordinate
systems.
Fortunately, the tensor notation not only warns us of a potential problem, but also presents us
with a clear path forward. Namely, no rules of the tensor notation would be violated if we called a
system
symmetric when
This works on the notational level,
but what is the system
and how is it related to ?
The tensor framework provides the answer to this question, as well: is
with a lowered first index and a raised second index. In other words,
where we are using the inner product
matrix (or, in the language of tensors, the metric tensor) for juggling indices. This enables us
to define what it means for a mixed system to
be symmetric, albeit not in isolation but in the context of an inner product. In other words, in
the context of the associated Euclidean space framework.
Specifically, a mixed system is
said to be symmetric if
where is
the result of index juggling on .
It is worth reiterating that this new definition of symmetry does not correspond to the concept of
symmetry as applied to matrices. In particular, the matrix corresponding to a system
that satisfies the identity above, will not be symmetric, except in very special circumstances.
Also note that the identity
indicates that the matrices
corresponding to
and
are not equal, but rather the transposes of each other. Note, however, that for a system
that satisfies this identity, the placeholder may be safely omitted, i.e. the symbol
can be replaced with .
After all, the above identity tells us that it does not matter whether is
interpreted as or
.
Also be reminded that, unlike systems with two subscripts or two superscripts, the concept of
symmetry, as applied to mixed systems, requires the availability of an inner product, i.e. the
context of a Euclidean space. To make the above definition more explicit, we could have written it
in the form
although this form seems to
obfuscate, rather than clarify, the matter. However, this form does represent an explicit recipe
for testing whether a system is
symmetric. For example, suppose that
We will now show that if the system
then it is symmetric. In order to
apply the definition
note that the combination
corresponds to , i.e.
The resulting matrix is the
transpose of the matrix corresponding to
proving that is
symmetric.
The symmetry criterion can be simplified by applying a single index juggling step to the equation
Lowering the index on both sides, we find
Thus, is
symmetric if the related system is
symmetric. This is a simpler criterion since a system with two superscripts is symmetric if it
corresponds to a symmetric matrix. To illustrate this criterion with the example above, note that
which corresponds to the matrix product , i.e.
Since the resulting matrix is
symmetric, is
symmetric and therefore is
symmetric.
19.6Linear transformations
The concept of a linear transformation is one of the most fundamental elements in Linear
Algebra. Recall that a linear transformation can be represented by a matrix in the component space. In this regard, linear
transformations and inner products are similar in that both are represented by matrices in the
component space: the matrix in the case of linear transformations and the matrix
in the case of inner products.
However, this similitude appears stronger than it really is and the tensor notation alerts us to a
fundamental difference between the two matrices.
A linear transformation maps one vector to another. The
original vector is known as the preimage and the result of the transformation is known as
the image. Denote the preimage by and its image by , i.e.
The role of the matrix is to convert the components of
the preimage into the components
of the image. Thus, in the tensor notation, the matrix naturally corresponds to a mixed system
which enables us to write
An inner product, on the other hand, takes two vectors and as inputs and produces a scalar. This
is why the matrix naturally appears with two
subscripts, so as to enable the combination
that results in .
Thus, under a change of basis, the matrices and transform by different rules. By the
quotient theorem, discussed in Chapter 14, the
variant is
a tensor of the type indicated by its indicial signature and therefore transforms according to
By the same theorem, the variant
is
a tensor of the type indicated by its indicial signature and therefore transforms according
to
If is
the matrix representing the linear transformation in the alternative basis , then
the equation
tells us that
where denotes the matrix corresponding to
the Jacobian .
This formula can be found in equation of Chapter of I.M. Gelfand's Lectures on Linear Algebra.
Similarly, if
represents the same inner product in the alternative basis , then
the equation
tells us that
This formula can be found in
equation of Chapter of the same book.
Thus, it may be said that the matrix notation blurs the distinction between the component space
representations of inner products -- or, more generally, bilinear forms -- and linear
transformations since both concepts are represented by matrices. In fact, this similitude creates
the temptation to consider a one-to-one correspondence between bilinear forms and linear
transformation. Such a correspondence indeed exists but, in order to avoid internal contradictions,
it needs to be carefully framed. The tensor notation can be instrumental in helping us establish
the right correspondence -- the task to which we now turn.
19.7An equivalence between bilinear forms and linear transformations
An inner product is a special case of a more general operation known as a bilinear form
defined as an operation that takes two
vectors and produces a number subject only to linearity in each argument, i.e.
and
Thus, an inner product is a
symmetric positive definite bilinear form, where the term symmetric refers to the
commutative property. Rather than introduce a special letter to distinguish inner products from
general bilinear forms, Linear Algebra simply drops the letter altogether for inner products
resulting in the symbol .
Much like inner products, bilinear forms are represented by matrices in the component space. If
and
, then,
by linearity,
Thus, if the matrix with entries is
defined by
then
or, in the language of matrices,
Naturally, transforms from under a change of
basis by the rule
As we have already discussed, the fact that both linear transformations and bilinear forms are
represented by matrices in the component space creates the temptation to propose a one-to-one
correspondence between linear transformations and bilinear forms. The naive way of doing this is to
state that a linear transformation corresponds to a bilinear form if the two are represented by the
same matrix. However, this approach is clearly self-contradictory since matrices representing
linear transformations and bilinear forms transform differently under a change of basis. As a
result, a linear transformation and a bilinear form that happen to be represented by the same
matrix with respect to a particular basis, will most certainly be represented by two different
matrices with respect to another basis.
As it is often the case, the remedy is found by going back to the Euclidean framework where one can
develop the idea in a geometric setting. For a linear transformation , define the corresponding bilinear
form by the equation
In other words, is the inner product between and .
Let us now use the tensor approach to determine the relationship between the matrix representing the bilinear form with
respect to a basis and
the matrix representing the linear transformation with respect to
the same basis. We have
while
Therefore,
Switch the names of the indices
and so that
appears as , i.e.
Since, by construction, equals for all and , we conclude that
for all and
.
Therefore,
which is precisely the relationship
we set out to determine. Invoking index jugging by the "metric tensor" , we
arrive at the simple relationship
Thus, somewhat ironically, the naive
approach of associating a linear transformation with the bilinear form represented by the same
matrix is not too far off. The only adjustment that one needs to make is to require -- to use
tensor terminology -- the lowering of the superscript on the matrix representing the linear
transformation.
19.8Self-adjoint transformations
A linear transformation is called self-adjoint if the
identity
holds for all vectors and . In other words, the inner product of
and is the same as that of and , i.e. it does not matter whether
is applied to or . The concept of a self-adjoint matrix
is interesting for its "reverse commute": while most ideas in Linear Algebra were developed for
vectors and subsequently extended to the component space, the concept of a self-adjoint matrix
clearly arose out of symmetric matrices and their special properties. But are self-adjoint linear
transformations necessarily represented by symmetric matrices? We will now answer this question.
As we discovered in the previous Section, the inner product is given by
Similarly, is given by
Since, for a self-adjoint , equals for all and , we conclude that
for all and
.
Therefore
which is precisely the
characterization of the matrix we were looking for. In other words, the linear
transformation is self-adjoint if it is represented
by a
that is symmetric in the sense of the definition given in Section 19.5. To see this, allow the inner product matrix (i.e. the "metric tensor")to
lower the superscript on both sides, leading to the more
recognizable form
Raising on both sides also yields
As we have already pointed out, in
the language of matrices, the criterion for a matrix representing a self-adjoint transformation
reads
19.9Exercises
Exercise 19.1Show that the tensor equation
corresponds to the matrix equation
or, equivalently,
Exercise 19.2Show that the tensor equation
corresponds to the matrix equation
Exercise 19.3Show that, for a symmetric matrix and any matrix , the combination
is symmetric. Thus, if a tensor is symmetric in a component space corresponding to one basis, it is symmetric in a component space corresponding to any basis.
Exercise 19.4If the component space basis is orthonormal, show that a symmetric system is represented by a symmetric matrix.