Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Symmetric matrices
Here Tr stands for the trace – the sum of diagonal elements of a (square) matrix. With this inner
product (called the Frobenius inner product), Mm,n becomes a legitimate Euclidean space, and
we may use in connection with this space all notions based upon the Euclidean structure, e.g.,
the (Frobenius) norm of a matrix
v
u∑ √
√ um ∑ n
∥X∥2 = ⟨X, X⟩ = t Xij2 = Tr(X T X)
i=1 j=1
and likewise the notions of orthogonality, orthogonal complement of a linear subspace, etc.
The same applies to the space Sm equipped with the Frobenius inner product; of course, the
Frobenius inner product of symmetric matrices can be written without the transposition sign:
⟨X, Y ⟩ = Tr(XY ), X, Y ∈ Sm .
333
334 APPENDIX C. SYMMETRIC MATRICES
In connection with Theorem C.2.1, it is worthy to recall the following notions and facts:
of A.
Theorem C.2.1 states, in particular, that for a symmetric matrix A, all eigenvalues are real,
and the corresponding eigenvectors can be chosen to be real and to form an orthonormal basis
in Rn .
Representation (C.2.2) with orthogonal U and diagonal Λ is called the eigenvalue decomposition
of A. In such a representation,
C.2.C. Vector of eigenvalues. When speaking about eigenvalues λi (A) of a symmetric n×n
matrix A, we always arrange them in the non-ascending order:
Ai = U Λi U T , i = 1, ..., k.
You are welcome to prove this statement by yourself; to simplify your task, here are two simple
and important by their own right statements which help to reach your target:
VCE says that to get the largest eigenvalue λ1 (A), you should maximize the quadratic form
xT Ax over the unit sphere S = {x ∈ Rn : xT x = 1}; the maximum is exactly λ1 (A). To get
the second largest eigenvalue λ2 (A), you should act as follows: you choose a linear subspace E
of dimension n − 1 and maximize the quadratic form xT Ax over the cross-section of S by this
336 APPENDIX C. SYMMETRIC MATRICES
subspace; the maximum value of the form depends on E, and you minimize this maximum over
linear subspaces E of the dimension n − 1; the result is exactly λ2 (A). To get λ3 (A), you replace
in the latter construction subspaces of the dimension n − 1 by those of the dimension n − 2,
and so on. In particular, the smallest eigenvalue λn (A) is just the minimum, over all linear
subspaces E of the dimension n − n + 1 = 1, i.e., over all lines passing through the origin, of the
quantities xT Ax, where x ∈ E is unit (xT x = 1); in other words, λn (A) is just the minimum of
the quadratic form xT Ax over the unit sphere S.
Proof of the VCE is pretty easy. Let e1 , ..., en be an orthonormal eigenbasis of A: Aeℓ =
λℓ (A)eℓ . For 1 ≤ ℓ ≤ n, let Fℓ = Lin{e1 , ..., eℓ }, Gℓ = Lin{eℓ , eℓ+1 , ..., en }. Finally, for
x ∈ Rn let ξ(x) be the vector of coordinates of x in the orthonormal basis e1 , ..., en . Note
that
xT x = ξ T (x)ξ(x),
since {e1 , ..., en } is an orthonormal basis, and that
∑ ∑
xT Ax = xT A ξi (x)ei ) = xT λi (A)ξi (x)ei =
∑ i i
λi (A)ξi (x) (xT ei )
i | {z } (C.3.2)
ξi (x)
∑
= λi (A)ξi2 (x).
i
Now, given ℓ, 1 ≤ ℓ ≤ n, let us set E = Gℓ ; note that E is a linear subspace of the dimension
n − ℓ + 1. In view of (C.3.2), the maximum of the quadratic form xT Ax over the intersection
of our E with the unit sphere is
{ n }
∑ ∑n
2 2
max λi (A)ξi : ξi = 1 ,
i=ℓ i=ℓ
and the latter quantity clearly equals to max λi (A) = λℓ (A). Thus, for appropriately chosen
ℓ≤i≤n
E ∈ Eℓ , the inner maximum in the right hand side of (C.3.1) equals to λℓ (A), whence the
right hand side of (C.3.1) is ≤ λℓ (A). It remains to prove the opposite inequality. To this end,
consider a linear subspace E of the dimension n − ℓ + 1 and observe that it has nontrivial
intersection with the linear subspace Fℓ of the dimension ℓ (indeed, dim E + dim Fℓ =
(n − ℓ + 1) + ℓ > n, so that dim (E ∩ F ) > 0 by the Dimension formula). It follows that there
exists a unit vector y belonging to both E and Fℓ . Since y is a unit vector from Fℓ , we have
∑ℓ ∑ℓ
y= ηi ei with ηi2 = 1, whence, by (C.3.2),
i=1 i=1
∑
ℓ
y T Ay = λi (A)ηi2 ≥ min λi (A) = λℓ (A).
1≤i≤ℓ
i=1
Since E is an arbitrary subspace form Eℓ , we conclude that the right hand side in (C.3.1) is
≥ λℓ (A).
A simple and useful byproduct of our reasoning is the relation (C.3.2):
Corollary C.3.1 For a symmetric matrix A, the quadratic form xT Ax is weighted sum of
squares of the coordinates ξi (x) of x taken with respect to an orthonormal eigenbasis of A; the
weights in this sum are exactly the eigenvalues of A:
∑
xT Ax = λi (A)ξi2 (x).
i
C.3. VARIATIONAL CHARACTERIZATION OF EIGENVALUES 337
We have proved the left inequality in (C.3.3). Applying this inequality to the matrix −A, we
get
−λℓ (Ā) = λn−k−ℓ (−Ā) ≥ λn−ℓ (−A) = −λℓ (A),
or, which is the same, λℓ (Ā) ≤ λℓ (A), which is the first inequality in (C.3.3).
338 APPENDIX C. SYMMETRIC MATRICES
A ≽ 0 ⇔ {A = AT and xT Ax ≥ 0 ∀x}.
A is called positive definite (notation: A ≻ 0), if it is positive semidefinite and the corresponding
quadratic form is positive outside the origin:
Theorem C.4.1 Let A be a symmetric n × n matrix. Then the following properties of A are
equivalent to each other:
(i) A ≽ 0
(ii) λ(A) ≥ 0
(iii) A = DT D for certain rectangular matrix D
(iv) A = ∆T ∆ for certain upper triangular n × n matrix ∆
(v) A = B 2 for certain symmetric matrix B;
(vi) A = B 2 for certain B ≽ 0.
The following properties of a symmetric matrix A also are equivalent to each other:
(i′ ) A ≻ 0
(ii′ ) λ(A) > 0
(iii′ ) A = DT D for certain rectangular matrix D of rank n
(iv′ ) A = ∆T ∆ for certain nondegenerate upper triangular n × n matrix ∆
(v′ ) A = B 2 for certain nondegenerate symmetric matrix B;
(vi′ ) A = B 2 for certain B ≻ 0.
as required in (vi).
(vi)⇒(v): evident.
(v)⇒(iv): Let A = B 2 with certain symmetric B, and let bi be i-th column of B. Applying
the Gram-Schmidt orthogonalization process (see proof of Theorem A.2.3.(iii)), we can find an
∑
i
orthonormal system of vectors u1 , ..., un and lower triangular matrix L such that bi = Lij uj ,
j=1
or, which is the same, BT= LU , where U is the orthogonal matrix with the rows u1 , ..., uTn .
T
We
now have A = B 2 = B T (B T )T = LU U T LT = LLT . We see that A = ∆T ∆, where the matrix
∆ = LT is upper triangular.
(iv)⇒(iii): evident.
C.4. POSITIVE SEMIDEFINITE MATRICES AND THE SEMIDEFINITE CONE 339
C.4.B. The semidefinite cone. When adding symmetric matrices and multiplying them by
reals, we add, respectively multiply by reals, the corresponding quadratic forms. It follows that
C.4.B.2: n × n positive semidefinite matrices form a cone Sn+ in the Euclidean space
Sn of symmetric n×n matrices, the∑Euclidean structure being given by the Frobenius
inner product ⟨A, B⟩ = Tr(AB) = Aij Bij .
i,j
The cone Sn+ is called the semidefinite cone of size n. It is immediately seen that the semidefinite
cone Sn+ is “good,” specifically,
• Sn+ is closed: the limit of a converging sequence of positive semidefinite matrices is positive
semidefinite;
• Sn+ is pointed: the only n × n matrix A such that both A and −A are positive semidefinite
is the zero n × n matrix;
Note that the relation A ≽ B means exactly that A − B ∈ Sn+ , while A ≻ B is equivalent to
A − B ∈ int Sn+ . The “matrix inequalities” A ≽ B (A ≻ B) match the standard properties of
340 APPENDIX C. SYMMETRIC MATRICES
A≽A [reflexivity]
A ≽ B, B ≽ A ⇒ A = B [antisymmetry]
A ≽ B, B ≽ C ⇒ A ≽ C [transitivity]
A ≽ B, C ≽ D ⇒ A + C ≽ B + D [compatibility with linear operations, I]
A ≽ B, λ ≥ 0 ⇒ λA ≽ λB [compatibility with linear operations, II]
Ai ≽ Bi , Ai → A, Bi → B as i → ∞ ⇒ A ≽ B [closedness]
A ≽ B, C ≻ D ⇒ A + C ≻ B + D,
etc. Along with these standard properties of inequalities, the inequality ≽ possesses a nice
additional property:
Theorem C.4.2 A symmetric matrix Y has nonnegative Frobenius inner products with all pos-
itive semidefinite matrices if and only if Y itself is positive semidefinite.
Proof. “if” part: Assume that Y ≽ 0, and let us prove that then Tr(Y X) ≥ 0 for every X ≽ 0.
Indeed, the eigenvalue decomposition of Y can be written as
∑
n
Y = λi (Y )ei eTi ,
i=1
where the concluding equality is given by the following well-known property of the trace:
C.4.B.4: Whenever matrices A, B are such that the product AB makes sense and
is a square matrix, one has
Tr(AB) = Tr(BA).
C.4. POSITIVE SEMIDEFINITE MATRICES AND THE SEMIDEFINITE CONE 341
Indeed, we should verify that if A ∈ Mp,q and B ∈ Mq,p , then Tr(AB) = Tr(BA). The
∑
p ∑
q
left hand side quantity in our hypothetic equality is Aij Bji , and the right hand side
i=1 j=1
∑
q ∑
p
quantity is Bji Aij ; they indeed are equal.
j=1 i=1
Looking at the concluding quantity in (C.4.1), we see that it indeed is nonnegative whenever
X ≽ 0 (since Y ≽ 0 and thus λi (Y ) ≥ 0 by P.7.5).
”only if” part: We are given Y such that Tr(Y X) ≥ 0 for all matrices X ≽ 0, and we should
prove that Y ≽ 0. This is immediate: for every vector x, the matrix X = xxT is positive
semidefinite (Theorem C.4.1.(iii)), so that 0 ≤ Tr(Y xxT ) = Tr(xT Y x) = xT Y x. Since the
resulting inequality xT Y x ≥ 0 is valid for every x, we have Y ≽ 0.
342 APPENDIX C. SYMMETRIC MATRICES