Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Thomas Baird
March 26, 2014
Contents
1 Introduction
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
1
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
4
4
4
5
7
7
8
8
13
13
14
14
15
16
17
18
19
20
22
23
23
24
26
.
.
.
.
.
28
28
30
31
33
33
3.4
3.5
3.6
3.7
3.8
4 Exterior Algebras
4.1 Multilinear algebra . . . . . . . .
4.2 The exterior algebra . . . . . . .
4.3 The Grassmanian and the Pl
ucker
4.4 Decomposability of 2-vectors . . .
4.5 The Klein Quadric . . . . . . . .
. . . . . . .
. . . . . . .
embedding
. . . . . . .
. . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
36
37
40
42
44
46
47
.
.
.
.
.
52
52
55
59
59
61
5 Extra details
64
65
68
Introduction
Projective geometry has its origins in Renaissance Italy, in the development of perspective
in painting: the problem of capturing a 3-dimensional image on a 2-dimensional canvas.
It is a familiar fact that objects appear smaller as the they get farther away, and that the
apparent angle between straight lines depends on the vantage point of the observer. The
more familiar Euclidean geometry is not well equipped to make sense of this, because in
Euclidean geometry length and angle are well-defined, measurable quantities independent
of the observer. Projective geometry provides a better framework for understanding how
shapes change as perspective varies.
The projective geometry most relevant to painting is called the real projective plane,
and is denoted RP 2 or P (R3 ).
Definition 1. The real projective plane, RP 2 = P (R3 ) is the set of 1-dimensional subspaces of R3 .
This definition is best motivated by a picture. Imagine an observer sitting at the origin
in R3 looking out into 3-dimensional space. The 1-dimensional subspaces of R3 can be
understood as lines of sight. If we now situate a (Euclidean) plane P that doesnt contain
the origin, then each point in P determines unique sight line. Objects in (subsets of) R3
can now be projected onto the plane P , enabling us to transform a 3-dimensional scene
into a 2-dimensional image. Since 1-dimensional lines translate into points on the plane,
we call the 1-dimensional lines projective points.
Of course, not every projective point corresponds to a point in P , because some 1dimensional subspaces are parallel to P . Such points are called points at infinity. To
2
motivate this terminology, consider a family of projective points that rotate from projective points that intersect P to one that is parallel. The projection onto P becomes a
family of points that diverges to infinity and then disappears. But as projective points
they converge to a point at infinity. It is important to note that a projective point is only
at infinity with respect to some Euclidean plane P .
One of the characteristic features of projective geometry is that every distinct pair
of projective lines in the projective plane intersect. This runs contrary to the parallel
postulate in Euclidean geometry, which says that lines in the plane intersect except when
they are parallel. We will see that two lines that appear parallel in a Euclidean plane will
intersect at a point at infinity when they are considered as projective lines. As a general
rule, theorems about intersections between geometric sets are easier to prove, and require
fewer exceptions when considered in projective geometry.
According to Dieudonnes History of Algebraic Geometry, projective geometry was
among the most popular fields of mathematical research in the late 19th century. Projective geometry today is a fundamental part of algebraic geometry, arguably the richest
and deepest field in mathematics, of which we will gain a glimpse in this course.
2.1
Fields
The rational numbers Q, the real numbers R and the complex numbers C are examples of
fields. A field is a set F equipped with two binary operations F F F called addition
and multiplication, denoted + and respectively, satisfying the following axioms for all
a, b, c R.
1. (commutativity) a + b = b + a, and ab = ba.
2. (associativity) (a + b) + c = a + (b + c) and a(bc) = (ab)c
3. (identities) There exists 0, 1 F such that a + 0 = a and a 1 = a
4. (distribution) a(b + c) = ab + ac and (a + b)c = ac + bc
5. (additive inverse) There exists a F such that a + (a) = 0
6. (multiplicative inverse) If a 6= 0, there exists
1
a
F such that a
1
a
= 1.
Two other well know operations are best defined in terms of the above properties. For
example
a b = a + (b)
and if b 6= 0 then
1
a/b = a .
b
We usually illustrate the ideas in this course using the fields R and C, but it is
worth remarking that most of the results are independent of the base field F . A perhaps
unfamiliar example that is fun to consider is the field of two elements Z/2. This is the
field containing two elements Z/2 = {0, 1} subject to addition and multiplication rules
0 + 0 = 0,
0 + 1 = 1 + 0 = 1,
1+1=0
0 0 = 0,
0 1 = 1 0 = 0,
11=1
Let F denote a field. A vector space V over F is set equipped with two operations:
Addition: V V V , (v, w) 7 v + w.
Scalar multiplication: F V V , (, v) 7 v
satisfying the following list of axioms for all u, v, w V and , F .
(u + v) + w = u + (v + w) (additive associativity)
u + v = v + u (additive commutativity)
There exists a vector 0 V called the zero vector such that, 0 + v = v for all v V
(additive identity)
For every v V there exists v V such that v + v = 0 (additive inverses)
( + )v = v + v (distributivity)
(u + v) = u + v (distributivity)
(v) = ()v (associativity of scalar multiplication)
1v = v ( identity of scalar multiplication)
I will leave it as an exercise to prove that 0v = 0 and that 1v = v.
Example 1. The set V = Rn of n-tuples of real numbers is a vector space over F = R in
the usual way. Similarly, F n is a vector space over F . The vector spaces we consider in
this course will always be isomorphic to one of these examples.
Given two vector spaces V, W over F , a map : V W is called linear if for all
v1 , v2 V and 1 , 2 F , we have
(1 v1 + 2 v2 ) = 1 (v1 ) + 2 (v2 ).
5
2.1.3
Matrices
n
X
Ai,k Bk,j
(1)
k=1
where the sum is indexed by permutations of the set {1, ..., n} and (1) = 1 is the
sign of the permutation.
Theorem 2.2. Let A be an n n-matrix. The following properties are equivalent (i.e. if
one holds, they all hold).
A is invertible.
A has non-zero determinant.
A has rank n.
The simplest way to check that A is invertible is usually to check if the rank is n by
row reducing.
2.1.4
Given any vector space V , the dual vector space V is the set of linear maps from V to
F = F1
V := { : V F | is linear}.
7
The dual V is naturally a vector space under point-wise addition and multiplication
(1 + 2 )(v) = 1 (v) + 2 (v)
and
()(v) = ((v)).
Example 3. Suppose V
= F n is the set of 1 n column vectors. Then V is identified
with the set of n 1 row vectors under matrix multiplication.
Suppose that v1 , ..., vn V is a basis. Then the dual basis v1 , ..., vn V is the unique
set of linear maps satisfying
vi (vj ) = ij ,
where ij is the Kronecker delta defined by ii = 1 and ij = 0 if i 6= j.
Example 4. Let v1 , ..., vn F n be a basis of column vectors and let A = [v1 ...vn ] be the
n n-matrix made up of these columns. The rules of matrix multiplication imply that
the dual basis v1 , ..., vn are the rows of the inverse matrix
v1
v2
A1 =
.: .
vn
2.2
To develop some intuition about projective spaces, we consider some low dimensional
examples.
If V is one-dimensional, then the only one dimensional subspace is V itself, and consequently P (V ) is simply a point (a one element set).
Now consider the case of the real projective line. Let V = R2 with standard coordinates
x, y and let L R2 be the line defined x = 1. With the exception of the y-axis, each
1-dimensional subspace of R2 intersects L in exactly one point. Thus RP 1 = P (R2 ) is
isomorphic as a set with the union of L with a single point. Traditionally, we identify L
with R and call the extra point {}, the point at infinity. Thus we have a bijection of
sets
RP 1
= R {}
8
Another approach is to observe that every 1-dimensional subspace intersects the unit
circle S 1 in a pair of antipodal points. Thus a point in RP 1 corresponds to a pair of
antipodal points in S 1 . Topologically, we can construct RP 1 by taking a closed half circle
and identifying the endpoints. Thus RP 1 is itself a circle.
By similar reasoning we can see that the complex projective line CP 1 is in bijection
with C {}
CP 1
= C {}.
Topologically, CP 1 is a isomorphic to the 2-dimensional sphere S 2 and is sometimes called
the Riemann Sphere. The correspondence between C {} and S 2 can be understood
through stereographic projection, pictured in Figure 2.2.1.
We turn next to the real projective plane RP 2 = P (R3 ). Let x, y, z be the standard
coordinates on R3 . Then every 1-dimensional subspace of R3 either intersects the plane
z = 1 or lies in the plane z = 0. The first class of 1-dimensional subspaces is in bijection
with R2 , and the second class is in bijection with RP 1 . Thus we get a bijection
RP 2
= R2 R {}.
= R2 RP 1
It is not hard to see that this reasoning extends inductively to every dimension n to
determine the following bijection
and similarly
RP n
= Rn RP n1
= Rn Rn1 ... R0
(2)
CP n
= Cn CP n1
= Cn Cn1 ... C0 .
(3)
Returning now to the projective plane, observe that every 1-dimensional subspace of
R3 intersects the unit sphere S 2 in a pair of antipodal points.
10
11
12
2.2.2
Homogeneous coordinates
n
X
xi vi
i=0
2.3
Linear subspaces
2.3.1
Theorem 2.3. Given two distinct points [v1 ], [v2 ] P (V ), there is a unique projective
line containing them both.
Proof. Since [v1 ] and [v2 ] represent distinct points in P (V ), then the representative vectors
v1 , v2 V must be linearly independent, and thus span a 2-dimensional subspace W V .
Thus P (W ) is a projective line containing both [v1 ] and [v2 ].
Now suppose that there exists another projective line P (W 0 ) P (V ) containing [v1 ]
and [v2 ]. Then v1 , v2 W , so W W 0 . Since both W and W 0 are 2-dimensional it follows
that W = W 0 and P (W ) = P (W 0 ).
Example 5. Consider the model of RP 2 as pairs of antipodal points on the sphere S 2 . A
projective line in this model is represented by a great circle. Proposition 2.3 amounts to
the result that any pair of distinct and non-antipodal points in S 2 lie on a unique great
circle (by the way, this great circle also describes the shortest flight route for airplanes
flying between two points on the planet).
2.3.2
Theorem 2.4. In a projective plane, two distinct projective lines intersect in a unique
point.
Proof. Let P (V ) be a projective plane where dim(V ) = 3. Two distinct projective lines
P (U ) and P (W ) correspond to two distinct subspaces U, W V each of dimension 2.
Then we have an equality
P (U ) P (W ) = P (U W )
so proposition amounts to showing that U W V is a vector subspace of dimension 1.
Certainly
dim(U W ) dim(U ) = 2.
Indeed we see that dim(U W ) 1 for otherwise dim(U W ) = 2 = dim(U ) = dim(W )
so U W = U = W which contradicts the condition that U, W are distinct.
14
2.4
So far we have explored how a vector space V determines a projective space P (V ) and
how a vector subspace W V determines a linear subspace P (W ). How do linear maps
fit into projective geometry?
It is not true in general that a linear map T : V V 0 induces a map between the
projective spaces P (V ) and P (V 0 ).1 This is because if T has a non-zero kernel, then
it sends some 1-dimensional subspaces of V to zero in V 0 . However, if ker(T ) = 0, or
equivalently if T is injective, then this works fine.
Definition 5. Let T : V V 0 be an injective linear map. Then the map
: P (V ) P (V 0 )
defined by
([v]) = [T (v)]
is called the projective morphism or projective transformation induced by T .
Projective transformations from a projective space to itself : P (V ) P (V ) are called
projective automophisms and are considered to be the group of symmetries of projective
space.
If T is non-zero but not injective, then it does define a map from a subset of P (V ) to P (V 0 ). This is
an example of a rational map which play an important role in algebraic geometry. We will not consider
rational maps in this course though.
1
15
2.4.1
Erlangen Program
To better understand the meaning of this phrase, we make a brief digression to consider
the more familiar Euclidean geometry in the plane. This kind of geometry deals with
certain subsets of R2 and studies their properties. Especially important are points, lines,
lengths and angles. For example, a triangle is an object consisting of three points and three
line segments joining them; the line segments have definite lengths and their intersections
have definite angles that add up to 180 degrees. A circle is a subset of R2 of points lying a
fixed distance r from a point p R2 . The length r is called the radius and circumference
of the circle is 2r.
The group of symmetries of the Euclidean plane (or Euclidean transformations) consists of translations, rotations, and combinations of these. All of the objects and quantities studied in Euclidean geometry are preserved by these symmetries: lengths of line
segments, angles between lines, circles are sent to circles and triangles are sent to triangles. Two objects in the plane are called congruent if they are related by Euclidean
transformations, and this is the concept of isomorphism in this context.
In 1872, Felix Klein initiated a highly influential re-examination of geometry called the
Erlangen Program. At the time, several different types of non-Euclidean geometries had
been introduced and it was unclear how they related to each other. Kleins philosophy
was that a fundamental role was played by the group of symmetries in the geometry;
that the meaningful concepts in a geometry are those that preserved by the symmetries
and that geometries can be related in a hierarchy according to how much symmetry they
possess.
With this philosophy in mind, let us explore the relationship between projective and
Euclidean geometry in more depth. Recall from 2.2.2 that we have a subset U0 RP n
consisting of those points with homogeneous coordinates [1 : y1 : ... : yn ]. There is an
obvious bijection
[1 : y1 : ... : yn ] (y1 , ..., yn )
U0
= Rn ,
Proposition 2.5. Any Euclidean transformation of U0
= Rn extends uniquely to a pron
jective transformation of RP .
Proof. We focus on the case n = 2 for notational simplicity. It is enough to show that
translations and rotations of U0 extend uniquely.
Consider first translation in R2 by a vector (a1 , a2 ). That is the map (y1 , y2 ) 7
(y1 + a1 , y2 + a2 ). The corresponding projective transformation is associated to the linear
transformation
1
1 0 0
1
1
y1 7 a1 1 0 y1 = y1 + a1
y2
a2 0 1
y2
y 2 + a2
A rotation by angle of U0 = R2 extends to the projective transformation induced by the
linear transformation
16
1
1
0
0
1
1
y1
7 0 cos() sin() y1 = cos()y1 sin()y2
y2
0 sin() cos()
y2
cos()y2 + sin()y1
Uniqueness will be clear after we learn about general position.
From the point of view of the Erlangen Program, Proposition 2.5 tells us that projective geometry more basic than Euclidean geometry because it has more symmetry.
This means that theorems and concepts in projective geometry translate into theorems
and concepts in Euclidean geometry, but not vice-versa. For example, we have lines and
points and triangles in projective geometry, but no angles or lengths. Furthermore, objects in Euclidean geometry that are not isomorphic (or congruent) to each other can
be isomorphic to each other in projective geometry. We will show for example that all
projective triangles are isomorphic to one another. We will also learn that circles, ellipses,
parabolas and hyperbolas all look the same in projective geometry, where they are called
conics.
2.4.2
Next, we should clarify the relationship between linear maps and projective morphisms.
Proposition 2.6. Two injective linear maps T : V W and T 0 : V W determine the
same projective morphism if and only if T = T 0 for some non-zero scalar .
Proof. Suppose that T = T 0 . Then for [v] P (V ),
[T (v)] = [T 0 (v)] = [T 0 (v)],
so T and T 0 define the same projective morphism.
Conversely, suppose that for all [v] P (V ) that [T (v)] = [T 0 (v)]. Then for any basis
v0 , ..., vn V , there exist non-zero scalars 0 , ..., n such that
T (vi ) = i T 0 (vi ).
However it is also true that [T (v0 + ... + vn )] = [T 0 (v0 + ... + vn )] so for some non-zero
scalar we must have
T (v0 + ... + vn ) = T 0 (v0 + ... + vn ).
Combining these equations and using linearity, we get
n
X
T 0 (vi ) = T 0 (
i=0
n
X
i=0
n
n
n
X
X
X
vi ) = T (
vi ) =
T (vi ) =
i T 0 (vi )
i=0
Pn
i=0
i=0
2.4.3
c + dy
].
a + by
c + dy
a + by
6= 0
1
y
Proof. Recall from linear algebra that any invertible matrix can be obtained from the
identity matrix using elementary row operations: multiplying a row by a non-zero scalar,
permuting rows, and adding one row to another. Equivalently, every invertible matrix
can be written as a product of matrices of the form
1 0
0 1
1 0
,
,
0
1 0
1 1
and these correspond to scaling, inversion and translation respectively. For instance,
0 1
1
y
1
=
=y
1 0
y
1
1/y
so corresponds to the inversion y 7
1
y
It is also informative to try and picture these transformations using our model where
RP 1 is a circle and CP 1 is a sphere (leave this to class discussion). A delightful video
about Moebius transformations is available here.
18
c + D
y
a + b y
Direct sums
Let V and W be vector spaces over the same field F . The direct sum V W is the vector
space with vector set equal to the cartesian product
V W := V W := {(v, w)|v V, w W }
with addition and scalar multiplication defined by (v, w) + (v 0 , w0 ) = (v + v 0 , w + w0 ) and
(v, w) = (v, w). Exercise: show that
dim(V W ) = dim(V ) + dim(W ).
There are natural linear maps
i : V V W,
i(v) = (v, 0)
p : V W V,
p(v, w) = v
and
called inclusion and projection maps respectively. Similar maps exist for W as well.
Proposition 2.9. Let V be a vector space and let U, W be vector subspaces of V . Then
the natural map
: U W V,
(u, w) 7 u + w
has image
im() = span{U, W }
with kernel
ker() = U W.
In particular, if span{U, W } = V and U W = {0}, then defines an isomorphism
U W
= V and we say that V is the internal direct sum of U and W .
19
1
u = w u0
General position
We gain a firmer grasp of the freedom afforded by projective transformations using the
following concept.
Definition 6. Let P (V ) be a projective space of dimension n. A set of n + 2 points in
P (V ) are said to lie in general position if the representative vectors of any proper subset
are linearly independent in V (A subset S 0 S is called a proper subset if S 0 is non-empty
and S 0 does not equal S).
Example 7. Any pair of distinct points in a projective line are represented by linearly
independent vectors, so any three distinct points on a projective line lie in general position.
Example 8. Four points in a projective plane lie in general position if and only if no
three of them are collinear (prove this).
20
Theorem 2.10. If P (V ) and P (W ) are projective spaces of the same dimension n and
both p1 , ..., pn+2 P (V ) and q1 , ..., qn+2 P (W ) lie in general position, then there is a
unique projective transformation
: P (V ) P (W )
sending (pi ) = qi for all i = 1, 2, ..., n + 2.
Proof. Choose representative vectors v1 , ..., vn+2 V such that [vi ] = pi for all i. Because
the points pi lie in general position, it follows that the first n + 1 vectors v1 , ..., vn+1 are
linearly independent and thus form a basis of V . It follows that vn+2 can be written as a
linear combination
n+1
X
i vi = vn+2
i=1
Moreover, all of the scalars i are non-zero, for otherwise the non-trivial linear expression
n+1
X
i vi vn+2 = 0
i=1
would imply that some proper subset of {v1 , ..., vn+2 } is linearly dependent, in contradiction with the points lying in general position. By replacing our representative vectors vi
with the alterative representing vectors i vi , we may assume without loss of generality
that
n+1
X
vi = vn+2 .
i=1
wi = wn+2 .
i=1
Since v1 , ..., vn+1 V are linearly independent, they form a basis for V and we can define
a linear map T : V W by sending T (vi ) = wi for i = 1, ..., n + 1 and extending linearly.
By linearity we have
n+1
n+1
n+1
X
X
X
T (vn+2 ) = T (
vi ) =
T (vi ) =
wi = wn+2
i=1
i=1
i=1
n+1
X
i=1
21
vi ) =
n+1
X
i=1
i w i
so
n+1
X
wi = wn+2 =
i=1
i
n+2
n+1
X
i
wi
n+2
i=1
vi = vn+2 .
i=1
This is a useful fact that we will apply repeatedly in the rest of the course.
2.4.6
The Cross-Ratio
R(a, b; c, d) =
(a c)(b d)
.
(a d)(b c)
2.5
Classical Theorems
2.5.1
Desargues Theorem
23
p = c + c0
It follows that a + a0 = b + b0 so
x := a b = b0 a0
must represent the intersection point X = AB A0 B 0 , because x is a linear combination
of both a, b and of a0 , b0 . Similarly we get
z := b c = c0 b0
and y := c a = a0 c0
Pappus Theorem
Desargues Theorem is our first example of a result that is much easier to prove using
projective geometry than using Euclidean geometry. Our next result, named after Pappus
of Alexandria (290-350) is similar.
Theorem 2.12. Let A, B, C and A0 , B 0 , C 0 be two pairs of collinear triples of distinct
points in a projective plane. Then the three points X = BC 0 B 0 C, Y := CA0 C 0 A and
Z := AB 0 A0 B are collinear.
24
Proof. Without loss of generality, we can assume that A, B, C 0 , B 0 lie in general position.
If not, then two of the three required points coincide, so the conclusion is trivial. By
Theorem 2.10, we can assume that
A = [1 : 0 : 0],
B = [0 : 1 : 0],
C 0 = [0 : 0 : 1],
B 0 = [1 : 1 : 1]
The line AB corresponds to the two dimensional vector space spanned be (1, 0, 0) and
(0, 1, 0) in F 3 , so the point C AB must have the form
C = [1 : c : 0]
for some c 6= 0 (because it is not equal to A or B). Similarly, the line B 0 C 0 corresponds
to the vector space spanned by (0, 0, 1) and (1, 1, 1) so
A0 = [1, 1, a]
for some a 6= 1.
Next compute the intersection points of lines
BC 0 = span{(0, 1, 0), (0, 0, 1)} = {(x0 , x1 , x2 ) | x0 = 0}
B 0 C = span{(1, 1, 1), (1, c, 0)}
so we deduce that BC 0 B 0 C is represented by (1, 1, 1) (1, c, 0) = (0, 1 c, 1).
AC 0 = span{(1, 0, 0), (0, 0, 1)} = {(x0 , x1 , x2 ) | x1 = 0}
A0 C = span{(1, 1, a), (1, c, 0)}
so we deduce that A0 C AC 0 is represented by (1, c, 0) c(1, 1, a) = (1 c, 0, ca).
AB 0 = span{(1, 0, 0), (1, 1, 1)} = {(x0 , x1 , x2 ) | x1 = x2 }
A0 B = span{(1, 1, a), (0, 1, 0)}
so we deduce that AB 0 A0 B is represented by (a 1)(0, 1, 0) + (1, 1, a) = (1, a, a).
25
Finally, we must check that the intersection points [0 : 1 c : 1], [1 c : 0 : ca] and
[1 : a : a] are collinear, which is equivalent to showing that (0, 1 c, 1), (1 c, 0, ca) and
(1, a, a) are linearly dependent. This can be accomplished by row reducing the matrix,
1
a
a
1
a
a
1c
0
ca 0 (c 1)a a
0
1c 1
0 1c
1
1
a
a
0
0
0
0 1c 1
which has rank two, so the span of the rows has dimension two and the vectors are linearly
dependent.
2.6
Duality
In this section we explore the relationship between the geometry of P (V ) and that of
P (V ) where V denotes the dual vector space. Recall that given any vector space V , the
dual vector space V is the set of linear maps from V to F
V := { : V F } = Hom(V, F ).
The dual V is naturally a vector space under the operations
(1 + 2 )(v) = 1 (v) + 2 (v)
()(v) = ((v)).
Suppose that v1 , ..., vn V is a basis. Then the dual basis 1 , ..., n V is the
unique set of linear maps satisfying i (vj ) = 1, if i = j and i (vj ) = 0 otherwise.
If T : V W is a linear map, then there is a natural linear map T : W V
(called the transpose) defined by
T (f )(v) = f (T (v)).
Although the vector spaces V and V have the same dimension, there is no natural isomorphism between them. However there is a natural correspondence between subspaces.
Definition 7. Let U V be a vector subspace. The annihilator U 0 V is defined by
U 0 := { V | (u) = 0, for all u U }.
To check that U 0 is a subspace of V , observe that if 1 , 2 U 0 and u U then
(1 1 + 2 2 )(u) = 1 1 (u) + 2 2 (u) = 0 + 0 = 0
so (1 1 + 2 2 ) U 0 , so U0 is closed under linear combinations.
Proposition 2.13. If U1 U2 V then U10 U20 .
26
Proof. This is simply because if (v) = 0 for all v U2 then necessarily (v) = 0 for all
v U1 .
We also have the following result.
Proposition 2.14. dim(U ) + dim(U 0 ) = dim(V )
Proof. Choose a basis u1 , ..., um U and extend to a basis u1 , ..., um , vm+1 , ..., vn of V .
Let 1 , ..., n be the dual basis. For a general element
=
n
X
i i V
i=1
3.1
Given a set {x1 , ..., xn } of variables, a polynomial f (x1 , ..., xn ) is an expression of the form
X
f = f (x1 , ..., xn ) :=
aI x I
I
where
28
For some fields, two different polynomials can define the same function. For example x2 + x defines
the zero function over Z2 .
29
3.2
What is the right notion of an algebraic set in projective geometry? We begin with a
definition.
Definition 10. A polynomial is called homogeneous if all non-zero monomials have the
same degree.
Example 16.
x2 + y + 7 is a non-homogeneous polynomial.
yz 2 + 2x3 xyz is a homogeneous polynomial of degree three.
Projective algebraic sets in F P n are defined using homogeneous polynomials in n + 1
variables, which we normally denote {x0 , x1 , ..., xn }. The crucial property that makes
homogenous polynomials fit into projective geometry is the following.
Proposition 3.1. Let f (x0 , ..., xn ) be a homogeneous polynomial of degree d. Then for
any scalar F , we have
f (x0 , ..., xn ) = d f (x0 , ..., xn ).
Proof. For a monomial xI = xi00 ...xinn of degree d = |I| = i0 + ... + in , we get
(x)I = (x0 )i0 ...(xn )in = i0 +...+in xi00 ...xinn = d xI .
The same will hold true for a linear combination of monomials of degree d.
It follows from Proposition 3.1 that for a non-zero scalar, f (x0 , ..., xn ) = 0 if and
only if f (x0 , ..., xn ) = 0. This makes the following definition well-defined.
Definition 11. Let f1 , ..., fk be a set of homogeneous polynomials in the variables x0 , ..., xn .
The zero set
Z(f1 , ..., fk ) := {[x0 : ... : xn ] F P n | fi (x0 , ..., xn ) = 0 for all i=1,...k}
is a called a projective algebraic set. Notice that
Z(f1 , ..., fk ) = Z(f1 ) ... Z(fk ).
The main example we have encountered so far are linear subsets of projective space.
These are always of the form Z(l1 , ..., lk ) for some set of homogeneous linear polynomials
l1 , ..., lk .
To understand the correspondence between projective and algebraic sets we recall that
within F P n , we have a copy of the affine plane U0 := {[1 : y1 , ... : yn ] F P n }
= F n.
Under this correspondence, we have an equality
Z(f1 , ..., fk ) U0 = Zaf f (g1 , ..., gk )
where gi (x1 , ..., xn ) := f (1, x1 , ..., xn ).
30
3.3
n
X
xi y i .
i=1
n
X
xi yj B(vi , vj ).
i,j=1
This means that the bilinear form is completely determined by the n n-matrix with
entries i,j = B(vi , vj ).
Proof.
B(u, w) = B(x1 v1 + ...xn vn , w)
n
n
X
X
=
xi B(vi , w) =
xi B(vi , y1 v1 + ... + yn vn )
i=1
n X
n
X
i=1
xi yj B(vi , vj ) =
i=1 j=1
n
X
xi yj i,j
i,j=1
Lemma 3.2 provides a nice classification of bilinear forms: Given a vector space V
equipped with a basis v1 , ..., vn , then there is a one-to-one correspondence between symmetric bilinear forms B on V and symmetric n n matrices , according to the rule
B(v, w) = vT w
where v, w denote the vectors v, w expressed as column vectors and T denotes transpose,
so that vT is v as a row vector.
Proposition 3.3. The bilinear form B is non-degenerate if and only if the matrix has
non-zero determinant.
Proof. Suppose that det() is zero if and only if there exists a nonzero row vector vT such
that vT = 0 if and only if vT w = 0 for all column vectors w
if and only if there exists
v V \ {0} such that B(v, w) = 0 for all w V if and only if B is degenerate.
32
3.3.1
Quadratic forms
i,j xi xj
i,j
a b
b c
ax2
ax2 + 2bxy + cy 2
a b d
b c e ax2 + 2bxy + cy 2 + 2dxy + 2exz + f z 2
d e f
3.3.2
Change of basis
Weve seen so far that a symmetric bilinear form B can be described using a symmetric
matrix with respect to a basis. What happens if we change the basis?
Let v1 , ..., vn and w1 , ..., wn be two different bases of V . Then we can find scalars
{Pi,j }ni,j=1 such that
X
wi =
Pi,j vj
(4)
j
The matrix P with entries Pi,j is called the change of basis matrix from w1 , ..., wn to
v1 , ..., vn .
33
Proposition 3.5. The change of basis matrix P is invertible and the inverse P 1 is the
change of basis matrix from v1 , ..., vn to w1 , ..., wn .
Proof. Let Q be the change of basis matrix from v1 , ..., vn to w1 , ..., wn . Then
vi =
n
X
j=1
Qi,j wj =
n
X
Qi,j (
j=1
n
X
Pj,k vk ) =
n X
n
n
X
X
(
Qi,j Pj,k )vk =
(QP )i,k vk
k=1 j=1
k=1
k=1
T
Pi,k k,l Pl,j
k,l
k,l
T
= (P P )i,j
Alternatively, we can argue more conceptually as follows. Let ei denote the ith standard basis row vector. Then in terms of the basis v1 , ..., vn , wi corresponds to the row
vector ei P . Then
0
i,j
= B(wi , wj ) = (ei P )(ej P )T = ei P P T eTj = (P P T )i,j .
Our next task is to classify bilinear forms up to change of basis. First we need a
lemma.
Lemma 3.7. Let V be a vector space of dimension n equipped with a symmetric bilinear
form B. For w V define
w := {v V |B(v, w) = 0}.
Then w V is a vector space of dimension n 1.
34
Proof. Consider the map : V F , defined by (v) = B(v, w). Then is linear because
(1 v1 + 2 v2 ) = B(1 v1 + 2 v2 , w) = 1 B(v1 , w) + 2 B(v2 , w) = 1 (v1 ) + 2 (v2 )
so w is equal to the kernel of , hence it is a vector subspace. By the rank nullity
theorem,
(
n
im() = {0}
(5)
(6)
B(vi , vi ) =
c
= 1.
|c|
Corollary 3.9. Let V and B be as above and let Q(v) = B(v, v). Then
35
m
X
Pn
i=1 zi vi
zi2
i=1
p
X
zi2
i=1
p+q
X
zi2
i=p+1
Proof. Choose the basis v1 , ..., vn so that B(vi , vj ) satisfy the equations (5) and (6). Then
n
X
zi zj B(vi , vj ) =
i,j=1
n
X
zi2 B(vi , vi ),
i=1
f
|(a ,...,a ) = 0,
xi 1 n
If the the second order partials are defined and continuous then the n n-matrix H with
entries
2f
|(a ,...,a )
Hi,j = fxi xj (a1 , ..., an ) =
xj xi 1 n
is a symmetric matrix (Hi,j = Hj,i ) defining a symmetric bilinear form on the set of
tangent vectors called the Hessian. The corresponding quadratic form is equal to the
second order Taylor polynomial of the function.
The Hessian can be used to determine if f has a local maximum or minimum at
(a1 , ..., an ) (using the Second Derivative Test). Specifically, if H is non-degenerate then
f has a local maximum at (a1 ..., an ) if H is negative definite, has a local minimum if H
is positive definite, and has neither a max or min if H is neither positive nor negative
definite.
36
3.4
The first case Z(x2 ) is a single point of multiplicity 2, while Z(x2 +y 2 ) = Z((xiy)(x+iy))
consists of the two points [i : 1] and [i : 1]. This corresponds to the fact that over the
complex numbers 0 = ax2 + bx + c with a 6= 0 has either two distinct roots or a single
root of order two (complex polynomials in one variable can always be factored!).
We now consider the case F = R and P (V ) is a projective plane (in this case a quadric
is called a conic). We may choose coordinates so that X = Z(f ) where f is one of:
x2 ,
x2 + y 2 ,
x2 + y 2 + z 2 ,
x2 y 2 ,
x2 + y 2 z 2 ,
up to multiplication by 1 which doesnt change Z(f ) = Z(f ).
The first example
x2 = xx = 0
has the same solution set as x = 0, which is projective line {[0 : a : b]}
= RP 1 , however
this line is counted twice in a sense we wont make precise (similar two a double root of
single variable polynomial). The second example
x2 + y 2 = 0
has a single solution [0 : 0 : 1]. The third example x2 + y 2 + z 2 = 0, has no projective
solutions (remember (0, 0, 0) does not represent a projective point!). The example
x2 y 2 = (x y)(x + y)
factors as a product of linear polynomials. It follows that Z(x2 y 2 ) is a union of two
distinct lines intersecting at one point.
The last example Z(x2 + y 2 z 2 ) is (up to coordinate change) the only non-empty
and non-degenerate real conic, up to projective transformation. The apparent difference
between the affine conics: ellipses, parabolas, hyperbolas, can be understood as how the
conic relates to the line at infinity. A conic looks like an ellipse if it is disjoint form the
line at infinity, looks like parabola if it is tangent to the line at infinity, and looks like a
hyperbola if it intersects the line at infinity at two points.
To illustrate, consider what happens when we set z = 1 and treat x and y like affine
coordinates (so the equation z = 0 is the line at infinity). Then the polynomial equation
x2 + y 2 z 2 = 0 becomes x2 + y 2 1 = 0 or
x2 + y 2 = 1
which defines the unit circle in the x-y plane.
38
Now instead, set y = 1 and consider x, z as affine coordinates (so y = 0 defines the
line at infinity). Then we get the equation
x2 + 1 z 2 = 0,
or
z 2 x2 = (z x)(z + x) = 1
which is the equation of a hyperbola asymptotic to the lines x = z and x = z.
Finally, consider performing the linear change of variables z 0 = y + z. In these new
coordinates, the equation becomes
x2 + y 2 (z 0 y)2 = x2 + 2z 0 y z 02 = 0.
Passing to affine coordinates x, y by setting z 0 = 1 we get
x2 + 2y 1 = 0
or
y = x2 /2 + 1/2
which defines a parabola.
Since ellipses, hyperbolas, and parabolas can all be obtained by intersecting an affine
plane with the cone in R3 defined by x2 + y 2 z 2 , they are sometimes called conic
sections.
x2 + y 2 + z 2 = 0.
The equation x2 = 0 defines a projective line that is counted twice or has multiplicity
two (in a sense we wont make precise). The equation
x2 + y 2 = (x + iy)(x iy) = 0
defines the union of two distinct lines determined by the equations x + iy = 0 and
x iy = 0. The last equation x2 + y 2 + z 2 = 0 defines the non-degenerate quadric curve.
We study what this curve looks like in the next section.
3.5
:LC
such that, for X L, the points A, X, (X) are collinear.
Proof. Suppose the conic is defined by the symmetric bilinear form B. Let a V represent
the fixed point A. Because A C, we know that B(a, a) = 0. Let x V represent a
point X L. Because A 6 L, it follows that X 6= A so the vectors a, x are linearly
independent, so we can extend them to a basis a, x, y V .
The restriction of B to the span of a, x is not identically zero. If it were then the
matrix for B associated to the basis a, x, y would be of the form
0 0
0 0
40
which has determinant zero, and this is impossible because B is non-degenerate. It follows
that either B(a, x) or B(x, x) is non-zero.
An arbitrary point on the line AX has the form [a + x] for some scalars and .
Such a point lies on the conic C if and only if
B(a + x, a + x) = 2 B(a, a) + 2B(a, x) + 2 B(x, x)
= (2B(a, x) + B(x, x)) = 0
for which there are two solutions in the homogeneous coordinates and : The solution
= 0 corresponding the point A and solution to 2B(a, x) + B(x, x) = 0 corresponding
to the point represented by the vector
w := 2B(a, x)x B(x, x)a
which is non-zero because one of the coefficients 2B(a, x) or B(x, x) is non-zero.
We define the map : L C by
(X) = [w].
It remains to prove that is a bijection. We do this by proving the preimage of any
point is a paint. If Y is a point on C distinct from A, then by Theorem 2.4 there is a
unique point of intersection AY L, so 1 (Y ) = AY L.
Though Theorem 3.10 only says that is a bijection of sets, it is in fact an isomorphism
of algebraic varieties. In particular, this means that a real projective conic is has the
topology to a real projective line, which in turn has the topology of a circle - a fact that
weve already seen. A second consequence is that the complex projective conic is has the
topology of CP 1 which in turn has the topology of the two sphere S 2 .
3.5.1
Let us now describe a special case of the map in coordinates. Let C = Z(x2 + y 2 z 2 )
corresponding to the matrix
1 0 0
= 0 1 0 .
0 0 1
Choose A = [1 : 0 : 1] and let L = Z(x). Then points in L have the form [0 : t : 1] or
[0 : 1 : 0] and the map : L C satisfies
([0 : t : 1]) = [2B((1, 0, 1), (0, t, 1))(0, t, 1) B((0, t, 1), (0, t, 1))(1, 0, 1)]
= [2(0, t, 1) (t2 1)(1, 0, 1)]
= [(t2 + 1, 2t, t2 1)]
= [t2 1 : 2t : t2 + 1]
and
([0 : 1 : 0]) = [2B((1, 0, 1), (0, 1, 0))(0, 0, 1) B((0, 1, 0), (0, 1, 0))(1, 0, 1)]
= [0(0, 0, 1) + 1(1, 0, 1)]
= [1 : 0 : 1] = A
These formulas can be used to parametrize the circle using rational functions. For t a
real number 1 + t2 > 0 is non-zero so we can divide
[t2 1 : 2t : t2 + 1] = [
2t
t2 1
: 2
: 1]
2
t +1 t +1
(t) = 2
t + 1 t2 + 1
that hits all points except (1, 0).
These formulas have an interesting application in the case F = Q is the field of rational
numbers (those of the form a/b where a and b are integers).
Since we have focused on the cases F = R and F = C so far, we take a moment to
discuss informally linear algebra and projective geometry over F = Q. To keep matters
simple, we concentrate on the standard dimension n vector space over Q, Qn consisting
of n-tuples of rational numbers. It is helpful geometrically to consider Qn as a subset
42
of Rn in the standard way. Elements in Qn can be added in the usual way and can be
multiplied by scalars from Q, so it makes sense to form linear combinations
1 v1 + ...k vk Qn ,
where 1 , ..., k Q and v1 , ..., vk Qn . The span of a set of vectors span(v1 , ..., vk ) Qn
is the set of all linear combinations of v1 , ..., vk . In particular, a one-dimensional subspace
of Qn is the span of a single non-zero vector. We define the projective space
QP n = P (Qn+1 )
to be the set of one dimensional subspaces in Qn+1 . Geometrically, we can regard
QP n RP n
consisting of those projective points that can be represented by [v] RP n where v Qn+1 .
Many of the results about projective geometry we have developed so far generalize
immediately to geometry in QP n (one notable exception is the classification of symmetric
bilinear forms because the proof depends on taking square roots but the square root of a
rational number is not necessarily rational!).
In particular the formulas parametrizing the circle remain valid. That is, solutions to
the equation
x2 + y 2 = z 2
for x, y, z Q must be of the form
[x : y : z] = [1 t2 : 2t : 1 + t2 ]
for some rational number t Q. If we let t = a/b where a and b are integers sharing no
common factors, then
[x : y : z] = [1 (a/b)2 : 2(a/b) : 1 + (a/b)2 ] = [b2 a2 : 2ab : b2 + a2 ].
It follows in particular that every integer solution to the equation x2 + y 2 = z 2 must have
the form
x = c(b2 a2 ),
y = c(2ab),
z = c(b2 + a2 )
for some integers a, b, c. Such solutions are called Pythagorean triples because they define
right triangles with integer side lengths. For example, a = c = 1, b = 2 determines
x = 3, y = 4, z = 5, while a = 2, b = 3, c = 1 determines x = 5, y = 8, z = 13.
43
3.6
Polars
Proof. Let U V be the 2-dimensional subspace defining the projective line P (U ) and
let u, v U be a basis for U . A general point in P (U ) has the form [u + v] for scalars
, C not both zero. A point [u + v] lies on the conic C if
0 = B(u + v, u + v) = 2 B(u, u) + 2B(u, v) + 2 B(v, v).
(8)
We think of and as homogeneous coordinates on the projective line, so that (8) defines
the roots of a quadratic polynomial. From 3.4 we know that (8) has one or two solutions,
establishing (i). Property (ii) is a direct consequence of Proposition 3.11.
Now let [a] denote a point not lying on C, and let P (U ) denote the polar of [a]. From
the calculation above, it follows that P (U ) meets C in two distinct points [u] and [w] with
[u] 6= [w]. Because [u] lies on C we know B(u, u) = 0. Because [u] lies on the polar of [a]
we know B(u, a) = 0. It follows that for all points on the line joining [u] and [a] that
B(u, u + a) = B(u, u) + B(u, a) = 0
so the line joining [u] and [a] is the polar of [u] hence also the tangent to C at [u]. The
same argument applies to [w] and this establishes (iii).
45
Proposition 3.14 establishes a duality between circumscribed polygons and inscribed polygons for complex conics, as depicted in the real plane below.3
3.7
Quadrics in projective spaces P (V ) of dimension greater than 2 often contain linear spaces.
However, such a linear space can only be so big.
Proposition 3.15. Suppose that Q P (V ) is a non-degenerate quadric over field F in
which 0 6= 2. Then if some subspace P (U ) is a subset of Q, then
dim(P (U )) (dim(P (V )) 1)/2.
Proof. Because of dimension shifts, the statement is equivalent to
dim(U ) dim(V )/2
Suppose that P (U ) Q. Then for all u U , B(u, u) = 0. Furthermore, for any pair
u, u0 U we have by Proposition 3.4 that
1
B(u, u0 ) = (B(u + u0 , u + u0 ) B(u, u) B(u0 , u0 )) = 0
2
from which it follows that U U . According to Proposition 2.14 this means that
2 dim(U ) dim(U ) + dim(U ) = dim(V )
(9)
This duality is explored in the suggested project topic Pascals Theorem, Brianchons Theorem and
duality
46
Indeed, over the complex numbers the maximum value of (9) is always realized. If n
is odd then we can factor
x20 + ... + x2n = (x0 + ix1 )(x0 ix1 ) + ... + (xn1 + ixn )(xn1 ixn )
so the non-degenerate quadric Z(x20 + ... + x2n ) in CP n contains the linear space
Z((x0 ix1 ), ..., (xn1 ixn ))
of dimension (n 1)/2.
In the real case, the maximum value of (9) is achieved when n is odd and n + 1 =
2p = 2q (we say B has split signature) . In this case,
x21 + ... + x2p x2p+1 ... x22p = (x1 + xp+1 )(x1 xp+1 ) + ... + (xp + x2p )(xp x2p )
so the quadric contains the linear subspace Z(x1 xp+1 , ..., xp x2p ).
This is best pictured in the case n = 3:
x2 + y 2 z 2 w2 = (x z)(x + z) + (y w)(y + w) = 0
passing to affine coordinates by setting w = 1 gives
x2 + y 2 = z 2 + 1
which defines a surface called a hyperboloid or cooling tower.
Indeed, the hyperboloid can be filled or ruled by a family of lines satisfying the
equations
(x z) = (y 1)
(x + z) = (y + 1)
for some homogeneous coordinates [ : ]. It can also by ruled by the family of lines
satisfying
(x z) = (y + 1)
(x + z) = (y 1)
for some homogeneous coordinates [ : ].
3.8
In this section, we move beyond the study of individual quadrics to the study of families
of quadrics. The proof of the following proposition is straightforward and will be omitted.
Proposition 3.16. The set of symmetric bilinear forms on a vector space V forms a
vector space under the operation
(1 B1 + 2 B2 )(v, w) = 1 B1 (v, w) + 2 B2 (v, w).
47
48
(10)
According to Theorem 3.8, we may choose the basis v1 , ..., vn so that for some 0 p n
0
= 0
we have 1,1 = 2,2 = ... = p,p = 1 and all other entries i,j = 0 . By (10), i,j
unless (i, j) {(1, 1), (2, 2), ..., (p, p)} . If p 1 then we are done. If p 2, then for any
1 i p,
0
1,1 0
1 0
0
. = 1,1
.
0
0 i,i
0 1
0
so = 0 with = 1,1
and thus B = B 0 .
Remark 5. Proposition 3.17 does not hold over general fields. For example, the quadratic
forms x2 + y 2 and x2 + 2y 2 are not scalar multiples of each other, but Z(x2 + y 2 ) =
Z(x2 + 2y 2 ) = over the real numbers.
Definition 15. Let B, B 0 be two linearly independent symmetric bilinear forms on a
vector space V . The pencil of quadrics generated by B and B 0 is the set of quadrics
associated to bilinear forms of the form B + B 0 , for , not both zero. For complex
V , the pencil forms a projective line in P (S 2 (V )).
Proposition 3.18. Let V be a vector space of dimension n. A pencil of quadrics contains
at most n degenerate quadrics. If F = C, then the pencil contains at least one degenerate
quadric.
49
Proof. Choose a basis for V so that B and B 0 are represented by matrices and 0 . The
bilinear for B + B 0 is degenerate if and only if the determinant det( + 0 ) equals
zero. Expanding out the expression det( + 0 ) is a degree n homogeneous polynomial
in the homogeneous variables [ : ], so it has at most n roots. Over C, such a polynomial
has at least one root by the Fundamental Theorem of Algebra.
Theorem 3.19. Two non-degenerate conics in a complex projective plane intersect in at
least one and no more than four points.
Proof. First we show that all pairs of conics on the pencil intersect at the same points.
Let [v] P (V ) be an intersection point of two non-degenerate conics Z(B) and Z(B 0 ). If
B 00 := B + B 0 is a linear combination then
B 00 (v, v) = B(v, v) + B 0 (v, v) = 0 + 0 = 0
so [v] Z(B 00 ). It follows that Z(B) Z(B 0 ) Z(B 00 ) so
Z(B) Z(B 0 ) Z(B) Z(B 00 ).
If 6= 0, then the pencil determined by B and B 0 coincides with the one determinded by
B and B 00 . In particular,
B0 =
1
(B + B 0 ) B = B 00 B
(11)
Corollary 3.20. Two conics in a complex projective plane have between one and four
tangent lines in common.
Proof. Follows from Theorem 3.19 by duality (see Proposition ??).
Figure 21:
which is equal to the pencil spanned by the two (degenerate) conics Z(x(y z)) and
Z(y(x z)). This contains also a third degenerate conic Z(z(x y)).
From the above argument, we see that if two conics intersect in four points lying in
general position, then the pencil they determine contains three degenerate conics (the
maximum allowed by Proposition 3.19), corresponding to the 42 /2 = 3 ways to contain
four points in general position in a union of two lines:
Exterior Algebras
In this chapter we will always assume that fields F satisfy 2 6= 0 and consequently that
1 6= 1.
4.1
Multilinear algebra
Definition 16. Let V be a vector space over a field F . A multilinear form of degree k
(k 1) or k-linear form is a function
M : V k = V ... V F
that is linear in each entry. That is, for each i = 1, 2, ..., k we have
M (u1 , u2 , ...., ui + u0i , ...., uk ) = M (u1 , ..., ui , ..., uk ) + M (u1 , ..., u0i , ..., uk )
for all u1 , ..., uk , u0i V and , F . The set of k-linear forms is denoted T k (V ). By
convention T 0 (V ) = F .
Weve seen many example of multilinear forms already.
A multilinear form of degree 1 is simply a linear map : V F , so T 1 (V ) = V .
52
Figure 22: Degenerate Conics determining pencil, viewed in the z = 1 affine plane.
53
j=1
n
X
(12)
j1 ,...,jk =1
(13)
Of the examples above, dual vectors are alternating because the condition is vacuous
in degree one. The determinant is alternating, because transposing columns (or rows for
that matter) changes the determinant by minus one. Symmetric bilinear forms are of
course not alternating, because transposing the entries leaves it unchanged.
A consequence of (13) is that an alternating multilinear form will vanish if two entries
are equal. From (12) it follows that there are no nonzero multilinear forms of degree k
greater than n the dimension of V . More generally, we can say
Proposition 4.1. Let V be a vector space of dimension n over a field F in which 0 6= 2.
For k a non-negative integer, then k (V ) is a vector space of dimension nk .
Proof. Let M1 and M2 be k-linear alternating forms. The vector space structure is defined
by
(1 M1 + 2 M2 )(u1 , ..., uk ) = 1 M1 (u1 , ..., uk ) + 2 M2 (u1 , ..., uk ).
(14)
That this makes the set of k-linear forms into a vector space is a routine verification.
54
(15)
n
4.2
Definition 17. The kth exterior power k (V ) is the dual of the vector space of k-linear
alternating forms. Elements of k (V ) are sometimes called k-vectors.
It follows immediately from
results in the last section that 0 (V ) = F , that 1 (V ) = V
and that dim(k (V )) = nk .
Definition 18. If u1 , ...uk V , define the exterior product u1 .... uk k (V ) to be
the dual vector satisfying
(u1 .... uk )(M ) = M (u1 , ..., uk ).
That u1 ... uk is in fact an element of k (V ) is immediate from (14).
An element of k (V ) is called decomposable if it equals u1 ... uk for some choice of
vectors u1 , ..., uk . The following proposition tells us that every k-vector in k (V ) is equal
to a linear combination of decomposable k-vectors.
Proposition 4.3. Let V be a vector space with basis v1 , ..., vn . The set
S := {vi1 ... vik | 1 i1 < ... < ik n}
is a basis for k (V ).
Proof. It follows from the proof of Proposition 4.1 that the vector space of k-linear alternating forms has a basis {MI } indexed by k-subsets I {{i1 < ... < ik } | i1 , ..., ik
{1, 2, ..., n}}, defined by the condition MI (vi1 , ..., vik ) = 1 and MI (vj1 , ..., vjk ) = 0 for all
other k-tuples. The set S is simply the dual basis.
55
(16)
It is a straightforward consequence of the definition that (16) is linear in each factor (i.e.
it is k-linear) and that transposing two factors introduces a minus sign. Consequently, if
ui = uj for some i 6= j, then u1 ... uk = 0.
Proposition 4.4. A set of vectors u1 , ..., uk V is linearly independent if and only if
u1 ... uk k (V ) is non-zero.
Proof. Suppose that u1 , ..., uk V are linearly dependent. Then it is possible to write
one of them as a linear combination of the others
X
ui =
j uj .
j6=i
j u1 ... uj ... uk
j6=i
= 0
which equals zero because it is a sum of terms containing repeated factors.
On the other hand, if u1 , ..., uk is linearly independent, then it may be extended to a
basis of V . By Proposition 4.3, this means that u1 ... uk is part of a basis for k (V ),
so in particular it is non-zero.
Suppose that T : V V is linear transformation from V to itself. If M is a k-linear
form on V , then the pull-back T M is the k-linear form on V defined by the formula
T M (u1 , ..., uk ) = M (T (u1 ), ..., T (uk )).
If M is alternating, then T M is alternating because
T M (u1 , ..., ui , ..., uj , ..., uk ) = M (T (u1 ), ..., T (ui ), ...T (uj ), ..., T (uk ))
= M (T (u1 ), ..., T (uj ), ...T (ui ), ..., T (uk ))
= T M (u1 , ..., uj , ..., ui , ..., uk ).
The transpose of T is the linear map (abusing notation)
k T : k (V ) k (V )
that sends u1 ... uk to T (u1 ) ... T (uk ).
In the case k = n = dim(V ), we get something familiar (the following proof involves
material outside of the prerequisites and will not be tested).
56
j1 =1
j1 ,...,jn =1
To proceed further, observe that the product vj1 ... vjn equals zero if any of the indices
are repeated and equals v1 ... vn if they are all distinct. The sign is determined by
whether it takes an odd or an even number of transpositions get from the list (j1 , ..., jn ) to
the list (1, 2, ..., n), because a minus sign is introduced for each transposition. Generally, a
permutation is a list of the numbers {1, ..., n} written without repetition and we define
sign() = (1)t , where t is how many transpositions it takes to reorder to (1, ..., n).
With this notation, we have
T (v1 ... vn ) =
n
X
j1 ,...,jn =1
Proposition 4.6. There exists a unique bilinear map, called the exterior product,
: k (V ) l (V ) k+l (V ),
satisfying the property that for decomposable elements, we have
(u1 ... uk ) (w1 ... wl ) = u1 ... uk w1 ... wl ,
57
(17)
Proof. By choosing a basis and applying Proposition 4.3, it is clear that bilinearity combined with (17) uniquely determines the map. It remains to show the map exists and is
independent of any choice of basis. We do this by supplying a definition of the exterior
product that is independent of any basis.
Let b l (V ) be an arbitrary element. Define a map
b : k+l (V ) k (V )
by the rule that for vectors u1 , ..., uk V ,
(b (M ))(u1 , ..., uk ) = b(M (u1 , ..., uk , , , ...)).
Observe that when k = 0, then k (V ) = 0 (V ) = F and b is simply b.
It is easy to check that b (M ) inherits multi-linearity and the alternating property
from M so ib is well-defined. Given a k (V ) and b l (V ), we define the exterior
product a b k+l (V ) by the formula
(a b)(M ) = a b (M ) 0 (V ) = F,
for all alternating (k + l)-forms M .
Example 20. Suppose we have vectors (v1 v2 ) V = 1 (V ) and (v2 v3 + v1 v3 )
2 (V ). Then the product satisfies
(v1 v2 ) (v2 v3 + v1 v3 ) =
=
=
=
=
v1 (v2 v3 + v1 v3 ) v2 (v2 v3 + v1 v3 )
v1 v2 v3 + v1 v1 v3 v2 v2 v3 + v2 v1 v3
v1 v2 v3 + v2 v1 v3
v1 v2 v3 v1 v2 v3
0
u1 ... uk w1 ... wl
(1)k w1 u1 ... uk w2 ... wl
(1)2k w1 w2 u1 ... uk w3 ... wl
...
(1)kl w1 w2 .... wl u1 ... uk
(linearity)
(associativity)
(graded commutativity).
58
4.3
(U ) = k (U )
4.4
Decomposability of 2-vectors
= (
n1
X
ain vi ) vn +
i=1
ai,j vi vj
1i<jn1
= u vn + a0
where U = Span{v1 , ..., vn1 } V , u U and a0 2 (U ). Then
0 =
=
=
=
aa
(u vn + a0 ) (u vn + a0 )
u vn u vn + a0 u vn + u vn a0 + a0 a0
2a0 u vn + a0 a0
60
where in the last step we use the fact that products with repeated factors vanish and
graded commutativity of exterior multiplication. Because vn does not appear in the
expansion of a0 u or a0 a0 , they must both vanish separately and we get
a0 u = 0,
a0 a0 = 0.
4.5
A special case of Theorem 4.8 when k = 2 identifies Gr2 (V ), the set of lines in projective
space, with the set of points in P (2 (V )) represented by decomposable vectors. Let us
consider the implications of our algebraic characterization of decomposability (Proposition
4.9) in various dimensions of V .
If dim(V ) = 2, then 2 (V ) is one-dimensional. In this case P (V ) is itself a projective
line so it contains only a single line and P (2 (V )) is a single point.
If dim(V ) = 3, then P (V ) is a projective plane. Since dim(2 (V )) = 32 = 0, it follows
from Proposition 4.9 that every 2-vector is decomposable so Theorem 4.8 identifies the
set of lines in P (V ) with the set of points in P (2 (V )). Of course, we have already know
(2.6) that the set of lines in a projective plane P (V ) is in one-to-one correspondence
with the dual projective plane P (V ), so this result implies a natural bijection
P (2 (V ))
= P (V ).
To see where this comes from consider the wedge product
: V 2 (V ) 3 (V ).
Since the product is bilinear, 2-vectors in 2 (V ) determine linear maps form V to 3 (V ).
Since 3 (V ) is one-dimensional, there is an isomorphism 3 (V )
= F , which is uniquely
determined up to a non-zero scalar. Combining these allows us to consider 2-vectors as
defining linear maps V F which are determined up to a non-zero scalar, and this leads
to the natural bijection P (2 (V ))
= P (V ).
61
Now consider the case when dim(V ) = 4, so that P (V ) is projective 3-space. Choose
a basis v1 , ..., vn V . An arbitrary 2-vector in 2 (V ) has the form
a = x1,2 v1 v2 + x1,3 v1 v3 + x1,4 v1 v4 + x2,3 v2 v3 + x2,4 v2 v4 + x3,4 v3 v4
for some scalars xi,j which we treat as homogeneous variables for P (2 (V )). This gives
rise to a formula
a a = B(a, a)v1 v2 v3 v4
where B(a, a) is the quadratic form 2(x1,2 x3,4 x1,3 x2,4 + x1,4 x2,3 ). Combing with Proposition 4.9 we deduce the following.
Corollary 4.10. The set of lines in a three dimensional projective space P (V ), is in oneto-one correspondence with the points on a quadric X (called the Klein Quadric) lying
in the five dimensional projective space P (2 (V )). With respect to the standard basis of
2 (V ) coming from a basis v1 , v2 , v3 , v4 we have X = Z(f ) where
f = x1,2 x3,4 x1,3 x2,4 + x1,4 x2,3
Homework exercise: Prove that for dim(V ) = n 4, that the set of lines in P (V ) can
be identified with an intersection of quadrics in P (2 (V )).
In the remainder of this section we study the Klein Quadric Q. Recall from 3.7 that a
quadric in complex projective 5-space P (2 (V )) is a filled by families of projective planes.
Our next task is to characterize these planes in terms of the the geometry of lines in P (V ).
Proposition 4.11. Let L1 and L2 be two lines in a projective 3-space P (V ). Then L1
and L2 intersect if and only if they correspond to points (L1 ), (L2 ) P (2 (V )) that
are joined by a line contained in the Klein Quadric Q.
Proof. The lines L1 and L2 correspond to 2-dimensional subspaces U1 , U2 V . The lines
intersect if and only if U1 U2 contains an non-zero vector u U1 U2 .
Suppose that the lines do intersect. Then we may choose bases u1 , u U1 and u2 , u
U2 . The lines are represented by points [u1 u], [u2 u] P (2 (V )). These determine a
line represented by
u1 u + u2 u = (u1 + u2 ) u
which are all decomposable and thus are contained in Q.
Suppose that the lines do not intersect. Then U1 U2 = {0} and thus by ... U1 U2
=
V . In particular, there is a basis v1 , v2 , v3 , v4 V such that span{v1 , v2 } = U1 and
span{v3 , v4 } = U2 . The line containing (L1 ) and (L2 ) has points represented by
v1 v2 + v3 v4 ,
so
(v1 v2 + v3 v4 ) (v1 v2 + v3 v4 ) = 2v1 v2 v3 v4
is non-zero except when either or is zero, hence does not lie in Q.
62
Proposition 4.12. The set of lines passing through a fixed point X P (V ) corresponds
to a projective plane contained in the Klein Quadric Q P (2 (V )).
Proof. Let x V represent the point X P (V ), and extend to a basis x, v1 , v2 , v3 V .
Then a projective line L containing X corresponds to span{x, u} for some u = x+1 v1 +
2 v2 + 2 v2 with at least one of i non-zero. Then
(L) = [x (x + 1 v1 + 2 v2 + 2 v2 )] = [1 x v1 + 2 x v2 + 3 x v3 ],
which lies in the plane in P (2 (V ) spanned by the points [x v1 ], [x v2 ], [x v3 ]. Conversely, any point lying in this plane has form 1 x v1 + 2 x v2 + 3 x v3 and thus
corresponds to the line span{x, 1 v1 + 2 v2 + 3 v3 }.
We call a plane in the Klein Quadric of type described above an -plane. There is
another type of plane as well.
Proposition 4.13. The set of lines lying in a fixed plane P P (V ) determines a plane
in the Klein quadric Q P (2 (V )). Such a plane in Q is called a -plane.
Proof. Let [u1 ], [u2 ], [u3 ] P be three non-collinear points lying on the plane. An arbitrary
line in P is determined by a pair of points [u1 + u2 + u3 ] and [0 u1 + 0 u2 + 0 u3 ], and
maps to the point in Q
[(u1 + u2 + u3 ) (0 u1 + 0 u2 + 0 u3 )]
= [(0 0 )u1 u2 + ( 0 0 )u1 u3 + ( 0 0 )u2 u3 ]
contained in the plane determined by u1 u2 , u1 u3 and u2 u3 . Conversely, points in
P (2 (V )) of the form [1 u1 u2 + 2 u1 u3 + 3 u2 u3 ] must be decomposable, because
they are contained in 2 (span{u1 , u2 , u3 }) and 2-vectors over a 3-dimensional vector space
are always decomposable.
It is interesting to note that -planes and -planes are related by duality. A point
X P (V ) corresponds by duality to a plane X 0 P (V ). The set of lines in P (V )
containing the point X corresponds via duality to the set of lines contained in X 0 .
Proposition 4.14. Every plane contained in the Klein quadric is either an -plane or a
-plane.
Proof. Let P Q be a plane in the quadric and choose three non-collinear points in P
corresponding to three lines L1 , L2 , L3 P (V ). By Proposition 4.11, the intersections
Li Lj are non-empty. There are two cases:
Case 1: L1 L2 L3 is a point X. Then all (Li ) lie in the -plane associated to X
which must equal P .
Case 2: L1 L2 L3 is empty. Then the pairs of lines intersect three distinct points
X, Y, Z P (V ). These three points cant be collinear, because that would imply L1 =
L2 = L3 . Thus X, Y, Z determine a plane containing L1 , L2 , L3 so P must equal the
corresponding -plane.
63
Extra details
64
Linear sets.
Let V be a vector space with basis v1 , ..., vn V . The dual basis x1 , ..., xn V are
the linear maps xi : V F for which xj (vi ) = i,j . Equivalently, the dual basis is defined
by the property that
v = x1 (v)v1 + .... + xn (v)vn
for all v V . The general element of V is a linear combination
f (x1 , ..., xn ) = a1 x1 + .... + an xn .
It follows that V is equal to the set of homogenous polynomials of degree one for V .
Let U V be a vector subspace of dimension n. Choose a basis u1 , ..., uk U and
extend to a basis u1 , ..., uk , vk+1 , ..., vn V and let x1 , ..., xn V be the dual basis. Then
U = Zaf f (xk+1 , ..., xn ) V and P (U ) = Z(xk+1 , ..., xn ) P (V ).
If U1 and U2 are a pair of 2-dimensional subspaces of V , then there exist linear polynomials f1 and f2 such that P (U1 ) = Z(f1 ) and P (U2 ) = Z(f2 ) and thus
Z(f1 f2 ) = P (U1 ) P (U2 )
is a quadric in P (V ).
Distinct lines to not always intersect (but can intersect at most once).
Hyperbolic geometry is distinguished from Euclidean geometry in the following ways:
If L is a line and P is a point not on L, then there are many lines passing through
P that do not intersect L (multiple parallel lines).
The sum of angles in a triangle is strictly less than radians.
There are several different models of the hyperbolic plane.
1. The Poincare half plane model.
In this case we pass to affine C, where lying on the equator, the rest of the
equator passing to R C, and H passing to the upper half plane of C.
Figure 24: Poincare half plane model (blue lines and a tiling by 7-gons).
In this model, lines are represented by half circles centred on the real line, or by
vertical lines (note that distinct lines intersect at most once). Hyperbolic angles are
the same as Euclidean angles in this model.
2. The Poincare disk model.
We can also pass to affine C by choosing the south pole of CP 1 as and sending
the equator RP 1 to the unit circle and H is sent to the unit disk in C.
Hyperbolic angles are the same as Euclidean angles in this model. Lines correspond
to circular arcs that intersect the boundary at right angles, or straight lines that
bisect the disk. You can see why the angles of a triangle add to less than radians,
because triangles are bent in in this geometry.
3. The Klein disk model.
In this version, H is represented by the unit disk in R2 and lines a represented by
straight Euclidean lines.
In this model, the hyperbolic angles are not the same as Euclidean angles, but
distance has a nice interpretation. Given points P, Q in the interior of the disk, they
66
Figure 25: Poincare disk model (lines and a tiling by triangles with vertices on the boundary).
Figure 26: Klein disk model (lines and tiling by 7-gons and triangles).
67
determine a line that intersects the unit circle in two points A, B. The hyperbolic
distance of between P and Q is equal to the logarithm of the cross ratio:
1
d(P, Q) = log(R(A, B; P, Q)).
2
Below, we have an arrangement of lines depicted in the three different models:
69