MATH3031 Ch 2 notes
MATH3031 Ch 2 notes
MATH3031 Ch 2 notes
When we write, for example, x ∈ Rn , this means that “x” stands for such an n-tuple, and
we will denote by xi ∈ R the ith entry of x. Similarly for a ∈ Rn , p ∈ Rn , etc. (we will
not put arrows over the letters or lines under them).
Using this notation, we can write the formulas defining addition of two elements
x, y ∈ Rn in the following space-saving way: (x + y)i = xi + y i for all i = 1, . . . , n.
These operations make Rn into a vector space, and the notation gives us a convenient
way of defining the standard basis vectors of Rn : e1 , . . . , en ∈ Rn are defined, for each
i = 1, . . . , n and each j = 1, . . . , n, by
(
j 1 if i = j
(ei )j = δi :=
0 otherwise
(this function δij is called the Kronecker delta; it will come up a lot in later chapters).
We also give the formulas for the Euclidean inner product and Euclidean norm in
terms of this notation: n
X
hx, yi := xi y i
i=1
is the norm (or magnitude) of any vector x ∈ Rn (note that we’ve dropped the double
lines used in previous courses; this shouldn’t bother you or make you worry whether norm
or absolute value is meant, since the norm and absolute value are exactly the same thing
when n = 1). The inner product and norm have the usual properties, and geometric
interpretation, which we’ll review as needed.
holds for all x, y ∈ Rn and all α ∈ R. You probably recall that a linear map L : Rn → Rm
corresponds to a m × n matrix (meaning, a matrix with m rows and n columns). The
1
(standard) correspondence is given as follows, using the notation introduced above: For
each i = 1, . . . , n we consider the basis vector ei ∈ Rn . Since it gets mapped to some
vector L(ei ) ∈ Rm , we can express this as a linear combination of the standard basis
vectors e1 , . . . , em of Rm , i.e.
m
X
L(ei ) = Lji ej , for some L1i , . . . , Lm
i ∈ R.
j=1
This gives us a matrix of real numbers, denoted (Lji ), where the subscript i denotes which
column an entry is in (and ranges from 1 to n) and the superscript j denotes which row
an entry is in (and ranges from 1 to m).
Now, you are probably not used to thinking of the derivative this way, and certainly
you wouldn’t usually use Definition 2.1 to calculate the derivative of most maps you are
given. Instead, you would use the notion of partial derivatives:
2
Definition 2.4: (a) If U ⊂ Rn is open and f : U → R is a real-valued function and
1 ≤ i ≤ n, then the ith partial derivative of f at p ∈ U , if it exists, is the number given
by the limit
f (p + hei ) − f (p)
lim .
h→0 h
∂f
If it exists, we denote this number by Di f (p) (the notation ∂xi
(p) is also used).
The following theorem tells us that these two notions of derivatives are equivalent
under reasonable assumptions:
Di f (p) = Df (p)(ei ),
On the other hand, if all partial derivatives Di f (x) exist for x in a neighborhood of p,
and they are continuous in x, then f is differentiable at p, and the derivative of f at p is
given in matrix form by the Jacobian above.
(For a proof, see [Spivak], Theorem 2-7.) We will generally work with functions that
have continuous partial derivatives (and f is said to be C 1 ), and moreover assume that the
functions given by these partial derivatives, Di f : U → Rm , themselves have continuous
partial derivatives, written Dj (Di f ) : U → Rm for each j = 1, . . . , n, and so on. If this
can be repeated r times (with any combination of partial derivatives), then f is said to
be C r . If this is true, then we can take partial derivatives up to order r in any order: for
example, if f is C 2 , then Di (Dj f ) = Dj (Di f ) for all i, j = 1, . . . , n. A function is said to
be smooth, or C ∞ , iff it is C r for all r = 1, 2, . . ..
3
A prime example is the inverse function theorem, which answers the question, for
U ⊂ Rn open and p ∈ U , when it is possible to “locally invert” a function f : U → Rn
near p. In other words, we ask when it is possible to find open subsets V ⊂ Rn with
p ∈ V and W ⊂ Rn with f (p) ∈ W , such that the restriction f : V → W is one-to-
one and has a differentiable inverse g = f −1 : W → V (so g ◦ f (x) = x for all x ∈ V
and f ◦g(y) = y for all y ∈ W . If this is true, we say that f is a local diffeomorphism near p.
We will see, in the next section, that a direct consequence of the Inverse Function
Theorem is the Implicit Function Theorem, which allows us to generate a lot of (impor-
tant) examples of manifolds as level sets of functions, analogous to the level surfaces in
R3 that were discussed in Chapter 1.1. For now, let us discuss a few examples of how the
Inverse Function Theorem is applied to aid us with changes of coordinates in Rn :
4
(the change from Polar-Spherical to Cartesian coordinates in three dimensions). Then we
have
cos θ cos ϕ −r sin θ cos ϕ −r cos θ sin ϕ
Df (r, θ, ϕ) = sin θ cos ϕ r cos θ cos ϕ −r sin θ sin ϕ ,
sin ϕ 0 r cos ϕ
and from this we can calculate
det(Df (r, θ, ϕ)) = r2 cos ϕ.
Therefore, the Inverse Function Theorem tells us that f is a local diffeomorphism near
all points (r, θ, ϕ) with r 6= 0 and ϕ 6= 2k+1
2
π for any integer k. In other words, except
along the z-axis, we can find a differentiable inverse to f transforming (locally) Cartesian
coordinates (x, y, z) back into Polar-Spherical coordinates (r, θ, ϕ). (For points along the
z-axis, we need to modify the formula for f .)
2. For any map f : R2 → R2 , write f = (u, v), where u, v : R2 → R are the component
functions. Let J : R2 → R2 be the linear map with matrix
0 −1
J= .
1 0
(a) Calculate J 2 .
5
2.2: Manifolds
The Chinese mathematician, astronomer and geographer Zhang Heng (78-139 AD) is
credited as the one of the first to make use of a mathematical grid reference in cartography.
The Book of Later Han states that Zhang “cast a network of coordinates about heaven
and earth, and reckoned on the basis of it”. In this section, we introduce the modern
mathematical notion of spaces for which it is possible to “cast a network of coordinates”
that can be used for “reckoning”: these are manifolds, the generalisation of surfaces, on
which differential and integral calculus can be developed.
Definition 2.8: (a) For n any non-negative integer, a parametrised n-manifold is given
by a map
r : U → Rn+k , with M = r(U ),
where U ⊂ Rn is a domain (open, connected subset) and k is some non-negative integer.
(The case n = 2, k = 1 just gives us a parametrised surface as in Definition 1.1.)
(b) In this course, we almost always work with regular parametrisations (often without
explicitly mentioning this). This means that the map r : U → Rn+k is differentiable; the
vectors D1 r(x), . . . , Dn r(x) ∈ Rn+k are linearly independent for all x ∈ U ; and the map
r : U → M is one-to-one, with continuous inverse.
(c) A (general, regular) n-dimensional manifold is any subset M ⊂ Rn+k that can be
“patched together” out of regular parametrised n-manifolds. More precisely, this means
that for any point p ∈ M , it is possible to find ε > 0 and a regular parametrisation
r : U → Rn+k of the subset M ∩ B(p, ε), for B(p, ε) = {y ∈ Rn+k | |y − p| < ε} the
open ball of radius ε around p. For a point p ∈ M , such a parmetrisation is called a local
coordinate system near p.
Example 2.9: (a) The easiest example is if M ⊂ Rn is any open subset of Rn . Then
we can take U = M and r : U → Rn the identity map (in other words, r(x) = x for all
x ∈ U ). This shows us that M is a regular parametrised n-manifold. It includes the case
M = Rn , since Rn itself is an open subset of Rn (we never said that subsets have to be
proper subsets), thus showing that Rn is a n-dimensional manifold. (That “ought to” be
true, since n-dimensional manifolds are in some sense “modelled on” Rn .) Note that in
this example k = 0.
(b) Slightly more interesting than example (a), consider the case where M ⊂ Rn+k is
a linear subspace of dimension n. By definition, this means that M has a basis consisting
of n vectors, so let’s suppose that v1 , . . . , vn ∈ M is a basis. We can use this to define a
parametrisation of M , with map r : Rn → Rn+k : for any x ∈ Rn , define
n
X
1 2 n
r(x) = x v1 + x v2 + . . . + x vn = xi vi .
i=1
(Remember, xi is the ith coordinate of the vector x ∈ Rn for each i = 1, . . . , n.) You
should try on your own to verify that this is a regular parametrisation of M ; some details
6
will be discussed in the lecture video.
(c) We can generalise explicit surfaces (see Example 1.2(b) in Chapter 1.1) to get
explicit n-manifolds M ⊂ Rn+k for any non-negative integers n and k: Suppose that
f : U → Rk is a differentiable map, for U ⊂ Rn some domain. This defines a subset
M ⊂ Rn+k , by letting
This subset is just the graph of the function f , as is clear from writing it in the more
compact notation
M = {(x, f (x)) | x ∈ U }.
(With this notation, it should be kept in mind that x ∈ Rn and f (x) ∈ Rk for any x ∈ U ,
so (x, f (x)) is technically an element of the Cartesian product Rn × Rk , which we are
silently identifying with Rn+k .) You should try, yourself, to use the function f to write
down a formula for a parametrisation r : U → Rn+k of M ; and then verify that this
parametrisation is regular. We can discuss this more in lecture videos if there’s a desire
to do so.
(d) We can also generalise level surfaces (Example 1.2(d)); for now, we’ll just discuss
how this is done for k = 1 (k > 1 will be discussed later in Chapter 2). Let g : W → R
be a differentiable function, for W ⊂ Rn+1 an open subset, and let c ∈ R be a value such
that the subset
M = g −1 (c) := {y ∈ W | g(y) = c}
is non-empty. Suppose, moreover, that the gradient ∇g(y) is non-zero for all y ∈ M .
Then it is a theorem that M ⊂ Rn+1 is a (general, regular) n-dimensional manifold.
Theorem 2.10 (The n-Sphere): For any integer n ≥ 0 and any real number R > 0,
the n-sphere of radius R, SRn ⊂ Rn+1 , defined as
is a n-dimensional manifold.
for any y ∈ Rn+1 . This function has gradient given by ∇g(y) = 2y for all y ∈ Rn+1 , which
is clearly non-zero for all vectors y 6= 0. Hence, since SRn = g −1 (R2 ), and this set does not
contain the 0-vector of Rn+1 , it follows from the theorem mentioned in Example 2.9(d)
that SRn is a n-manifold. (QED)
7
Theorem 2.11 (The Special Linear Group): For any integer n ≥ 1, let SL(n, R)
denote the set of real n × n matrices with determinant 1, which we consider as a subset
2
of Rn by identifying n × n matrices with vectors with n2 coordinates. Then SL(n, R) is a
manifold of dimension n2 −1 (it is also group under the operation of matrix multiplication,
called the special linear group).
2 2
Proof: Identifying n × n matrices with vectors in Rn , we have the map det : Rn → R
given by taking the determinant of a matrix. Using Question 1 from the Chapter 2.1 TUT
2
Problems, you can derive the following formula for the derivative Ddet(m) : Rn → R
2 2
at a point m ∈ Rn . We think of m ∈ Rn as a n × n matrix with column vectors
2
m1 , . . . , mn ∈ Rn . Then, using the same notation for any other vector x ∈ Rn , we can
show (see TUT problems for this chapter) that
n
X
Ddet(m)(x) = det(m1 , . . . , mi−1 , xi , mi+1 , . . . , mn ),
i=1
where (m1 , . . . , mi−1 , xi , mi+1 , . . . , mn ) denotes the n×n matrix with the indicated column
vectors. In particular, if m ∈ SL(n, R) then
n
X
Ddet(m)(m) = det(m) = n.
i=1
the theorem mentioned in Example 2.9(d) tells us that SL(n, R) is a (n2 − 1)-manifold.
(QED)
8
2.3: Implicit Manifolds
In this section, we’re going to state and prove the theorem mentioned in Example
2.9(d), in its general form. This theorem gives us a large and important class of exam-
ples, called implicit manifolds (they generalise implicit surfaces).
Linear Algebra Example 2.12: Before talking about the general theory of implicit
manifolds, let’s consider the “baby example” from linear algebra. Let L : Rn+k → Rk be
a linear map, and a ∈ Rk a fixed vector. Then the level set
This shows that y0 := y1 − y2 ∈ Ker(L), and clearly with y0 thus defined we have
y2 = y1 + y0 , as claimed.
Now, from linear algebra we know that Ker(L) ⊂ Rn+k is a linear subspace. (You
should prove this, as an exercise, if you’re unsure.) Furthermore, its dimension is given
by the following formula, known as the “Rank-Nullity Theorem” of linear algebra (ditto
for this):
dim(Ker(L)) = (n + k) − rank(L),
where the integer rank(L) is given by any of the following three (equivalent) definitions:
Using any of these definitions, it follows that rank(L) ≤ k. The case where equality
holds is the main one we’re interested in (we often say L has maximal rank ), and we get:
9
Proof: Exercise, using the facts from Example 2.12 and adapting (slightly) the method
from Example 2.9(b).
Theorem 2.14 (Implicit Manifold Theorem): Let V ⊂ Rn+k be an open subset and
g : V → Rk a differentiable map, and suppose that a ∈ Rk is a choice of vector for which
the level set
M = g −1 (a) = {y ∈ Rn+k | g(y) = a}
is non-empty. If p ∈ M is a point where the derivative Dg(p) : Rn+k → Rk has maximal
rank, then M has a regular parametrisation as a n-manifold near p. If Dg(p) has maximal
rank for all p ∈ M , then M ⊂ Rn+k is a n-dimensional manifold.
Idea of the Proof: The philosophy of this theorem, like for the Inverse Function Theorem,
is that we can “carry over” the result from the linear algebra version above, using the
philosophy that since the derivative Dg(p) is a “good linear approximation” of g near p,
what holds for this linear approximation should be transferrable to g, at least “near p”.
In fact, this works and we use the Inverse Function Theorem to make the proof go through.
To get a feel for what’s actually going on in the proof given below (and in exam-
ples where we apply it), it helps to think of it in the following way: the level set is
defined by k equations (which may now be non-linear) depending on n + k variables.
We can prove the result if we manage, in each of these k equations, to make a different
one of the variables the subject of that equation. For example, we could try to make
y n+1 the subject of the first equation g 1 (y) = a1 (this means rewriting the equation as
y n+1 = f 1 (y 1 , . . . , y n , y n+2 , . . . , y n+k ) for some real-valued function f 1 ); then try to make
y n+2 the subject of the second equation g 2 (y) = a2 ; etc., until we get to making y n+k the
subject of the last equation g k (y) = ak . If this can be done, in some neighbourhood of
p, then it means we have succeeded in writing M (locally) as the graph of the function
f = (f 1 , . . . , f k ), and this gives us a regular parametrisation using the solution to Exam-
ple 2.9(c).
Of course, we can’t do all of this explicitly, since we don’t know the form of the func-
tion g. We only know that rank(Dg(p)) = k. So we will use this fact to build a certain
function that does the job for us (actually, its inverse does the job, using the Inverse
Function Theorem) and then we get the function f almost as if by magic.
It’s also clear that we won’t necessarily be able to make the last k variables into the
subjects of our equations near a given point p. (To see this, and get some more intuition
for the abstract proof, consider the case of an implicit surface in R3 , i.e. the case where
we have V ⊂ R3 open and g : V → R, so n = 2 and k = 1. Then, for well-known examples
like a cylinder, with g(x, y, z) = x2 + y 2 , or a sphere, with g(x, y, z) = x2 + y 2 + z 2 , there
are many points where it’s not possible to make z the subject of the defining equation.)
However, it’s enough to make ANY k different variables into the subjects of the equa-
tions. This only differs from making the last k variables the subject by a combination
of rotations of Rn+k needed to interchange the variables. Since such a rotation doesn’t
change the regularity of our parametrisation, we will make our lives easier by starting off
with the assumption (which does not limit the generality) that g and p are of the form
10
where a solution for the last k variables is possible. How this works will be clear below.
Actual Proof: We use the definition of rank(Dg(p)) = k that tells us that the matrix of
Dg(p) has k linearly independent column vectors. Without loss of generality, we assume
that the last k columns are linearly independent (this is the assumption discussed above,
which will make it possible for us to solve for the last k variables). Since the ith column
of Dg(p), for 1 ≤ i ≤ n + k, is given by
Dg(p)(ei ) = Di g(p)
(in other words, the ith partial derivative of g at p), this means that the vectors
Taking partial derivatives of the component functions, we see that this function is differ-
entiable and its derivative is given, in block form, by
In×n 0n×k
Dϕ(y) = n+k , y ∈ V.
(Di g(y))ni=1 (Dj g(y))j=n+1
(Note that the blocks on top each have n rows, the blocks on bottom each have k rows,
while each block on the left has n columns and each block on the right has k columns, so
that overall the size of the matrix is (n + k) × (n + k).) This shows that the determinant
is given by
det(Dϕ(y)) = det(Dn+1 g(y) · · · Dnk g(y)),
which shows in particular that det(Dϕ(p)) 6= 0, since Dn+1 g(p), . . . , Dn+k g(p) are linearly
independent and hence the k × k matrix having them as column vectors has non-zero
determinant (linear algebra).
Thus, the Inverse Function Theorem lets us conclude that ϕ is a local diffeomorphism
near p. We will assume that the open set V is taken to be small enough that ϕ : V → W
is a diffeomorphism for some open subset W ⊂ Rn+k containing ϕ(p). Let ψ : W → V be
its inverse, so ψ(ϕ(y)) = y for all y ∈ V and ϕ(ψ(w)) = w for all w ∈ W . Then define an
open subset U ⊂ Rn and a map f : U → Rk by letting
U = {x ∈ Rn | (x, a) ∈ W },
Finally, we verify that M is the graph of f near p, meaning that we need to show
M ∩ V = {(x, f (x)) | x ∈ U }.
11
To see this, let y ∈ M ∩ V . Then y ∈ M ⇒ g(y) = a, and y ∈ V means that ϕ(y) ∈ W
and y = ψ(ϕ(y)). But
which shows that (y 1 , . . . , y n ) ∈ U for all y ∈ M ∩ V . On the other hand, using the
definition we have
f (y 1 , . . . , y n ) = π(ψ(y 1 , . . . , y n , a))
= π(ψ(ϕ(y)), by (2)
= π(y), since ψ, ϕ are inverses.
On the other hand, if x ∈ U is any point, then we see that (x, f (x)) ∈ M ∩ V as
follows: Since (x, a) ∈ W ⇒ ψ(x, a) ∈ V . Writing y := ψ(x, a), then since ϕ and ψ are
inverses,
for any point (x1 , y 1 , . . . , xn , y n ) ∈ R2n . Fix the vector a = (1, 1, . . . , 1) ∈ Rn , and define
T n = g −1 (a).
Solution: To prove this, we just have to show that the hypotheses of Theorem 2.14 are
satisfied. The derivative of g is easily calculated (note that the size of the matrix is n
rows and 2n columns):
2x1 2y 1 0 0 ··· 0 0
0
0 2x2 2y 2 ··· 0 0
1 1 n n .. .
. . . . . . . .
. .
.
Dg(x , y , . . . , x , y ) = . .
. . . . . .
n−1 n−1
0 0 · · · 2x 2y 0 0
0 0 ··· 0 0 2xn 2y n
12
This shows that the derivative has maximal rank (i.e. rank(Dg)(p) = n) at all points
p = (x1 , y 1 , . . . , xn , y n ) ∈ R2n for which we have (xi , y i ) 6= (0, 0) for all i = 1, . . . , n. But
this is clearly the case for all points of T n , since
Hence, Dg(p) : R2n → Rn has maximal rank for all p ∈ T n , which by Theorem 2.14
implies that T n ⊂ R2n is a n-dimensional manifold.
T n = S 1 × . . . × S 1.
We will see, later in this chapter, how to identify the 2-torus T 2 = S 1 × S 1 (note: T 2 ⊂ R4
by this description) with the more familiar “torus of revolution” which you know as a
surface in R3 . This is one example showing that we can have two n-manifolds that are
basically the same, even though they “live in” Euclidean spaces of different dimension.
In other words, the integer k in our definition of a manifold is not really fundamental
to the manifold M that we are actually interested in. In technical terms, k is called the
codimension and we have defined M as an embedded submanifold of Euclidean space, with
codimension k.
13
2.4: Tangent Spaces and Derivatives
In this section, we’ll see how the coordinate systems on a manifold allow us to define
tangent spaces and to make sense of the derivative of a map between two manifolds.
Before discussion all of that, we need to be able to talk about what happens if we
change parametrisation of “change coordinates” on our manifold. The key to this is the
following result, which we will not worry about proving:
hαβ := r−1
α ◦ rβ : Vβ → Vα
Definition 2.17 (Tangent Space): For M ⊂ Rn+k a manifold and p ∈ M a point, the
tangent space of M at p, Tp M , is the set of tangent vectors of curves through p in M :
Proof: First, note that since Dr(q) : Rn → Rn+k is a linear map, its image is a linear
subspace of Rn+k and satisfies
Dr(q)(Rn ) = {Dr(q)(v) | v ∈ Rn }
= span{Dr(q)(e1 ), . . . , Dr(q)(en )}
= span{D1 r(q), . . . , Dn r(q)}.
Therefore, since the vectors D1 r(q), . . . , Dn r(q) ∈ Rn+k are always linearly independent
for a regular parametrisation r, this linear subspace is n-dimensional.
Second, we can show that this linear subspace is contained in Tp M with an argument
analogous to what was done in the lectures and TUT problems to prove part of Theorem
14
1.4 in Chapter 1.1. This means we let α1 , . . . , αn ∈ R be any real coefficients and show
that there is a curve γ = γα : (−ε, ε) → M with γ(0) = p and
and letting γ = r ◦ h : (−ε, ε) → M . Then γ(0) = r(h(0)) = r(q) = p, and its derivative
γ 0 (0) is shown to satisfy (2) using the Chain Rule.
The remainder of the proof will be done only in the case that M is an implicit manifold.
In fact, this is not a restriction since ANY manifold can be written, near a point p ∈ M ,
as an implicit manifold. This means that for a suitable function g : V → Rk , a ∈ Rk
and an open set V ⊂ Rn+k with p ∈ V , we have M ∩ V = g −1 (a) (the proof does not
involve any major new ideas, but is too “technical” to justify including it). So, suppose
M = g −1 (a) for some g : V → Rk and a ∈ Rk satisfying the hypotheses of Theorem 2.14.
We show that Tp M ⊂ Ker(Dg(p)) as follows: suppose γ : (−ε, ε) → M is a curve in M
through p, and define h = g ◦ γ : (−ε, ε) → Rk . Then h(t) = a for all t ∈ (−ε, ε) since
M = g −1 (a), so h0 (0) = 0 (the 0-vector in Rk ). But by the Chain Rule,
Dr(q)(Rn ) ⊂ Tp M ⊂ Ker(Dg(p)).
However, the sets on the left and the right are both linear subspaces of Rn+k , and they
both have dimension n. (We discussed the dimension of Dr(q)(Rn ) above, and the dimen-
sion of Ker(Dg(p)) is seen to be n using the Rank-Nullity Theorem of linear algebra, and
the fact that Dg(p) has rank k.) But this implies that the two linear subspaces are in fact
equal, and therefore Tp M must also equal both of them, which concludes the proof. (QED)
15
local coordinate systems, this might make you worried about the possibility of ever figur-
ing out if a given function is differentiable. Luckily, that’s not a problem, as a result of
the following Lemma, which also tells us how the tangent space is related to differentiable
functions.
(a) If rα : Uα → Rn+k and rβ : Uβ → Rn+k are two local coordinate systems near p,
then f ◦ rα is differentiable at qα = r−1
α (p) ∈ Uα if and only if f ◦ rβ is differentiable at
−1
qβ = rβ (p) ∈ Uβ .
where π i : Rn → R denotes projection onto the ith coordinate in Rn , and define the ith
partial derivative with respect to this parametrisation as
∂f
i
(p) := Di (f ◦ r)(q), q = r−1 (p) ∈ U.
∂x
Then
(b) The coordinate functions xi : M → R are differentiable, and the ith partial
derivatives exist at a point p ∈ M if and only of f is differentiable at p. For two different
local coordinate systems rα , rβ near p, the partial derivatives of a differentiable function
are related by
n
X ∂xj
∂f α ∂f
i
(p) = i
(p) j (p) (3)
∂xβ j=1
∂xβ ∂xα
Proof: Suppose rα and rβ are two parametrisations near a point p. Then for Vα , Vβ ⊂ Rn
as in Lemma 2.16, we know that the change of coordinates map
hαβ = r−1
α ◦ rβ : Vβ → Vα
f ◦ rβ = f ◦ rα ◦ r−1
α ◦ rβ
= f ◦ rα ◦ hαβ .
Thus, since qβ ∈ Vβ and qα ∈ Vα , the fact that hαβ is a diffeomorphism, together with the
Chain Rule, implies that f ◦ rβ is differentiable at qβ if and only if f ◦ rα is differentiable
at qα , procing (a).
16
To prove (b), we note first that xi ◦ r = π i : U → R, so since the projection map
π i : Rn → R is linear (and hence differentiable), this shows that xi is differentiable using
(a). Moreover, writing f ◦ rβ = f ◦ rα ◦ hαβ as above, then xjα ◦ rβ = π j ◦ hαβ for any
j = 1, . . . , n and hence
n
X ∂xjα
Dhαβ (qβ )(ei ) = (p)ej .
j=1
∂xiβ
Thus, using the Chain Rule for the derivative, we calculate
∂f
(p) = Di (f ◦ rβ )(qβ )
∂xiβ
= D(f ◦ rβ )(qβ )(ei )
= D(f ◦ rα )(hαβ (qβ )[Dhαβ (qβ )(ei )]
" n #
X ∂xj
α
= D(f ◦ rα )(qα ) i
(p)ej
j=1
∂xβ
n
X ∂xj α ∂f
= i
(p) j (p).
j=1
∂xβ ∂xα
(b) Given a point p ∈ Mf , write p = (q, f (q)) for a unique point q ∈ U . Use Theorem
2.18 to show that the tangent space of Mf at p is given by
Tp Mf = {(v, Df (q)(v)) | v ∈ Rn }.
(i) Using Theorem 2.14 (Implicit Manifold Theorem), prove that for a given value
a ∈ R, Si,a ⊂ R3 is either empty or a 2-dimensional manifold (i.e. a general regular
surface).
17
(ii) Let a, b ∈ R be two values such that S1,a and S2,b have nonempty intersection, and
denote their intersection by Γa,b ⊂ R3 :
Use Theorem 2.14 to prove that if the gradient vectors ∇g1 (x, y, z) and ∇g2 (x, y, z) are
linearly independent for all (x, y, z) ∈ Γa,b , then Γa,b is a 1-dimensional manifold (i.e.
a regular differentiable curve). Use Theorem 2.18 to show that in this case, for any
p = (x, y, z) ∈ Γa,b , we have
(iii) Use the results of (i) and (ii) to prove that the subset
H = {(x, y, z) | x2 + y 2 − z 2 = 1}
g2 (x, y, z) = n1 x + n2 y + n3 z,
3. The Special Linear Group (SL(n, R)): For this problem and the next, let n ≥ 2
be a given integer and let Mn = Mn (R) denote the set of n × n matrices with real
2
entries. We identify Mn with Rn whenever needed, and we will also use the notation
m = (m1 , . . . , mn ) for any m ∈ Mn , where m1 , . . . , mn denote the columns of a matrix
m, which identifies Mn = (Rn )n (see Question 1 of Chapter 2.1 TUT problems).
(b) Define
SL(n, R) = det−1 (1) = {m ∈ Mn | det(m) = 1}.
Use the result of (a), and Theorem 2.14, to show that this is a manifold of dimen-
sion n2 − 1. It is also a group, under the operation of matrix multiplication, since
det(mp) = det(m) det(p) for m, p ∈ Mn , from Linear Algebra. Moreover, the opera-
tions of taking inverses and matrix multiplication are differentiable, and hence we have
what is called a Lie group.
18
(c) Let e ∈ SL(n, R) denote the n × n identity matrix (which has 1s along the diagonal
entries and 0s elsewhere). Use Theorem 2.18 and the result of part (a) to show that the
tangent space of SL(n, R) at e consists of all trace-free matrices:
Sn = {m ∈ Mn | mT = m}
= {m = (mji ) | mji = mij for all i, j = 1, . . . , n},
where mT denotes the transpose of a matrix m, which you should remember from Linear
Algebra (if not, please ask for further clarification). Since a symmetric matrix is com-
pletely determined by just its entries that are on the diagonal and above it (i.e. by the
entries mji with i ≥ j), we see that the dimension of Sn (as a vector space) is
n(n + 1)
dim(Sn ) = 1 + 2 + . . . + n = ,
2
and hence we can identify Sn = Rn(n+1)/2 if we want to.
(a)* Given a matrix m ∈ Mn , let g(m) = mmT , the (matrix) product of m with its
transpose. Show that this defines a map
g : Mn → Sn ,
2
(which, using the above remarks, we can think of as a map g : Rn → Rn(n+1)/2 ). Moreover,
show that the derivative of this map is given, at any m ∈ Mn , by the linear map Dg(m) :
Mn → Sn defined, for any x ∈ Mn , by
where e ∈ Sn is the n × n identity matrix (as in Question 3), which is clearly symmetric.
Use the result of (a) and Theorem 2.14 to prove that O(n) is a manifold of dimension
n(n − 1)
dim(O(n)) = .
2
Show that O(n) is closed under matrix multiplication, using the fact that (mp)T = pT mT
for any m, p ∈ Mn (this is a fact from Linear Algebra; please ask if you don’t believe it).
Hence, O(n) is another example of a Lie group.
19
(c) Using the result of part (a), and Theorem 2.18, show that the tangent space of
O(n) at the identity e ∈ O(n) consists of all anti-symmetric n × n matrices:
Te O(n) = {x ∈ Mn | xT = −x}.
(d)* By the above, O(2) is a manifold of dimension 1. Can you identify it with any
1-dimensional manifold we have already met? (Hint: Consider 2 × 2 matrix that rotates
R2 about the origin. What form does it have, and why should it be in O(2)? On the
other hand, consider a matrix that reflects R2 (say about the horizontal axis). Is this in
O(2)? Is it in the family of matrices describing rotations?)
20
2.5: Differentiable Maps Between Manifolds
Definition/Lemma 2.21 (Differentiable Maps): Let M ⊂ Rn+k be a n-manifold,
N ⊂ Rm+l a m-manifold, ϕ : M → N a map between them, and p ∈ M a point. Then
the following are equivalent, and if any one of them holds we say that ϕ is differentiable
at p (as usual, we say ϕ is differentiable iff it is differentiable at all points p ∈ M ):
(a) For some local coordinate system r = rM : U → Rn+k for M near p, the function
ϕ ◦ r : U → Rm+l is differentiable at the point q = r−1 (p) ∈ U ;
(b) For each j = 1, . . . , k + l, and π j : Rm+l → R the jth coordinate function, the
real-valued function π j ◦ ϕ : M → R is differentiable at p (in the sense of Definition 2.19);
(c) For some choices of local coordinate systems r = rM : U → Rn+k for M near p,
and s = sN : V → Rm+l for N near ϕ(p), the map
ϕsr := s−1 ◦ ϕ ◦ r : U → V
We will not go through the proof, since it is very similar to the proof of Lemma 2.20.
Note that Lemma 2.20 immediately implies that the definition is independent of which
choices of local coordinate coordinate systems one makes in either (a) or (c).
21
with ϕsr : U → V as defined in part (c) of Def./Lemma 2.21.
Using these equivalent definitions, it can be proven that the derivatives of differentiable
maps between manifolds have all the usual properties, such as the Chain Rule, etc.
22