Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
IRA
Sets are the most basic building blocks in mathematics, and it is in fact not easy to give a
precise definition of the mathematical object set. Once sets are introduced, however, one
can compare them, define operations similar to addition and multiplication on them, and
use them to define new objects such as various kinds of number systems. In fact, most of
the topics in modern analysis are ultimately based on sets.
Therefore, it is good to have a basic understanding of sets, and we will review a few
elementary facts in this section. Most, if not all, of this section should be familiar and its
main purpose is to define the basic notation so that there will be no confusion in the
remainder of this text.
Definition 1.1.1: Sets and Operations on Sets
A set is a collection of objects chosen from some universe. The universe is
usually understood from the context. Sets are denoted by capital, bold letters or
curly brackets.
A B: A is a subset of B means that every element in A is also contained
in B.
A B: A union B is the set of all elements that are either in A or in B or
in both.
A B: A intersection B is the set of all elements that are in both sets A
and B.
A \ B: A minus B are all elements from A that are not in B.
comp(A): The complement of A consists of all elements that are not in
A.
Two sets are disjoint if A B = 0 (the empty set)
Two sets A and B are equal if A B and B A
The most commonly used sets are the sets of natural numbers, integers, rational and
real numbers, and the empty set. They are usually denoted by these symbols:
N = {1, 2, 3, 4, ... } = natural numbers (sometimes 0 is considered part of the
natural numbers as well)
Z = {... -3, -2, -1, 0, 1, 2, 3, ... } = integers
Q = {p / q : p, q Z} (read as "all number p / q, such that p and q are elements of
Z") = rational numbers
R = real numbers
0 = empty set (the set that contains no elements)
All of the number systems (except the natural numbers) will be defined in a
mathematically precise way in later sections. First, some examples:
Examples 1.1.2:
3.
Find O E, O I, comp(I).
Sets can be combined using the above operations much like adding and multiplying
numbers. Familiar laws such as associative, commutative, and distributive laws will be true
for sets as well. As an example, the next result will illustrate the distributive law; other
laws are left as exercises.
Proposition 1.1.3: Distributive Law for Sets
A (B
A
(B
C) = (A
C) = (A
B)
(A C)
B)
(A
C)
Proof
Many results in set theory can be illustrated using Venn diagram, as in the above proof.
However, such diagrams do not represent mathematically rigorous proofs. Nonetheless,
before an actual proof is developed, it is first necessary to form a mental picture of the
assumptions, conclusions, and implications of a theorem. For this process a Venn diagram
can be very helpful. You can practice Venn diagrams by using them for some of the
true/false statements in the exercises.
There are many other theorems dealing with operation on sets. One that is particularly
interesting is the theorem about de Morgan's Laws, because it deals with any number of
sets (even infinitely many). Drawing a Venn diagram in such a situation would be
impossible, but a mathematical proof can easily deal with this situation:
Theorem 1.1.4: de Morgan Laws
So far, we have reviewed a few basic facts from set theory, and also got an idea about how
a course in Real Analysis will proceed:
First, there are definitions, stating precisely what we are talking about. From those
definitions we derive new results, based on old results, notation, and logic. The new results
are called Theorems (if they are important or broad), Propositions (if they are interesting,
but not so broadly applicable) and Corollaries (which are usually restatements of theorems
or propositions in special situations). We will proceed that way throughout the text.
The most difficult part of Real Analysis is trying to understand the proofs of new results, or
even developing your own proofs. While there are a few 'general' methods for proofs, a lot
of experience and practice is needed before you will feel familiar with giving your own
proofs. However, only a few proofs require real ingenuity, and many other proofs can be
understood by carefully reviewing the definitions of terms involved. Therefore, as a rule:
write down the precise mathematical definitions of all terms involved before
starting a proof
In following that rule, one often gets ideas about how to start a proof by starting to
manipulate the mathematical symbols involved in the precise definitions of the terms.
Keep in mind that a proof can (almost) never be given by means of examples. Working out
a few examples can certainly be helpful - and should in fact always be done before starting
a proof - but they can not constitute a rigorous proof of a general statement.
Two types of proofs will be encountered frequently, and deserve special attention:
Proof by Induction: This type of proof is introduced in detail in the next chapter.
Proof by Contradiction: In this type of proof one assumes that the proposition (i.e.
what one actually would like to proof) is false. Then one derives a contradiction,
i.e. a logical impossibility. If that can be accomplished, then one has shown that the
negation of a statement will result in an illogical situation. Hence, the original
statement must be true.
Examples 1.1.5:
Prove that when two even integers are multiplied, the result is an even
integer, and when two odd integers are multiplied, the result is an odd
integer.
Prove that if the square of a number is an even integer, then the
original number must also be an even integer. (Try a proof by
contradiction)
Euclid's Theorem states that there is no largest prime. A proof by
contradiction would start out by assuming that the statement is false, i.e.
there is a largest prime. The advantage now is that if there was a largest
prime, there would be only finitely many primes. This seems easier to
handle than the original statement which implies the existence of
infinitely many primes. Finish the proof.
IRA
After introducing some of the basic elements of set theory (sets), we will move on to the
second most elementary concept, the concept of relations and functions.
Definition 1.2.1: Relation
Let A and B be two sets. A relation between A and B is a collection of ordered
pairs (a, b) such that a A and b B. Often we use the notation a ~ b to
indicated that a and b are related, rather then the order pair notation (a, b).
Note that this does not mean that each element from A needs to be associated with one (or
more) elements from B. It is sufficient if some associations between elements of A and B
are defined. In contrast, there is the definition of a function:
Definition 1.2.2: Function, Domain, and Range
Let A and B be two sets. A function f from A to B is a relation between A and B
such that for each a A there is one and only one associated b B. The set A is
called the domain of the function, B is called its range.
Often a function is denoted as y = f(x) or simply f(x), indicating
the relation { (x, f(x)) }.
Examples 1.2.3:
{(x,sin(x))}
Which of those relations are functions ?
The outcomes of a function (i.e. the elements from the range associated to elements in the
domain) do not only depend on the rule of the function (such as x is associated with sin(x))
but also on the domain of the function. Therefore, we need to specify those outcomes that
are possible for a given rule and a given domain:
Definition 1.2.4: Image and Preimage
Let A and B be two sets and f a function from A to B. Then the image of f is
defined as
imag(f) = {b B : there is an a A with f(a) = b}.
Let A and B be two sets and f a function from A to B. If C is a subset of the
range B then the preimage, or inverse image, of C under the function f is the set
defined as
f -1(C) = {x A : f(x) C }
As an example, consider the following functions:
Example 1.2.5:
A function f from A to B is called one to one (or one- one) if whenever f(a) =
f(b) then a = b. Such functions are also called injections.
A function f from A to B is called onto if for all b in B there is an a in A such
that f(a) = b. Such functions are also called surjections.
A function f from A to B is called a bijection if it is one to one and onto, i.e.
bijections are functions that are injective and surjective.
Examples 1.2.7:
3. g(x) = sin(x)
4. h(x) = 2x3 + 5x2 - 7x + 6
IRA
Just as there were different classes of functions (bijections, injections, and surjections),
there are also special classes of relations. One of the most useful kind of relation (besides
functions, which of course are also relations) are those called equivalence relations.
Definition 1.3.1: Equivalence Relation
Let S be a set and r a relation between S and itself. We call r an equivalence
relation on S if r has the following three properties:
1. Reflexivity: Every element of S is related to itself
2. Symmetry: If s is related to t then t is related to s
3. Transitivity: If s is related to t and t is related to u, then s is related to u.
Examples 1.3.2:
Proof
Example 1.3.4:
What kind of equivalence classes do you get when x and y are defined
to be related if their difference is divisible by m ? How could you add
those ?
Consider the set R x R \ {(0,0)} of all points in the plane minus the
origin. Define a relation between two points (x,y) and (x, y) by saying
that they are related if they are lying on the same straight line passing
through the origin. Then:
o
Check that this relation is an equivalence relation and find a
graphical representation of all equivalence classes by picking an
appropriate member for each class.
IRA
In this section we will define some number systems based on those numbers that anyone is
familiar with: the natural numbers. In a course on logic and foundation even the natural
numbers can be defined rigorously. In this course, however, we will take the numbers {1,
2, 3, 4, ...} and their basic operations of addition and multiplication for granted. Actually,
the most basic properties of the natural numbers are called the Peano Axioms, and are
defined (being axioms they need not be derived) as follows:
Definition 1.4.1: Peano Axioms
1 is a natural number
For every natural number x there exists another natural number x called
the successor of x.
1 # x for every natural number x (x being the successor of x)
If x = y then x = y
If Q is a property such that:
1. 1 has the property Q
2. if x has property Q then x has property Q
then these operations are well-defined and the resulting set of all equivalence
classes has all of the familiar properties of the integers (it therefore serves to
define the integers based only on the natural numbers).
Proof
When defining operations on equivalent classes, one must prove that the operation is welldefined. That means, because the operation is usually defined by picking particular
representatives of a class, one needs to show that the result is independent of these
particular representatives. Check the above proof for details.
The above proof is in fact rather abstract. The basic question, however, is: why would
anyone even get this idea of defining those classes and their operations, and what does this
really mean, if anything ? In particular, why is the theorem entitled The Integers ? Try to
go through the few examples below:
Examples 1.4.3:
Which elements are contained in the equivalence classes of, say, [(1,
2)], [(0,0)] and of [(1, 0)] ? Which of the pairs (1, 5), (5, 1), (10, 14), (7,
3) are in the same equivalence classes ?
What do you get when adding, say, [(1,2)] + [(4, 6)] ? How about
[(3,1)] + [(1,3)] ? What about multiplying [(5,4)] * [(7, 4)] and [(1,2)] *
[(2,1)] ?
Can you think of a better notation for denoting the classes containing,
say, (5, 7) ? How about the class containing (7, 5) ? Do you see why the
above theorem is called 'The Integers' ?
An obvious question about multiplication is: why did we not define
[(a, b)] * [(a, b)] = [(a * a, b * b)]
It certainly looks a lot easier. But, keep in mind that all pair (a,b) are in the same class, if
their difference b - a is the same. Hence, we think of the class [(a,b)] as represented by b a and of the class [(a, b)] as the class represented by b - a. Then, in view of the fact that
(b - a) * (b - a) = b * b + a * a - (a * b + a * b)
our original definition of multiplication does make some sense.
Incidentally, now it is clear why multiplying two negative numbers together does give a
positive number: it is precisely the above operation on equivalent classes that induces this
interpretation. For example:
(-2) * (-4) = 8
This now has a precise mathematical interpretation.
(-2) is a representative of the equivalence class of pair of natural numbers (a. b)
(which we understand completely) whose difference b - a = -2.
One such pair representing the whole class is, for example, (4, 2).
To represent the class that might be denoted by -4, we might choose the pair (8, 4).
But then, according to our definition of multiplication, we have:
[(4, 2)] * [(8, 4)] = [ (4 * 4 + 2 * 8) , (4 * 8 + 2 * 4)] = [(32, 40)]
and the appropriate notation for that class is 40 - 32 = 8. Hence, we now understand
completely why (-2) * (-4) = +8.
Next, we will define the rational numbers in much the same way, leaving the proof as an
exercise. To motivate the next theorem, think about the following:
a / b = a / b if and only if a * b = a * b
a / b + a / b = (a b + b * a) / (b * b)
a / b * a / b = (a * a) / (b * b)
Theorem 1.4.4: The Rationals
Let A be the set N x N - {0} and define a relation r on N x N - {0} by saying
that (a, b) is related to (a, b) if a * b = a * b. Then this relation is an
equivalence relation.
If [(a, b)] and [(a, b)] denotes the equivalence classes containing (a, b) and
(a, b), respectively, and if we define the operations
1. [(a, b)] + [(a', b')] = [(a * b' + a * b, b * b')]
2. [(a, b)] * [(a, b)] = [(a * a, b * b)]
then these operations are well-defined and the resulting set of all equivalence
classes has all of the familiar properties of the rational numbers (it therefore
serves to define the rationals based only on the natural numbers).
Proof
Note that the second component of a pair of integers can not be zero (otherwise the relation
would not be an equivalence relation). As before, this will yield, in a mathematically
rigorous way, a new set of equivalence classes commonly called the 'rational numbers'. The
individual equivalent classes [(a,b)] are commonly denoted by the symbol a / b, and are
often called a 'fraction'. The requirement that the second component should not be zero is
the familiar restriction on fractions that their denominator not be zero.
As a matter of fact, the rational numbers are much nicer than the integers or the natural
numbers:
A natural number has no inverse with respect to addition or multiplication
An integer has an inverse with respect to addition, but none with respect to
multiplication.
A rational number has an inverse with respect to both addition and multiplication.
Example 1.4.5:
IRA
One of the more obvious features of the three number systems N, Z, and Q that were
introduced in the previous chapter is that each contains infinitely many elements. Before
defining our next (and last) number system, R, we want to take a closer look at how one
can handle 'infinity' in a mathematically precise way. We would like to be able to answer
questions like:
1. Are there more even than odd numbers ?
2. Are there more even numbers than integers ?
3. Are there more rational numbers than negative integers ?
While most people would probably agree that there are just as many even than odd
numbers, it might be surprising that the answer to the last two questions is no as well. All
of the sets mentioned have the same number - albeit infinite - of elements. The person who
first established a rigorous 'theory of the infinite' was G. Cantor.
The basic idea when trying to count infinitely large (or otherwise difficult to count) sets
can roughly be described as follows:
Suppose you are standing in an empty classroom, with a lot of students waiting to
get in. How could you know whether there are enough chairs for everyone? You
can not count the students, because they walk around too much. So, you simply let
in the students, one by one, and take a seat. If all the seats are taken, and no
students are left standing, then there was the same number of students as chairs.
This simple idea of matching two sets element by element is the basis for comparing two
sets of any size, finite or infinite. Since 'matching elements from one set with those in
another set ' seems related to the concept of a function, we have arrived at the following
definition:
Definition 2.1.1: Cardinality
Let A and B be two sets. We say that A and B have the same
cardinality if there is a bijection f from A to B. We write card(A) =
card(B).
If there exists a function f from A to B that is injective (i.e. one-to-one)
we say that card(A) card(B).
By the above examples, the set of even integers, odd integers, all positive and negative
integers are all countable.
Note that there is a difference between finite and countable, but we will often use the word
countable to actually mean countable or finite (even though it is not proper). However,
here is a nice result that distinguishes the finite from the infinite sets:
Theorem 2.1.4: Dedekind Theorem
A set S is infinite if and only if there exists a proper subset A of S which has the
same cardinality as S.
Proof
Examples 2.1.5:
Use Dedekind's Theorem to show that the set of integers Z and the
interval of real numbers between 0 and 2, [0, 2], are both infinite (which
is of course not surprising).
The surprising fact when dealing with countably infinite sets is that when combining two
countable sets one gets a new set that contains no more elements than each of the previous
sets. The next result will illustrate that.
Proposition 2.1.6: Combining Countable Sets
Think about these propositions carefully. It seems to be contrary to ones beliefs. To see
some rather striking examples for the above propositions, consider the following:
Examples 2.1.7:
IRA
The last section raises the question whether it is at all possible to have sets that contain
more than countably many elements. After all, the examples of infinite sets we encountered
so far were all countable. It was Georg Cantor who answered that question: not all infinite
sets are countable.
Proposition 2.2.1: An Uncountable Set
The open interval (0, 1) is uncountable.
Proof
Note that this proposition assumes the existence of the real numbers. At this stage,
however, we have only defined the integers and rationals. We are not supposed to know
anything about the real numbers. Therefore, this proposition should - from a strictly logical
point of view - be rephrased:
if there are 'real numbers' in the interval (0, 1) then there must be uncountably
many.
However, the real numbers, of course, do exist, and are thus uncountable. As for a more
elementary uncountable set, one could consider the following:
the set of all infinite sequences of 0's and 1's is uncountable
The proof of this statement is similar to the above proposition, and is left as an exercise.
What about other familiar sets that are uncountable ?
Examples 2.2.2:
If S = {1,2,3}, then what is P(S) ? What is the power set of the set S
= {1, 2, 3, 4} ? How many elements does the power set of S = {1, 2, 3,
4, 5, 6} have ?
Show that card(P(S)) card (S) for any set S (you dont have to
proof strict inequality, only greater or equal).
numbers card(S), where S is any finite or infinite set, are called cardinal
numbers, and one can indeed establish rigorous rules for adding and
subtracting them.
Try to establish the following definitions for dealing with cardinalities:
1. Definition of a cardinal number
2. Comparing cardinal numbers
3. Addition of two cardinal numbers
Examples 2.2.8:
Using your above definitions, find the answers for the following
examples:
1. What is card(N) + card(N) ?
2. What is card(N) - card(N) ?
3. What is card(R) + card(N) ?
4. What is card(R) + card(R) ?
This theorem can be used to show, for example, that R x R has the same cardinality as R
itself, which is one of the exercises. It can also be used to prove that Q is countable by
showing that Q and Z x Z have the same cardinality (another exercise).
IRA
In this section we will briefly review a common technique for many mathematical proofs
called the Principle of Induction. Based on this principle there is a constructive method
called Recursive Definition that is also used in several proofs. Both principles, in fact, can
be applied to many well-ordered set.
Definition 2.3.1: Ordered and Well-Ordered Set
A set S is called partially ordered if there exists a relation r (usually denoted by
the symbol ) between S and itself such that the following conditions are
satisfied:
1. reflexive: a a for any element a in S
2. transitive: if a b and b c then a c
3. anti-symmetric: if a b and b a then a = b
A set S is called ordered if it is partially ordered and every pair of elements x
and y from the set S can be compared with each other via the partial ordering
relation.
A set S is called well-ordered if it is an ordered set for which every non-empty
subset contains a smallest element.
Examples 2.3.2:
Determine which of the following sets and their ordering relations are
partially ordered, ordered, or well-ordered:
1. S is any set. Define a b if a = b
2. S is any set, and P(S) the power set of S. Define A B if A B
3. S is the set of real numbers between [0, 1]. Define a b if a is less
than or equal to b (i.e. the 'usual' interpretation of the symbol )
4. S is the set of real numbers between [0, 1]. Define a b if a is
greater than or equal to b.
Which of the following sets are well-ordered ?
1. The number systems N, Z, Q, or R ?
2. The set of all rational numbers in [0, 1] ?
3. The set of positive rational numbers whose denominator equals
3?
Recall the this is very similar to parts of the Peano Axioms, and it is easy to see that the
principle of induction applies to the well-ordered set of natural numbers.
To use the principle of induction for the natural numbers one has to proceed in four steps:
1. Define a property that you believe to be true for some ordered set (such as N)
2. Check if the property is true for the smallest number of your set (1 for N)
3. Assume that property is true for an arbitrary element of your set (n for N)
4. Prove that the property is still true for the successor of that element (n+1 for N)
Examples 2.3.4:
1.
2.
Suppose the induction principle defined above does not contain the
assumption that every element except for the smallest has an immediate
predecessor. Then show that it could be proved that every natural
number must be even (which is, of course, not true so the additional
assumption on the induction principle is necessary).
A somewhat more complicated, but very useful theorem that can be proved by induction is
the binomial theorem:
Theorem 2.3.6: Binomial Theorem
Let a and b be real numbers, and n a natural number. Then:
Proof
Based on the Induction principle is the principle of Recursive Definition that is used
frequently in computer science.
Definition 2.3.7: Recursive Definition
Let S be a set. If we define a function h from N to S as follows:
1. h(1) is a uniquely defined element of S
2. h(n) is defined via a formula that involves at most terms h(j) for 0 < j <
n
Then this construction determines a unique function h from N to S.
Examples 2.3.8:
looking at an induction proof, and therefore, one might use induction to prove anything
(Incidentally, such proofs are often called non-constructive proofs).
Exercise 2.3.9:
Try to use induction to prove that the sum of the first n square
numbers is equal to (n + 2)/3. (This induction proof should fail, since
the statement is false - or is it true ?)
Here is a more elaborate example of an invalid induction proof:
Example 2.3.10:
"Proof"
The property is clearly true for n = 1, because one bird is of the same color.
Assume that n birds are of the same color.
Now take (n+1) birds. Put one aside. There are n birds left, which, by assumption,
are of the same color. For simplicity, say they are all black. Put the one bird back
into the group, and take out another one. Again, there are n birds remaining, which
by assumption must have the same color. In particular, the bird that was taken out
in the first place must now have the same color as all the other birds, namely black.
The bird taken outside was also black. Therefore, the n+1 birds must all be black.
But that means that we have proved property Q for all natural numbers by
induction. Yet, the statement is obviously wrong. Where does this proof actually
break down ?
A similar word of caution applies to Recursive Definitions. While that principle can be
very useful, one has to be careful not to get into logical difficulties.
Example 2.3.11:
A classical example for a recursive definition that does not work is the
paradox of the barber of Seville: The barber of Seville is that inhabitant
of Seville who shaves every man in Seville that does not shave himself.
IRA
Right now, however, it will be more important to describe those properties of R that we
will need for the remainder of this class.
The first question is: why do we need the real numbers ? Arent the
rationals good enough ?
Theorem 2.4.1: No Square Roots inQ
There is no rational number x such that x2 = x * x = 2.
Proof
Thus, we see that even simple equations have no solution if all we knew were rational
numbers. We therefore need to expand our number system to contain numbers which do
provide a solution to equations such as the above.
There is another reason for preferring real over rational numbers: Informally speaking,
while the rational numbers are all 'over the place', they contain plenty of holes (namely the
irrationals). The real numbers, on the other hand, contain no holes. A little bit more formal,
we could say that the rational numbers are not closed under the limit operations, while the
real numbers are. More formally speaking, we need some definitions.
Definition 2.4.2: Upper and Least Upper Bound
Let A be an ordered set and X a subset of A. An element b is called
an upper bound for the set X if every element in X is less than or
equal to b. If such an upper bound exists, the set X is called bounded
above.
Let A be an ordered set, and X a subset of A. An element b in A is
called a least upper bound (or supremum) for X if b is an upper
bound for X and there is no other upper bound b' for X that is less
than b. We write b = sup(X). By its definition, if a least upper bound
exists, it is unique.
Examples 2.4.3:
In fact, this last fact is exactly the property which distinguishes the real numbers from the
other number systems, and makes it useful to us. We will state this as a theorem:
Theorem 2.4.4: Least Upper Bound Property
There exists an ordered field of numbers R with the properties:
1. it contains all rational numbers
2. it has the property that any non-empty subset which has an
upper bound has a least upper bound.
Proof
Note that we have not defined 'ordered field', but that is not so important for us right now.
The importance for us is that this property is one of the most basic properties of the real
numbers, and it distinguishes the real from the rational numbers (which do not have this
property).
In order to prove this theorem we need to know what exactly the real numbers are, and we
have indeed given two possible constructions at the beginning of this section.. However, it
is more important to understand these properties of R, and to know about the differences
between R and the other number systems N, Z, and Q.
We can use this theorem to illustrate another property of the real numbers that makes them
more useful than the rational numbers:
Theorem 2.4.5: Square Roots inR
There is a positive real number x such that x2 = 2
Proof
There are several other properties that will be of importance later on. Two of those are the
Archimedean and the Density property. Again, as for the Least Upper Bound property, it is
more important to understand what these properties mean than to follow the proof exactly.
Theorem 2.4.6: Properties ofR and Q
The set of real numbers satisfies the Archimedean Property:
Let a and b be positive real numbers. Then there is a natural
number n such that n * a > b
The set of rational numbers satisfies the following
Density Property:
Let c < d be real numbers. Then there is a rational number q
with c < q < d.
Proof
This concludes the 'elementary' part of this text. We have now defined much of our basic
notation, learned how to count to infinity, introduced all number systems that we will be
using, and we have seen several different types of proofs. We are now ready for some more
complicated topics.
IRA
Pi and e (Euler's number) are two numbers that occur everywhere in mathematics. Both are
irrational and in fact transcendental numbers. In this little appendix we will prove that both
numbers are irrational. We will not show that they are also transcendental numbers. The
proofs will actually use tools from much later sections, but they might be interesting
nevertheless, and the results fit well in this section.
First will now prove that e (Euler's number) is irrational. In fact, it is a transcendental
number, but that requires a lot of work to prove.
Theorem 2.5.3: e is irrational
Euler's number e is irrational
Proof: We know that for any integer n
(check the definition of Euler's number and the definition of Euler' series) where
Now suppose that e was rational, i.e. there are two positive integers a and b with e = a / b.
Choose n > b. Then, using the above representation, we have:
Since n > b, the left side of this equation represents an integer, and hence n! Rn is also an
integer. But we know that
so that if n is large enough the left side is less than 1. But then n! Rn must be a positive
integer less than 1, which is a contradiction.
Next we prove that Pi is irrational, which is a lot harder to do. We need a preliminary
result:
Lemma 2.5.1:
Define a function fn(x) =
. Then this function has the
following properties:
0 < fn(x) < 1 / n! for 0 < x < 1
and
Proof: Using the binomial theorem, we see that when the numerator of the function is
multiplied out, the lowest power of x will be n, and the highest power is 2n. Therefore, the
function can be written as
fn(x) =
where all coefficients are integers. It is clear from this expression that
n and for k > 2n.
Also, looking at the sum more carefully, we see that
= 0 for k <
...
we also have
Therefore
is an integer, by assumption on
. We also know that
well. Hence, G(0) and G(1) are both integers.
Differentiating G(x) twice, we get:
and
are integers as
The last term in this equation, the (2n+2)-th derivative, is zero, by the properties of the
function f. Adding G(x) and G''(x) we get:
G''(x), we have:
=
=
Now we can use the second fundamental theorem of calculus to
conclude:
= H(1) - H(0)
=
=
[ G(1) + G(0)]
is an integer. But we also know that 0 < fn(x) < 1 / n! for 0 < x < 1. Therefore, estimating
the above integral, we get
0<
<
0<
<
<1
Here we have used the fact that the last fraction approaches zero if n is large enough. But
know we have a contradiction, because that integral was supposed to be an integer. Since
there is no positive integer less than 1, our assumption that
contradiction. Hence,
must be irrational.
This prove is, admittedly, rather curious. Aside from the assumption that
Pi is rational (leading to the contradiction) the only other properties of Pi
that are really involved in this proof are that
sin( ) = 0, cos( ) = 1
and we need the defining properties of the trig. functions that
sin'(x) = cos(x), cos'(x) = - sin(x), sin(0) = 0, cos(0) = 1
That does not make the proof much clearer, but it illustrates that some essential properties
of are indeed used along the way, and this proof will not work for any other number.
Taken from Calculus (2nd Edition), by Michael Spivak
Publish or Perish, Inc, 1980, pages 307 - 310
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
3.1. Sequences
So far we have introduced sets as well as the number systems that we will use in this text.
Next, we will study sequences of numbers. Sequences are, basically, countably many
numbers arranged in an order that may or may not exhibit certain patterns. Here is the
formal definition of a sequence:
Definition 3.3.1: Sequence
A sequence of real numbers is a function f: N R. In other words, a
sequence can be written as f(1), f(2), f(3), ..... Usually, we will denote
such a sequence by the symbol
, where aj = f(j).
For example, the sequence 1, 1/2, 1/3, 1/4, 1/5, ... is written as
. Keep in mind that
despite the strange notation, a sequence can be thought of as an ordinary function. In many
cases that may not be the most expedient way to look at the situation. It is often easier to
simply look at a sequence as a 'list' of numbers that may or may not exhibit a certain
pattern.
We now want to describe what the long-term behavior, or pattern, of a
sequence is, if any.
Definition 3.1.2: Convergence
A sequence
of real (or complex) numbers is said to converge
to a real (or complex) number c if for every > 0 there is an integer
N > 0 such that if j > N then
| aj - c | <
and we
Example 3.1.3:
. It converges to zero.
does not converge. Prove it.
The sequence
converges to zero Prove it.
Convergent sequences, in other words, exhibit the behavior that they get closer and closer
to a particular number. Note, however, that divergent sequence can also have a regular
pattern, as in the second example above. But it is convergent sequences that will be
particularly useful to us right now.
Example 3.1.5:
The Fibonacci numbers are recursively defined as x1 = 1,
x2 = 1, and for all n > 2 we set xn = xn - 1 + xn - 2. Show that the
sequence of Fibonacci numbers {1, 1, 2, 3, 5, ...} does not
converge.
Convergent sequences can be manipulated on a term by term basis, just as one would
expect:
Proposition 3.1.6: Algebra on Convergent Sequences
Suppose
and
are converging to a and b, respectively.
Then
1. Their sum is convergent to a + b, and the sequences can be
added term by term.
2. Their product is convergent to a * b, and the sequences can be
multiplied term by term.
3. Their quotient is convergent to a / b, provide that b # 0, and
the sequences can be divided term by term (if the
denominators are not zero).
4. If an bn for all n, then a b
Proof
This theorem states exactly what you would expect to be true. The proof
of it employs the standard trick of 'adding zero' and using the triangle
inequality. Try to prove it on your own before looking it up.
Note that the fourth statement is no longer true for strict inequalities. In
other words, there are convergent sequences with an < bn for all n, but
strict inequality is no longer true for their limits. Can you find an
example ?
While we now know how to deal with convergent sequences, we still
need an easy criteria that will tell us whether a sequence converges.
The next proposition gives reasonable easy conditions, but will not tell
us the actual limit of the convergent sequence.
First, recall the following definitions:
Definition 3.1.7: Monotonicity
A sequence
j.
A sequence
aj aj + 1 for all j.
In other words, if every next member of a sequence is larger than the previous one, the
sequence is growing, or monotone increasing. If the next element is smaller than each
previous one, the sequence is decreasing. While this condition is easy to understand, there
are equivalent conditions that are often easier to check:
Monotone increasing:
1. aj + 1 aj
2. aj + 1 - aj 0
3. aj + 1 / aj 1, if aj > 0
Monotone decreasing:
1. aj + 1 aj
2. aj + 1 - aj 0
3. aj + 1 / aj 1, if aj > 0
Examples 3.1.8:
Is the sequence
decreasing ?
Is the sequence
decreasing ?
monotone increasing or
monotone increasing or
If a sequence
is bounded above, then c = sup(xk) is finite. Moreover, given
any > 0, there exists at least one integer k such that xk > c - , as illustrated in the
picture.
If a sequence
is bounded below, then c = inf(xk) is finite. Moreover, given
any > 0, there exists at least one integer k such that xk < c + , as illustrated in the
picture.
If
is a monotone increasing sequence that is bounded above,
then the sequence must converge.
If
is a monotone decreasing sequence that is
bounded below, then the sequence must converge.
Proof
Example 3.1.12:
Show that the sequence sin(n) / n and cos(n) / n both converge to
zero.
What is slightly annoying for the mathematician (in theory and in praxis) is that we refer to
the limit of a sequence in the definition of a convergent sequence when that limit may not
be known at all. In fact, more often then not it is quite hard to determine the actual limit of
a sequence.
We would prefer to have a definition which only includes the known
elements of the particular sequence in question and does not rely on the
unknown limit. Therefore, we will introduce the following definition:
Definition 3.2.1: Cauchy Sequence
Let
be a sequence of real (or complex) numbers. We say that
the sequence satisfies the Cauchy criterion (or simply is Cauchy) if
for each > 0 there is an integer N > 0 such that if j, k > N then
| a j - ak | <
This definition states precisely what it means for the elements of a sequence to get closer
together, and to stay close together. Of course, we want to know what the relation between
Cauchy sequences and convergent sequences is.
Theorem 3.2.2: Completeness Theorem in R
Let
be a Cauchy sequence of real numbers. Then the
sequence is bounded.
Let
be a sequence of real numbers. The
sequence is Cauchy if and only if it converges to some
limit a.
Proof
IRA
3.3. Subsequences
So far we have learned the basic definitions of a sequence (a function from the natural
numbers to the Reals), the concept of convergence, and we have extended that concept to
one which does not pre-suppose the unknown limit of a sequence (Cauchy sequence).
Unfortunately, however, not all sequences converge. We will now
introduce some techniques for dealing with those sequences. The first is
to change the sequence into a convergent one (extract subsequences)
and the second is to modify our concept of limit (lim sup and lim inf).
Definition 3.3.1: Subsequence
Let
be a sequence. When we extract from this sequence only
certain elements and drop the remaining ones we obtain a new
sequences consisting of an infinite subset of the original sequence.
That sequence is called a subsequence and denoted by
One can extract infinitely many subsequences from any given sequence.
Examples 3.3.2:
If
is a convergent sequence, then every subsequence of that
sequence converges to the same limit
If
is a sequence such that every possible
subsequence extracted from that sequences converge
to the same limit, then the original sequence also
converges to that limit.
Proof
exists a subsequence
that converges.
Proof
Example 3.3.5:
Does
converge ? Does there exist a convergent
subsequence ? What is that subsequence ?
sequence
that converges to L. This is difficult to
prove.
Next, we will broaden our concept of limits.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
While we know how to deal with convergent sequences, we don't know much about
divergent sequences. One possibility is to try and extract a convergent subsequence, as
described in the last section. In particular, Bolzano-Weierstrass' theorem can be useful in
case the original sequence was bounded. However, we often would like to discuss the limit
of a sequence without having to spend much time on investigating convergence, or
thinking about which subsequence to extract. Therefore, we need to broaden our concept of
limits to allow for the possibility of divergent sequences.
Definition 3.4.1: Lim Sup and Lim Inf
Let
and let c = lim (Aj). Then c is called the limit inferior of the sequence
.
Let
and let c = lim (Bj). Then c is called the limit superior of the
sequence
.
In short, we have:
1. lim inf(aj) = lim(Aj) , where Aj = inf{aj , aj + 1 , aj + 2 , ...}
2. lim sup(aj) = lim(Bj) , where Bj = sup{aj , aj + 1 , aj + 2 , ...}
When trying to find lim sup and lim inf for a given sequence, it is best to find the first few
Aj's or Bj's, respectively, and then to determine the limit of those. If you try to guess the
answer quickly, you might get confused between an ordinary supremum and the lim sup, or
the regular infimum and the lim inf.
Examples 3.4.2:
?
?
lim sup and lim inf always exist (possibly infinite) for any sequence
of real numbers.
Proof
lim sup
If c and d are both finite, then: given any > 0 there are arbitrary
large j such that aj > c - and arbitrary large k such that ak < d +
Proof
IRA
Proof:
This seems an obvious statement: if a number is in absolute values less than one, it gets
smaller and smaller when raised to higher and higher powers. Proving something 'obvious',
however, is often difficult, because it may not be clear how to start. To prove the statement
we have to resort to one of the elementary properties of the real number system: the
Archimedian principle.
Case a > 1:
Take any real number K > 0 and define
x=a-1
Since a > 1 we know that x > 0. By the Archimedian principle there exists a
positive integer n such that nx > K - 1. Using Bernoulli's inequality for that n we
have:
an = (1 + x)n 1 + nx > 1 + (K - 1) = K
But since K was an arbitrary number, this proves that the sequence {an} is
unbounded. Hence it can not converge.
Case 0 < a < 1:
Take any > 0. Since 0 < a < 1 we know that 1/a > 1, so that by the previous proof
we can find an N with
Proof:
Write n a = e a ln(n). Then:
if a > 0 then as n approaches infinity, the function ea ln(n) approaches infinity as well
if a < 0 then as n approaches infinity, the function ea ln(n) approaches zero
if a = 0 then the sequence is the constant sequence, and hence convergent
Is this a good proof ? Of course not, because at this stage we know nothing about the
exponential or logarithm function. So, we should come up with a better proof.
But actually we first need to understand exactly what na really means:
If a is an integer, then clearly na means to multiply n with itself a times
If a = p/q is a rational number, then np/q means to multiply n p times, then take the
q-root
It is unclear, however, what na means if a is an irrational number. For example,
what is n , or n ?
One way to define na for all a is to resort to the exponential function:
na = ea ln(n)
In that light, the original 'proof' was not bad after all, but of course we now need to know
how exactly the exponential function is defined, and what its properties are before we can
continue. As it turns out, the exponential function is not easy to define properly. Generally
one can either define it as the inverse function of the ln function, or via a power series.
Another way to define na for a being rational it to take a sequence of
rational numbers rn converging to a and to define na as the limit of the
sequence {nrn}. There the problem is to show that this is well-defined,
i.e. if there are two sequences of rational numbers, the resulting limit
will be the same.
In either case, we will base our proof on the simple fact:
if p > 0 and x > y then xp > yp
which seems clear enough and could be formally proved as soon as the exponential
function is introduced to properly define xp.
Now take any positive number K and let n be an integer bigger than K1/a.
Since
n > K1/a
we can raise both sides to the a-th power to get
na > K
which means - since K was arbitrary - that the sequence {na} is unbounded.
The case a = 0 is clear (since n0 = 1). The second case of a < 0 is
related to the first by taking reciprocals (details are left as an exercise).
Since we have already proved the first case we are done.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
If n > 1, then
0 an
But that implies that an converges to zero as n approaches to infinity, which
means, by the definition of an that
converges to 1 as n goes to infinity.
That is what we wanted to prove.
Proof:
Case a > 1:
If a > 1, then for n large enough we have 1 < a < n. Taking roots on both sides we
obtain
1<
<
But the right-hand side approaches 1 as n goes to infinity by our statement of the
root-n sequence. Then the sequence {
} must also approach 1, being squeezed
between 1 on both sides (Pinching theorem).
Case 0 < a < 1:
If 0 < a < 1, then (1/a) > 1. Using the first part of this proof, the reciprocal of the
sequence {
} must converge to one, which implies the same for the original
sequence.
Incidentally, if a = 0 then we are dealing with the constant sequence, and the limit is of
course equal to 0.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Note that both numerator and denominator tend to infinity. Our goal will be to show that
the denominator grows faster than the k-th power of n, thereby 'winning' the race to infinity
and forcing the whole expression to tend to zero.
The name of this sequence indicates that we might try to use the
binomial theorem for this. Indeed, define x such that
b=1+x
Since b > 1 we know that x > 0. Therefore, each term in the binomial theorem is positive,
and we can use the (k+1)-st term of that theorem to estimate:
But this expression is true for all n > 2k + 1 as well, so that, with k fixed, we can take the
limit as n approaches infinity and the right hand side will approach zero. Since the lefthand side is always greater than or equal to zero, the limit of the binomial sequence must
also be zero.
If k is replaced by any real number, see the exponential sequence to
find out how nk could be defined for rational and irrational values of k.
But perhaps some simple estimate will help if k is not an integer. Details
are left as an exercise.
Proof:
We will show that the sequence is monotone increasing and bounded above. If that was
true, then it must converge. Its limit, by definition, will be called e for Euler's number.
Euler's number e is irrational (in fact transcendental), and an
approximation of e to 30 decimals is e ~
2.71828182845904523536028747135.
First, we can use the binomial theorem to expand the expression
The first expression has (n+1) terms, the second expression has (n+2) terms. Each of the
first (n+1) terms of the second expression is greater than or equal to each of the (n+1)
terms of the first expression, because
But then the sequence is monotone increasing, because we have shown that
Next, we need to show that the sequence is bounded. Again, consider the expansion
1+
so that, finally,
for all n.
But then, putting everything together, we have shown that
1+
1 + Sn 3
Proof:
We will use a simple substitution to prove this. Let
x/n = 1/u, or equivalently, n = u x
Then we have
But the term inside the square brackets is Euler's sequence, which converges to Euler's
number e. Hence, the whole expression converges to ex, as required.
In fact, we have used a property relating to functions to make this proof
work correctly. What is that property ?
If we did not want to use functions, we could first prove the statement
for x being an integer. Then we could expand it to rational numbers, and
then, approximating x by rational number, we could prove the final
result.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
So far we have learned about sequences of numbers. Now we will investigate what may
happen when we add all terms of a sequence together to form what will be called an
infinite series.
The old Greeks already wondered about this, and actually did not have
the tools to quite understand it This is illustrated by the old tale of
Achilles and the Tortoise.
Example 4.1.1: Zeno's Paradox (Achilles and the Tortoise)
This continuous for a while, but whenever Achilles manages to reach the spot where the
tortoise has just been a split-second ago, the tortoise has again covered a little bit of
distance, and is still ahead of Achilles. Hence, as hard as he tries, Achilles only manages to
cut the remaining distance in half each time, implying, of course, that Achilles can actually
never reach the tortoise. So, the tortoise wins the race, which does not make Achilles very
happy at all.
Obviously, this is not true, but where is the mistake ?
Now let's return to mathematics. Before we can deal with any new
objects, we need to define them:
Definition 4.1.2: Series, Partial Sums, and Convergence
Let { a n } be an infinite sequence.
1. The formal expression
A series
values
Let
be an absolutely convergent series. Then any
rearrangement of terms in that series results in a new series that is
also absolutely convergent to the same limit.
Let
be a conditionally convergent series. Then, for
any real number c there is a rearrangement of the
series such that the new resulting series will converge
to c.
Proof
and
be two absolutely convergent series. Then:
1. The sum of the two series is again absolutely convergent. Its
limit is the sum of the limit of the two series.
2. The difference of the two series is again absolutely
convergent. Its limit is the difference of the limit of the two
series.
3. The product of the two series is again absolutely convergent.
Its limit is the product of the limit of the two series (Cauchy
Product).
Proof
We will give one more rather abstract result on series before stating and proving easy to
use convergence criteria. The one result that is of more theoretical importance is
Theorem 4.1.9: Cauchy Criteria for Series
The series
converges if and only if for every > 0 there is a
positive integer N such that if m > n > N then
|<
Proof
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
In this section we will list many of the better known tests for convergence or divergence of
series, complete with proofs and examples. You should memorize each and every one of
those tests. The most useful tests are marked with a start (*). Click on the question marks
below to learn more about that particular test.
Divergence Test (*)
Comparison Test
Limit Comparison Test (*)
Cauchy Condensation Test
Geometric Series Test (*)
p Series Test (*)
Root Test
Ratio Test (*)
Abel's Convergence Test
Alternating Series Test (*)
Integral Test
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
This is a brief listing of some of the most important series. Each of the convergence tests
given in the previous section contains even more examples.
Geometric Series
converges absolutely for |a| < 1 to 1 / 1-a, diverges otherwise
p Series
converges for p > 1, diverges otherwise
Harmonic Series
diverges
Alternating Harmonic Series
converges conditionally
Euler's Series
converges to e (Euler's number)
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Divergence Test
If the series
Equivalently:
If the sequence
series
converges to zero.
This test can never be used to show that a series converges. It can only be used to show
that a series diverges. Hence, the second version of this theorem is the more important,
applicable statement.
Examples 4.2.2:
Does the Divergence test apply to show that the series
Proof:
Suppose the series does converge. Then it must satisfy the Cauchy
criterion. In other words, given any > 0 there exists a positive integer
N such that whenever n > m > N then
|<
converges to zero.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Comparison Test
Suppose that
is a sequence of
If the series
converges to positive infinity, and
is
a sequence of numbers for which an bn for all n > N. Then
the series
also diverges.
Context
This is a useful test, but the limit comparison test, which is rather similar, is a much easier
to use, and therefore more useful. However, this comparison test is very easy to memorize:
Assuming that everything is positive, for simplicity, say we know that:
|bn| |an|
for all n. then just sum both sides to see what you get formally:
Then:
Does
Does
converge or diverge ?
converge or diverge ?
Proof:
The proof, at first glance, seems easy: Suppose that
converges
absolutely, and | b n | | a n | for all n. For simplicity, assume that all
terms in both sequences are positive. Let
SN=
and T N =
converges, it satisfies
|<
| |
|<
Suppose
and
r = lim | a n / b n |
Context
This test is more useful than the "direct" comparison test because you
do not need to compare the terms of two series too carefully. It is
sufficient if the two terms behave similar "in the long run".
Examples 4.2.6:
Use the limit convergence test to decide whether the
following series converge or diverge. Note that you need to
know convergence of the p-series .
converge or diverge ?
converge of diverge ?
if n > N. Assume
c | b n | < | an |
|a n | < C | b n |
Suppose
and
r = lim | a n / b n |
Context
This test is more useful than the "direct" comparison test because you
do not need to compare the terms of two series too carefully. It is
sufficient if the two terms behave similar "in the long run".
Examples 4.2.6:
Use the limit convergence test to decide whether the
following series converge or diverge. Note that you need to
know convergence of the p-series .
converge or diverge ?
converge of diverge ?
if n > N. Assume
c | b n | < | an |
|a n | < C | b n |
converges.
Context
This test is rather specialized, just as Abel's Convergence Test. The main
purpose of the Cauchy Condensation test is to prove that the p-series
converges if p > 1.
Example 4.2.8:
converges: We have
Now the partial sums on the right are bounded, by assumption. Hence the partial sums on
the left are also bounded. Since all terms are positive, the partial sums now form an
increasing sequence that is bounded above, hence it must converge. Multiplying the left
sequence by 2 will not change convergence, and hence the series
Assume that
converges: We have
converges.
But now the sequence of partial sums on the right is bounded, by assumption. Therefore,
the left side forms an increasing sequence that is bounded above, and therefore must
converge.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
=
Context
Note that the index for the geometric series starts at 0. This is not
important for the convergence behavior, but it is important for the
resulting limit.
Examples 4.2.10:
?
?
and
The series
is called a p Series.
if p > 1 the p-series converges
if p 1 the p-series diverges
Context
Examples 4.2.12:
converge or diverge ?
converge or diverge ?
Proof:
If p < 0 then the sequence
converges to infinity. Hence, the
series diverges by the Divergence Test.
If p > 0 then consider the series
=
The right hand series is now a Geometric Series.
if 0 < p 1 then 2 1-p 1, hence the right-hand series diverges
if 1 < p then 2 1-p < 1, hence the right-hand series converges
Now the result follows from the Cauchy Condensation test .
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Root Test
Compare this test with the Ratio test. Although this root test is more difficult to apply, it is
better than the ratio test in the following sense: there are series for which the ratio test give
no information, yet the root test will be conclusive. You can also use the root test to prove
the ratio test , but not visa versa.
It is important to remember that when the root test gives 1 as the
answer for the lim sup, then no conclusion at all is possible.
The use of the lim sup rather than the regular limit has the advantage
that we do not have to be concerned with the existence of a limit. On
the other hand, if the regular limit exists, it is the same as the lim sup,
so that we are not giving up anything using the lim sup.
Examples 4.2.14:
converge or diverge ?
Proof:
Assume that lim sup | a n | 1/n < 1: Because of the properties of the limit
superior, we know that there exists > 0 and N > 1 such that
| a n | 1/n < 1 for n > N. Raising both sides to the n-th power we have:
| a n | < (1 - ) n
for n > N. But the terms on the right hand side form a convergent geometric series. Hence,
by the comparison test the series with terms on the left-hand side will converge absolutely.
The proof for the second case if left as an exercise.
Ratio Test
The ratio test is easier to use than the Root test . However, there are series for which the
ratio tests gives no information, but the root test will. In that sense, the ratio test is weaker
than the root test. In addition, the ratio test can be proved using the root test, but not visa
versa.
Using the lim sup rather than the regular limit has the advantage that
we don't have to worry about existence of the limit. However, if the
regular limit exists, the lim sup yields the same number. Therefore, we
loose nothing by looking at the limit superior.
Example 4.2.16:
converge ?
Proof:
The proof is very similar to the proof of the root test:
Assume that lim sup | a n+1 / a n | < 1: because of the properties of the
lim sup, we know that there exists > 0 and N > 1 such that
| a n+1 / a n | < 1 for n > N. Multiplying both sides by a n we obtain
. Suppose that
is decreasing
converges.
Context
This test is rather sophisticated. Its main application is to prove the Alternating Series test,
but one can sometimes use it for other series as well, if the more obvious tests do not work.
Examples 4.2.18:
converge or diverge ?
converge or diverge ?
Proof:
First, we need a lemma, called the Summation by Parts Lemma:
Lemma: Summation by Parts
Proof
Assuming this lemma is proved, we will use it as follows for Abel's Test:
First, let's assume that the partial sums S N are bounded by, say, K.
Next, since the sequence
converges to zero, we can choose an
integer N such that | b n | < / 2K. Using the Summation by Parts lemma,
we then have:
But by our choice of N this is less than if we choose n and m larger than the
predetermined N. This proves Abel's Test.
What remains to do is the proof of the lemma, which can be found here.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
This test does not prove absolute convergence. In fact, when checking for absolute
convergence the term 'alternating series' is meaningless.
It is important that the series truly alternates, that is each positive term
is followed by a negative one, and visa versa. If that is not the case, the
alternating series test does not apply (while Abel's Test may still work).
Examples 4.2.20:
converge or diverge ?
Proof:
Let a n = (-1) n. Then the formal sum
has bounded partial sums
(although the sum does not itself converge. Why not ?). Then, with the
given choice of
indeed converges.
Integral Test
Suppose that f(x) is positive, continuous, decreasing function on the interval
[N, ). Let a n = f(n). Then
converges
Context
Note that this test is much different from all the others. We have not yet formally
introduced the concept of an Integral - or even of a continuous function - so that we can not
prove this test here. However, for completeness it is included as a test that is sometimes
useful to apply.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
In the previous chapters we dealt with collections of points: sequences and series. Each
time, the collection of points was either finite or countable.and the most important property
of a point, in a sense, was its location in some coordinate or number system. Now will deal
with points, or more precisely with sets of points, in a more abstract setting. The location
of points relative to each other will be more important than their absolute location, or size,
in a coordinate system. Therefore, concepts such as addition and multiplication will not
work anymore, and we will have to start, in a sense, at the beginning again.
All of the previous sections were, in effect, based on the natural
numbers. Those numbers were postulated as existing and all other
properties - including other number systems - were deduced from those
numbers and a few principles of logic.
We will now proceed in a similar way: first, we need to define the basic
objects we want to deal with, together with their most elementary
properties. Then we will develop a theory of those objects and called it
topology.
Definition 5.1.1: Open and Closed Sets
A set U R is called open, if for each x U there exists and > 0
such that the interval ( x - , x + ) is contained in U. Such an
interval is often called an - neighborhood of x, or simply a
neighborhood of x.
A set F is called closed if the complement of F, R \ F,
is open.
Examples 5.1.2:
The set {1, 1/2, 1/3, 1/4, 1/5, ...} and {1, 1/2, 1/3,
1/4, ...} {0}
It is fairly clear that when combining two open sets (either via union or intersection) the
resulting set is again open, and the same statement should be true for closed sets. What
about combining infinitely many sets ?
Proposition 5.1.3: Unions of Open Sets, Intersections of Closed Sets
How complicated can an open or closed set really be ? The basic open
(or closed) sets in the real line are the intervals, and they are certainly
not complicated. As it will turn out, open sets in the real line are
generally easy, while closed sets can be very complicated.
The worst-case scenario for the open sets, in fact, will be given in the
next result, and we will concentrate on closed sets for much of the rest
of this chapter.
Proposition 5.1.4: Characterizing Open Sets
Let U R be an arbitrary open set. Then there are countably many
pairwise disjoint open intervals Un such that U = Un
Proof
Next we need to establish some relationship between topology and our previous studies, in
particular sequences of real numbers. We shall need the following definitions:
Definition 5.1.5: Boundary, Accumulation, Interior, and Isolated
Points
Let S be an arbitrary set in the real line R.
1. A point b R is called boundary point of S if every nonempty neighborhood of b intersects S and the complement of
S. The set of all boundary points of S is called the boundary of
S, denoted by bd(S).
2. A point s S is called interior point of S if there exists a
neighborhood of S completely contained in S. The set of all
interior points of S is called the interior, denoted by int(S).
3. A point t S is called isolated point of S if there exists a
neighborhood U of t such that U S = {t}.
4. A point r S is called accumulation point, if every
neighborhood of r contains infinitely many distinct points of
S.
Examples 5.1.6:
What is the boundary and the interior of (0, 4), [-1, 2], R,
and O ? Which points are isolated and accumulation points, if
any ?
Finally, here is a theorem that relates these topological concepts with our previous notion
of sequences.
Theorem 5.1.8: Closed Sets, Accumulation Points, and Sequences
is any sequence in S,
of
that
Proof
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
The above definition of compact sets using sequence can not be used in more abstract
situations. We would also like a characterization of compact sets based entirely on open
sets. We need some definitions first.
Definition 5.2.4: Open Cover
Let S be a set of real numbers. An open cover of S is a collection C of
open sets such that S
C. The collection C of open sets is said to
cover the set S.
A subset of sets from the collection C that still covers
the set S is called a subcovering of S.
Examples 5.2.5:
Let S = [0, 1], and C = { (-1/2, 1/2), (1/3, 2/3), (1/2, 3/2)}.
Is C an open cover for S ?
Let S = [0, 1]. Define
= { t R : | t - | < and
S}
for a fixed > 0. Is the collection of all {
},
S, an open
cover for S ? How many sets of type
are actually needed to
cover S ?
Let S = (0, 1). Define a collection C = { (1/j, 1), for all j >
0 }. Is C an open cover for S ? How many sets from the
collection C are actually needed to cover S ?
Here is the characterization of compact sets based only on open sets:
Theorem 5.2.6: Heine-Borel Theorem
Proof
Compact sets share many properties with finite sets. For example, if A and B are two nonempty sets with A B then A B # 0. That is, in fact, true for finitely many sets as well,
but fails to be true for infinitely many sets.
Examples 5.2.7:
Consider the collection of sets (0, 1/j) for all j > 0. What is
the intersection of all of these sets ?
Can you find infinitely many closed sets such that their
intersection is empty and such that each set is contained in its
predecessor ? That is, can you find sets Aj such that Aj+1 Aj
and Aj = 0 ?
Compact sets, on the other hand, have the following nice property, which will be used in
some of the following chapters:
Proposition 5.2.8: Intersection of Nested Compact Sets
Find a perfect set. Find a closed set that is not perfect. Find
a compact set that is not perfect. Find an unbounded closed set
that is not perfect. Find a closed set that is neither compact
nor perfect.
Is the set {1, 1/2, 1/3, ...} perfect ? How about the set {1,
1/2, 1/3, ...} {0} ?
As an application of the above result, we will see that perfect sets are closed sets that
contain lots of points:
Proposition 5.2.11: Perfect sets are Uncountable
This can yield a quick, but rather sophisticated proof of the fact that the
interval [a, b] is uncountable: the interval [a, b] is a perfect set, hence,
it must be uncountable.
Another, rather peculiar example of a closed, compact, and perfect set
is the Cantor set.
Definition 5.2.12: Cantor Middle Third Set
Start with the unit interval
S0 = [0, 1]
Remove from that set the middle third and set
S1 = S0 \ (1/3, 2/3)
Remove from that set the two middle thirds and set
S2 = S1 \ { (1/9, 2/9)
(7/9, 8/9) }
Show that the Cantor set does not contain any open set
Think about this set. It seems surprising that
a set of length zero can contain uncountably many points.
a perfect set does not have to contain an open set
Therefore, the Cantor set shows that closed subsets of the real line can be more
complicated than intuition might at first suggest. It is in fact often used to construct
difficult, counter-intuitive objects in analysis.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
In the last two section we have classified the open sets, and looked at two classes of closed
set: the compact and the perfect sets. In this section we will introduce two other classes of
sets: connected and disconnected sets.
Definition 5.3.1: Connected and Disconnected
An open set S is called disconnected if there are two open, nonempty sets U and V such that:
1. U V = 0
2. U V = S
A set S (not necessarily open) is called disconnected
if there are two open sets U and V such that
1. (U S) # 0 and (V S) # 0
2. (U S) (V S) = 0
3. (U S) (V S) = S
If S is not disconnected it is called connected.
Note that the definition of disconnected set is easier for an open set S. In principle,
however, the idea is the same: If a set S can be separated into two open, disjoint sets in
such a way that neither set is empty and both sets combined give the original set S, then S
is called disconnected.
To show that a set is disconnected is generally easier than showing
connectedness: if you can find a point that is not in the set S, then that
point can often be used to 'disconnect' your set into two new open sets
with the above properties.
Examples 5.3.2:
Hence, as with open and closed sets, one of these two groups of sets
are easy:
open sets in R are the union of disjoint open intervals
connected sets in R are intervals
The other group is the complicated one:
closed sets are more difficult than open sets (e.g. Cantor set)
disconnected sets are more difficult than connected ones (e.g. Cantor set)
In fact, a set can be disconnected at every point.
Definition 5.3.4: Totally Disconnected
6.1. Limits
IRA
We now want to combine some of the concepts that we have introduced before: functions,
sequences, and topology. In particular, if we have some function f(x) and a given sequence
{ an }, then we can apply the function to each element of the sequence, resulting in a new
sequence. What we would want is that if the original sequence converges to some number
L, then the new sequence { f( an )} should converge to f(L), and if the original sequence
diverges, the new one should diverge also. This seems not too much to ask for, but is quite
simple minded.
Example 6.1.1:
f(x) = L
Examples 6.1.3:
if x
D and | x - c | <
Example 6.1.5:
Consider the function f with f(x) = 1 if x is rational and f(x)
= 0 if x is irrational. Does the limit of f(x) exist at an arbitrary
number x ?
Regardless of which of the two definitions might be considered easier to use in a particular
situation, the basic problem right now is that we have two different definitions for the same
concept. We therefore have to show that both definitions are actually equivalent to each
other.
Proposition 6.1.6: Equivalence of Definitions of Limits
In other words, both definitions of continuity are equivalent, and we can use which ever
seems the easiest. Here are some basic properties of limits of functions.
Proposition 6.1.7: Properties for limits of Functions
If
f(x) and
and
g(x) exist.
[ f(x) g(x) ] =
f(x) +
f(x)
f(x)
g(x) exist.
[ f(x) / g(x) ] =
f(x) and
f(x) /
Sometimes a function may not have a limit using the above definitions, but when the
domain of the function is restricted, then a limit exists. This leads to the concepts of onesided limits.
Definition 6.1.8: One-Sided Limits of a Function
If f is a function with domain D and c closure(D). Then:
1. f has a left-hand limit L at c if for every > 0 there exists
> 0 such that if x D and c - < x < c then | f(x) - L | < .
We write
f(x) = L.
f(x) = L and
f(x) = L
Proof
Now that we have some idea about limits of functions, we will move to the next question:
if some sequence converges to c, and the function converges to L as x approaches c, then
when is it true that f(c) = L ? This will be the contents of the next section, continuity.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
IRA
If one looks up continuity in a thesaurus, one finds synonyms like perpetuity or lack of
interruption. Descartes said that a function is continuous if its graph can be drawn without
lifting the pencil from the paper.
Example 6.2.1:
f(x) = f(c).
If
f(x) = f(c) for every sequence { xn } of points in
D converging to c, then f is continuous at the point c.
Proof
Again, as with limits, this proposition gives us two equivalent mathematical conditions for
a function to be continuous, and either one can be used in a particular situation.
Example 6.2.4:
Continuous functions can be added, multiplied, divided, and composed with one another
and yield again continuous functions.
Proposition 6.2.5: Algebra with Continuous Functions
While this proposition seems not very important, it can be used to quickly prove the
following:
Examples 6.2.6:
IRA
1.
2.
3.
4.
Which of these functions, without proof, has a 'fake'
discontinuity, a 'regular' discontinuity, or a 'difficult'
discontinuity ?
Of course, we need some mathematical description of the various types of discontinuities
that a function could have.
Definition 6.3.3: Classification of Discontinuities
D is a point of
2. if
f(x) does not exist, but both
f(x) and
f(x)
exit, then c is called a discontinuity of the first kind , or
jump discontinuity
3. if either
f(x) or
f(x) does not exist, then c is called
a discontinuity of the second kind, or essential discontinuity.
Examples 6.3.4:
x=0
4.
This theorem also states that if a function wants to have a discontinuity of the second kind
at a point x = c, then it can not be monotone in any neighborhood of c.
Corollary 6.3.7: Discontinuities of Second Kind
If f has a discontinuity of the second kind at x = c, then f must change
from increasing to decreasing in every neighborhood of c.
Proof
In other words, f must look pretty bad if it has a discontinuity of the second kind.
Examples 6.3.8:
IRA
While the definition of continuity suffices for functions on the real line, there are other,
more abstract spaces for which this definition will not work. In particular, our continuity
definitions relies on the presence of an absolute value. There are spaces which do not have
such a distance function, yet we still might want to study continuous functions on those
abstract spaces. In this section we will investigate some topological properties of
continuity which will, in fact, apply equally well to more general settings. In addition, this
section will contain several important theoretical results on continuous function on the real
line.
Proposition 6.4.1: Continuity and Topology
Let f be a function with domain D in R. Then the following
statements are equivalent:
1. f is continuous
2. If D is open, then the inverse image of every open set under f
is again open.
3. If D is open, then the inverse image of every open interval
under f is again open.
4. If D is closed, then the inverse image of every closed set
under f is again closed.
5. If D is closed, then the inverse image of every closed interval
under f is again closed.
6. The inverse image of every open set under f is the intersection
of D with an open set.
7. The inverse image of every closed set under f is the
intersection of D with a closed set.
Proof
This proposition can be used to prove that a function is continuous, and is especially nice if
the domain of the function is either open or closed. This is true in particular for function
defined on all of R (which is both open and closed).
Examples 6.4.2:
Since compact sets in the real line are characterized by being closed and bounded, we
should note while it is not true that the image of a closed set is closed, one must look at an
unbounded closed set for a counterexample. If the set was closed and bounded, then its
image would be closed again, because the image of a compact set is, in particular, closed,
by the above theorem.
Examples 6.4.5:
If
then:
1. what is the image of [-2, 1] ?
2. find a closed set whose image is not closed
Find examples for the following situations:
1. A continuous function and a set whose image is not
connected.
2. A continuous function and a disconnected set whose
image is connected.
3. A function such that the image of a connected set is
disconnected.
f'(c) =
Let f be a function defined on (a, b) and c any number in (a, b). Then
f is differentiable at c if and only if there exists a constant M such that
f(x) = f(c) + M ( x - c ) + r(x)
where the remainder function r(x) satisfies the condition
=0
Proof
This theorem provides a suitable method to generalize the concept of derivative to other
spaces: a function defined in some general space is called differentiable at a point c if it can
be approximated by a linear function at that point. On the real line the linear function M ( x
- c ) + f(c), of course, is the equation of the tangent line to f at the point c. In higher
dimensional real space this concept is known as the total derivative of a function.
Examples 6.5.4:
Examples 6.5.6:
o
(f(x) + g(x)) = f'(x) + g'(x)
Product Rule: If f and g are differentiable at x = c then f(x)
g(x) is differentiable at x = c, and
o
(f(x) g(x)) = f'(x) g(x) + f(x) g'(x)
Quotient Rule: If f and g are differentiable at x = c, and g(c)
# 0 then then f(x) / g(x) is differentiable at x = c, and
( f(x) / g(x) ) =
Chain Rule: If g is differentiable at x = c, and f is
differentiable at x = g(c) then f(g(x)) is differentiable at x = c,
and
o
An extension of Rolle's theorem that removes the conditions on f(a) and f(b) is the MeanValue-Theorem. It is actually a 'shifted' version of Rolle's theorem, as its proof illustrates.
A more general version of the Mean Value theorem is also mentioned which is sometimes
useful.
Theorem 6.5.9: Mean Value Theorem
If f is continuous on [a, b] and differentiable on (a, b), then there
exists a number c in (a, b) such that
f'(c) =
Proof
Examples 6.5.10:
Rolle's theorem and the Mean Value theorem allow us to develop the familiar test for local
extrema of a function, as well as increasing and decreasing functions. Recall the definition
of local extremum:
Definition 6.5.11: Local Extremum
Let f be a function defined on a domain D, and c a point in D.
1. If there exists a neighborhood U of c with f(c) f(x) for all x
in U, then f(c) is called a local maximum for the function f
that occurs at x = c.
2. If there exists a neighborhood U of c with f(c) f(x) for all x
in U, then f(c) is called a local minimum for the function f
that occurs at x = c.
3. If f(x) has either a local minimum or a local maximum at x =
c, then f(c) is called local extremum of the function f.
You can find possible local extrema by applying the following theorem:
Theorem 6.5.12: Local Extrema and Monotonicity
This theorem suggests the following table in order to find local minima and maxima:
Suppose you have found a point c such that f'(c) either does not exist or f'(c) = 0. For each
c (called a critical point of f) we may have one of these four situations:
Loc. Max
interval
Loc. Min
(a, c) (c , b) interval
sign of f'(x)
dir. of f(x)
up
sign of f'(x)
No Extremum
interval
(a, c) (c , b)
-
down
up
No Extremum
(a, c) (c , b) interval
(a, c) (c , b)
sign of f'(x)
sign of f'(x)
dir. of f(x)
up
up
dir. of f(x)
down down
These results above are the cornerstones of Calculus 1 in most colleges. As a review, you
may enjoy the following examples:
Examples 6.5.14:
exists.
The same result holds for one-sided limits.
If f and g are differentiable and
,then
f(x) =
g(x) =
Proof
There are other situations where l'Hospital's rule may apply, but often
expressions can be rewritten so that one of these two cases will apply.
Examples 6.5.16:
3.
Dirichlet Function
The Dirichlet Function is not continuous at any point on the real line.
Back
Proof:
Left as Exercise. Show that g can not be continuous at a rational number by considering
sequences of irrational numbers. Then show that g can not be continuous at an irrational
number by considering sequences of rational numbers.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Countable Discontinuities
Proof:
The actual proof is left as an exercise. However, you may want to make use of the
following fact:
Lemma
If r = p / q is a rational number (in lowest terms) define a function with
domain Q via f(r) = 1 / q. Then the limit of f(r) as r approaches any real
number is zero.
Proof of Lemma:
Take any sequence { rn } of rational numbers converging to a fixed number a (which could
be rational or irrational). Since the sequence converges, it is bounded. For simplicity
assume that all rational numbers rn are in the interval [0, K] for some integer K. Now take
any > 0 and pick an integer M such that 1 / M < . Because each rational number is the
quotient of two positive integers, we have:
at most K of the rational numbers rn in the interval [0, N] can have denominator
equal to 1
at most 2 K of the rational numbers rn in the interval [0, N] can have denominator
equal to 2
...
at most M * K of the rational numbers rn in the interval [0, N] can have
denominator equal to M
In total, at most finitely many of the rn can have a denominator less than or equal to K.
That means, however, that there exists an integer N such the denominator of rn is bigger
than M for all n > N. But then
| f(rn) | < 1 / M < for all n > N
Since > 0 was arbitrary, that means that the limit of f(rn) must be zero, as needed. Our
assumption that the numbers rn were all positive can easily be dropped, and a similar proof
would work again.
Using this lemma it should not be too hard to prove the original assertion.
C(1) Function
Proof:
We have studied this function before. The details are left as an exercise, but here is the
idea:
g is differentiable for all non-zero x (apply some theorems)
find the derivative for non-zero x by using chain and product rules
find the derivative of g at zero by looking at the limit of the difference quotient at
zero
show that the derivative you have just found is not continuous at zero by
considering left and right limits as x approaches zero
C(n) Function
The function g is n-times differentiable, the n-th derivative is continuous,
and g is not (n+1)-times differentiable.
On the right are the graphs of this root function for n = 0, 1 and 2. Can you decide which
graph goes with what n ? Click on Options and choose different powers of x to see how the
function gets smoother and smoother at 0, which is necessary if it wants to be more and
more times differentiable.
Proof:
This proof is left as an exercise. As a hint, start with n = 1. The example, if correct, says
that for n = 1 the function g is once differentiable, the derivative is continuous, but
differentiable again. Why ? Then try n = 2. Then, maybe, try an induction argument to
prove the general case.
C(inf) Function
of g at x = 0 equals zero.
Can a polynomial have that property ? In other words, can you find a polynomial of any
(fixed) degree such that it is not identically zero, yet all derivatives at x = 0 are zero ?
Proof:
This function is special, as we will see when dealing with power series representations: if
one wants to find a power series representation, one could apply Taylor's formula to find
the coefficients of the power series, say, centered at zero. That requires that the function
under consideration has to be infinitely often differentiable. This function is. Taylor's
formula involves derivatives of the function at the origin. In this case, they are all zero.
Hence, the power series associated with this function would be identically equal to zero.
But then it does not represent the original function.
In other words, there are functions for which you can use Taylor's theorem to find a
convergent power series, but this power series is not equal to the original function.
In any case, we will first prove that this function is once differentiable, with g'(0) = 0.
Obviously, if x # 0, then g'(x) exists by the chain rule, and we have
This limit looks hard, but we can make the substitution u = 1 / x2 and use 'Hospital's rule:
Note that we do have to look at u approaching positive and negative infinity, but because
of the square term in the denominator we can deal with both cases in one line.
Therefore, g is differentiable at x = 0 and the derivative is zero. Next we need to show that
g'(x) is continuous. It is obviously continuous for all x but zero, so we only need to check
continuity at x = 0:
Again, we have used the above substitution, and we have applied l'Hospital's rule several
times in our head. But the result is that g(x) is differentiable everywhere, and g'(0) = 0.
Now we could treat the case N = 2, N = 3, and so forth. But we need to prove this for all N,
so eventually we will have to employ an induction argument. Also, the computations of
higher derivatives will become more and more complicated, since the product rule will
introduce additional terms. Therefore, we have to look at the problem in a more abstract
way.
While the details are left as an exercise, you might want to make use of (and first prove)
the following steps:
Let p(x) be a polynomial in 1 / x, i.e.
. Then show the following:
the n-th derivative of g(x) for x not zero is of the form p(x)
the limit of the difference quotient at zero for functions of the form p(x)
zero
equals zero
is
Weierstrass Function
The graph of the n-th partial sum for Weierstrass's monster. To get a better idea about the
limit, try larger values of n (values larger than 30 do not work due to roundoff errors). Try
to zoom into the function using the mouse. Can you believe that this function is continuous,
but differentiable nowhere ?
There are many other, similarly constructed. Here is another example which is the same as
the function above but without the absolute values:
Proof:
This function, incidentally, is the one which was used to generate the logo at the title
pages. To show that it is continuous will be easy - if we assume knowledge from the next
chapter. To show it is not differentiable will be somewhat more difficult.
We'll do it later, sorry.
Cantor Function
The Cantor function is a function that is continuous, differentiable,
increasing, non-constant, and the derivative is zero everywhere except at a
set with length zero.
This is the most difficult function in our repertoire and can be found, for example, in
Kolmogorov and Fomin.
Recall the definition of the Cantor set: Let
= [ 1/3, 2/3]
be the middle third of the interval [0, 1]. Let
= [ 1/9, 2/9],
= [ 7/9, 8/9 ]
= [7/27, 8/27],
= [19/27, 20/27],
= [ 25/27, 26,27]
from [0, 1]. Continue in this fashion so that at the n-th stage we have the
, ...,
, ...,
if t is in the interval
Then, for example, we have that
if 1/3 t 2/3
Then F(t) is defined everywhere in [0, 1] except at the Cantor set minus the end points 0,
1, 1/3, 2/3, 1/9, 2/9, 7/9, 8/9, ... If t is a number where F is not defined, then there exists an
increasing sequence { xn } of these endpoints converging to t, and a decreasing sequence
{ xn' } of these endpoints converging to t. Since F is defined at those endpoints xn and xn',
we define
F(t) =
F(xn) =
F(xn')
Now we have defined completely the Cantor function. It has the following properties:
F is defined everywhere in the interval [0, 1]
F is not constant
F is increasing
F is continuous in the interval [0, 1]
F is differentiable in the interval [0, 1]
In particular, F' is zero at points of total length 1 in the interval [0, 1], yet F it is not
constant.
Proof
Some of these properties are obvious, and some require more thought. In particular, why
does the above limit of endpoints exist ? That is a crucial point, because we used this limit
to extend the function to the whole interval [0, 1]. For details and hints about the Cantor
Function, please consult Kolmogorov and Fomin, p 334 ff.
IRA
In a calculus class integration is introduced as 'finding the area under a curve'. While this
interpretation is certainly useful, we instead want to think of 'integration' as more
sophisticated form of summation. Geometric considerations, in our situation, will not be so
fruitful, whereas the summation interpretation of integration will make many of its
properties easy to remember.
First, as usual, we need to define integration before we can discuss its
properties. We will start with defining the Riemann integral, and we will
move to the Riemann-Stieltjes and the Lebesgue integral later.
Definition 7.1.1: Partition of an Interval
A partition P of the closed interval [a, b] is a finite set of points P =
{ x0, x1, x2, ..., xn} such that
a = x0 < x1 < x2 < ... < xn-1 < xn = b
The maximum difference between any two consecutive points of the
partition is called the norm or mesh of the partition and denoted as |
P |, i.e.
| P | = max { xj - xj-1, j = 1 ... n }
A refinement of the partition P is another partition P' that contains
all the points from P and some additional points, again sorted by
order of magnitude.
Examples 7.1.2:
If P = { x0, x1, x2, ..., xn} is a partition of the closed interval [a, b] and
f is a function defined on that interval, then the n-th Riemann Sum
of f with respect to the partition P is defined as:
R(f, P) =
Examples 7.1.4:
Suppose f(x) = x2 on [0, 2]. Find
1. the fifth Riemann sum for an equally spaced partition, taking
always the left endpoint of each subinterval
2. the fifth Riemann sum for an equally spaced partition, taking
always the right endpoint of each subinterval
3. the n-th Riemann sum for an equally spaced partition, taking
always the right endpoint of each subinterval.
Riemann sums have the practical disadvantage that we do not know which point to take
inside each subinterval. To remedy that one could agree to always take the left endpoint
(resulting in what is called the left Riemann sum) or always the right one (resulting in the
right Riemann sum). However, it will turn out to be more useful to single out two other
close cousins of Riemann sums:
Definition 7.1.5: Upper and Lower Sum
Let P = { x0, x1, x2, ..., xn} be a partition of the closed interval [a, b]
and f a bounded function defined on that interval. Then:
the upper sum of f with respect to the partition P is defined
as:
U(f, P) =
cj (xj - xj-1)
L(f, P) =
dj (xj - xj-1)
Examples 7.1.6:
Suppose P = { x0, x1, x2, ..., xn} is a partition of the closed interval [a,
b], f a bounded function defined on that interval. Then we have:
The lower sum is increasing with respect to refinements of
partitions, i.e. L(f, P') L(f, P) for every refinement P' of the
partition P
The upper sum is decreasing with respect to refinements of
partitions, i.e. U(f, P') U(f,P) for every refinement P' of the
partition P
L(f, P) R(f, P) U(f, P) for every partition P
Proof
In other words, the lower sum is always less than or equal to the upper
sum, and the upper sum is decreasing with respect to a refinement of
the partition while the lower sum is increasing with respect to a
refinement of the partition. Hence, a natural question is: will the two
quantities ever coincide ?
Definition 7.1.8: The Riemann Integral
Suppose f is a bounded function defined on a closed, bounded interval
[a, b]. Define the upper and lower Riemann integrals, respectively, as
I*(f) = inf{ U(f,P): P a partition of [a, b]}
I*(f) = sup{ L(f,P): P a partition of [a, b]}
Then if I*(f) = I*(f) the function f is called Riemann integrable and
the Riemann integral of f over the interval [a, b] is denoted by
f(x) dx
Note that upper and lower sums depend on the particular partition chosen, while the upper
and lower integrals are independent of partitions. However, this definition is very difficult
for practical applications, since we need to find the sup and inf over any partition.
Examples 7.1.9:
Examples 7.1.11:
f(x) dx
c f(x) + d g(x) dx = c
f(x) dx + d
g(x) dx
3. | f(x) dx |
| f(x) | dx
4. If g is another function defined on [a, b] such that g(x) < f(x)
on [a, b], then g(x) dx
f(x) dx
5. If g is another Riemann integrable function on [a, b] then f(x)
.
g(x) is integrable on [a, b]
Proof
Examples 7.1.13:
M (a - b)
f(x) dx M (b - a)
Now we can illustrate the relation between Riemann integrable and continuous functions.
Theorem 7.1.14: Riemann Integrals of Continuous Functions
Every continuous function on a closed, bounded interval is Riemann
integrable. The converse is false.
Proof
IRA
This section provides integration techniques, i.e. methods for finding the actual value of an
integral. We have already found the most basic technique (the Integral Evaluation Shortcut
or First Fundamental Theorem of Calculus): to evaluate an integral over an interval [a, b],
find an antiderivate F of the integrand f and compute F(b) - F(a).
But different from finding a derivative, where a couple of rules can be
used to find virtually every derivative, finding an antiderivative is
basically a guessing game.
Example 7.2.1: Standard Antiderivatives
Find the following antiderivatives:
(a) xr dx
(b) 1/x dx
(c) ex dx
(f) tan(x) dx
(g)
(h)
(i)
As is typical for finding the antiderivatives, the answers for the above examples can not be
deduced. They must be guessed, or learned. However, there are two standard mechanisms
that can be useful: substitution and integration by parts.
Theorem 7.2.2: Substitution Rule
If f is a continuous function defined on [a, b], and s a continuously
differentiable function from [c, d] into [a, b]. Then
Proof
This theorem, in words, says that if you can identify a composition of functions as well as
the derivative of one of the composed functions, there's a good chance you can find the
antiderivative and evaluate the corresponding integral.
Think about the substitution rule as a "change of variable": if a
particular expression in x makes an integrand difficult, change it to u.
Then compute the derivative of u, i.e. du/dx = u'(x) and "solve" it for du:
du = u'(x) dx. If you can use u'(x) dx to change the integrand into an
expression involving only u and no x, then the substitution worked (and
has hopefully simplified the integrand). Otherwise try a different
substitution or a different technique altogether.
Example 7.2.3: Applying the Substitution Rule
f(cx + d) dx
x f(cx2) dx
f'(x) / f(x) dx
Compute the area of a circle with radius r.
Find
and
Another theorem that is commonly used is "integration by parts".
Theorem 7.2.4: Integration by Parts
f'(x) g(x) dx
Proof
x ex dx
x2 cos(x) dx
ln(x) dx
sin5(x) dx
Integration by parts can be used to provide an effective means to compute the value of an
integral numerically. Of course, Riemann sums can be used to approximate an integral, but
convergence is usually slow. A much faster convergence scheme is based on the trapezoid
rule. To prove it, we need the Mean Value Theorem for Integration:
Theorem 7.2.6: Mean Value Theorem for Integration
g(x) dx
Proof
[a, b] }
The Trapezoid Rule is useful because the error R(n) depends on the square of h. If h is
small, h2 is a lot smaller so that the Trapezoid Rule provides a good approximation to the
numeric value of an integral (as long as f is twice continuously differentiable).
Example 7.2.8: Application of the Trapezoid Rule
Compare the numeric approximations to the integral
sin(x) cos(x) dx
obtained by using (a) a left Riemann sum and (b) the Trapezoid Rule,
using a partition of size 5 and of size 100.
Our final integration technique uses partial fraction decomposition of a polynomial to
simplify rational functions so that they can be integrated.
Theorem 7.2.9: Partial Fraction Decomposition
7.3. Measures
IRA
The Riemann integral is certainly useful (and complicated) enough, but it does have a few
limitations and oddities:
Examples 7.3.1:Oddities of Riemann Integral
l(An) }
interval An.
Examples 7.3.3: Outer Measure of Intervals
Find the outer measure of the empty set O, and prove that
m*(A) m*(B) for all A B.
Find the outer measure of a closed interval [a, b]
Find the outer measure of an open interval (a, b)
Find the outer measure of an infinite interval
An)
m*(An)
Proof
Since outer measure is defined for all sets in R, our first key condition above is satisfied.
But outer measure is not quite similar to a length, because it is only subadditive ( m*(A
B) m*(A) + m*(B) ), not additive ( m*(A B) = m*(A) + m*(B) for disjoint sets A and B).
Therefore one of the requirements of the second key condition is not satisfied.
Examples 7.3.5: Properties of Outer Measure
m*(A) = m*(A
E) + m*(A
comp(E))
E) + m*(A
comp(E))
The motivation for this definition is that it ensures that for two disjoint measurable sets E
and F we have that
m(E
F) = m(E) + m(F)
Show that the empty set, the set R, and the complement of
a measurable set are all measurable.
Show that every set with outer measure 0 is Lebesgue
measurable.
Show that the union of two measurable sets is measurable.
Show that the intersection of two measurable sets is
measurable.
There are sets that are not (Lebesgue) measurable, i.e. not every set is
(Lebesgue) measurable.
Proof
To summarize, we introduced a new concept called measure in two stages, each of which
had a typical "good news, bad news" property.
1. In the first stage we defined outer measure.
o good news: outer measure is defined for all sets.
o bad news: outer measure is not additive, i.e. it is not quite comparable to a
length.
2. In the second stage we defined measure by restricting outer measure to the
measurable sets.
o good news: measure is additive, i.e. it is a good generalization of length.
o bad news: measure is not defined for all sets
Before concluding this section, we want to prove one more result that will be
important for the next section:
Proposition 7.3.11: Monotone Sequences of Measurable Sets
If { An } is a sequence of measurable sets that is decreasing, i.e. Aj
Aj+1 for all j, and m(A1) is finite, then
lim m(Aj) = m( Aj)
If { An } is a sequence of measurable sets that is increasing in the
sense that Aj+1 Aj for all j, then
lim m(Aj) = m( Aj)
Proof
IRA
ai XEi(x)
an XAn(x)
where
An = { x
A: f(x) = an }
Therefore simple functions can be thought of as dividing the range of f, where the resulting
sets An may or may not be intervals.
Examples 7.4.2: Simple Functions
an m(An)
Just as step functions were used to define the Riemann integral of a bounded function f
over an interval [a, b], simple functions are used to define the Lebesgue integral of f over a
set of finite measure.
Definition 7.4.5: Lebesgue Integral for Bounded Function
Suppose f is a bounded function defined on a measurable set E with
finite measure. Define the upper and lower Lebesgue integrals,
respectively, as
I*(f)L = inf{ s(x) dx: s is simple and s f }
I*(f)L = sup{ s(x) dx: s is simple and s f }
If I*(f)L = I*(f)L the function f is called Lebesgue integrable over E and
the Lebesgue integral of f over E is denoted by
E
f(x) dx
f(x) dx =
[a,b]
f(x) dx
Proof
IRA
In progress ...
If f is L-integrable, so is |f|, but the converse is not true (I think).
There is a function f that is R-integrable but |f| is not R-integrable
Riemann integral for bounded functions only, Lebesgue for bounded and
unbounded functions.
Riemann integral can be extended to improper Riemann integrals, but
can not allow functions that are extended real valued.
An improper R-integral may exist without the function being L-integrable
(see f(x) = sin(x) / x for x > 0.
If f is L-integrable and the improper R-integral exists, then both agree.
Mention sequences of R-integrable function vs. L-integrable
Lebesgue's Theorem: a bounded function f on [a, b] is Riemann
integrable if and only if the set of points where f is not continuous has
measure zero.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.1.2(a):
If E = { x: x = 2n for n N}, O = {x : x = 2n - 1 for n }, and I = {x
= -2}, then what, in your own words, do these sets represent ?
R: x2
back
(B
C) = (A
B)
C) = (A
(A C)
B)
(A
C)
Context
Proof:
These relations could be best illustrated by means of a Venn Diagram.
C)
Examples 1.1.2(b):
If A = {x R : -4 < x < 3} and B = {x
B, A \ B, and comp(A).
A B={x
R : -4 < x < 7 }
A B={x
R : -1 < x < 3 }
A\B={ x
R : -4 < x -1 }
comp( A ) = { x
R:-
< x -4 } { x
R:3 x<
Examples 1.1.5(a):
Prove that when two even integers are multiplied, the result is an even
integer, and when two odd integers are multiplied, the result is an odd
integer.
back
To prove this we first need to know what exactly an even and odd integer is:
an integer x is even if x = 2n for some integer n
an integer x is odd if x = 2n + 1 for some integer n
Now that we have a precise definition, the actual proof is easy: Take x and y two even
numbers. Then
x = 2n for some integer n
y = 2m for some integer m
Multiplying these numbers together we get
xy = (2n)(2m) = 4 nm = 2 (2nm) = 2 k
where k = 2nm. Hence, xy is again even.
If x and y are two odd numbers, then
x = 2n + 1 for some integer n
y = 2m + 1 for some integer m
Multiplying these numbers together we get
xy = (2n+1)(2m+1) = 4nm + 2(n + m) + 1 = 2 (2nm + n + m) + 1 = 2k + 1
where k = 2nm + n + m. Hence, xy is again odd.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.1.5(b):
Prove that if the square of a number is an even integer, then the original
number must also be an even integer. (Try a proof by contradiction).
back
To prove this we first need to know what exactly an even and odd integer is:
an integer x is even if x = 2n for some integer n
an integer x is odd if x = 2n + 1 for some integer n
Now that we have a precise definition, the actual proof is easy: Suppose x is a number such
that x2 is even. To start a proof by contradiction we will assume the opposite of what we
would like to prove: assume that x is odd (but x2 is still even). Then, because x is odd, we
can write it as
x = 2n + 1
But then the square of x is
x2 = (2n + 1)(2n + 1) = 4 n2 + 4n + 1 = 2 (2 n2 + 2n) + 1 = 2k + 1
with k = 2 n2 + 2n. Therefore x2 is odd. But that is contrary to our assumption that the
square of x is even. Hence, if the square of x is supposed to be even, x itself must be 'not
odd'. But 'not odd' means even. Therefore, the proof is finished.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.1.5(b):
Prove that if the square of a number is an even integer, then the original
number must also be an even integer. (Try a proof by contradiction).
back
To prove this we first need to know what exactly an even and odd integer is:
an integer x is even if x = 2n for some integer n
an integer x is odd if x = 2n + 1 for some integer n
Now that we have a precise definition, the actual proof is easy: Suppose x is a number such
that x2 is even. To start a proof by contradiction we will assume the opposite of what we
would like to prove: assume that x is odd (but x2 is still even). Then, because x is odd, we
can write it as
x = 2n + 1
But then the square of x is
x2 = (2n + 1)(2n + 1) = 4 n2 + 4n + 1 = 2 (2 n2 + 2n) + 1 = 2k + 1
with k = 2 n2 + 2n. Therefore x2 is odd. But that is contrary to our assumption that the
square of x is even. Hence, if the square of x is supposed to be even, x itself must be 'not
odd'. But 'not odd' means even. Therefore, the proof is finished.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Euclid Theorem:
There is no largest prime number.
Context
Proof
Suppose there was a largest prime number; call it N. Then there are only finitely many
prime numbers, because each has to be between 1 and N. Let's call those prime numbers a,
b, c, ..., N. Then consider this number:
M = a * b * c * ... * N + 1
Is this new number M a prime number? We could check for divisibility:
M is not divisible by a, because M / a = b * c * ... * N + 1 / a
M is not divisible by b, because M / b = a * c * ... * N + 1 / b
M is not divisible by c, because M / c = a * b * ... * N + 1 / c
.....
Hence, M is not divisible by a, b, c, ..., N. Since these are all possible prime numbers, M is
not divisible by any prime number, and therefore M is not divisible by any number. That
means that M is also a prime number. But clearly M > N, which is impossible, because N
was supposed to be the largest possible prime number. Therefore, our assumption is wrong,
and thus there is no largest prime number.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.2.3:
Let A = {1, 2, 3, 4}, B = {14, 7, 234}, C = {a, b, c}, and R = real numbers.
Define the following relations:
1. r relates A and B via: 1 ~ 234, 2 ~ 7, 3 ~ 14, 4 ~ 234, 2 ~ 234
2. f relates A and C via: {(1,c), (2,b), (3,a), (4,b)}
3. g relates A and C via: {(1,a), (2,a), (3,a)}
4. h relates R and itself via: {(x,sin(x))}
Back
1. The relation r is not a function, because the element 2 from the set A is associated
with two elements from B.
2. The relation f is a function, because every element from A has exactly one relation
from the set C.
3. The relation g is not a function, because the element {4} from the domain A has no
element associated with it.
4. The relation h is a function with domain R, because every element {x} in R has
exactly one element {sin(x)} associated with it.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 1.2.5(a):
Let f(x) = 0 if x is rational and f(x) = 1 if x is irrational. This function is
called Dirichlets Function. The range for f is R. Find the image of the
domain of the Dirichlet Function when:
1. the domain of f is Q
2. the domain of f is R
3. the domain of f is [0, 1] (the closed interval between 0 and 1)
Back
1. When the domain is Q, we have that f(x) = 0 for any x, because x must be a rational
number. Hence, the image of the domain Q is the set consisting of the single
element {0}.
2. When the domain is R, we have that f(x) could be 0 or 1, because x could be
rational or irrational. f(x) can not be any other number. Hence, the image of the
domain R is the set consisting of the two elements {0, 1}.
3. The interval [0, 1] contains irrationals as well as rational numbers. Therefore, f(x)
could be equal to 0 or 1, and the image of [0, 1] under f is the set with the two
elements {0, 1}.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 1.2.5(b):
Let f(x) = 0 if x is rational and f(x) = 1 if x is irrational. This function is
called Dirichlets Function. The domain and range for f is R.
What is the preimage of R ? What is the preimage of [-1/2, 1/2] ?
Back
The preimage of R is the set of all elements such that f(x) is contained in R. Since R
includes the numbers 0 and 1, the preimage of R under f is everything in the domain, i.e.
R.
The preimage of [-1/2, 1/2] consists of all elements such that f(x) is
contained in [-1/2, 1/2]. This set does not contain 1, so the preimage of
that set can not contain any irrational number. Hence, the preimage of [1/2, 1/2] is the set of rational numbers Q.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.2.7(a):
If the graph of a function is known, how can you decide whether a function
is one-to-one (injective) or onto (surjective) ?
Back
Injection
If a horizontal line intersects the graph of a function in more than one place, then there are
two different points a and b for which f(a) = f(b), but a # b. Then the function is not oneto-one.
Surjection
If you place a light on the left and on the right hand side of the coordinate system, then the
shadow of the graph on the y axis is the image of the domain of the function. If that
shadow covers the range of the function, then the function is onto.
Examples 1.2.7(b):
Which of the following functions are one-one, onto, or bijections ? The
domain for all functions is R.
1. f(x) = 2x + 5
2. g(x) = arctan(x)
3. g(x) = sin(x)
1. Linear Function
This function is linear. The equation y = 2x + 5 has a unique solution for every x, so that
the function is one-one and onto, i.e. a bijection. In fact, all linear functions are bijections.
f(x) = 2x + 5 is a bijection.
2. Inverse Functions
This function is the inverse function of the tangent function. As such, it must be one-toone. It is onto if the range is [- / 2, / 2], but not if the range is R. In fact, all inverse
functions are one-to-one.
3. Periodic Function
Since this function is periodic, it can not be one-to-one. It is onto, if the range is the
interval [-1, 1], but not onto if the range is R. In fact, every periodic function is not one-toone.
This is an odd-degree polynomial. Hence, the limit as x approaches plus or minus infinity
must be plus or minus infinity, respectively. That means that the function is onto. Since
most third degree equations have three zero, this function is probably not one-to-one. A
look at the graph confirms this. In fact, every odd-degree polynomial is onto while no even
degree polynomial is onto.
Proof:
We have to decide what the equivalence classes
should be: Since by property (1) two
elements a and b are supposed to be related if and only if they are in the same class, it
seems natural to define:
={b A:b~ }
To simplify notation a little, we denote this class by A( ) . Now we need to check whether
this is a good definition, and whether it satisfies the above properties.
Because of reflexivity of the equivalence relation, the class A(a)
contains the element a and is therefore not empty.
Next, we will show that two classes that are different can not have any
elements in common, i.e. they are disjoint. Recall that disjoint means
that two sets have either an empty intersection or are the same sets.
Take two elements a and a and suppose that A(a) and A(a) have a nonempty intersection; say both classes contain the element c. Then
c ~ a, because c is in A(a)
a ~ c, by symmetry and the above line
c ~ a because c is in A(a)
Therefore, by transitivity, a ~ a
Now take any element b in A(a). Then b ~ a and a ~ a, so that b ~ a by transitivity, and
hence b is contained in A(a). Therefore A(a) is a subset of A(a). The same argument
works the other way around as well (by symmetry), showing that A(a) is a subset of A(a).
Hence, A(a) = A(a). This shows that two different classes must be disjoint.
Finally, we need to show that all classes A(a) make up the original set A.
But that is clear, because every element a in A belongs to the class
A(a), hence the union of all classes A(a) must be contained in A, and
therefore must be equal to A.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 1.3.4(a):
Consider the set Z of all integers. Define a relation r by saying that x and y
are related if their difference y - x is divisible by 2. Then
1. Check that this relation is an equivalence relation
2. Find the two equivalence classes, and name them appropriately.
3. How would you add these equivalence classes, if at all ?
Back
1. Equivalence Relation
reflexive:
x - x is equal to zero, which is divisible by two. Hence, every element is related to
itself.
symmetry:
if x ~ y, then y - x is divisible by 2. But then - (y - x) = x - y is divisible by two.
Hence, y ~ x
transitivity:
if x ~ y then y - x = 2n for some integer n
if y ~ z then z - y = 2m for some integer m
But then z - x = (2m + y) - (y - 2n) = 2 (m + n), so that z - x is divisible by 2. In
other words, x and z are related.
2. Equivalence Classes
Two elements are in the same equivalence class if and only if they are
related. If x and y are in the same class, then y - x = 2m for some
integer n.
If y was even, then y = 2m for some integer m, and x = 2m - 2n must also be even.
If y was odd, then y = 2m + 1 for some integer m, and x = 2m + 1 - 2n must also be
odd
Therefore, there are two equivalence classes, and they are appropriately labeled:
E = even numbers: E = [(2)] contains all even numbers
O = odd numbers: O = [(3)] contains all odd numbers.
3. Adding Equivalence Classes
Define [x] + [y] = [x + y]. We need to show that this is well-defined, i.e.
independent of the particular representative of the equivalence classes
of [x] and [y]. Take x ~ x' and y ~ y'.
Then x' - x = 2m and y' - y = 2n for some integers n and m.
Then (x' + y') - (x + y) = x' - x + y' - y = 2m + 2n = 2 (m + n)
That means that (x' + y') and (x + y) are related if x ~ x' and y ~ y'. But then
[x] + [y] = [x + y] = [x' + y'] = [x'] + [y']
Hence, addition does not depend on the particular representative from a class, so that
addition as defined above is indeed a well-defined operation.
A better method would be the following: Define two classes 0 and 1 by
saying:
all numbers divisible by 2 with no remainder are in class 0
all numbers divisible by 2 with remainder 1 are in class 1
Then add equivalence classes by adding numbers modulo 2.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 1.3.4(b):
Consider the set Z of all integers. Define a relation r by saying that x and y
are related if their difference y - x is divisible by m. Then we have:
Back
This relation is an equivalence relation (i.e. the three conditions are satisfied)
There are m equivalence classes:
1. all numbers divisible by m with no remainder are in class 0
2. all numbers divisible by m with remainder 1 are in class 1
3. all numbers divisible by m with remainder 2 are in class 2
4. ...
5. all numbers divisible by m with remainder m-1 are in class m-1
Addition can be defined by adding modulo m. That is, if we consider the
equivalence classes obtained by dividing the differences by, say, 5, then we have, as
an example:
Example 1.3.5:
Consider the set R x R \ {(0,0)} of all points in the plane minus the origin.
Define a relation between two points (x, y) and (x, y) by saying that they
are related if they are lying on the same straight line passing through the
origin.
This relation is an equivalence relation, and the resulting
space of equivalence classes is called projective space. By
picking an appropriate member for each class one can find a
graphical representation for this space.
Back
As usual, we need to check to first verify that this is a relation, and then check the three
conditions that turn a relation into an equivalence relation: This clearly is a relation (most
things are) so that we will have to check:
reflexive:
every point is related to itself, since it lies on a line through the origin.
symmetry:
if (x, y) ~ (x', y'), then they are both on the same line through the origin. This is a
symmetric definition, so symmetry holds as well.
transitivity:
here we should use some math. Recall that two points (x, y) and (x', y') are on the
same line through the origin, if and only if there is a non-zero real number t such
that (x, y) = t (x', y').
take (x, y) ~ (x', y').
Then there exists t # 0 with (x, y) = t (x', y').
take (x', y') ~ (x'', y'')
Then there exists s # 0 with (x', y') = s (x'', y''). But then (x'', y'') = 1/s (x', y') = 1 /
(s t) (x, y), hence (x'', y'') ~ (x, y), proving transitivity.
This proves that the relation is indeed an equivalence relation. Next, we have to try to find
a graphical representation of P by picking an appropriate member for each class.
Since the relation is an equivalence relation we know that there are
disjoint equivalence classes [(x, y)], and we can pick any member of [(x,
y)] as a representative of that class. Define a point p on the unit circle to
be equal to the whole equivalence class [(x, y)] containing p. That is, we
identify each equivalence class with a point on the unit circle by drawing
the line through that point and the origin. That line contains every
member of the equivalence class.
Numbers on the unit circle that are diagonally opposite each other,
however, are in the same equivalence class, since they are on the same
line through the origin.Thus we identify the space P with the unit circle,
where diagonally opposite numbers are identical. Here is a poor picture
of this space:
This space is called projective space, and can be generalized to higher dimensions as well.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
We first have to prove reflexivity, symmetry, and transitivity to show that the relation is an
equivalence relation. The first two properties are easy, and are left as an exercise. As for
transitivity:
Take (a,b) ~ (a', b), i.e. a + b = a + b
Take (a', b) ~ (a, b), i.e. a + b = a + b
Adding, we get (a + b) + (a + b) = (a + b) + (a + b). Canceling a and b on both
sides yields
a + b = a + b, i.e. (a, b) ~ (a, b)
Hence, transitivity is proved.
Next, we will have to show that the way that the definition of addition
and multiplication is well-defined. In particular, we need to show that
the definition of these operations does not depend on the particular
representative of the equivalence classes that we chose. The idea of
that proof is clear: pick different members of a class, and show that their
sum or product results in the same class. Thus, suppose:
(a, b) and (c, d) are related
(a, b) and (c, d) are related.
Then [(a,b)] + [(a, b)] = [(a + a, b + b)]
and [(c, d)] + [(c, d)] = [(c + c, d + d)]
Because (a, b) ~ (c, d) we know: a + d = c + b
Because (a, b) ~ (c, d) we know that a + d = c + b
Adding both of those equations we get that:
(a + a) + (d + d) = (c + c) + (b + b)
which means, in other words, that the two elements are related in turn. Hence:
[(a + a), (b + b)] = [(c + c), (d + d)]
which is exactly what we wanted to show.
Finally, we need to show that multiplication is also well-defined.
Therefore, suppose:
(a, b) and (c, d) are related, i.e. a + d = c + b
(a, b) and (c, d) are related. i.e. a + d = c + b
Examples 1.4.3(a):
Let A be the set N x N and define an equivalence relation r on N x N and
addition of the equivalence classes as follows:
1. (a,b) is related to (a,b) if a + b = a + b
2. [(a,b)] + [(a',b')] = [(a + a', b + b')]
3. [(a,b)] * [(a, b)] = [(a * b + b * a, a * a + b * b)]
Which elements are contained in the equivalence classes of, say, [(1, 2)],
[(0,0)] and of [(1, 0)] ? Which of the pairs (1, 5), (5, 1), (10, 14), (7, 3) are in
the same equivalence classes ?
Back
The elements in the equivalence class of [(1, 2)] are all numbers (x,y) that are related to (1,
2), i.e. all (x,y) such that
1 + y = x + 2 or
y-x=1
In other words, the difference of the second and the first entry is one. Some members of
this equivalence class are therefore
(2, 3), (3, 4), (100, 101) [(1, 2)]
Some of the elements of the class [(0,0)] and of [(1, 0)] are:
(x, y) [(0, 0)] if (0, 0) ~ (x,y)
0 + y = x + 0 or y = x
Hence, (1, 1), (5, 5), (100, 100) [(0, 0)]
(x, y) (1, 0) if (1, 0) ~ (x, y)
1 + y = 0 + x or x - y = 1
Hence, (2, 1), (6, 5), (101, 100) [(1, 0)]. To determine which of the pairs (1, 5), (5, 1),
(10, 14), (7, 3) are in the same equivalence classes, all we have to do is compare the
differences between the second and the first entry:
(1,5): the difference y - x = 4
(5, 1): the difference y - x = -4
(10, 14): the difference y - x = 4
(7, 3): the difference y - x = -4
Therefore (1,5) and (10, 14) are in the same class, and (5,1) and (7,3) are also in the same
class.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.4.3(b):
Let A be the set N x N and define an equivalence relation r on N x N and
addition of the equivalence classes as follows:
1. (a,b) is related to (a,b) if a + b = a + b
2. [(a,b)] + [(a',b')] = [(a + a', b + b')]
3. [(a,b)] * [(a, b)] = [(a * b + b * a, a * a + b * b)]
Below are some examples for addition and multiplication of these
equivalence classes.
Back
If you add [(1,2)] + [(4, 6)] you would get the following:
[(1,2)] + [(4, 6)] = [(1 + 4, 2 + 6)] = [(5, 8)]
By the above examples, that implies
[(1,2)] contains all pairs whose difference y - x = 1
[(4,6)] contains all pairs whose difference y - x = 2
[(1,2)] + [(4,6)] contains all pairs whose difference y - x = 3
Adding [(3,1)] + [(1,3)] gives the following:
[(3,1)] + [(1,3)] = [(3+1, 1+3)]
This is, by the above example, equivalent to the following:
[(3,1)] contains all pairs whose difference y - x = -2
[(1,3)] contains all pairs whose difference y - x = 2
Multiplying the equivalence classes [(5,4)] * [(7, 4)] we get the following:
[(5,4)] * [(7, 4)] = [(5*4 + 4*7, 5*7 + 4*4)] = [(48,51)]
This is, by the above example, equivalent to the following:
[(5,4)] contains all pairs whose difference y - x is -1
[(7,4)] contains all pairs whose difference y - x is -3
[(5,4)] * [(7, 4)] = [(48,51)] contains all pairs whose difference y - x = 3
Multiplying the equivalence classes [(1,2)] * [(2,1)] we get the following:
[(1,2)] * [(2,1)] = [(1*1 + 2*2, 1*2 + 2*1)] = [(5,4)]
This is, by the above example, equivalent to the following:
[(1,2)] contains all pairs whose difference y - x is 1
[(2,1)] contains all pairs whose difference y - x is -1
[(1,2)] * [(2,1)] = [(5,4)] contains all pairs whose difference y - x is -1.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 1.4.3(c):
Let A be the set N x N and define an equivalence relation r on N x N and
addition of the equivalence classes as follows:
1. (a,b) is related to (a,b) if a + b = a + b
2. [(a,b)] + [(a',b')] = [(a + a', b + b')]
3. [(a,b)] * [(a, b)] = [(a * b + b * a, a * a + b * b)]
What is the best symbol to use for the resulting equivalence classes ?
Back
Example 1.4.5:
Why does one need another number system more complicated than the
rational numbers Q?
Back
There are several reasons, many of which are explored in detail in the next chapters. Here
are a few of them (which may use terms we are not yet familiar with).
a simple equation like x2 - 9 = 0 does have a solution in Q, but another, just as
simple equation x2 - 2 = 0 does not have a solution in Q.
If we construct a right triangle for which two sides have length 1, then we could not
measure the length of the remaining side if all we knew were rational numbers.
We could not measure the circumference of any circle if all we knew were rational
numbers.
if we set x0 = 2 and then for each integer n > 0 compute the number
successively, then each resulting number is a rational number, the sequence of
numbers is getting smaller and smaller, but they seem to get closer and closer to
some limit. However, this sequence of numbers does not converge to a rational
number. The sequence looks like this (do you know its limit ?):
o x0 = 2
o x1 = 3/2 = 1.5
o x2 = 17/12 = 1.416...,
o x3 = 577/408 = 1.414215686 , ...
o and so on ...
There are sets consisting of rational numbers that are bounded, but do not have a
least upper bound in Q.
Equations such as sin(x) = 1/2 or cos(x) = 0 do not have solutions in Q.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
We can not count the students, since they are moving around too much. Therefore, we set
up a function f that associates each chair with a student by simply asking each student to
find a chair and sit down. If all chairs are taken, and no students are left standing, then
what does this mean for our function f ?
the domain of f is the space of all students
the range of f is the space of all chairs
the function f is one-to-one, because: if f(a) = f(b), then student a and student b are
occupying the same chair. This can not happen unless student a and student b is the
same. Hence, f is injective.
the function f is onto, because: the range is the space of all chairs. Since all chairs
are occupied, there is a student associated with each chair. Hence, f is surjective.
Therefore, the function f is a bijection between the domain and the range, and by definition
of cardinality the number of students matches the number of chairs, i.e. both sets have the
same cardinality.
On the other hand, if the two sets did not contain the same number of
elements, the following could happen:
if f was one-to-one, but not onto, then there would be empty chairs. Hence,
cardinality of the students is less than the cardinality of the chairs.
if f was onto, but not one-to-one, then all chairs are taken, but some chairs hold
more than one student. Hence, cardinality of the students is greater than the
cardinality of the chairs.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.1.2:
Find the bijections to prove the following statements:
1. Let E be the set of all even integers, O be the set of odd integers.
Then card(E) = card(O)
2. Let E be the set of even integers, Z be the set of all integers. Then
card(E) = card(Z)
3. Let N be the set of natural numbers, Z be the set of all integers. Then
card(N) = card(Z)
Back
In each of the three cases we have to find a bijection between the two pairs of sets.
1. Define the function f(n) = n + 1 with domain E and range O. Then the function f is
clearly one-to-one and onto, hence it is a bijection. Now f is a bijection between E
and O, so that card(E) = card(O).
2. Define the function f(n) = 2n with domain Z and range E. Then it is straightforward to show that this function is one-to-one and onto, giving the required
bijection. Hence, card(Z) = card(E).
3. Define the following function: f(n) = n / 2 if n is even and f(n) =- (n-1) / 2 if n is
odd, with domain N and range Z. Again, it is not hard to show that this function is
one-to-one and onto, and therefore card(N) = card(Z).
The actual details of proving that the functions are bijections are left as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The proof contains some gaps that you are supposed to fill in as an exercise.
Assume that S is countable. Then, by definition,
there exists a bijection f from S to N
Examples 2.1.5:
The set of integers Z and the interval of real numbers between 0 and 2, [0,
2], are both infinite.
Back
According to Dedekind's theorem, we need to find a proper subset of each set that has the
same cardinality as the original set. In other words, we need to find a bijection from the
original set into a proper subset of itself.
Define the function f(n) = 2n. Then this function is a bijection between Z
and the even integers. Hence, Z has the same cardinality as a proper
subset of itself, and therefore Z is infinite.
Define the function f(x) = x / 2. Then the function is a bijection between
the interval [0, 2] and the interval [0, 1]. Hence, the interval [0, 2] is
infinite.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The first statement seems obvious. Nonetheless, it needs to be proved - as an exercise.
To prove the second statement, we list all ordered pairs as follows:
(1,1)
(1,2)
(1,3)
(1,4) ...
(2,1)
(2,2)
(2,3)
(2,4) ...
(3,1)
(3,2)
(3,3)
(3,4) ...
(4,1)
(4,2)
(4,3)
(4,4) ...
...
...
...
... ...
Next, we have do devise a method for counting these elements in order to enumerate all
pairs. Here is one possible solution:
Diagonalized Counting of N x N
Since we are able to enumerate all elements in the above table, the set
must be countable.
The proof of the next statement - that the countable union of countable
sets is again countable - is very similar. Try to duplicate the above proof
by writing down the elements from all sets in a table similar to the
above table. The details are left as another exercise.
Finally, we need to prove the last statement: Suppose S(1), S(2),
S(3), ..., S(n) are all countable. Then the cross product of those sets is
defined to be
S := {(x(1), x(2), ..., x(n)) : x(j)
If there were only two such sets, then the proof is exactly the same as the proof of the first
statement. If there are more than two sets, we can group them. That is, consider
S(1) x S(2) x S(3) x ... x S(n) = ( S(1) x (S2) ) x S(3) x ... x S(n)
The first cross product is countable, because its a cross product of two countable sets. Then
consider
S(1) x S(2) x S(3) x ... x S(n) = ( ( S(1) x (S2) ) x S(3) ) x ... x S(n)
The cross product ((S(1) x (S2)) x S(3)) is countable, because it can be considered again as
the cross product of only two sets, each of which is countable. Clearly we can continue in
this fashion, and since there are only finitely many sets, we will have proved the third
statement.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Every rational number has the form p / q, and we may assume that q is always positive.
Define the height of a rational number p / q as |p| + q. Then we can list all rational
numbers as follows:
Height 1:
0/1
But then each row is countable (in fact, finite), and there are countably many rows. Hence,
we have written the rationals as a countable union of countable sets, so that the rationals
are indeed countable. As a matter of fact, the above list contains many rational numbers
more than once. But if that set is countable, then the subset of rational numbers of that set
must also be countable, because it is certainly not finite.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.1.7(b):
The set of all polynomials that have integer coefficients and degree n is
countable.
Back
Let P(n) be the set of all polynomials with integer coefficients and degree n. Then a
particular element of P(n) is
pn(x) = anxn + an - 1xn - 1 + an - 2xn - 2 + ... + a1x + a0
Define a function f as follows:
domain of f is P(n), range of f is Z x Z x ... x Z (n+1 times)
f(pn) = f( anxn + an - 1xn - 1 + ... a1x + a0) = (an, an - 1, ..., a1, a0)
Because all coefficients are integers, this functions is onto, and is clearly one-to-one.
Hence it is a bijection between the domain and the range. But because the finite cross
product of countable sets is countable, this implies that P(n) is also countable.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.1.7(c):
The set of all polynomials with integer coefficients is countable.
Back
Let P be the set of all polynomials with integer coefficients, and define the set P(n) to be
the set of all polynomials with integer coefficients and degree n. From before we already
know that P(n) is countable. But
Hence, P is the countable union of countable sets, and must therefore be countable itself by
our result on countable unions of countable sets.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Any number x in the interval (0, 1) can be expressed as a unique, never-ending decimal.
Actually, this is not quite true: 0.1499999... is the same number as 0.15000.... But when we
simply discard those numbers with a non-ending tail of 9's we still get the open interval (0,
1), and now every number has a unique decimal representation. If these numbers were
countable, we could list them in a two-way infinite table:
1. number: x11, x21, x31, x41, ...
2. number: x12, x22, x32, x42, ...
3. number: x13, x23, x33, x43, ...
4. number: x14, x24, x34, x44, ...
...
where each expression in parenthesis represents all decimals in the decimal representation
of a particular number without the leading '0.'.
In this list, what would be the number associated to the following
element:
Let x be the number represented by (x1, x2, x3, x4, ...), where we let:
o x1 = 0 if x11 = 1 and x1 = 1 if x11 = 0
o x2 = 0 if x22 = 1 and x2 = 1 if x22 = 0
o x3 = 0 if x33 = 1 and x3 = 1 if x33 = 0
o ...
This new element x is different from the first one in our list, because they differ in their
first entry; x is different from the second one in the list, because they differ in the second
entry; x is different from the third one because they differ in the third entry, etc. But now it
is clear that x can not be in the above list, because it differs with the n-th element of that
list in the n-th entry. But this element represents a number in the interval (0, 1). Hence, we
have found that we were unable to list all numbers in (0,1), and therefore the interval is
indeed uncountable.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.2.2:
Show that all of the following sets are uncountable:
1. The open interval (-1, 1) is uncountable
2. Any open interval (a, b) is uncountable
3. The set of all real numbers R is uncountable
Back
This is a bijection between those two intervals, and therefore both intervals have the same
cardinality.
2. A similar proof can show that any open interval (a, b) is uncountable.
What is the appropriate bijection (try a linear function that maps 0 to a
and 1 to b) ?
3. Define a function
f(x) = x - / 2
Then this function is a bijection between the open intervals (0, 1) and (take the function
g(x) = tan(x)
/ 2,
/ 2). Next,
This function is a bijection between (- / 2, / 2) and R. But then the composition of the
two function will be a bijection from (0, 1) to R, and hence both sets must have the same
cardinality.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.2.4(a):
If S = {1,2,3}, then what is P(S) ? What is the power set of the set S = {1, 2,
3, 4} ? How many elements does the power set of S = {1, 2, 3, 4, 5, 6}
have ?
Back
Hence:
Therefore, we have:
card(S) = 3 and card(P(S)) = 8 = 23
For the set S = {1,2,3,4} this means:
subsets with 0 elements: 0 (the empty set)
subsets with 1 element: {1}, {2}, {3}, {4}
subsets with 2 elements: {1,2}, {1,3}, {1,4}, {2,3}, {2,4}, {3,4}
subsets with 3 elements: {1,2,3}, {1,2,4}, {1,3,4} {2,3,4}
subsets with 4 elements: S
Therefore, we have:
card(S) = 4 and card(P(S)) = 16 = 24
Finally, if S = {1,2,3,4,5,6} then, based on the above examples, we would suspect that
card(S) = 6, therefore card(P(S)) = 26 = 64
In fact, if a set S contains n elements, then its power set will contain 2n elements. This can
be proved by induction as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.2.4(b):
The cardinality of the power set of S is always greater than or equal to the
cardinality of a set S for any set S.
Back
To show that the cardinality of P(S) is greater than or equal to that of a set S we have to
find either a surjection from P(S) to S or an injection from S to P(S).
Define the function f from S to P(S) as follows:
i.e. every element s in S is mapped to the set {s} containing the single element s. Note that
the set {s} is an element of the power set P(S).
This map is clearly one-to-one, and therefore card (S) card(P(S)).
This does not prove that card(P(S)) > card(S). However, that statement
is also true, but its proof is more complicated.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
The cardinality of the power set P(S) is always bigger than the cardinality of
S for an set S.
Context
Proof:
The proof will be similar to proof about the uncountablility of the open interval (0,1): we
will attempt to list all sets of P(S) next to all elements of S, and then exhibit a set that can
not be in that list. However, we will do this in a more abstract fashion this time.
Let us denote the elements of the set S by small letters a, b, c, ..., and
the elements of P(S) by capital letters A, B, C, .... Note that each
member of P(S) is again a set in itself. Suppose there was a function f
between S and P(S). That means, that we may have the following
correspondence:
a corresponds to A, i.e. A = f(a)
b corresponds to B, i.e. B = f(b)
c corresponds to B, i.e. C = f(c)
and so on. The fact that small letters and capital letters are the same is merely for
convenience. Now define the set X as follows:
X={s
S: {s}
f(s) = 0 }
Since X contains exactly those elements that are not in its associated set,
and x is assumed to be not in X = f(x), then x must be contained in X.
This is clearly not possible, showing that our assumption that f is onto is false. Hence,
whatever map between S and P(S) exists, it can not be onto. But from a previous example
we already know that the cardinality of P(S) is at least as large as the cardinality of S, i.e.
there is a one-to-one map from S to P(S). Since this map can not also be onto, we have
proved that card(P(S)) > card(S).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
This definition seems to make sense, because a set could be an entity of its own, as well as
an element of another set. For example, we could define two sets
A = { {1}, {1,3}, A }
B = { {1}, {1,3} }
Then A is a set that is also a member of itself, whereas B is not a member of itself.
Therefore, we could consider the set of all those sets that are not members of itself. Call
this set S. The above set A would not be an element of S, whereas B is an element of S.
While this, albeit strange, does seem to make sense, we might ask:
Is S an element of itself or not ?
But this question will give a contradiction, because:
If S is an element of itself, then - since by definition S contains those sets only that
are not part of itself - S is not an element of itself. That's not possible.
If S is not an element of itself, then - since S does contain those sets that are not
part of itself - S is a member of S. That's not possible either.
Hence, we have arrived at a logical impossibility, and the set S does indeed not exist.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Two sets A and B are called equivalent if there exists a bijection between A and B.
The two sets are said to have the same cardinality, or power.
The cardinality of a set A is denoted by card(A).
The cardinal numbers of two sets are equal if the sets are equivalent.
Note that for finite sets the cardinality is just the number of elements in that set. That is,
card({a, b, 2} ) = card({1, 2, 9}) = 3. However, the concept of cardinality also applies to
infinite sets.
A cardinal number c is less than or equal to another cardinal number d if there exist
two sets A and B with card(A) = c and card(B) = d and card(A) card(B)
Note that according to this definition, we have that the cardinality of the natural numbers is
strictly less than the cardinality of the real numbers. In fact, those cardinalities have a
special name:
3. Special Cardinalities
The cardinality of the real numbers is called the cardinality (or power) of the continuum,
and is denoted by
c = card(R).
The cardinality of the natural number is called aleph null and is denoted by
= card(N)
Example:
Let A = {1,2,3} and B = {1,2}. Then card(A) = 3 and card(B) = 2. However,
card(A B) = 3
because A B = {1,2,3}. Using the above definition, we get:
A' = { (1,0), (2,0), (3,0)}
B' = { (1,1), (2,1) }
therefore
Examples 2.2.8:
We want to add or subtract the following cardinal numbers:
1. card(N) + card(N) = card(N)
2. card(N) - card(N) = undefined
3. card(R) + card(N) = card(R)
4. card(R) + card(R) = card(R)
Back
+ =
- is undefined
Using our notation for the cardinalities of the natural numbers and the continuum we can
rephrase this equation as:
c+
=c
We would actually need to show that the cardinality of A B can not be strictly larger that
the cardinality of A to establish this. That, however, is left as an exercise.
We need to find a set whose cardinality is bigger than N and less that that of R. The most
obvious candidate would be the power set of N. However, one can show that
card(P(N)) = card(R)
In fact, this is a deep question called the continuum hypothesis. This question results in
serious problems:
In the 1940's the German mathematician Goedel showed that if one denies the
existence of an uncountable set whose cardinalities is less than the cardinality of
the continuum, no logical contradictions to the axioms of set theory would arise.
One the other hand, it was shown recently that the existence of an uncountable set
with cardinality less than that of the continuum would also be consistent with the
axioms of set theory.
Hence, it seems impossible to decide this question with our usual methods of proving
theorems.
Such undecidable questions do indeed exist for any reasonably complex
logical system (such as set theory), and in fact one can even prove that
such 'non-provable' statements must exist. To read more about this
fascinating subject, look at the book Goedel's Proof or Goedel, Escher,
Bach as mentioned in the reference section of the glossary.
Can you find sets with cardinality strictly bigger than that of the
continuum ?
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.3.2(a):
Which of the following sets and their ordering relations are partially
ordered, ordered, or well-ordered:
1. S is any set. Define a b if a = b
2. S is any set, and P(S) the power set of S. Define A B if A B
3. S is the set of real numbers between [0, 1]. Define a b if a is less
than or equal to b (i.e. the 'usual' interpretation of the symbol )
4. S is the set of real numbers between [0, 1]. Define a b if a is greater
than or equal to b.
Back
Examples 2.3.2(b):
Which of the following sets are well-ordered ?
1. The number systems N, Z, Q, or R ?
Proof:
It is not quite obvious what it is that we have to prove: we need to show that if a property
holds for the smallest element of a well-ordered set S, and it holds for every successor of a
general element of that set, then it is indeed true for the whole set S.
We will prove this by contradiction: Suppose property Q is true for the
smallest element of a well-ordered set S, denoted by the symbol 1.
Suppose also that property Q is true for the successor n + 1 of the
general element n of that set, if it is assumed to be true for that
element n.
Denote by E the set of all elements from S for which property Q is not true
Since S is well-ordered, the subset E contains a smallest element, say n. If n = 1, we would
have a contradiction immediately. If n > 1, then it must have an immediate predecessor,
denoted by the element n-1.
For that element the property Q holds true, and therefore, being true for
n - 1 it must be true for (n-1) + 1 = n, by assumption. Hence, the set E
must be empty, and property Q is therefore true for all of S.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.3.4(a):
Use induction to prove the following statements:
1. The sum of the first n positive integers is n (n+1) / 2.
2. If a, b > 0, then (a + b)n an + bn for any positive integer n.
Back
Proof:
The proof goes by induction.
Property Q(n):
If x -1 then (1 + x) n 1 + nx
Check Q(1):
(1 + x) 1 + x is true.
Assume Q(n) is true:
If x -1 then (1 + x) n 1 + nx
Check Q(n+1):
(1 + x) n + 1 = (1 + x) n (1 + x) (1 + nx) (1 + x) =
1 + nx 2 + nx + x = 1 + nx 2 + (n + 1)x 1 + (n + 1)x
so Q(n+1) is true (where have we used that x -1 ?).
As an alternative proof, this statement follows directly from the binomial theorem (do it !).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 2.3.5(a):
Impose a new ordering labeled << on the natural numbers as follows:
if n and m are both even, then define n << m if n < m
if n and m are both odd, then define n << m if n < m
if n is even and m is odd, we always define n << m
Is the set of natural numbers, together with this new ordering << wellordered ? Does it have the property that every element has an immediate
predecessor ?
Back
The natural numbers, ordered by the ordering <<, could be listed in order as follows:
2, 4, 6, 8, ....., 1, 3, 5, 7, 9, ..... ,
To show it is well-ordered, take any subset A of natural numbers.
If it contains only odd numbers, then the smallest number in the usual ordering is
the smallest element of A
If it contains only even numbers, then the smallest number in the usual ordering is
the smallest element of A
If it contains both even and odd numbers, then the smallest of the even numbers in
the usual ordering is the smallest element of A
Hence, the set is well-ordered.
But, not every element has an immediate predecessor. For example, the
set:
A = {1, 3, 5, 7, ...}
has a smallest element (namely 1), but 1 does not have an immediate predecessor, since
every even number is smaller than 1 by definition.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 2.3.5(b):
Suppose the induction principle defined above does not contain the
assumption that every element except for the smallest has an immediate
predecessor. Then show that it could be proved that every natural number
must be even (which is, of course, not true so the additional assumption on
the induction principle is necessary).
Back
In other words, we assume that the induction principles was stated as follows:
Let S be a well-ordered set Then: if Q is a property such that:
1. the smallest element of S has the property Q
2. if s S has property Q then the successor of s also has property Q
Then the property Q holds for every element in S
If this principle was true, we could prove that every natural number must be even as
follows:
Consider the natural numbers with the ordering << defined as follows:
if n and m are both even, then define n << m if n < m
if n and m are both odd, then define n << m if n < m
if n is even and m is odd, we always define n << m
We have already proved that this set is well-ordered. We want to show that every number is
even. Therefore:
Q is the property that every element is even.
The smallest element of our set in the << ordering is 2, which is even.
Also, if s has property Q then so does the successor of s. That is because in our
ordering, the successor of an even number is always the next even number, never
an odd number, and if s has property Q, then s must be even.
Examples 2.3.8:
Which of the following two recursive definition is valid ?
1. Let x0 = x1 = 1 and define xn = xn - 1 + xn - 2 for all n > 1.
2. Select the following subset from the natural numbers:
o x0 = smallest element of N
o xn = smallest element of N - {x0 , x1 , x2 , ..., xn + 1}
Back
1. Suppose we let x0 = x1 = 1 and define xn = xn - 1 + xn - 2 for all n > 1. Then the first
element is uniquely determined, and each following element to be selected depends
only on the ones that have already been selected. Hence, this is a valid recursive
definition. Note that to compute, say, x5 , we need to compute all previous numbers
first. The numbers defined in this way are called Fibonacci numbers.
2. This is not a valid recursive definition, because to find the element x2, say, we need
to find the smallest element of the set N - {x0 , x1 , x2 , x3}. Since this definition
involves the element x2 itself (and also x3) which we do not know at this stage, this
recursive definition is not valid.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Exercise 2.3.9:
Is the sum of the first n square numbers equal to (n + 2)/3 ?
Back
Example 2.3.11:
The barber of Seville is that inhabitant of Seville who shaves every man in
Seville that does not shave himself
Back
This is a paradox, because who will shave the barber of Seville himself ?
if the barber does not shave himself, he - according to definition - must shave
himself, because he shaves all who do not shave themselves
if the barber does shave himself, he - according to definition - can not shave
himself, because the barber only shaves those who do not shave themselves
This is actually and example of an invalid recursive definition: In the definition the barber
appears twice: one as the defined entity "barber of Seville" and again as a member of all
men living in Seville. Such a recursive definition is invalid, and in fact leads to a
contradiction in this case.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
1. We actually have already proved this statement in example 2.3.4, but we should mention
another proof of this statement that does not use induction.
Actually, as the story goes, the young Gauss, who later became a
famous mathematician, was once in grade school. His teacher, who did
not want to teach anything, asked his students to compute the sum of
the first 100 integers, thinking that this would give him plenty of time to
read a newspaper. However, after a few seconds the young Gauss came
up with the correct answer, much to the dismay of his teacher.
How was Gauss able to find the answer so quickly ? He certainly did not
know about induction, so he used the following trick. We want to find 1
+ 2 + 3 + ... + n-1 + n. Instead of finding that sum once, we compute it
twice, calling the unknown answer S. We list the numbers once in
forward and once in backward order:
1+ 2+
3 + ... + (n-1) + n = S
n + (n-1) + (n-2) + ... + 2 + 1 = S
Each column except the last adds up to n+1, and there are n of them. The last column adds
up to 2 S, so that we have the equation:
n (n+1) = 2 S
+ (n+1)2 =
=
The last expression is exactly property Q(n+1), which finishes the induction proof.
3. We will - of course - leave the third statement as an exercise. It may
be that the last step requires some factorization that may or may not be
easy to do. Instead, try working your way 'backward'. For example,
instead of trying to factor an expression such as 2 n2 + 7n + 6, you
could start with the answer you are hoping to get, in this case (n + 2)
(2n + 3), and work your way back.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Suppose there was such an x. Being a rational number, we can write it as
x = a / b (with no common divisors)
Since x2 = x * x = 2 we have
a2 = 2 b2
In other words, a2 is even, and therefore a must be even as well. (Can you prove this ?).
Hence,
a = 2 c for some integer c.
But then we have that
4 c2 = 2 b2, or 2 c2 = b2
As before, this means that b is even.. But then both a and b are divisible by 2. That's a
contradiction, because a and b were supposed to have no common divisors.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.4.3(a):
Consider the set S of all rational numbers strictly between 0 and 1. Then this
set has many upper bounds, but only one least upper bound. That supremum
does not have to be part of the original set.
Back
An upper bound for the set S is any number that is greater than or equal to any number in
the set S. Five different upper bounds for S are, for example:
1, 10, 100, 42, and e (Euler's number)
Note that an upper bound is therefore not unique. All that is required is to find number
bigger than all other numbers in the set S.
To find the least upper bound for S we need to find a number x such that
x is an upper bound for S
there is no other upper bound lower than x
Clearly, this upper bound is 1. Note that the supremum, or least upper bound, is unique, but
it is not part of the original set S.
If we include 0 and 1 in the original set S, then the least upper bound is
again 1. This time the unique supremum is part of the set S, which may
or may not happen in general.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 2.4.3(b):
Consider the set of rational numbers {1, 1.4, 1.41, 1.414, 1.4142, ...}
converging to the square root of 2. If all we knew were rational numbers,
this set would have no supremum. If we allow real numbers, there is a
unique supremem.
Back
If we consider the universe to consist only of rational numbers, then this set does not have
a least upper bound.
No number bigger than
is the least upper bound (although each of these numbers
is an upper bound), because if x was that least upper bound, then we can find a
rational number between
and x. That rational number would then be an upper
bound smaller than x, which is a contradiction.
No number less than
is the least upper bound, because if x was that least upper
bound, there is some element of the set between x and
. But then x is not an
upper bound, which is a contradiction.
Hence, there is no supremum for this set S in Q.
If we consider this set as a subset of the real numbers, then the least
upper bound of this set is
.
Proof:
Since the above equation is not true in Q, we have to use a property of the real numbers
that is not true for the rational numbers. Define the set
S = { t R: t > 0 and t2 2 }
Our hope is that the supremum of this set should be the desired solution to the above
equation. However, we first need to make sure that the supremum indeed exists before
showing that it is the desired solution.
S is not empty, because 1 is contained in S, and S is bounded above by,
say, 5. Hence, using the least upper bound property of the real
numbers, S has a least upper bound s:
s = sup(S)
Note that this would not necessarily be true if we restricted ourselves to the rational
numbers.
Now we hope that s2 = 2, i.e. s is the desired solution. Since 1 is in S, we
know
s>1
Now s either is the solution, or one of the following two cases are true:
Is s2 < 2 ?
Let
Hence, s + is also in S, in which case s can not be an upper bound for S. This is a
contradiction, so this case is not possible.
2
Is s > 2 ?
Let
Hence, s - is another upper bound for S, so that s is not the least upper bound for
S. This is a contradiction, so that this case is not possible.
Having eliminated these two cases, we are left with s2 = 2, which is what we wanted to
prove.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.1.3(a):
The sequence
converges to zero.
Back
= {1, 1/2, 1/3, 1/4, ... }, which seems to indicate that the terms
are getting closer and closer to zero. According to the definition of
convergence, we need to show that no matter which > 0 one chooses,
the sequence will eventually become smaller than this number. To be
precise: take any > 0. Then there exists a positive integer N such that
1 / N < . Therefore, for any j > N we have:
| 1/j - 0 | = | 1/j | < 1/N <
whenever j > N. But this is precisely the definition of the sequence {1/j} converging to
zero.
While it looks like this prove is easy, it is a good indication for ' arguments' that will appear again and again. In most of those cases the
proper choice of N will make it appear as if the proof works like magic.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.1.3(b):
The sequence
Example 3.1.3(c):
The sequence
converges to zero.
Back
N = max{3, 1/ }
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Let's prove uniqueness first. Suppose the sequence has two limits, a and a'. Take any > 0.
Then there is an integer N such that:
| aj - a | <
if j > N. Also, there is another integer N' such that
| aj - a' | <
if j > N'. Then, by the triangle inequality:
| a - a' | < | a - aj + aj - a' |
|aj - a | + | aj - a' |
< + =2
if j > max{N,N'}. Hence | a - a' | < 2 for any > 0. But that implies that a = a', so that the
limit is indeed unique.
Next, we prove boundedness. Since the sequence converges, we can
take, for example, = 1. Then
| aj - a | < 1
if j > N. Fix that number N. We have that
| aj | | aj - a | + | a | < 1 + |a|
for all j > N. Define
M = max{|a1|, |a2|, ...., |aN|, (1 + |a|)}
Then | aj | < M for all j, i.e. the sequence is bounded as required.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.1.5:
We will show by induction that the sequence of Fibonacci numbers is unbounded. If that is
true, then the sequence can not converge, because every convergent sequence must be
bounded.
As for the induction process: The first terms of the Fibonacci numbers
are
{1, 1, 2, 3, 5, 8, 13, 21, ...}
We will show that the n-th term of that sequence is greater or equal to n, at least for n > 4.
Property Q(n):
xn n for all n > 4
Check Q(5) (the lowest term):
x5 = x4 + x3 = 3 + 2 = 5 5 is true.
Assume Q(n) true:
xn n for all n > 4
Check Q(n+1):
xn + 1 = xn + xn - 1 n + xn - 1 n + 1 n
Hence, by induction the Fibonacci numbers are unbounded and the sequence can not
converge.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The proofs of these statements involve the triangle inequality, as well as an occasional trick
of adding and subtracting zero, in a suitable form. A proof of the first statement, for
example, goes as follows.
Take any > 0. We know that an a, which implies that there exists an
integer N1 such that
| an - a | < / 2
| bn - b | < / 2
if n > N2. But then we know that
| (an + bn) - (a + b) | = | (an - a) + (bn - b) |
| an - a | + | b n - b |
< /2 + /2 =
if n > max(N1, N2), which proves the first statement.
Proving the second statement is similar, with some added tricks. We
know that { bn } converges, therefore there exists an integer N1 such
that
| bn | < |b| + 1
if n > N1. We also know that we can find integers N2 and N3 so that
| an - a | < / (|b| + 1)
if n > N2, and
| bn - b | < / (|a| + 1)
if n > N3, because |a| and |b| are some fixed numbers. But then we have:
| an bn - a b | = | an bn - a bn + a bn - a b |
= | bn(an - a) + a (bn - b) |
| bn| |an - a | + | a | | bn - b |
< (| b | + 1) / (|b| + 1) + | a | / (|a| +1) < 2
if n > max(N1, N2, N3), which proves the second statement.
The proof of the third statement is similar, so we will leave it as an
exercise.
The last statement does require a new trick: we will use a proof by
contradiction to get that result:
Assume that an bn for all n, but a > b.
We now need to work out the contradiction: the idea is that since a > b there is some
number c such that b < c < a.
<----------[b]-------[a]-------->
<----------[b]--[c]--[a]-------->
Since an converges to a, we can make the terms of the sequence fall between c and a, and
the terms of bn between b and c. But then we no longer have that an bn, which is our
contradiction. Now let's formalize this idea:
Let c = (a + b)/2. Then clearly b < c < a (verify!). Choose N1 such that
bn < c if n > N1. That works because b < c. Also choose N2 such that an
> c if n > N2. But now we have that
bn < c < an
for n > max(N1, N2). That is a contradiction to the original assumption that an bn for all n.
Hence it can not be true that a > b, so that the statement is indeed proved.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.1.8(a):
Is the sequence
One can start to investigate this statement without having to suspect the correct answer. We
will simply compare the quotient of two consecutive terms to check whether the answer is
greater or less than one:
(1/n) / (1/ (n+1) ) = (n+1) / n > 1
Hence, the n-th term of the sequence divided by the (n+1) term is always greater than 1, or,
in other words, the n-th term is greater than the (n+1)-th term.
That is the definition of a decreasing sequence so that the sequence is
decreasing. Checking a graphical representation of this sequence
confirms that.
Examples 3.1.8(b):
Is the sequence
We will again not guess what the correct answer might be ahead of time. We will instead
look at the difference between two consecutive terms and see if that comes out greater or
less than zero.
This means that the n-th term minus the (n+1)-th term of the sequence is less than 0, so
that the n-th term is less than then (n+1)-th term.
That means, by definition, that the sequence is increasing. Checking a
graphical representation of this sequence confirms that.
Examples 3.1.8(c):
Is it true that a bounded sequence converges ? How about monotone
increasing sequences ?
Back
Both statements are false. As a counter-example to the first statement, consider the
sequence
{ (-1) j }
Each term of this sequence is bounded by -1 or +1, so that the sequence is indeed bounded.
But, as we have seen before, the sequence does not converge.
As for the second statement, consider the simple sequence {n}, i.e. the
sequence consisting of the numbers {1, 2, 3, 4, ...}. It is obviously
increasing, but does not converge to a finite number.
It does, however, get closer and closer to infinity, but we do not, at this
time, consider this convergent.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.1.8(c):
Is it true that a bounded sequence converges ? How about monotone
increasing sequences ?
Back
Both statements are false. As a counter-example to the first statement, consider the
sequence
{ (-1) j }
Each term of this sequence is bounded by -1 or +1, so that the sequence is indeed bounded.
But, as we have seen before, the sequence does not converge.
As for the second statement, consider the simple sequence {n}, i.e. the
sequence consisting of the numbers {1, 2, 3, 4, ...}. It is obviously
increasing, but does not converge to a finite number.
It does, however, get closer and closer to infinity, but we do not, at this
time, consider this convergent.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.1.8(c):
Is it true that a bounded sequence converges ? How about monotone
increasing sequences ?
Back
Both statements are false. As a counter-example to the first statement, consider the
sequence
{ (-1) j }
Each term of this sequence is bounded by -1 or +1, so that the sequence is indeed bounded.
But, as we have seen before, the sequence does not converge.
As for the second statement, consider the simple sequence {n}, i.e. the
sequence consisting of the numbers {1, 2, 3, 4, ...}. It is obviously
increasing, but does not converge to a finite number.
It does, however, get closer and closer to infinity, but we do not, at this
time, consider this convergent.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
If
is a monotone decreasing sequence that is bounded below, then
the sequence must converge (see picture).
Context
Proof:
Let's look at the first statement, i.e. the sequence in monotone increasing. Take an > 0
and let c = sup(xk). Then c is finite, and given > 0, there exists at least one integer N such
that xN > c - . Since the sequence is monotone increasing, we then have that
xk > c for all k > N, or
| c - xk | <
for all k > N. But that means, by definition, that the sequence converges to c.
The proof for the infimum is very similar, and is left as an exercise.
Examples 3.1.10(a):
The sequences
and
both converge.
Back
. It is decreasing because:
Examples 3.1.10(b):
Define x1 = b and let xn = xn - 1 / 2 for all n > 1. Then this sequence converges
for any number b.
Back
The proof is very easy using the theorem on monotone, bounded sequences:
b > 0: the sequence is decreasing and bounded below by 0.
b < 0: the sequence is increasing and bounded above by 0
b = 0: the sequence is constantly equal to zero
In either case the sequence converges. As to finding the actual limit, we proceed as
follows: we already know that the limit exists. Call that limit L. Then we have:
lim xn = L = lim xn + 1
Before giving the proof, let's see how this recursive sequence can be used to compute a
square root very efficiently. Let's say we want to compute
. Let's start with x0 = 2. Note
that the 'true' value of
with 12 digits after the period is 1.414213562373. Our recursive
sequence would approximate this value quickly as follows:
Term
Exact Value
Approximate Value
x0
2
1
1.5
17
1.416666667
577
/408
1.414215686
1.414213562375
After only 4 steps our sequence has approximated the 'true' value of
with 11 digits accuracy.
Now that we have seen the usefulness of this particular recursive
sequence we need to prove that it really does converge to the square
root of a.
First, note that xn > 0 for all n so that the sequence is bounded below.
Next, let's see if the sequence is monotone decreasing, in which case it
would have to converge to some limit. Compute
xn =
xn+1 =
Proof:
The statement of the theorem is easiest to memorize by looking at a diagram:
All bj are between aj and cj, and since aj and cj converge to the same limit L the bj have no
choice but to also converge to L.
Of course this is not a formal proof, so here we go: we want to show that
given any > 0 there exists an integer N such that | bj - L | < if j > N.
We know that
aj bj cj
Subtracting L from these inequalities gives:
aj - L bj - L cj - L
But there exists an integer N1 such that | aj - L | < or equivalently
- < aj - L <
and another integer N2 such that | cj - L | < or equivalently
- < cj - L <
if j > max(N1, N2). Taking these inequalities together we get:
- < aj - L bj - L cj - L <
But that means that
- < bj - L <
or equivalently | bj - L | < as long as j > max(N1, N2). But that means that {bj} converges
to L, as required.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.1.12:
Show that the sequence sin(n) / n and cos(n) / n both converge to zero.
Back
This might seem difficult because trig functions such as sin and cos are
often tricky. However, using the Pinching theorem the proof will be very
easy.
We know that | sin(x) | 1 for all x. Therefore
-1 sin(n) 1
for all n. But then we also know that:
-1/n sin(n)/n 1/n
The sequences {1/n} and -1/n both converge to zero so that the Pinching theorem applies
and the term in the middle must also converge to zero.
To prove the statement involving the cos is similar and left as an
exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Let
be a sequence of real numbers. The sequence is
Cauchy if and only if it converges to some limit a.
Context
Proof:
The proof of the first statement follows closely the proof of the corresponding result for
convergent sequences. Can you do it ?
To prove the second, more important statement, we have to prove two
parts:
First, assume that the sequence converges to some limit a. Take any >
0. There exists an integer N such that if j > N then | aj - a | < /2. Hence:
| aj - ak | | aj - a | + | a - ak| < 2 / 2 =
if j, k > N. Thus, the sequence is Cauchy.
Second, assume that the sequence is Cauchy (this direction is much
harder). Define the set
S = {x
Since the sequence is bounded (by part one of the theorem), say by a constant M, we know
that every term in the sequence is bigger than -M. Therefore -M is contained in S. Also,
every term of the sequence is smaller than M, so that S is bounded by M. Hence, S is a
non-empty, bounded subset of the real numbers, and by the least upper bound property it
has a well-defined, unique least upper bound. Let
a = sup(S)
We will now show that this a is indeed the limit of the sequence. Take any > 0 , and
choose an integer N > 0 such that
| a j - ak | < / 2
if j, k > N. In particular, we have:
| a j - aN + 1 | < / 2
if j > N, or equivalently
- / 2 < a j - aN + 1 < / 2
Hence we have:
aj > aN + 1 - / 2
for j > N. Thus, aN + 1 - / 2 is in the set S, and we have that
a aN + 1 - / 2
It also follows that
aj < aN + 1 < / 2
for j > N. Thus, aN + 1 < / 2 is not in the set S, and therefore
a aN + 1 < / 2
But now, combining the last several line, we have that:
|a - aN + 1 | < / 2
and together with the above that results in the following:
| a - aj | < |a - aN + 1 | + | aN + 1 - aj | < 2 / 2 =
for any j > N.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.3.2(a):
Take the sequence
. Extract every other member, starting with the
first. Then do the same, starting with the second.
Back
Examples 3.3.2(b):
Take the sequence
. Extract three different subsequences of your
choice and look at the convergence behavior of these subsequences.
Back
If
is a sequence such that every possible subsequence
extracted from that sequences converge to the same limit,
then the original sequence also converges to that limit.
Context
Proof:
The first statement is easy to prove: Suppose the original sequence {aj} converges to some
limit L. Take any sequence nj of the natural numbers and consider the corresponding
subsequence of the original sequence For any > 0 there exists an integer N such that
| an - L | <
as long as n > N. But then we also have the same inequality for the subsequence as long as
nj > N. Therefore any subsequence must converge to the same limit L.
The second statement is just as easy. Suppose {aj} is a sequence such
that every subsequence extracted from it converges to the same limit L.
Now take any > 0. Extract from the original sequence every other
element, starting with the first. The resulting subsequence converges to
L by assumption, i.e. there exists an integer N such that
| aj - L | <
where j is odd and j > N. Now extract every other element, starting with the second. The
resulting subsequence again converges to L, so that
| aj - L | <
where j is even and j > N. But now we take any j, even or odd, and assume that j > N
if j is odd, then | aj - L | < because aj is part of the first subsequence
if j is even, then | aj - L | < because aj is part of the second subsequence
Hence, the original sequence must also converge to L.
Note that we can see from the proof that if the "even" and "odd"
subsequence of a sequence converge to the same limit L, then the full
sequence must also converge to L. It is not enough to just say that the
"even" and "odd" subsequence simply converge, they must converge to
the same limit.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
a subsequence
that converges.
Context
Proof:
Since the sequence is bounded, there exists a number M such that | aj | < M for all j. Then:
either [-M, 0] or [0, M] contains infinitely many elements of the sequence
Say that [0, M] does. Choose one of them, and call it
either [0, M/2] or [M/2, M] contains infinitely many elements of the
(original) sequence.
Say it is [0, M/2]. Choose one of them, and call it
either [0, M/4] or [M/4, M/2] contains infinitely many elements of the
(original) sequence
This time, say it is [M/4, M/2]. Pick one of them and call it
Keep on going in this way, halving each interval from the previous step
at the next step, and choosing one element from that new interval. Here
is what we get:
| < M / 2k-1
Now we can see the choice for N: we want to make is so large, such that whenever k, m >
N, the difference between the members of the subsequence is less than the prescribed .
What is therefore the right choice for N to finish the proof ?
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.3.5(a):
The sequence
subsequence.
Since | sin(x) | < 4, the sequence is clearly bounded above and below (the sequence is also,
of course, bounded by 1).
Therefore, using the Bolzano-Weierstrass theorem, there exists a
convergent subsequence.
However, it is nearly impossible to actually list this subsequence. The
Bolzano-Weierstrass theorem does guaranty the existence of that
subsequence, but says nothing about how to obtain it.
The original sequence { sin(j) }, incidentally, does not converge. The
proof of this is not so easy, but if we assume that the second part of this
example has been proved, it would be easy. Remember that the second
part of this example states that given any number L with |L| < 1 there
exists a subsequence of
that converges to L. If that was true
the original sequence can not converge, because otherwise all its
subsequences would have to converge to the same limit.
Of course this proof is only valid if this - more complicated - statement
can be proved.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.3.5(b):
Given any number L between -1 and 1, it is possible to extract a
subsequence from the sequence
that converges to L.
Examples 3.4.2(a):
?
Back
Clearly, the infimum of the sequence is -1, and the supremum is +1. To find lim inf and lim
sup, we will first find the sequence of numbers Aj and Bj mentioned in the definition.
Let's find the numbers Aj = inf{aj, aj + 1, aj + 2, ...} for the sequence {-1, 1,
-1, 1, ...}.
= -1
=1
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Back
For this statement the Bolzano-Weierstrass theorem does not help at all. We need to
proceed differently - but not at this time, sorry.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.4.2(b):
What is inf, sup, lim inf and lim sup for
?
Back
Since this sequence is {1, 1/2, 1/3, 1/4, ...} the infimum is zero, while the supremum is 1.
As for lim inf and lim sup, we find first the sequence of numbers Aj and Bj mentioned in the
definition.
A1 = inf{1, 1/2, 1/3, 1/4, ...} = 0
A2 = inf{1/2, 1/3, 1/4, 1/5, ...} = 0
A3 = inf(1/3, 1/4, 1/5, 1/6, ...} = 0
lim inf
=0
lim sup
=0
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 3.4.2(c):
What is inf, sup, lim inf and lim sup for
Back
This sequence is {-1, 2, -3, 4, -5, 6, -7, ...}. You can quickly check, by looking at the
definition of lim inf and lim sup and working out the numbers Aj and Bj that:
inf { (-1) j j } = lim inf { (-1) j j } = sup{ (-1) j j } =
lim sup{ (-1) j j } =
Note that while lim sup and lim inf are not real numbers, they are uniquely defined as plus
or minus infinity. The limit of the original sequence, on the other hand, does not exist at all.
Hence, there is a difference between a limit not existing, and a limit that
approaches infinity. In the latter sense, lim inf and lim sup will always
exist, which is their most useful property.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The sequence
Aj = inf{aj , aj + 1 , aj + 2 , ...}
is monotone increasing (which you should prove yourself). Hence, lim inf exists (possibly
positive infinity).
The sequence
Bj = sup{aj , aj + 1 , aj + 2 , ...}
is monotone decreasing (which you should prove yourself). Hence, lim sup exists (possibly
negative infinity).
Here we have to allow for a limit to be positive or negative infinity,
which is different from saying that a limit does not exist.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
lim sup
If c and d are both finite, then: given any > 0 there are arbitrary large j
such that aj > c - and arbitrary large k such that ak < d + .
Context
Proof:
First let's assume that c = lim sup{aj} is finite, which implies that the sequence {aj} is
bounded. Recall the properties of the sup (and inf) for sequences:
If a sequence is bounded above, then given any > 0 there exists at least
one integer k such that ak > c Now take any > 0. Then
Ak = sup{ak, ak+1, ...}
so by the above property there exists an integer jk > k such that
Ak >
or equivalently
> Ak - / 2
| Ak -
|< /2
We also have by definition that Ak converges to c so that there exists an integer N such that
| Ak - c | < / 2
But now the subsequence {
|
-c|=|
- Ak + Ak - c |
|
- Ak | + | Ak - c |
< /2+ /2= /2
if jk > N. Hence, this particular subsequence of {an} converges to c.
The proof to find a subsequence converging to the lim inf is similar and
is left as an exercise.
Statement (3) is pretty simple to prove: For any sequence we always
have that
inf{ak, ak+1, ... } sup{ak, ak+1, ... }
Taking limits on both sides gives lim inf(an) lim sup(an) for any sequence, so it is true in
particular for any subsequence.
Next take any subsequence of {an}. Then:
inf(ak, ak+1, ...) inf(
, ...)
because an infimum over more numbers (on the left side) is less than or equal to an
infimum over fewer numbers (on the right side). But then
d lim inf(
lim sup
} that converges to c
-c|<
<c+
as long as k > N. But that of course means that there are arbitrarily large indices - namely
those jk for which k > N - with the property that
> c - as required. Hence, we have
shown the last statement involving the lim sup, and a similar proof would work for the lim
inf.
All our proofs rely on the fact that the lim sup and lim inf are bounded. It
is not hard to adjust them for unbounded values, but we will leave the
details as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 3.4.5:
If
is the sequence of all rational numbers in the interval [0, 1],
enumerated in any way, find the lim sup and lim inf of that sequence.
Back
Since the numbers 1 and 0 are itself rational numbers, it is clear that
sup{ an } = 1 and inf{ an } = 0
Therefore, we already know that
0 lim inf{ an } lim sup{ an } 1
To find the lim sup, we will construct a subsequence that converges to 1:
These numbers exist because the rational numbers in the interval [0, 1] are
arbitrarily close to any real number in that interval, according to the Density
principle.
The subsequence { ajk } constructed in the above way
converges to 1. We already know that any limit of any
convergent subsequence must be less than or equal to 1.
Therefore, since the lim sup is the greatest limit of any
convergent subsequence, we have
lim sup { an } = 1
Similarly, we can extract a subsequence { ajk } that converges to 0. We also
know that every limit of any convergent subsequence must be greater or
equal to zero. Therefore, since the lim inf is the smallest possible limit of all
convergent subsequence, we have:
lim inf { an } = 0
Proof:
Let c = lim sup aj. From before we know that there exists a subsequence of {aj} that
converges to c. But since the original sequence converges, every subsequence must
converge to the same limit. Hence
c = lim sup aj = lim aj
To prove that lim inf aj = c is similar.
The converse of this statement can be proved by noting that
Bj = inf(aj, aj+1, ...) aj sup(aj, aj+1, ...) = Aj
Noting that lim Bj = lim inf(aj) = lim sup(aj) = lim Aj we can apply the Pinching Theorem
to see that the terms in the middle must converge to the same value.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
This seems an obvious statement: if a number is in absolute values less than one, it gets
smaller and smaller when raised to higher and higher powers. Proving something 'obvious',
however, is often difficult, because it may not be clear how to start. To prove the statement
we have to resort to one of the elementary properties of the real number system: the
Archimedian principle.
Case a > 1:
Take any real number K > 0 and define
x=a-1
Since a > 1 we know that x > 0. By the Archimedian principle there exists a
positive integer n such that nx > K - 1. Using Bernoulli's inequality for that n we
have:
an = (1 + x)n 1 + nx > 1 + (K - 1) = K
But since K was an arbitrary number, this proves that the sequence {an} is
unbounded. Hence it can not converge.
Case 0 < a < 1:
Take any > 0. Since 0 < a < 1 we know that 1/a > 1, so that by the previous proof
we can find an N with
Proof:
Write n a = e a ln(n). Then:
if a > 0 then as n approaches infinity, the function ea ln(n) approaches infinity as well
if a < 0 then as n approaches infinity, the function ea ln(n) approaches zero
if a = 0 then the sequence is the constant sequence, and hence convergent
Is this a good proof ? Of course not, because at this stage we know nothing about the
exponential or logarithm function. So, we should come up with a better proof.
But actually we first need to understand exactly what na really means:
If a is an integer, then clearly na means to multiply n with itself a times
If a = p/q is a rational number, then np/q means to multiply n p times, then take the
q-root
It is unclear, however, what na means if a is an irrational number. For example,
what is n , or n ?
One way to define na for all a is to resort to the exponential function:
na = ea ln(n)
In that light, the original 'proof' was not bad after all, but of course we now need to know
how exactly the exponential function is defined, and what its properties are before we can
continue. As it turns out, the exponential function is not easy to define properly. Generally
one can either define it as the inverse function of the ln function, or via a power series.
Another way to define na for a being rational it to take a sequence of
rational numbers rn converging to a and to define na as the limit of the
sequence {nrn}. There the problem is to show that this is well-defined,
i.e. if there are two sequences of rational numbers, the resulting limit
will be the same.
In either case, we will base our proof on the simple fact:
if p > 0 and x > y then xp > yp
which seems clear enough and could be formally proved as soon as the exponential
function is introduced to properly define xp.
Now take any positive number K and let n be an integer bigger than K1/a.
Since
n > K1/a
Proof:
If n > 1, then
0 an
But that implies that an converges to zero as n approaches to infinity, which
means, by the definition of an that
converges to 1 as n goes to infinity.
That is what we wanted to prove.
Proof:
Case a > 1:
If a > 1, then for n large enough we have 1 < a < n. Taking roots on both sides we
obtain
1<
<
But the right-hand side approaches 1 as n goes to infinity by our statement of the
root-n sequence. Then the sequence {
} must also approach 1, being squeezed
between 1 on both sides (Pinching theorem).
Case 0 < a < 1:
If 0 < a < 1, then (1/a) > 1. Using the first part of this proof, the reciprocal of the
sequence {
} must converge to one, which implies the same for the original
sequence.
Incidentally, if a = 0 then we are dealing with the constant sequence, and the limit is of
course equal to 0.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Note that both numerator and denominator tend to infinity. Our goal will be to show that
the denominator grows faster than the k-th power of n, thereby 'winning' the race to infinity
and forcing the whole expression to tend to zero.
The name of this sequence indicates that we might try to use the
binomial theorem for this. Indeed, define x such that
b=1+x
Since b > 1 we know that x > 0. Therefore, each term in the binomial theorem is positive,
and we can use the (k+1)-st term of that theorem to estimate:
But this expression is true for all n > 2k + 1 as well, so that, with k fixed, we can take the
limit as n approaches infinity and the right hand side will approach zero. Since the lefthand side is always greater than or equal to zero, the limit of the binomial sequence must
also be zero.
If k is replaced by any real number, see the exponential sequence to
find out how nk could be defined for rational and irrational values of k.
But perhaps some simple estimate will help if k is not an integer. Details
are left as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Euler's sequence
Proof:
We will show that the sequence is monotone increasing and bounded above. If that was
true, then it must converge. Its limit, by definition, will be called e for Euler's number.
Euler's number e is irrational (in fact transcendental), and an
approximation of e to 30 decimals is e ~
2.71828182845904523536028747135.
First, we can use the binomial theorem to expand the expression
The first expression has (n+1) terms, the second expression has (n+2) terms. Each of the
first (n+1) terms of the second expression is greater than or equal to each of the (n+1)
terms of the first expression, because
But then the sequence is monotone increasing, because we have shown that
Next, we need to show that the sequence is bounded. Again, consider the expansion
1+
If we define Sn =
, then
so that, finally,
for all n.
But then, putting everything together, we have shown that
1+
1 + Sn 3
Proof:
We will use a simple substitution to prove this. Let
x/n = 1/u, or equivalently, n = u x
Then we have
But the term inside the square brackets is Euler's sequence, which converges to Euler's
number e. Hence, the whole expression converges to ex, as required.
In fact, we have used a property relating to functions to make this proof
work correctly. What is that property ?
If we did not want to use functions, we could first prove the statement
for x being an integer. Then we could expand it to rational numbers, and
then, approximating x by rational number, we could prove the final
result.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
10 meters
t=1
5 = 10 / 2 meters
t = 1 + 1/2
2.5 = 10 / 4 meters
t = 1 + 1/2 + 1/4
1.25 = 10 / 8 meters
The tortoise runs 5 meters per second, and has an advantage of 10 meters.
Therefore, it also reaches the 20 meter mark after 2 seconds
Therefore, both are even after 2 seconds
Of course, Achilles will finish the race after 10 seconds, while the tortoise needs 18
seconds to finish, and Achilles will clearly win.
The problem with Zeno's paradox is that Zeno was uncomfortable with
adding infinitely many numbers. In fact, his basic argument was: if you
add infinitely many numbers, then - no matter what those numbers are you must get infinity. If that was true, it would take Achilles infinitely
long to reach the tortoise, and he would loose the race. However,
reducing the infinite addition to the limit of a sequence, we have seen
that this argument is false.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.1.3(a):
The series
= 1 + 1/2 + 1/3 + 1/4 + 1/5 + ... is called harmonic series.
It diverges to infinity.
Context
Proof:
We need to estimate the n-th term in the sequence of partial sums.
Examples 4.1.5(a):
The series
S n = -1 + 1 - 1 + 1 ... - 1 + 1 = 0
if n is even.
Examples 4.1.5(b):
The series
converges absolutely.
Back
and this series converges, as shown before. It is in fact a special case of the geometric
series.
Hence, the series of absolute values converges, which by definition
means that our series converges absolutely. Since absolute converge
implies 'regular' convergence, the series also converges in its original
form.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
The series
is called the Alternating Harmonic series. It converges
but not absolutely, i.e. it converges conditionally.
Context
Proof:
and
We have that
S 2n+2 - S 2n = 1 / (2n+1) - 1 / (2n+2) > 0
and
S 2n+3 - S 2n+1 = - 1 / (2n+2) + 1/ (2n+3) < 0
which means for the two subsequences
{ S 2n } is monotone increasing and { S 2n+1 } is monotone decreasing
For each sequence we can combine pairs to see that
S 2n 1 and S 2n+1 0
for all n. Hence, both subsequences are monotone and bounded and must therefore be
convergent. Define their limits as
lim S 2n = L and lim S 2n+1 = M
Then
| M - L | = | lim (S 2n+1 - S 2n) | = 1 / (2n+1)
which converges to zero. Therefore, M = L, i.e. both subsequences converge to the same
limit. But this common limit is the same as the limit of the full sequence, because: given
any > 0 we have
there exists an integer N such that | L - S 2n | < if n > N
there exists an integer M such that | L - S 2n+1 | < if n > M
Now set K = max(N, M). Then, for the above > 0 we have
|L-Sn|<
for n > K because n is either even or odd. Hence, the alternating harmonic series converges
conditionally.
If a series
converges absolutely, it converges in the ordinary sense.
The converse is not true.
Back
Suppose
If a series
converges absolutely, it converges in the ordinary sense.
The converse is not true.
Back
Suppose
Let
be an absolutely convergent series. Then any rearrangement of
terms in that series results in a new series that is also absolutely convergent
to the same limit.
Let
be a conditionally convergent series. Then, for any
real number c there is a rearrangement of the series such
that the new resulting series will converge to c.
Context
Proof:
Suppose
converges. In particular, it is bounded, i.e. |Sn| < K for some number K. If we take any
rearrangement of terms in the series and form a new sequence of partial sums:
Tn = |
|+|
| + ... |
then Tn is again bounded by the same number K. But since all terms in the partial sum Tn
are positive the sequence is monotone increasing. Therefore {Tn} is monotone increasing
and bounded and must therefore converge.
It remains to show that the limit of the rearrangement is the same as
the limit of the original series. That is left as an exercise.
Finally suppose the series
converges conditionally. Let's first collect
a few facts:
By the divergence test (which we will prove later) we know that the sequence of
general terms an converges to zero.
The series does not converge absolutely.
Since the sequence of partial sums of absolute values is increasing it means that the
series of absolute values must "converge" to positive infinity.
There must be infinitely many positive terms among the aj. If we call them bj and
collect them to form a series then that new series must "converge" to positive
infinity (why?).
There must be infinitely many negative terms among the aj. If we call them cj and
collect them to form a series then that new series must "converge" to negative
infinity (why?).
Now we can describe the idea of the proof, leaving the details as an exercise:
Collect enough bj so that they add up to a number just bigger than c
Add as many terms from cj as is necessary to make the resulting sum just less than
c
Add again terms from bj to be just bigger than c, then again terms from cj to be less
than c, and so on.
Then you can show that the resulting arrangment of bj's and cj indeed
forms a series that converges to c. As usual, the details of this proof are
left as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
This is a simple numeric computation, involving some trial and error. First, let's collect
positive odd terms so that they add up to something larger than 2:
1 + 1/3 + ... + 1/15 = 2.021800422
Examples 4.1.7(b):
that
Back
Let
and
be two absolutely convergent series. Then:
1. The sum of the two series is again absolutely convergent. Its limit is
the sum of the limit of the two series.
2. The difference of the two series is again absolutely convergent. Its
limit is the difference of the limit of the two series.
3. The product of the two series is again absolutely convergent. Its limit
is the product of the limit of the two series (Cauchy Product ).
Context
Proof:
The proof of the first statement is a simple application of the triangle inequality. Let
An = |a1| + |a2| + ... + |an|
and
Bn = |b1| + |b2| + ... + |bn|
Assume that An converges to A and Bn converges to B. Then
Sn = |a1 + b1| + |a2 + b2| + ... + |an + bn|
|a1| + |a2| + ... + |an| + |b1| + |b2| + ... + |bn|
A+B
Therefore the sequence {Sn} is bounded above by A + B. The sequence is also monotone
increasing so that it must converge to some limit. To find the limit, note that
| (a1 + b1) + (a2 + b2) + ... (an + bn) - (A + B)|
| a1 + a2 + ... + an - A | + | b1 + b2 + ... + bn - B |
< /2+ /2
if n is big enough. Therefore the sequence {Sn} converges to A + B, as required.
The proof for the difference of sums is similar. The proof for the Cauchy
product, on the other hand, is much more complicated and will be given
in the statement on the Cauchy Product.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
The series
|<
Context
Proof:
Suppose that the Cauchy criterion holds. Pick any > 0. Then
|Sn-Sm|=|
|<
But that means precisely that the sequence of partial sums { S N } is a Cauchy sequence,
and hence convergent.
Now suppose that the sum converges. Then, by definition, the sequence
of partial sums converges. In particular, that sequence must be a
Cauchy sequence: given any > 0, there is positive integer N such that
whenever n, m > N we have that
| S n -S m | = |
|<
But that, in turn, means that the Cauchy criterion for series holds.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.2:
Does the Divergence test apply to show that the series
diverges ? How about the series
converges or
?
Back
The first series converges: to apply the divergence test, we have to consider the sequence
which we have shown to diverge. In particular, the sequence does not converge to zero.
Hence, the above series can not converge.
One can easily show directly that this series does not converge by
looking at the sequence of partial sums.
For the second series, the divergence test does not apply. In order to
use the divergence test, we need to check the limit of the sequence
{ 1/n }
But this limit is indeed zero. Therefore, the divergence test does not give any information
about this series.
In fact, this series is the harmonic series, and it can be shown to
diverge.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.4(a):
The series
converges.
Back
But
<
Examples 4.2.4(b):
The series
diverges.
Back
=
Therefore, the series diverges.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.6:
Use the limit comparison test together with the results on p-series to
investigate the following series:
1.
2.
3. If r(n) = p(n) / q(n), where p and q are polynomials in n, can you
find general criteria for the series p(n) to converge or diverge ?
Back
Hence, both series have the same convergence behavior, and since the p-series
n 2 converges, so does the original series.
1/
, we want to limit-
Hence, both series have the same convergence behavior, and since the p-series
diverges, so does the original series.
1/
p(n)
The last series is left as an exercise. Here are some hints:
The p-series test tells you the convergence behavior of 1 / n k for different
k.
Check the limit of an expression like n k r(n) by comparing the degrees of
numerator and denominator
Depending on the answers above, use the limit comparison to find the
behavior of the original series, based on the degrees of numerator and
denominator.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.8:
Use the Cauchy Condensation criteria to answer the following questions:
1. In the sum
, list the terms a 4, a k, and a 2 k. Then show that this
series (called the harmonic series) diverges.
2. For which p does the series
converge or diverge ?
Back
a 4 = 1/4
a k = 1/k
a 2k = 1 / 2 k
and the last series diverges by the Divergence test. Hence, the original series also diverges.
Next, we investigate the series
for various p:
converges to infinity. Hence, the series diverges
=
The right hand series is now a Geometric Series, so that:
o
o
But this is exactly what we need, in conjunction with Cauchy's Condensation test, to finish
the proof of the statement.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.10:
Investigate the convergence properties of the following series:
1. What is the actual limit of the sum
2. What is the actual limit of the sum
3. Does the sum
?
?
converge ?
Back
1+
= 1 / (1 - 1/2)
= 1 / (1 - 1/2) - 1 = 1.
= 1 / (1 - 3/4) = 4
we have for our series:
Examples 4.2.12(a):
The series
diverges.
Back
This series is called harmonic series, and is a p-series with p = 1. Hence, by the p-series
test, the series diverges.
One can also proof divergence of this series directly by looking at the
limit of partial sums.
Examples 4.2.12(b):
The series
converges.
Back
Examples 4.2.12(c):
The series
diverge.
Back
and
1/
have the same convergence behavior. But the second series is a p-series with p = 1/2, so it
diverges by the p-series test. Hence, the original series also diverges.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.14(a):
lim sup
= lim sup 1 /
=1
by our result on the limit of the n-th root sequence. Hence, the root test, applied to the
above series, is inconclusive.
However, since this series is the harmonic series, we can prove either
directly that it diverges or apply the p-series test to show divergence.
Note that the root test also does not apply to the alternating harmonic
series, which turns out to converge conditionally.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.14(b):
To apply the root test, we have to compute the limit superior of the expression
(n2)1/n = n2/n = (n1/n)2
By our result on the limit of the n-th root sequence, the lim sup of this expression is equal
to 1. Hence, the root test is inconclusive.
However, the above series is a p-series with p = 2, and therefore
converges by the p-series test.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.14(c):
The series
By our result on the limit of the n-th root sequence, the lim sup of this expression is 1/2.
Hence, by the root test the above series converges absolutely.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
The series
Hence, the series converges by the ratio test. However, this test says nothing about the
actual limit of the series. We have to be a lot more sophisticated to find the actual limit of
this series.
Recall that we have defined Euler's number as the limit of the Euler
sequence. The proof that the above sum equals that limit is very similar
to the proof that the Euler sequence converges in the first place.
First, we use the binomial theorem to expand the expression
because each term in parenthesis is smaller than one. On the other hand, we also have:
for N < n, because each term in parenthesis is greater than zero. But then, taking the limit
as n approaches infinity, we have:
Hence, taking the limit as N approaches infinity, we have the sum squeezed in between the
limit of Euler's sequence, which we know is equal to e. That proves the result.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.16(b):
The ratio test as well as the root test show that the series
converges.
Back
Clearly, the lim sup of this expression is equal to 1/2. Hence, by the ratio test the series
converges absolutely.
The root test for this series, of course, gives the same result. However, it
is much easier to compute the limit of the above expression rather than
computing it for the expression resulting in the root test.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.16(c):
The following statements are not equivalent:
There exists an N such that | an+1 / an | 1 for all n > N
lim sup | an+1 / an | 1
In fact, the first statement implies the second, but not the other way around.
Back
a n = 2 if n is even
a n = 1 if n is odd
Then
a n+1 / a n = 1 / 2 if n is even
a n+1 / a n = 2 / 1 if n is odd
Therefore, the second statement above does not imply the first one.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.16(d):
Here is an alternative proof of the ratio test that uses the root test directly.
Note that this shows that the root test is 'better' than the ratio test, because
the ratio test can be deduced from the root test.
Back
Proof:
First, recall the ratio test
if lim sup | a n+1 / a n | < 1 then the series converges absolutely.
and the root test
if lim sup | a n | 1/n < 1 then the series converges absolutely.
We will use the root test to prove the ratio test
Assume that lim sup | a n+1 / a n | < 1. Using the properties of the limit
superior, there exists a number c with 0 < c < 1, such that
| a n+1 / a n | < c
for n > N or equivalently,
| a n+1 | < c | a n |
for n > N. Then, for any positive integer k, we have that:
| a N+k | c | a N+k-1 | c ( c | a N+k-2 |) ... c k |a N |
for all k > 1. Making the substitution N + k = n, for n > N, this is equivalent to:
| a n | c n-N | a N |
or
Taking the lim sup on both sides as n approaches infinity (N is fixed) we obtain, using the
result on the n-th Root Sequence:
lim sup | a n | 1/n c < 1
Hence, the sequence satisfies to root test, and therefore the series converges absolutely.
The proof of divergence is left as an exercise.
Examples 4.2.18(a):
The sum
converges conditionally.
Back
Examples 4.2.18(b):
The series
converges conditionally.
Back
cos(t) + i sin(t) = e it
for all real numbers t, where i is the basic complex number given as the square root of -1.
Taking this formula for granted, we have:
( cos(1) + i sin(1) ) + ( cos(2) + i sin(2) ) + ... + ( cos(n+1) + i sin(n+1) )
=
= e i + e 2i + ... + e (n+1)i =
= e i ( 1 + ( e i ) + ( e i ) 2 + ... + ( e i ) n ) =
Therefore, taking absolute values on both sides we can estimate the last expression using
the triangle inequality and the fact that | e it | = 1 to obtain:
for all n. Finally, using the fact that | a | | a + i b | for all real numbers a, b, we have
Putting the last two expressions together, we see that the sequence of partial sums
| cos(1) + cos(2) + ... + cos(n) |
is bounded for all n. Therefore, both conditions of Abel's test are verified, and hence the
original series converges.
Of course, this is an 'unfair' proof, since it uses complex analysis and we
have not even defined properly what a complex number is. Therefore,
another proof of this example is necessary that uses only real analysis.
Such a proof is left as an exercise, as well as the proof that the series
does not converge absolutely (not easy).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
. Let S N =
be the n-th
Context
Proof
The proof is simply a calculation, where the various sums are carefully reindexed:
Examples 4.2.20(a):
Proof:
There are many proofs of this fact. For example. the series of absolute values is a p-series
with p = 1, and diverges by the p-series test. The original series converges, because it is an
alternating series, and the alternating series test applies easily. However, here is a more
elementary proof of the convergence of the alternating harmonic series.
We already know that the series of absolute values does not converge
by a previous example. Hence, the series does not converge absolutely.
As for regular convergence, consider the following two partial sums:
We have that
S 2n+2 - S 2n = 1 / (2n+1) - 1 / (2n+2) > 0
S 2n+3 - S 2n+1 = - 1 / (2n+2) + 1/ (2n+3) < 0
which means for the two subsequences
{ S 2n } is monotone increasing
{ S 2n+1 } is monotone decreasing
For each sequence we can combine pairs to see that
S 2n 1 for all n
S 2n+1 0 for all n
Hence, both subsequences are monotone and bounded, and must therefore be convergent.
Define their limits as
lim S 2n = L and lim S 2n+1 = M
Then
| M - L | = | lim (S 2n+1 - S 2n) | = 1 / (2n+1)
which converges to zero. Therefore, M = L, i.e. both subsequences converge to the same
limit. But this common limit is the same as the limit of the full sequence, because: given
any > 0 we have
there exists an integer N such that | L - S 2n | < if n > N
there exists an integer M such that | L - S 2n+1 | < if n > M
Now set K = max(N, M). Then, for the above > 0 we have
|L-Sn|<
for n > K, because n is either even or odd. Hence, the alternating harmonic series
converges conditionally.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 4.2.20(b):
The series
converges conditionally.
Back
= e i ( 1 + ( e i ) + ( e i ) 2 + ... + ( e i ) n ) =
Therefore, taking absolute values on both sides we can estimate the last expression using
the triangle inequality and the fact that | e it | = 1 to obtain:
for all n. Finally, using the fact that | a | | a + i b | for all real numbers a, b, we have
Putting the last two expressions together, we see that the sequence of partial sums
| cos(1) + cos(2) + ... + cos(n) |
is bounded for all n. Therefore, both conditions of Abel's test are verified, and hence the
original series converges.
Of course, this is an 'unfair' proof, since it uses complex analysis and we
have not even defined properly what a complex number is. Therefore,
another proof of this example is necessary that uses only real analysis.
Such a proof is left as an exercise, as well as the proof that the series
does not converge absolutely (not easy).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Let { U n } be a collection of open sets, and let U = U n. Take any x in U. Being in the
union of all U's, it must be contained in one specific U n. Since that set is open, there exists
a neighborhood of x contained in that specific U n. But then that neighborhood must also be
contained in the union U. Hence, any x in U has a neighborhood that is also in U, which
means by definition that U is open.
To prove the second statement, simply use the definition of closed sets
and de Morgan's laws.
Now let U n, n=1, 2, 3, ..., N be finitely many open sets. Take x in the
intersection of all of them. Then:
x is in the first set: there exists an with ( x - , x + ) contained in the first set
x is in the second set: there is with ( x - , x + ) contained in the second set.
....
x is in the N-th set: there is
with ( x ,x+
) contained in the last set.
But then
let = min{
, ...,
Hence it is contained in their finite intersection, which is therefore open, since x was
arbitrary.
The last statement follows again from de Morgan's laws.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
This proposition is rather interesting, giving a complete description of any possible open
set in the real line. To prove it, we will make use of equivalence relations and classes
again. First, let us define a relation on U:
if a and b are in U, we say that a ~ b if the whole line segment between a and b is
also contained in U.
Is this relation indeed an equivalence relation ? Assuming that it is, we know
immediately that U equals the union of the equivalence classes, and the equivalence classes
are pairwise disjoint. Denote those equivalent classes by U n
Each U n is an interval: take any two points a and b in U n. Being in the
same equivalence classes, a and b must be related. But then the whole
line segment between a and b is contained in U n as well. Since a and b
were arbitrary, U n is indeed an interval.
Each U n is open: take any x U n. Then x U, and since U is open, there
exists an > 0 such that ( x - , x + ) is contained in U. But clearly
each point in that interval is related to x, hence this neighborhood is
contained in U n, proving that U n is open.
There are only countably many U n: This seems the hard part. But, each
U n must contain at least one different rational number.
Why ? Since
there are only countably many rational numbers, there can only be
countably many of the U n's (since they are disjoint).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.1.6(a):
What is the boundary and the interior of (0, 4), [-1, 2], R, and O ? Which
The boundary of (0, 4) is the set consisting of the two elements {0, 4}. Every
neighborhood of these two points contains points both from the interval (0,4) and
from the complement of that interval. Therefore, both form the boundary. The
interior of the set (0, 4) is the set (0, 4) (i.e. itself). No points of either set are
isolated, and each point of the either set is an accumulation point. The same is true,
incidentally for each of the sets (0, 4), [0, 4), (0, 4], and [0, 4].
The boundary of [-1, 2] is the two-element set {-1, 2}, and the interior is (-1, 2). No
points are isolated, and each point in either set is an accumulation point.
The boundary of the set R as well as its interior is the set R itself. No point is
isolated, all points are accumulation points.
The boundary of the empty set as well as its interior is the empty set itself. Since
the set contains no points, it can not contain isolated or accumulation points.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.1.6(b):
Find the boundary, interior, isolated and accumulation points, if any, for the
set
{1, 1/2, 1/3, ... } {0}
Back
The boundary of the set is the set itself, because any neighborhood of every point
contains that point itself and some irrational points. Therefore, any neighborhood of
every point contains points from within and from without the set, i.e. every point of
the set if a boundary point.
The interior of this set is empty, because if x is any point in that set, then any
neighborhood of x contains at least one irrational point that is not part of the set.
Therefore, this neighborhood is not contain in the set.
Every point except 0 is an isolated point. First, it is easy to find a small enough
neighborhood for any point of the form x = 1/n that does not contain any point
from the set but x = 1/n. Therefore, every point x = 1/n is isolated. On the other
hand, if (-a, a) is any small neighborhood of 0, then if n is large enough we have
that 1/n < a. But then the point 1/n which is different from 0, is also part of the
neighborhood, and hence 0 is not isolated.
The point 0 is the only accumulation point of the set. It is clear that no point of the
form x = 1/n is an accumulation point, and it is also clear that 0 is an accumulation
point. Simply look at the previous example, and recall the definition of
accumulation points and their related points.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.2(a):
Is the interval [0,1] compact ? How about [0, 1) ?
Back
The interval [0, 1] is compact. To see this, take any sequence of points in [0, 1].
Since the sequence must be bounded, we can extract a convergent subsequence by
the Bolzano Weierstrass theorem. Using the theorem on accumulation and
boundary points and noting that the set [0, 1] is closed, the limit of this
subsequence must be contained in [0, 1]. Hence, the set is compact by definition.
The interval [0, 1) is not compact. Consider the sequence { 1 - 1/n }. Then that
sequence is contained in [0, 1), and converges to 1. Therefore, every subsequence
of it must also converge to 1, which is not part of the original set. Therefore, the set
can not be compact.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.2(b):
Is the set {1, 2, 3} compact ? How about the set N of natural numbers ?
Back
The set {1, 2, 3} is compact. Take any sequence with elements from the set {1, 2,
3}. This sequence is bounded, so we can extract a convergent subsequence from it.
Since the subsequence converges, it forms a Cauchy sequence. Therefore,
consecutive numbers in this subsequence must eventually be closer than, say, 1/2.
But then the subsequence must eventually be constant, and that constant must be
either 1, 2, or 3. Therefore the subsequence converges to an element of the original
set. But then the set {1, 2, 3} is compact. Note that a similar argument applies to
any set of finitely many numbers.
The set of natural numbers N is not compact. The sequence { n } of natural numbers
converges to infinity, and so does every subsequence. But infinity is not part of the
natural numbers.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.2(c):
Is the set {1, 1/2, 1/3, 1/4, ...} compact ?
Back
This set is not compact. Take the sequence { 1/n } which is the set itself. It - and hence
every subsequence - converges to zero, which is not part of the set. Therefore, the set can
not be compact by our definition.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.2(d):
Is the set {1, 1/2, 1/3, 1/4, ...}
{0} compact ?
Back
This set is compact. To see this, first note that the set is bounded. Hence, any sequence of
elements from this set is also bounded, and using the Bolzano-Weierstrass theorem we can
extract a convergent subsequence. But the set is also closed, so that by the theorem on
accumulation and boundary points the limit in fact must be part of the set as well.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
First, suppose S is closed and bounded. Take a sequence
in S. Because S is
bounded, the sequence is bounded also, and by the Bolzano-Weierstrass theorem we can
extract a convergent subsequence from
. Using the theorem about closed sets,
accumulation points and sequences, we know that the limit of the subsequence is again in
S. Hence, S is compact.
Now assume that S is compact. We have to show that S is bounded and
closed.
Suppose S was not bounded. Then for every n there exists a number a n
S with | a n | > n. But then no subsequence of the sequence
will
converge to a finite number, hence S can not be compact. Thus, S must
be bounded.
Suppose S was not closed. Then there exists an accumulation point c for
the set S that is not contained in S. Since c is an accumulation point,
there exists a sequence
of elements in S that converges to c. But
then every subsequence of that sequence will also converge to c, and
since c is not in S, that contradicts compactness. Hence, S must be
closed.
Examples 5.2.5(a):
Let S = [0, 1], and C = { (-1/2, 1/2), (1/3, 2/3), (1/2, 3/2)}. Is C an open
cover for S ?
Back
Examples 5.2.5(b):
Let S = [0, 1]. Define
= { t R : | t - | < and
S} for a fixed > 0.
Is the collection of all {
},
S, an open cover for S ? How many sets of
type
are actually needed to cover S ?
Back
= k * , k = 0, 1, 2, ... N
Examples 5.2.5(c):
Let S = (0, 1). Define a collection C = { (1/j, 1), for all j > 0 }. Is C an open
cover for S ? How many sets from the collection C are actually needed to
cover S ?
Back
(1 / j, 1) = (0, 1)
so that C is indeed an open cover of S.
Not all sets from the original collection C are needed to cover S. For
example, the subcollection of intervals (1 / (2 j) , 1) is also an open
Proof:
First, assume that every open cover C of S can be reduced to a finite subcovering. We will
show that S must then be closed and bounded, which means by the previous result that S is
compact.
S must be bounded: Take the collection C = {
:
S}, where
=(
- 1, + 1). Then this collection is an open cover of S, and by
assumption can be reduced to a finite subcovering of S. But if aj1 is the
smallest of the centers of the sets
, and aj2 is the largest one, then S
is contained in the set ( aj1 - 1, aj2 + 1) and is therefore bounded.
S must be closed: Suppose S was not closed. Then there exists an
accumulation point s of S that is not contained in S. Since s is an
accumulation point of S we know:
for any n > 1 there exists an S with | s - an | < 1 / n
because every neighborhood of s must contain elements from S. The sequence { an }
clearly converges to s. Define the collection of sets
C = { comp([s - 1/n, s + 1/n]), n > 0 }
Then each set in C is open, being the complement of closed sets. They also cover S,
because the only point that this collection is missing is the point s, which is not part of S.
By assumption, a finite subcover already covers S. If N is the largest index of that finite
subcovering, then aN+1 is not part of that subcovering. However, aN+1 is an element of S, so
that this subcovering can not cover S. That is a contradiction, showing that if S was not
closed, not every covering of S can be reduced to a finite subcovering. Hence, S must be
closed.
Now we have to prove the other direction. Assume therefore that S
compact. Let C be any open cover of S. We need to reduce C to a finite
subcover. Since S is compact, we know it is closed and bounded. Then a
= inf(S) and b = sup(S) are both part of S (
Why ?). Define the set A
as
A = { x: x [a, b] and a finite subcollection of C covers [a, x] S }
A). Define
Since A is a subset of [a, b], we know that a < c b. Suppose c < b. Since S is closed,
comp(S) is open. Therefore, if c comp(S) then there exists an open neighborhood U of c
that is contained in [a, b] (because c < b) and disjoint from S. But then c can not be the
supremum of the set A. Therefore, if c < b then c S. Then c must be contained in some
set
from the open cover C of S. Choose two points y and z in
with y < c < z. As
before, there exists a finite subcollection of C whose members cover [a, y] S. Then these
sets, together with
cover [a, z] S. But then z A, which means again that c can not be
the upper bound for A. This means that assuming c < b leads to a contradiction, so that c =
b. But that will be exactly what we need. If sup(A) = c = b, then let
be that member of
the open cover C that contains b. There exists some open neighborhood (b - , b + )
contained in
. But b - is not an upper bound for A, so there exists x with x > b - and
x A. Then [a, x] S is covered by a finite number of members of C. Together with the
set
these sets form a finite open cover for S.
We have indeed reduced the open cover of S to a finite subcovering of
S, finishing the proof. I think.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.7(a):
Consider the collection of sets (0, 1/j) for all j > 0. What is the intersection
of all of these sets ?
Back
The intersection of all intervals (0, 1/j) is empty. To see this, take any real number x. If x
0 it is not in any of the intervals (0, 1/j), and hence not in their intersection. If x > 0, then
there exists an integer N such that 0 < 1 / N < x. But then x is not in the set (0, 1 / N) and
therefore x is not in the intersection. Therefore, the intersection is empty.
Note that this is an intersection of 'nested' sets, that is sets that are
decreasing: every 'next' set is a subset of its predecessor.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.2.7(b):
Can you find infinitely many closed sets such that their intersection is empty
and such that each set is contained in its predecessor ? That is, can you find
sets A j such that A j+1 A j and A j= 0 ?
Back
It is easy to simply find some closed sets with empty intersection. For example, the
intersection of all intervals of the form [n, n+1] is certainly empty.
To find sets contained in one another is slightly more complicated. We
might try sets of the form A j = [0, 1 / j] for all j. Then A j+1 A j, but their
intersection contains the point {1}.
Proof:
Each Aj is compact, hence closed and bounded. Therefore, A is closed and bounded as
well, and hence A is compact. Pick an aj Aj for each j.
Then the sequence { aj } is contained in A1. Since that set is compact,
there exists a convergent subsequence { ajk } with limit in A1.
But that subsequence, except the first number, is also contained in A2.
Since A2 is compact, the limit must be contained in A2.
Continuing in this fashion, we see that the limit must be contained in
every Aj, and hence it is also contained in their intersection A. But then
A can not be empty.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 5.2.10(a):
Find a perfect set. Find a closed set that is not perfect. Find a compact set
that is not perfect. Find an unbounded closed set that is not perfect. Find a
closed set that is neither compact nor perfect.
Back
A perfect set needs to be closed, such as the closed interval [a, b]. In fact, every
point in that interval [a, b] is an accumulation point, so that the set [a, b] is a
perfect set.
The simplest closed set is a singleton { b }.The element b in then set { b } is not an
accumulation point, so the set { b } is closed but not perfect.
The set { b } from above is also compact, being closed an bounded. Hence, it is
compact but not perfect.
The set {-1} [0, ) is closed, unbounded, but not perfect, because the element -1
is not an accumulation point of the set.
The set {-1} [0, ) from above is closed, not perfect, and also not compact,
because it is unbounded.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 5.2.10(b):
Is the set {1, 1/2, 1/3, ...} perfect ? How about the set {1, 1/2, 1/3, ...}
{0} ?
Back
Proof:
If S is perfect, it consists of accumulation points, and therefore can not be finite. Therefore
it is either countable or uncountable. Suppose S was countable and could be written as
S = { x1, x2, x3, ...}
The interval U1 = (x1 - 1, x1 + 1) is a neighborhood of x1. Since x1 must be an accumulation
point of S, there are infinitely many elements of S contained in U1.
Take one of those elements, say x2 and take a neighborhood U2 of x2
such that closure( U2 ) is contained in U1 and x1 is not contained in
closure( U2 ). Again, x2 is an accumulation point of S, so that the
neighborhood U2 contains infinitely many elements of S.
Select an element, say x3, and take a neighborhood U3 of x3 such that
closure( U3 ) is contained in U2 but x1 and x2 are not contained in
closure( U3 ) .
Continue in that fashion: we can find sets Un and points xn such that:
closure( Un+1 ) Un
xj is not contained in Un for all 0 < j < n
xn is contained in Un
Now consider the set
V = ( closure( Un )
S)
Then each set closure( Un ) S ) is closed and bounded, hence compact. Also, by
construction, ( closure( Un+1 ) S ) (closure( Un) S ). Therefore, by the above result, V
is not empty. But which element of S should be contained in V ? It can not be x1, because
x1 is not contained in closure( U2 ). It can not be x2 because x2 is not in closure( U3 ), and so
forth.
Hence, none of the elements { x1, x2, x3, ... } can be contained in V. But
V is non-empty, so that it must contain an element not in this list. That
means, however, that S is not countable.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
A n = A n-1 /
Each set
is open. Since A 0 is closed, the sets A n are all closed as well,
which can be shown by induction. Also, each set A n is a subset of A 0, so that all sets A n
are bounded.
Hence, C is the intersection of closed, bounded sets, and therefore C is
also closed and bounded. But then C is compact.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
A n = A n-1 \
subinterval (if x is equal to that endpoint, then let x n be equal to the right endpoint of that
subinterval). Since each subinterval has length 1 / 3 n, we have:
|x-xn|<1/3n
Hence, the sequence { x n } converges to x, and since all endpoints of the subintervals are
contained in the Cantor set, we have found a sequence of numbers contained in C that
converges to x. Therefore, x is a limit point of C. But since x was arbitrary, every point of
C is a limit point. Since C is also closed, it is then perfect.
Note that this proof is not yet complete. One still has to prove the
assertion that each set A n is indeed comprised of 2 n closed
subintervals, with all endpoints being part of the Cantor set. But that is
left as an exercise.
Since every perfect set is uncountable, so is the Cantor.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
An = An-1 \
Then the Cantor set is given as:
C=
An
from the unit interval [0, 1]. Since we remove a set of total length 1 from the unit interval,
the length of the remaining Cantor set must be 0.
The Cantor set contains uncountably many points because it is a perfect
set.
A n = A n-1 \
Examples 5.3.2(a):
Is the set { x R : | x | < 1, x # 0 } connected or disconnected ? What about
the set { x R : | x | 1, x # 0 }
Back
The first set is open, so we can try to use the easier definition of disconnected. It is indeed
disconnected; it can be 'split at 0'. More precisely, let U = (-1,0) and V = (0,1). Then U and
V have empty intersection and their union is the original set. But that means the original
set must be disconnected.
The second set is also disconnected, but since it is not open we must
use the second part of the definition of disconnected sets. Again, we can
'split' the set at zero. Let U = (-2,0) and V = (0, 2) and set S = [-1, 0)
(0, -1]. Then both sets U and V are open and we have that
U S = [-1, 0)
V S = (0, 1]
so that
( U S)
( U S)
(V
(V
S)=0
S)=S
Note that the actual choice of U and V was not important. We could have chosen different
intervals, e.g. (-50,0) and (0,10), to show both sets were disconnected. The important part
was that the sets could be split at zero into two disjoint pieces.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.3.2(b):
Is the set [-1, 1] connected or disconnected ?
Back
It seems obvious that this set is connected. How should we even try to split it into two
disjoint pieces whose union gives back the original set ? But, try to prove it ! You will find
that it is much easier to show that a set is disconnect than it is to show that a set is
connected.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.3.2(c):
Is the set of rational numbers connected or disconnected ? How about the
irrationals ?
Back
Both sets are disconnected. The idea is simple: The set of rational numbers can be 'split' at
any irrational number, while the set of irrationals can be 'split' at any rational number. Can
you provide the details ? Note that neither set is open, so that the more complicated
definition of disconnected must be used.
In fact, both of these sets are totally disconnected.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 5.3.2(d):
Is the Cantor set connected or disconnected ?
The Cantor set is disconnected. In fact, one of the properties of the Cantor
set states that it does not contain any open set. Using that fact one can see
that the Cantor set must in fact be totally disconnected, which implies
disconnected. Can you provide the details ?
Back
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
If S is not an interval, then there exists a, b
not in S. Then define the two sets
U = ( - , t ) and V = ( t, )
Example 5.3.5(a):
The Cantor set is disconnected. It is in fact totally disconnected.
Back
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.1.1:
Consider the function f, where f(x) = 1 if x 0 and f(x) = 2 if x > 0.
1. The sequence {1/n} converges to 0. What happens to the sequence
{ f( 1/n ) } ?
2. The sequence { 3 + (-1)n} is divergent. What happens to the sequence
{ f(3 + (-1)n)} ?
3. The sequence {(-1)n / n } converges to zero. What happens to the
sequence { f((-1)n / n ) } ?
Back
1.
For the first sequence, we clearly have that f(1/n) = 2 for all n. Hence, the limit of
f(1/n) equals 2. So a function applied to a sequence results in a new sequence that
can converge to a different number.
2. For the second sequence, we know that 3 + (-1)n > 0 for all n. Hence, f(3 + (-1)n) =
2 for all n. This time, a function applied to a divergent sequence results in a
convergent one.
3. For the third sequence we know that alternating terms switch signs. Hence, f((-1)n /
n) = 1 if n is odd, and 2 if n is even. But then the resulting new sequence is
divergent. Hence, we have an example where a function can turn a convergent
sequence into a divergent one.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.1.3:
1. Let f(x) = m x + b. Then does the limit of that function exist at an
arbitrary point x ?
2. Let g(x) = [x], where [x] denotes the greatest integer less than or
equal to x. Then does the limit of g exist at an integer ? How about at
numbers that are not integers ?
3. In the above definition, does c have to be in the domain D of the
function ? Is c in the closure(D) ? Do you know a name for c in
terms of topology ?
Back
1. Consider a sequence {xn} converging to c. Then we want to show that f(xn) converges as
well. So choose any > 0 and take a positive integer N so large so that
| xn - c| < / m for n > N
Then we have
| f(xn) - (m c + b) | = | m xn + b - m c - b | = m | xn - c | < m / m =
for n > N. But that means that the sequence f(xn) converges to the number (m c + b).
2. Let g(x) = [x] and assume that c is an integer. Then take the
sequence
xn = c + (-1)n / n
Then this sequence converges to c as n approaches infinity, but the sequence g(xn) does not
converge at all. Hence, the limit of g(x) does not exist at any integer. On the other hand, if
c is not an integer, then the function g(x) converges to the limit L = g(c). Can you prove it ?
3. Recalling our knowledge of topology, we remember that if {xn} is a
sequence in a set D converging to c, then c is called an accumulation
point of D. Would it be correct to say that a function f(x) with domain D
converges to a limit L if for every accumulation point c of D the number
L is also an accumulation point of the image of D ?
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.1.5:
Consider the function f with f(x) = 1 if x is rational and f(x) = 0 is x is
irrational. Does the limit of f(x) exist at an arbitrary number x ?
Back
Recall that f is called the Dirichlet function. Since f 'jumps wildly' up and down, we
suspect that the function does not have a limit at any point. This is indeed the case, as is
easy to show using our new definition of limit:
Let c be any real number and pick = 1/2. Suppose there was a
such that whenever | x - c | < then | f(x) - L | < = 1/2.
We can find a rational number q with | q - c | <
therefore
>0
Proof:
Suppose the first condition is true, but the second condition fails. Then
there exists an > 0 such that there is no > 0 with the property that if
| x - c | < then | f(x) - L | < . Therefore, if we let = 1 / n, then for
each n we can produce a number xn with
| xn- c | < but | f(xn) - L |
But then the sequence { xn} converges to c, but the sequence f(xn) does
not converge to L. That is contrary to the first condition being true, and
hence we have proved by contradiction that the first condition implies
the second.
Example 6.2.1:
Which of the two functions is intuitively continuous and which one is not ?
1. f(x) = 1 if x > 0 and f(x) = -1 if x < 0. Is this function continuous ?
2. f(x) = 5x - 6. Is this function continuous?
Back
Since Descartes claims that a function is continuous if its graph can be drawn without
lifting the pencil, we will look at the graph of each function:
f(x) = -1 if x < 0 and 1 if x > 0: Since we have to lift the pencil to draw this graph, this
function does not appear to be continuous.
f(x) = 5x - 6: Since this graph, being a straight line, does not require us to lift the pencil, we
would call this function continuous.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
f(x) = f(c).
If
f(x) = f(c) for every sequence {xn} of points in D
converging to c, then f is continuous at the point c.
Context
Proof:
The proof is very similar to the previous result about the equivalence of
the two definitions of limits for a function. It is therefore left as an
exercise. It would be good practice to see if you can modify the previous
proof and adapt it to this result.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.4(a):
Which of the following two functions is continuous:
1. If f(x) = 5x - 6, prove that f is continuous in its domain.
2. If f(x) = 1 if x is rational and f(x) = 0 if x is irrational, prove that x is
not continuous at any point of its domain.
Back
We have already seen from the graph that the first function seems to be continuous while
the second one does not. We have to formally prove it, though.
Pick any > 0. Take any sequence { xn } converging to c. Then there
exists an integer N such that
| xn - c | < / 5
for n > N. Then
| f(xn) - (5 c - 6) | = | 5 xn - 6 - 5 c + 6 | = 5 | xn - c | <
for n > N. But then the sequence {f(xn)} converges to 5 c - 6, or in other
words: if a sequence {xn} converges to c, then f(xn) converges to f(c).
That proves continuity of the first function.
As for the second one: if c is any real number we can find a sequence of
rational numbers an converging to c, as well as another sequence of
irrational numbers bn also converging to c. But then the sequence
{f(an)} is identically 1, and the sequence {f(bn)} is identically 0. But
then f does not have a limit at c, and hence can not be continuous at c
either (we have seen this argument - more formally - in a previous
example already).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.4(b):
If f(x) = x if x is rational and f(x) = 0 if x is irrational, prove that x is
continuous at 0.
Back
If one looks at this poor representation of the function, we see that it does not at all look
continuous. But if {xn} is any sequence of numbers (rational or irrational) that converges to
zero, then there exists an integer N such that |xn| < for n > N. But f(xn) is either zero or xn
itself, and in any case we have
| f(xn) | | xn| <
That proves that the sequence of {xn)} converges to 0 = f(0), which proves that the function
is continuous at zero.
As an exercise, prove that the function is not continuous for any other x.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.4(c):
If f(x) is continuous in a domain D, and {xn} is a Cauchy sequence in D, is
the sequence {f(xn)} also Cauchy ?
Back
This seems to be true. After all, if f is continuous at c, and {xn} is a sequence converging to
c, then f(xn) must converge to f(c). And since convergent sequences are Cauchy one would
assume that the statement should be true.
But of course life is not so simple. Consider the function f(x) = 1/x and
the sequence {1/n}. Then the sequence is convergent to zero, and thus
is Cauchy. But f(1/n) = n, which is not Cauchy. This function is
continuous in the domain D = (0, 2), say, the sequence {1/n} is Cauchy
in D, but the sequence f(1/n) fails to be Cauchy.
The point here is that the function does not need to be continuous at the
limit point of a sequence, and hence the above statement is false. While
the sequence {1/n} converges to zero, the function f(x) is not
continuous at zero. Yet the sequence {1/n} is Cauchy in (0, 2).
Is the statement true if the domain of the function is all of R? Can you
find other formulations for which the statement would become true ?
Proof:
Suppose f(x) = x. Then, given any > 0 choose
= / 2. Then, if
|x-c|<
it implies that
| f(x) - f(c) | = | x - c | <
= /2<
. Hence, the identity function is indeed continuous. Was it really necessary to take = /
2?
The sum of continuous functions is continuous follows directly from the
triangle inequality. Take any > 0. There exists 1 > 0 such that
whenever
|x-c|<
we know that
| f(x) - f(c) | <
(because f is continuous at c). There also exists
|x-c|<
we know that
| g(x) - g(c) | <
(because g is continuous at c). But then, if we let
|x-c|<
then
| (f(x) + g(x)) - (f(c) + g(c)) |
| f(x) - f(c) | + | g(x) - g(c) | <
+ =2
= min(
), we have: if
That finishes the proof. (That we don't get a simple should not bother us any more).
The product of two continuous functions is again continuous, which
follows from a simple trick. We will only look at the trick involved, and
leave the details to the reader:
| f(x) g(x) - f(c) g(c) |
= | f(x) g(x) - f(x) g(c) + f(x) g(c) - f(c) g(c) |
| f(x) | | g(x) - g(c) | + | g(c) | | f(x) - f(c) |
With this trick the rest of the proof should not be too difficult.
A similar trick works for the quotient. Here is the idea:
| f(x) / g(x) - f(c) / g(c) | = | 1 / g(x) g(c) | | f(x) g(c) - f(c) g(x) |
Can you see how to continue ? Adding and subtracting will help again.
As for composition of functions, we have to proceed somewhat different:
We know that f(x) is continuous at c, and g(x) is continuous at f(c).
Therefore, given any > 0 there exists 1 > 0 such that whenever
|t-d|<
then
| g(t) - g(d) | <
There also exists
|x-c|<
then
| f(x) - f(c) | <
and then
| g(f(x)) - g(f(c)) | <
In other words, f(g(x)) is continuous at x = c.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.2.6(a):
Every polynomial is continuous in R, and every rational function r(x) = p(x)
/ q(x) is continuous whenever q(x) # 0.
Back
Examples 6.2.6(b):
The absolute value of any continuous function is continuous.
Back
First we have to prove that the usual absolute value function f(x) = | x |
is continuous. But this is clear from the graph of that function
(nonetheless, can you prove it formally ?).
Now the statement follows immediately from the fact that the
composition of two continuous functions yields another continuous
function.
As in the case of proving that a polynomial is continuous, the abstract
case is much easier to prove that trying to prove, say, that the absolute
value of a particular polynomial is continuous.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.8(a):
The function f(x) = 1 / x is continuous on (0, 1). Is it uniformly continuous
there ?
Back
As the - interval 'slides' up the positive y-axis, the corresponding -interval on the x-axis
gets smaller and smaller. That indicates that the function is not uniformly continuous - but
it is of course not a proof. So:
Take any = 1. Does there exist any
if | t - s | <
such that
Example 6.2.8(b):
The function f(x) = x2 is continuous on [0, 1]. Is it uniformly continuous
there ?
Back
This function (as you could guess from its graph) is uniformly continuous
on the closed interval [0, 1]. To prove it, note that
| f(t) - f(s) | = | t - s | | t + s | < | t - s | 2
because s and t are in the interval [0, 1]. Hence, given any > 0 we can
simply choose = / 10 (or something similar) to prove uniform
convergence. Can you fill in the details ? A similar argument,
incidentally, would work on the interval [0, N] for any number N, but it
would fail for the interval [0, ). So, if this function then uniformly
continuous on the interval [0, ) ? That's the next example.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.8(c):
The function f(x) = x2 is continuous on [0,
there ?
). Is it uniformly continuous
Back
While this function is uniformly continuous on any interval [0, N] (N any number) it is no
longer uniformly continuous on the interval [0, ). To prove this, take = 1. Note that
| f(s) - f(t) | = | s - t | | s + t |
Can you see that if s = t + and if t is sufficiently large (depending on
the undetermined ) then no matter what is chosen, | f(s) - f(t) | > 1.
That would prove that the function is no longer uniformly continuous.
The details are left as an exercise.
Note that this argument no longer works on a bounded interval [0, N].
Here we can not make t 'sufficiently large', since it can be no larger than
N. And indeed, the function is uniformly continuous on those bounded
intervals. Later we will show that any function that is continuous on a
compact set is necessarily uniformly continuous.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Example 6.2.8(d):
If f(x) is uniformly continuous in R, and { xn} is a Cauchy sequence, is the
sequence { f( xn) } also Cauchy ?
Back
Note that this is not true if the function is only assumed to be continuous (as the example
f(x) = 1 / x on the interval (0, 1) illustrates). But if f is uniformly continuous, we have:
given any > 0 there exists > 0 such that if | t - s | < then | f(t) - f(s) | <
If { xn} is a Cauchy sequence, there exists an integer N such that
| xn - xj| < if n, j > N
But then, by uniform continuity, we have that
| f( xn) - f(xj) | < if n, j > N
Therefore, the sequence { f( xn) } is again a Cauchy sequence.
Why will this argument will not work for a continuous function ? As an
example, consider again the function f(x) = 1/x. It is continuous in the
open interval (0, 1). Therefore, if we consider a sequence { xn} that
converges to a fixed point inside (0, 1), the same argument as above
would to show that { f( xn) } is also Cauchy. We can not use this
argument, however, for a sequences that converge to 0, since f is not
even defined at zero. On the other hand, we can find Cauchy sequences
which converge to 0, without ever having to refer to the value of the
function at the limit of the sequence.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The first part of the proof is obvious. If | f(t) - f(s) | is small whenever |s t | is small, regardless of the particular location of s and t, then in
particular | f(x) - f( x0) | must be small when | x - x0| is small.
The second part is much more complicated, and relies on the structure
of compact sets on the real line. Recall that a set is compact if every
open cover has a finite subcover. We first want to define a suitable open
cover: pick an > 0. For every fixed x0in the compact set D there exists
a (possibly different) ( x0) > 0 such that
| f( x0) - f(x) | < if | x0- x | < 1/2 ( x0) (here depends on x0).
Define
U( x0) = { x: | x0- x | < 1/2
Let
= 1/2 min{
( x1),
( x2), ...
( xn)}
Since this is a minimum over a finite set, we know that > 0. Now take
any two numbers t, s in D such that | t - s | < . Since the finite
collection covers the compact set D we know that s is contained in, say,
U2. How far away from the center of U2 is t then ?
By the choice of
we know that
+ 1/2
( x2) +
( x2)) =
( x2)
then
The first difference on the right is less than epsilon because | t - x2| < (
x2). The second one is also less than epsilon because s is contained in
the set U2. Hence, the difference on the left is less than twice epsilon.
But now it should be easy for you to modify the proof so that we can
arrive at
| f(t) - f(s) | < whenever | t - s | < in D
That finishes the proof.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.2:
Which of the following functions, without proof, has a 'fake' discontinuity, a
'regular' discontinuity, or a 'difficult' discontinuity ?
Back
This function seems to have a 'fake' discontinuity at x = 3, since we could easily move the
single point at x = 3 to the right height, thereby filling in the discontinuity.
This function, while simple, seems to have a 'true' discontinuity at x = 0. We can not
change the function in a single point to make it continuous.
This function is unclear. It is hard to determine what exactly is going on as x gets closer to
zero. Assuming that the function does turn out to be discontinuous at x = 0, it definitely
seems to have a 'difficult' discontinuity at x = 0.
This function is impossible to graph. The picture above is only a poor representation of the
true graph. Nonetheless, given any point x, the function jumps between 1 and 0 in every
neighborhood of x. That seems to mean that the function has a difficult discontinuity at
every point.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.4(a):
Prove that k(x) has a removable discontinuity at x = 3, and draw the graph of
k(x).
Back
We can easily check that the limit as x approaches 3 from the right and from the left is
equal to 4. Hence, the limit as x approaches 3 exists, and therefore the function has a
removable discontinuity at x = 3. If we define k(3) = 4 instead of k(3) = 1 then the
function in fact will be continuous on the real line.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.4(b):
Prove that h(x) has a jump discontinuity at x = 0, and draw the graph of h(x)
Back
It is easy to see that the limit of h(x) as x approaches 0 from the left is -1, while the limit of
h(x) as x approaches 0 from the right is +1. Hence, the left and right handed limits exist
and are not equal, which makes x = 0 a jump discontinuity for this function.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.4(c):
Prove that f(x) has a discontinuity of second kind at x = 0
Back
Since f(x) is an odd function, the same argument shows that the limit of
f(x) as x approaches zero from the left does not exist.
Therefore, the function has an essential discontinuity at x = 0.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.4(d):
What kind of discontinuity does the function g(x) have at every point (with
proof).
Back
This function is impossible to graph. The picture above is only a poor representation of the
true graph. Nonetheless, take an arbitrary point x0 on the real axis. We can find a sequence
{xn} of rational points that converge to x0 from the right. Then g(xn) converges to 1. But we
can also find a sequence {xn} of irrational points converging to x0 from the right. In that
case g(xn) converges to 0. But that means that the limit of g(x) as x approaches x0 from the
right does not exist. The same argument, of course, works to show that the limit of g(x) as
x approaches x0 from the left does not exist. Hence, x0 is an essential discontinuity for g(x).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Suppose, without loss of generality, that f is monotone increasing, and
has a discontinuity at x0. Take any sequence xn that converges to x0from
the left, i.e. xn x0. Then f( xn) is a monotone increasing sequence of
numbers that is bounded above by f( x0). Therefore, it must have a limit.
Since this is true for every sequence, the limit of f(x) as x approaches x0
from the left exists. The same prove works for limits from the right.
Note: This proof is actually not quite correct. Can you see the mistake ?
Is it really true that if xn converges to x0from the left then f( xn) is
necessarily increasing ? Can you fix the proof so that it is correct ?
As for the second statement, we again assume without loss of generality
that f is monotone increasing. Define, at any point c, the jump of f at x =
c as:
j(c) =
f(x) -
f(x)
Note that j(c) is well-defined, since both one-sided limits exist by the
first part of the theorem. Since f is increasing, the jumps j(c) are all nonnegative. Note that the sum of all jumps can not exceed the number f(b)
- f(a). Now let J(n) be the set of all jumps c where j(c) is greater than
1/n., and let J be the set of all jumps of the function in the interval [a, b].
Since the sum of jumps must be smaller than f(b) - f(a), the set J(n) is
finite for all n. But then, since the union of all sets J(n) gives the set J,
the number of jumps is a countable union of finite sets, and is thus
countable.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
Suppose not, i.e. f has a discontinuity of the second kind at a point x =
c, and there does exist some (small) neighborhood of c where f, say, is
always decreasing. But then f is a monotone function, and hence, by the
previous theorem, can only have discontinuities of the first kind. Since
that contradicts our assumption, we have proved the corollary.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.3.8(a):
What kind of discontinuity does the function f(x) = exp(1/x) have at x = 0 ?
Back
f(x) = exp(1/x)
Examples 6.3.8(b):
What kind of discontinuity does the function f(x) = x sin(1/x) have at x = 0 ?
Back
f(x) = x sin(1/x)
Examples 6.3.8(c):
f(x) = cos(1/x)
Proof:
(1) => (2): Assume that f is continuous on an open set D. Let U be an
open set in the range of f. We need to show that f-1(U) D is again open.
Take any x0 f-1(U). That is equivalent to saying that f( x0) U. Since U is
open, we can find an > 0 such that the - neighborhood of f( x0) is
contained in U. For this fixed we can use the continuity of f to pick a
> 0 such that
if | x - x0| < then | f(x) - f( x0) | <
Examples 6.4.2(a):
Let f(x) = x2. Show that f is continuous by proving
1. that the inverse image of an open interval is open.
2. that the inverse image of a closed interval is closed.
Back
f(x) = x2
First, let's look at the inverse images of an open interval:
If 0 < a < b then the inverse image of the open interval (a, b) is (,
,-
If a < 0 < b then the inverse image of the open interval (a, b) is (,
), which
is again open.
If a < b < 0, then the inverse image of the open interval (a, b) is again open (which
set is it ?)
But it is now obvious that the inverse image of closed intervals is again
a closed set (note that the empty set is both open and closed).
Hence, we have proved that the function f(x) = x2 is continuous,
avoiding the tedious epsilon-delta proof.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.2(b):
Let f(x) = 1 if x > 0 and f(x) = -1 if x 0. Show that f is not continuous by
1. finding an open set whose inverse image is not open.
2. finding a closed set whose inverse image is not closed.
Back
Examples 6.4.3:
Is it true that if f is continuous, then the image of an open set is again open ?
How about the image of a closed set ?
Back
This is true for inverse images but not for images. Consider the example
of a parabola, which certainly represents a continuous function:
f(x) = x2
Then the image of the set (-1, 1) is the set [0, 1). That set is neither open nor closed; in
particular, it is not open.
To find a counterexample for images of closed sets, let's look at the
following function:
This function is continuous on the whole real line, and the image of the set [0, ) is the set
(0, 1]. Therefore we have found a closed set whose image under a continuous function is
not closed (nor open).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
1. Note that the image of a closed set is not necessarily closed for
continuous functions, and the image of a bounded set is not necessarily
bounded. However, the image of a close and bounded set is again
closed and bounded (under continuous functions). Despite this, the
proof is fairly easy: Recall that a set D is compact if every open cover of
D can be reduced to a finite subcover.
Let A be an open cover of the set f(D). Since f is continuous, the
collection
{f-1(U): U
A}
Examples 6.4.5(a):
If
then:
1. what is the image of [-2, 1] ?
2. find a closed set whose image is not closed
Back
Note that this function is continuous. The image of the interval [-2, 1] is [1/5, 1]. In
particular, the image of a compact set is compact.
To find a closed set whose image is not closed we must consider an
unbounded set. Otherwise, a closed and bounded set is compact, and
since the image of a compact set under continuous functions is
compact, it is in particular closed again.
But if we look at the closed, unbounded set [0, ), we see that the
image of that set under the above function is the set (0, 1], which is not
closed.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.5(b):
Find examples for the following situations:
1. A continuous function and a set whose image is not connected.
2. A continuous function and a disconnected set whose image is
connected.
3. A function such that the image of a connected set is disconnected.
4. Is it true that inverse images of connected sets under continuous
functions are again connected ?
Back
Proof:
With the work we have done previously, this proof is easy: Since K is
compact and f a continuous function, f(K) is compact also. The compact
set f(K) is bounded, so that f is bounded on K. The compact set f(K) also
contains its infimum and supremum, so that f has an absolute minimum
and maximum on K.
That's all !
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.7(a):
Find a continuous function on a bounded interval that is unbounded. How
about a continuous function on a bounded interval that does not have an
absolute maximum but is bounded ?
Back
If the function was defined on a bounded and closed set (i.e. a compact set), it would have
to have an absolute maximum and minimum, and would therefore be bounded. To
construct our counterexample, we again have to define a function on an open, bounded
interval. The rest is left as an exercise.
While at first glance confusing, all we have do remember is that if the
domain of the function was a compact (i.e. closed and bounded)
interval, the function would have to have an absolute maximum. Hence,
to construct a counterexample, we need to define a continuous function
on an open, bounded interval. The rest is easy, and left as an exercise.
These examples show that the closedness condition in the Max/Min
theorem for continuous functions is essential.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.7(b):
Does the function
have an absolute maximum and minimum
on [-2, 1] ? How about on the interval [0, ) ?
Back
This function does have an absolute maximum and minimum on the interval [-2, 1], as
predicted by the Max/Min theorem for continuous functions. The absolute maximum is 1,
and the absolute minimum is 1/5.
On the other hand, on the unbounded interval [0, ) the function fails to
possess both absolute maximum and minimum. While 1 is still the
absolute maximum, there no longer is an absolute minimum.
Thus, the boundedness condition in the Max/Min theorem for continuous
functions is essential.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
With the work we have done so far this proof is easy. Since (a, b) is
connected and f a continuous function, the interval f( (a,b) ) is also
connected. Therefore that set must contain the interval (f(a), f(b))
(assuming that f(a) < f(b) ). Since f(a) and f(b) have opposite signs, this
interval must include 0. Therefore, 0 is in the image of (a, b), or
equivalently: there exists a c in open interval (a, b) such that f(c) = 0.
That's it !
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.9(a):
Show that the equation cos(x) = x has a solution in the interval [-10, 10].
Back
Let's take a look at the two function f(x) = cos(x) and g(x) = x in one coordinate system:
So one can clearly see that there is exactly one solution. We can use Bolzano's theorem to
actually prove that there must be at least one solution: Let h(x) = cos(x) - x. Then h is a
continuous function and
h(- ) = -1 + > 0
h( ) = -1 - < 0
Hence, by Bolzano's theorem there must be at least one place x0 where h( x0) = 0, or
equivalently where cos( x0) = x0.
One can use Bolzano's theorem to construct an algorithm that will find
zeros of a function to a prescribed degree of accuracy in many cases. In
simple terms:
start with an interval [a, b] where h(a) * h(b) < 0 (i.e. h(a) and h(b) have opposite
signs)
find a point c - usually (a + b) / 2 - such that either h(a) * h(c) < 0 or h(b) * h(c) <
0
o if h(a) * h(c) < 0, repeat this procedure with b replaced by c
o if h(b) * h(c) < 0, repeat this procedure with a replaced by c.
Continue until the difference b - a is small enough.
Would this procedure find the zero of the function f(x) = x2 in the
interval [-1, 1] ?
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.4.9(b):
Show that the equation
in R
Using a computer it is simple enough to draw this function and to see the approximate
solution. However, it is even easier to prove that there must be a solution (without
specifying where the solution would be).
The function p(x) is an odd-degree polynomial. Therefore:
If c =
If c = -
, then
, then
p(x) =
p(x) = -
Proof:
With the work we have done so far this proof is easy. In fact, the easiest
proof is an application of Bolzano's theorem, and is left as an exercise.
Examples 6.5.2(a):
Find the derivative of f(x) = x and of f(x) = 1 / x
Back
1. If f(x) = x, then
x.
2. If f(x) = 1 / x, then
=
(1/x - 1/c) / (x - c) = (c - x) / (x - c) *
2
2
1 / (x c) = - 1 / c , so that f'(x) = - 1 / x for all x not equal to zero.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.2(b):
Find the derivative of
Back
=
=
=
Hence, f'(x) = 2 - 3 x2 for all x.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
=0
Context
Proof:
- f'(c)
= 0 Then
-M=
Examples 6.5.4(a):
Why might our original definition of differentiability not be suitable for
functions of, say, two or three real variables ?
Back
Suppose we have a function of three real variables X = (x, y, z), defined in all of three
dimensional space, with domain in R..
Let us try to use the original definition of differentiability: we would like
to call the function f(X) = f(x, y, z) differentiable in a point C = (a, b, c) if
the limit of the difference quotient
( f(X) - f(C) ) / (X - C)
Examples 6.5.4(b):
Use the characterization of differentiability via approximation by linear
functions to define the concept of 'derivative' for functions of n real
variables.
Back
As we have seen, we can not use the definition involving the limit of the difference
quotient as a definition of 'the derivative' of a function of n real variables. Therefore, we
want to try to use the 'linear approximation' characterization to give us a workable concept
of derivatives in higher dimensions. First, recall that a linear function in n variables is
given as matrix multiplication:
Proof:
Note that
f(x) - f(c) =
(x-c)
Examples 6.5.6(a):
The function f(x) = | x | is continuous everywhere. Is it also differentiable
everywhere ?
Back
f'(x) =
o
o
o
=
( | x | - | c | ) / (x - c)
If c > 0 then x > 0 eventually. Then there is no need for the absolute value.
The limit become +1.
If c < 0 then x < 0 eventually. The absolute values are resolved by an
additional negative sign. The limit becomes -1.
If c = 0, then the left and the right-handed limits will be different (-1 and
+1). Therefore, the function is not differentiable at 0.
Examples 6.5.6(b):
The function f(x) = x sin(1/x) is continuous everywhere except at x = 0,
where it has a removable discontinuity. If the function is extended
appropriately to be continuous at x = 0, is it then differentiable at x = 0 ?
Back
Note that
(f(x) - f(0) ) / x = x sin( 1 / x) / x = sin( 1 / x )
But we have seen before that this function does not have a limit as x
approaches 0. Can you recall the argument for this statement ?
Therefore, the function is not differentiable at 0. As a product and
composition of functions that are differentiable for all x but zero, this
function is differentiable everywhere except at x = 0.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.6(c):
The function f(x) = x2 sin(1 / x ) has a removable discontinuity at x = 0. If
the function is extended appropriately to be continuous at x = 0, is it then
differentiable at x = 0 ?
Back
Since | x sin( 1 / x ) | < | x | for all x, we see that the limit of the
difference quotient for c = 0 equals zero. Hence, f is differentiable at 0,
and
f'(0) = 0
( f(x) / g(x) ) =
Chain Rule: If g is differentiable at x = c, and f is differentiable at x = g(c)
then f(g(x)) is differentiable at x = c, and
Proof:
These proofs, except for the chain rule, consist of adding and
subtracting the same terms and rearranging the result. We will prove
the product and chain rule, and leave the others as an exercise.
Product rule:
o
o
=
= f(c) g'(c) + g(c) f'(c)
=
=
f'(s) = h(s) =
f(t) - f(s) = h(t) (t - s)
which proves the chain rule. This is, of course, the rigorous version of
the above 'simple substitution'.
Note that the chain rule and the product rule can be used to give a
quick proof of the quotient rule.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Context
Proof:
if x < c then
if x > c then
The first inequality implies that as x approaches c from the left, the limit
must be greater than or equal to zero. The second one says that as x
approaches c from the right, the limit must be less than or equal to zero.
But since f is differentiable at c we know that both right and left handed
limits exist and must agree. Therefore, f'(c) = 0.
The proof is similar if f(c) is a minimum. Can you see what would
change, if anything ?
Note: As a consequence of this proof we have shown that if a
differentiable function has a maximum or minimum in the interior of its
domain then the derivative at that point must be zero.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
f'(c) =
If f and g are continuous on [a, b] and differentiable on (a, b) and g'(x) # 0
in (a, b) then there exists a number c in (a, b) such that
Context
Proof:
The first version of the Mean Value theorem is actually Rolle's theorem
in disguise. A simple linear function can convert one situation into the
other:
geometric interpretation of
MVT
We need a linear function (linear so that we can easily compute its
derivative) that maps the line through the two points ( a, f(a) ) and ( b,
f(b) ) to the points ( a, 0 ) and ( b, 0 ). If we subtract that map from the
function we will be in a situation where we can apply Rolle's theorem.
To find the equation of such a line is easy:
y - f(a) =
(x-a)
h(x) = f(x) -
( x - a ) - f(a)
0 = h'(c) = f'(c) -
h(x) = f(x) -
Examples 6.5.10(a):
Does Rolle's theorem apply to
defined on (-3, 3) ? If so, find
the number guarantied by the theorem to exist.
Back
Your browser
can not handle
Java applets
This function is continuous on the interval [-3, 3], and differentiable on (3, 3). It is not differentiable at x = -3 and x = +3, but Rolle's theorem
does not require the function to be differentiable at the endpoints. Also,
f(3) = f(-3) = 0. Therefore, Rolle's theorem does apply.
Examples 6.5.10(b):
If f is differentiable on R and | f'(x) | M for all x, then | f(x) - f(y) | M | x y | for all numbers x, y.
Back
f'(x) =
Taking absolute values on both sides and moving the denominator to
the other side we have
| f(b) - f(a) | = | f'(x) | | b - a |
Since f'(x) is uniformly bounded by M, we therefore have
| f(b) - f(a) | M | b - a |
But that is exactly what we wanted to prove.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.10(c):
).
= f'(c)
for c between x and x + 1. But then
0<
= 1/2 * 1 /
As x goes to infinity, so does c (it is always bigger than x). The left side of this equation
goes to 0 as c goes to infinity. Therefore, the right side must also go to zero.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
This proof is left as an exercise. For a hint on the first part you might
want to look at the proof of Rolle's theorem. If you have proved the first
part, it should be more or less clear how the remaining parts can be
proved. Recall the definitions of an increasing or decreasing function
and compare it with the difference quotient involved in the derivative of
f.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
This corollary becomes obvious when we interpret what it means for the
function to have a positive or negative derivative, as in these tables:
Loc. Max
interval
Loc. Min
(a, c) (c , b) interval
sign of f'(x)
dir. of f(x)
up
No Extremum
interval
sign of f'(x)
(a, c) (c , b)
-
down
up
No Extremum
(a, c) (c , b) interval
sign of f'(x)
sign of f'(x)
dir. of f(x)
up
up
dir. of f(x)
(a, c) (c , b)
-
down down
Of course, these tables are no proof - which is once again left as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.14(b):
If f(x) = | 1 - x2|, then find all relative extrema
Back
This function is so easy that you should 'see' the answer right away
by looking at a mental picture of the graph of the function. There are
three local extrema ... all details are left to you.
provided the limit on the right exists. The same result holds for one-sided
limits.
If f and g are differentiable and
then
f(x) =
g(x) = -
Proof:
The first part can be proved easily, if the right hand limit equals f'(c) /
g'(c): Since f(c) = g(c) = 0 we have
Taking the limit as x approaches c we get the first result. However, the
actual result is somewhat more general, and we have to be slightly
more careful. We will use a version of the Mean Value theorem:
Take any sequence {xn} converging to c from above. All assumptions of
the generalized Mean Value theorem are satisfied (check !) on [c, xn].
Therefore, for each n there exists a number cn in the interval (c, xn) such
that
Taking the limit as n approaches infinity will give the desired result for
right-handed limits. The proof is similar for left handed limits and
therefore for 'full' limits.
The proof of the last part of this theorem is left as an exercise.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.16(a):
Find
Back
L'Hospital's Rule applies directly in this case. If g(x) = sin(x) and f(x)
= x, then g'(x) = cos(x) and f'(x) = 1. Hence, by l'Hospital's rule, we
have:
= cos(0) / 1 = 1
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 6.5.16(b):
Find
Back
L'Hospital's rule does not seem to apply in this case, since we have '0
* ', not '0 / 0'. But if we write this expression as xn / ex we see that
we can apply the second of l'Hospital's rules.
Examples 6.5.16(c):
Find
Back
First, we need to write this expression as a fraction before we can try to apply l'Hospital's
rule:
Written in this form we see that we can apply l'Hospital's rule. It leads to the expression
This is again an expression for which l'Hospital's rule applies. We get the expression
For this expression we get that the limit as x approaches zero is zero. Hence, according to
l'Hospital's rule applied twice we get
=0
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.2(a):
What is the norm of a partition of 10 equally spaced subintervals in the
interval [0, 2] ?
Back
To find a partition of the interval [0, 2] into 10 equally spaced subintervals means to find
points x0, x1, ... x10 inside that interval so that each point has the same distance from its
predecessor. Therefore, the points are:
x0 = 0/10 = 0 x1 = 2/10
x2 = 4/10
x3 = 6/10
x4 = 8/10
x5 = 10/10
x6 = 12/10
x7 = 14/10
x8 = 16/10
x9 = 18/10
x10 = 20/10 = 2
Therefore, the norm of this partition is 2/10.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.2(b):
What is the norm of a partition of n equally spaced points in the interval [a,
b] ?
Back
If we want to divide [a, b] into n subintervals so that each of them has the same width,
then we must choose the width of those subintervals to be b - a / n. Moreover, the partition
would consist of the following n+1 points:
x0 = a = a + 0 . b - a / n
x1 = a + 1 . b - a / n
...
xn-1 = a + (n-1) . b - a / n
xn = a + n . b - a / n = b
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.2(b):
What is the norm of a partition of n equally spaced points in the interval [a,
b] ?
Back
If we want to divide [a, b] into n subintervals so that each of them has the same width,
then we must choose the width of those subintervals to be b - a / n. Moreover, the partition
would consist of the following n+1 points:
x0 = a = a + 0 . b - a / n
x1 = a + 1 . b - a / n
...
xn-1 = a + (n-1) . b - a / n
xn = a + n . b - a / n = b
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.2(c):
Show that if P' is a refinement of P then | P' | | P |
Back
To prove this fact is more confusing than enlightening. It seems clear that if one or more
points are inserted into the partition P to form the refinement partition P', the largest
distance between the points of P' must now be less than (or equal to) that of the points of
P.
But alas, even things that "seem clear" still need formal proof, so ...
Since | P | is a maximum, there must be at least one integer j such that |
P | = xj+1 - xj. Take all such points from the partition P, i.e. all points such
that | P | = xj+1 - xj. Now consider the refinement P'.
Suppose none of the additional points are inside the intervals [xj, xj+1]. Then the
original maximum has not changed so that | P | = | P' |
Suppose at least one of the additional points is inside at least one of the subintervals
[xj, xj+1]. Then this subinterval can no longer contribute to the maximum of P' so
that | P' | | P |
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.4:
Suppose f(x) = x2 on [0, 2]. Find
1. the fifth Riemann sum for an equally spaced partition, taking always
the left endpoint of each subinterval
2. the fifth Riemann sum for an equally spaced partition, taking always
the right endpoint of each subinterval
3. the n-th Riemann sum for an equally spaced partition, taking always
the right endpoint of each subinterval.
Back
The right and left Riemann sums are illustrated in the Java applet below. To verify the
result of that applet, let's perform the computation manually as well.
The interval is [0, 2], and we want to find the fifth Riemann sum.
Therefore the partition we need is:
Examples 7.1.6(a):
Suppose f(x) = x2-1 for x in the interval [-1, 1]. Find:
1. The left and right sums where the interval [-1, 1] is subdivided into
10 equally spaced subintervals.
2. The upper and lower sums where the interval [-1, 1] is subdivided
into 10 equally spaced subintervals.
3. The upper and lower sums where the interval [-1,1] is subdivided
into n equally spaced subintervals.
Back
For the first two questions we can again use a Java applet that will perform the
computations for us. Here is are the results computing the left and right sums.
Note, in particular, that right and left Riemann sums are the same (an
accident ?). The upper and lower sums are similarly computed, noting
that in the interval [0, 1] the largest value of x2-1 over any subinterval
inside [0, 1] is always the right endpoint, while the smallest value occurs
on the left endpoint. For subintervals inside [-1, 0] it is just the other
way around.
Note that this time the values don't agree but we have that L(f, P) U(f,
p), which is again no accident.
To answer the last question we can not use the above Java applet
because we don't know a numeric value for n. Here is the appropriate
manual computation for the, say, the upper sum: We are looking for n
equally spaced subintervals of [-1, 1] so that our partition consists of the
points:
xj = -1 + j * 2/n
where j = 0, 1, ..., n and | P | = 2/n.
Case 1: n is even, i.e. n = 2 N for some integer N
The renumbered partition now is:
xj = -1 + j * 1/N
where j = 0, 1, ..., N, N+1, ..., 2 N. In that case the point 0 is part of the partition (take j =
N). The point to notice before we can begin the computation is that the function is
decreasing over the interval [-1, 0] and increasing over the interval [0, 1]. That implies
that for all partition points less than 0 the maximum of f occurs on the left endpoint, for
points bigger than 0 the maximum occurs on the right. Hence:
U(f, P) =
In appropriate "sigma" notation we therefore have:
U(f, P) =
=
=
=
=
=
=
=
We have used the results on the sum of square integers mentioned in the chapter on
Induction to simplify the various sums.
We can verify our formula by looking at the above applet. There the 10upper sum computes to -1.119999. In our formula we need to let N = 5
because 10 = n = 2 N. Substituting N = 5 gives the exact value of -1.12
for the 10-th upper sum.
The case where n is odd, as well as the computation of the lower sum, is
left as an exercise. There's nothing new in those computations (but for n
odd there's something special happening at 0), it's just somewhat
tedious. You can check whichever formulas you come up with against
the numeric answers of the above applet.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.6(b):
Why is, in general, an upper (or lower) sum not a special case of a Riemann
sum ? Find a condition for a function f so that the upper and lower sums are
actually special cases of Riemann sums.
Back
The reason is simple: a Riemann sum requires points where the function is defined, while
an upper/lower sum involves an sup/inf which may not correspond to points in the range of
the function.
What this means is best illustrated via an example. Take the function
Now take the simple partition consisting of the interval [-1, 1]. Since f(x)
> 0 for all x (which is not visible in the above applet) we must have that
any Riemann sum R(f, P) must give a value strictly bigger than 0. The
lower sum with respect to this partition, on the other hand, is truly zero
(which again is not visible in the above applet (why?)). Therefore this
particular lower sum can not be a Riemann sum.
So to find a condition that ensures that upper/lower sums are special
cases of Riemann sums we must ensure that the sup/inf that appears in
the definition of upper/lower sum is a max or min, respectively. In the
topology chapter we have shown that a continuous function over a
closed, bounded interval must have a max and a min.
Therefore, if f is continuous over the interval [a, b] then the upper and
the lower sum are both special cases of a Riemann sum.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.6(c):
Find conditions for a function so that the upper sum can be computed by
always taking the left endpoint of each subinterval of the partition, or
conditions for always being able to take the right endpoints.
Back
To ensure, for example, that the left endpoint will always be used for the computation of a
lower sum we must ensure that regardless of the partition that was chosen the function
takes its smallest value inside every partition subinterval on its left endpoint.
We will leave it to you to find the correct condition(s), but you can use
the applet below to experiment with various functions.
Click on Options and use the functions:
f(x) = x2
f(x) = 1 - x2
For one of them the right sum is identical to the upper sum, for the other it is identical to
the lower sum. Perhaps that helps finding the right conditions.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.6(d):
Suppose f is the Dirichlet function, i.e. the function that is equal to 1 for
every rational number and 0 for every irrational number. Find the upper and
lower sums over the interval [0, 1] for an arbitrary partition.
Back
Take an arbitrary partition P = { x0, x1, ..., xn } of the interval [0, 1].
Between any two points xj and xj+1 there is an irrational number. Therefore the inf
over [ xj, xj+1 ] must be 0. That means that L(f, P) = 0.
Between any two points xj and xj+1 there is a rational number. Therefore the sup
over [ xj, xj+1 ] must be 1. That means that
U(f, P) = (x1 - x0) + (x2 - x1) + ... + (xn - xn-1)
which is a telescoping sum so that
U(f, P) = xn - x0 = 1 - 0 = 1
Thus, we have shown that for the Dirichlet function and for any partition P we have that
L(f, P) = 0 and U(f, P) = 1.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The last statement is simple to prove: take any partition P = {x0, x1, ..., xn}. Then
inf{ f(x), xj-1 x xj } f(tj) sup{ f(x), xj-1 x xj }
where tj is an arbitrary number in [xj-1, xj] and j = 1, 2, ..., n. That immediately implies that
L(f, P) R(f, P) U(f, P)
The other statements are somewhat trickier. Let's first find out why they
should be true. To make it simple, let's say that P = {a, b} and P' = {a,
x0, b}. Then
U(f, P) = sup{ f(x), x
[a, b] } (b - a)
[x0, b] } (b - x0)
Geometrically, the upper sum for P corresponds to one large rectangle, the one for P' to
two smaller rectangles, where the smaller rectangles fit into the larger one but do not cover
it.
U(f, P) = 1.089 U(f, P') = 0.86
Since the area covered by the two rectangles is smaller than that covered by the first one,
we have U(f, P) > U(f, P').
Let's show this mathematically, in case one additional point t0 is added
to a particular subinterval [xj-1, xj]. Let:
cj be the sup of f(x) in the interval [xj-1, xj]
Aj be the sup of f(x) in the interval [xj-1, t0]
Bj be the sup of f(x) in the interval [t0, xj]
Then cj Aj and cj Bj so that
cj (xj - xj-1) = cj (xj - t0 + t0 - xj-1) = cj (xj - t0) + cj (t0 - xj-1)
Bj (xj - t0) + Aj (x0 - tj-1)
That shows that if P = {x0, ... xj-1, xj, ..., xn} and P' = {x0, ... xj-1, t0, xj, ..., xn} then U(f, P)
U(f, P').
The proof for a general refinement P' of P uses the same idea plus some
confusing indexing scheme. No more details should be necessary.
The proof for the statement regarding the lower sum is analogous.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.9(a):
Show that the constant function f(x) = c is Riemann integrable on any
interval [a, b] and find the value of the integral.
Back
We have to compute the upper and lower sums for an arbitrary partition, then find the
appropriate inf and sup to compute the lower and upper integrals. If they agree, we are
done and the common value is the answer. So, here we go:
Take an arbitrary partition P = {x0, x1, ..., xn}. The lower sum of f(x) = c
is:
L(f, P) = c (x1 - x0) + c (x2 - x1) + ... + c (xn - xn-1)
= c (xn - x0) = c (b - a)
because the inf over any interval (as well as the sup) is always c, and the above sum is
telescoping.
Similarly, we have that
U(f, P) = c (b - a)
Hence, the upper and lower sums are independent of the particular partition. Therefore f is
integrable and
I*(f) = I*(f) = c (b - a)
In particular,
f(x) dx = c (b - a)
Examples 7.1.9(b):
Is the function f(x) = x2 Riemann integrable on the interval [0,1] ? If so, find
the value of the Riemann integral. Do the same for the interval [-1, 1].
Back
2
First let's experiment with our "Integrator" applet. To "show" that f(x) = x is integrable we
need to take partitions with more and more points, compute the upper and lower sum, and
hope that the numeric answers will get closer and closer to one common value. If that's the
case, our guess is that this limit is the integral of f over the indicated interval.
But of course that's no proof. As it turns out, to prove that this simple
function is integrable will be difficult, because we do not have a simple
condition at our disposal that could tell us quickly whether this, or any
other function, is integrable. That's the bad news; the good news will be
that we should be able to generalize the proof for this particular
example to a wider set of functions.
Anyhow, here we go. First we should note that in the definition of upper
and lower integral it is not necessary to take the sup and inf over all
partitions. After all, if P is a partition and P' is a refinement of P then L(f,
P') L(f, P) and U(f, P') U(f, P). Therefore partitions with large norm
don't contribute to the sup or inf and it is enough to compute the upper
and lower integral by considering partitions with a small norm only.
Next, take any > 0 and a partition P with |P| < / 2. Then
where cj is the sup of f over [xj-1, xj] and dj is the inf over that interval.
Since f is increasing over [0, 1] we know that the sup is achieved on the
right side of each subinterval, the inf on the left side. Therefore:
| U(f, P) - L(f, P)|
|cj - dj| (xj - xj-1) =
To estimate this sum, we'll apply the Mean Value Theorem for f(x) = x2:
|f(x) - f(y)| |f'(c)| |x - y|
for c between x and y. Since |f'(c)| 2 for c in the interval [0, 1] we know that
|f(x) - f(y)| 2 |x - y|
But the partition P was chosen with |P| < / 2 so that
U(f, P) =
cj (xj - xj-1) =
f(xj) 1/n
x2 dx = 1/3
where a = 0 and b = 1.
As for the interval [-1, 1], we can first play our applet game to get an
idea about the answer, then set out to prove everything just as we did
above.
It seems that the applet says that the integral now should evaluate to
something close to 2/3. Of course we need some formal proof, but since
that would be similar to the above one we'll leave it as an exercise.
The proof is somewhat misleading, because it seems to be based on the
fact that f(x) = x2 is differentiable. As a generalization we might
conclude that differentiable functions are integrable. That's correct, but
another, more general concept can be substituted for differentiability
(what might it be?).
Examples 7.1.9(c):
Is the Dirichlet function Riemann integrable on the interval [0, 1] ?
Back
In a previous example we have shown that U(f, P) = 1 and L(f, P) = 0, regardless of the
partition. Therefore:
I*(f) = 1 and I*(f) = 0
Therefore, the Dirichlet function is not integrable over the interval [0, 1].
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
One direction is simple: If f is Riemann integrable, then I*(f) = I*(f) = L. By the properties
of sup and inf we know:
There exists a partition P such that L = I*(f) > U(f, P) - / 2
there exists a partition Q such that L = I*(f) < L(f, Q) + / 2
Take the partition P' that is the common refinement of P and Q. Then we know that:
U(f,P) U(f,P')
L(f,Q) L(f,P')
Taking this together we have:
L > U(f,P) - / 2 U(f,P') - / 2
L < L(f,Q) + / 2 L(f,P') + / 2
Multiplying the second inequality by -1 and adding it to the first gives:
0 > U(f,P') - L(f,P') or equivalently:
> U(f,P') - L(f,P') = | U(f, P') - L(f, P')|
Therefore we found a particular partition (namely P') such that
| U(f, P') - L(f, P')| <
When we proved before that the integral of f(x) = x2 exists and is equal to 1/3 we showed
that
(*) U(f, P) - L(f, P)
for every partition P with small enough norm. Using Riemann's Lemma it is enough to find
one partition such that inequality (*) holds. Therefore, take the partition
P = {j/n, j = 0, 1, 2, ..., n}
for the interval [0, 1]. On every subinterval [(j-1)/n, j/n] the maximum of f occurs on the
right, the minimum on the left. Therefore:
| U(f, P) - L(f, P) |
=
= 1 / n3 | j2 - j2 + 2 j - 1| 2 / n3
= 2 / n3 1/2 n (n+1) = (n+1) / n2
| j2 - (j-1)2 |
j
But since the last expression converges to zero as n goes to infinity, Riemann's Lemma
shows that f(x) = x2 is indeed integrable over the interval [0, 1].
The remainder of this example is just as before (exercise!).
So, Riemann's lemma has indeed simplified our computation, because
we are now able to pick one partition that is best suited for our
particular function and/or interval.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
function, i.e. f(-x) = -f(x). Show that the integral of f over [-a, a] is zero.
What can you say if f is an even function?
Back
By assumption f is Riemann integrable so we can use the previous example and take a
sequence of partitions with smaller and smaller mesh, compute a Riemann sum for each
partition, and find the limit.
For an evenly spaced partition that includes 0 it is easy to compute a
particular Riemann sum (such as a "middle Riemann sum") as long as
the function is odd.
The rest is left as an exercise.
For even functions, i.e. functions where f(x) = f(-x), we can show that
the integral of f from -a to a is twice the integral from 0 to a. The details
are left as an exercise again (sorry -:).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
c f(x) + d g(x) dx = c
f(x) dx + d
g(x) dx
3. | f(x) dx |
| f(x) | dx
4. If g is another function defined on [a, b] such that g(x) < f(x) on [a,
b], then g(x) dx
f(x) dx
5. If g is another Riemann integrable function on [a, b] then f(x) . g(x)
is integrable on [a, b]
Context
Proof:
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.13(a):
x dx
x sin(x) dx
x dx
-8
x sin(x) dx 8
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.13(b):
f(x) dx = 3 dx = 3 (2 - 1) = 3
Therefore
Examples 7.1.13(c):
If f is an integrable function defined on [a, b], which is bounded by M on
that interval, prove that
M (a - b)
f(x) dx M (b - a)
Back
-M dx
f(x) dx
M dx
or equivalently (we have computed the left and right integrals before):
-M (b - a)
f(x) dx M (b - a)
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
We have shown before that f(x) = x2 is integrable where we used the fact that f was
differentiable. We will now adjust that proof to this situation, using uniform continuity
instead of differentiability. We can actually simplify the previous proof because we now
have Riemann's lemma at our disposal.
where cj is the sup of f over [xj-1, xj] and dj is the inf over that interval. Since the function is
continuous, it assumes its maximum and minimum over each of the subintervals, so that
inf and sup can be replaced with min and max.
Since the norm of the partition is less than we know that |cj - dj| < for
all j.
But then
| U(f, P) - L(f, P)|
(xj - xj-1)
=
(xj - xj-1)
= (xn - x0)
= (b - a)
because the last sum is telescoping. That finishes the proof.
For the purist, we should have chosen such that |f(x) - f(y)| < / (b-a)
whenever |x - y| < so that at the end of our above computation we
could get a single . But that's just details...
It is easy to find an example of a function that is Riemann integrable but
not continuous. For example, the function f that is equal to -1 over the
interval [0, 1] and +1 over the interval [1, 2] is not continuous but
Riemann integrable (show it!).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.15:
Find a function that is not integrable, a function that is integrable but not
continuous, and a function that is continuous but not differentiable.
Back
Take the function that equals 1 over the interval [0, 1] and 2 over the
interval [1, 2]. It is clear that the function is not continuous, but we need to
prove that it is integrable.
Take a partition P of the interval [0, 2] with norm less than
such that the point xk = 1 is part of the partition. Then:
Proof:
To prove this is not easy; we will start with a simpler version of this theorem: if f is
continuous and bounded over the interval [a, b] except at one point xk, then f is Riemann
integrable over [a, b].
| U(f,P) - L(f,P) |
Examples 7.1.17(a):
Show that every monotone function defined on [a, b] is Riemann integrable.
Back
We have shown before that a monotone function f defined on a closed interval [a, b] has at
most countably many discontinuities.
Therefore such a function f is continuous except at countably many
points, so that by our previous theorem the function must be Riemann
integrable.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.17(b):
Let
where p, q relatively prime and q > 0, and assume g is restricted to [0, 1]. Is
g Riemann integrable ? If so, what is the value of the integral ?
Back
We have seen this function before, where we have shown that it is continuous at all
irrational numbers and discontinuous at the rationals. In particular, the function has
countably many points of discontinuity. Since the discontinuities are dense, i.e. they are
"all over" the interval [0, 1] it might seem that it is difficult to find the value of the integral
(if the function is Riemann integrable). But with the theoretical background we developed
so far it will be easy to compute the answer.
Having countably many discontinuities, we know by our previous
theorem that the function is Riemann integrable and it remains to find
the value of the integral.
Take any partition P = {x0, x1, ..., xn} and look at:
dj = inf{g(x): x
[xj-1, xj]}
Since every subinterval [xj-1, xj] contains irrational numbers we clearly have that dj = 0 for
all j. But then the lower integral I*(g) = sup{ L(g,P): P a partition of [a, b]} must also be 0.
Since g was integrable the upper and lower integral agree so that
g(x) dx = 0
for a = 0 and b = 1.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
F(x) = f(t) dt
Then F is continuous in [a, b]. Moreover, if f is also continuous, then F is
differentiable in (a, b) and
F'(x) = f(x) for all x in (a, b).
Note: In many calculus texts this theorem is called the
Second fundamental theorem of calculus. Those books also
define a First Fundamental Theorem of Calculus.
Context
Proof:
The first assumption is simple to prove: Take x and c inside [a, b]. Since f is bounded we
know that | f(x) | < M for some number M. By the properties of the Riemann integral we
know:
assuming without loss of generality that c > x. But then we can take the limit as x
approaches c to see that
| F(x) - F(c) | = 0
which implies continuity of F at x = c.
Now we want to prove that F(x) is differentiable with F' = f if f is
continuous.
Pick x inside the interval (a, b) and choose a number h so small that
x+h is also in (a, b). We compute the difference quotient:
and define
m = inf{ f(t), t [x, x+h] }
M = sup{ f(t), t [x, x+h] }
Clearly we have that m f(t) M so that by the properties of the Riemann integral we have
that:
Proof:
This being a corollary means that it must be easy to prove. We already know from the
previous theorem that if we define the function
F(x) = f(t) dt
so that H(x) = c for some constant c. But then F(x) = G(x) + c so that
F(b) - F(a) = (G(a) + c) - (G(b) + c) = G(b) - G(a)
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.20(a):
Define a function F(x) = t2 sin(t) dt for x in the interval [a, a + 10].
1. Find F(a)
2. Find F'(x)
3. Find F''(x)
4. Find all critical points of F(x) in [a, a + 10]
Back
1. Since F(x) = t2 sin(t) dt the value F(a) is an integral from a to a. But such an integral is
0 regardless of the integrand. Therefore F(a) = 0.
2. The second part is a direct application of the (second) Fundamental
Theorem of Calculus:
t2 sin(t) dt = x2 sin(x)
3. Since we have computed the first derivative already, it is easy to
compute the second derivative:
F''(x) =
F'(x) =
4. To find the critical points of F we need to find the points where F is not differentiable or
where F'(x) = 0. We know that F is differentiable on any closed interval so that the critical
points are those where
F'(x) = x 2 sin(x) = 0
Therefore the critical points are x = 0 and x = k
1, ... that are inside the interval [a, a+10].
Examples 7.1.20(b):
Find the value of the following integrals:
1.
2.
3.
This time we will use the Integral Evaluation Shortcut, or First Fundamental Theorem of
Calculus, which requires us to find the antiderivative for each of the functions.
x5 - 4 x2 dx = P(b) - P(a)
With our choices of a and b we can evaluate the integral to
P(2) - P(0) = 1/6 26 - 4/3 23 = 1/6 64 - 4/3 * 8 = 0
To compute the exact value, first note that 1/x2 + cos(x) is continuous over the interval [-1,
4] so that the function is Riemann integrable. If we define P(x) = -1/x + sin(x) + C then
P'(x) = 1/x2 + cos(x) so that again we found an antiderivative of the original integrand.
Using the Integral Evaluation Shortcut over [-1, 4] we can compute the value of the
integral to
P(4) - P(1) = -1/4 + sin(4) - (-1/1 + sin(1)) = -.8482734801
The integrand is continuous everywhere, so that it is Riemann integrable over any closed
bounded interval. To find an antiderivative of (1 + x2)-1 is less obvious than in the previous
cases until we remember that
arctan(x) = (1 + x2)-1
But then the integral evaluates to
arctan(1) - arctan(-1) = 2
/4 =
/2 = 1.570796327
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Examples 7.1.20(c):
Show that if one starts with an integrable function f in the Fundamental
Theorem of Calculus that is not continuous, the corresponding function F
may not be differentiable.
Back
Our Fundamental Theorem states that if we start with a continuous function f(t) over some
interval [a, b] then the new function F(x) obtained by integrating f from a to some variable
value x is differentiable:
and for x 0
But then
(b) 1/x dx
(c) ex dx
(f) tan(x) dx
(g)
(i)
(h)
Back
For each function to integrate we need to come up with another function whose derivative
is the integrand. Therefore we guess a function F, then differentiate it (for which there are
fixed rules). If indeed F' is equal to the integrand, we have found an antiderivative and we
are done.
Actually, if we can find one antiderivative F of a function f, the function
F(x) + c is also an antiderivative. To simplify notation, we will be content
with finding one antiderivative and not note the arbitrary constant
(which drops out anyway when evaluating F(b) - F(a)).
(a) xr dx
We need a function whose derivate is f(x) = xr. Since differentiating a power
reduces it by one, finding the opposite might mean increasing the power by
one. Therefore, our first guess for the antiderivative is F(x) = xr+1. A quick
check shows:
xr+1 = 1/r+1 xr
which has the correct power but an incorrect coefficient. Therefore we
guess again: let F(x) = 1/r+1 xr+1, where r # -1. Then
F'(x) = r+1/r+1 xr = xr
so that we have found the correct antiderivative, valid for all x and for r #
-1:
xr dx = 1/r+1 xr+1
(b) 1/x dx
This is x to the power -1, which we had to exclude from the example above.
Simple power functions, therefore, are not sufficient as antiderivatives and
we need to resort to more exotic functions. Thinking a little we remember
that there is the natural logarithm ln whose derivative is
ln(x) = 1/x
But that function is only defined for positive x. On the other hand, if x is
negative then -x is positive, and
(g)
This is again tricky. We could guess for quite a while here, not much seems
to work. But then - again an intuitive wonder - we remember the arctan
function, whose derivative is just what we need:
that:
arctan(x) = 1/ 1 + x2 so
= arctan(x)
(h)
Well, tricky again. But we already tried to find the antiderivative of the tan
in one of the earlier examples. While we were trying to find that answer, we
might have computed the derivative of the tan, which was at that time not
helpful. But it works now, as a quick application of the quotient rules
shows:
Therefore:
= tan(x)
(i)
Now it get's to be really exotic. We have already seen that the derivative of
the arctan function yields an expression with an x2 in the denominator.
Therefore we should recall the derivatives of the other two (less frequently
used) inverse trig functions arccos and arcsin.
Therefore we have:
= arcsin(x)
Incidentally, can you prove that the derivative for the arcsin is indeed as
shown above (or for the arccos or arctan, for that matter)?
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof
Proof:
f is continuous so that there exists a function F with F' = f (in other words, F is an
antiderivative of f). Differentiate the function F(s(x)) using the Chain Rule:
f'(x) g(x) dx
Back
Proof:
For the function G(x) = f(x) g(x) we have by the Product Rule:
G(x) =
Therefore the function G is an antiderivative of the function f'(x) g(x) + f(x) g'(x) which
means that
G(b) - G(a) =
=
f'(x) g(x) dx +
f(x) g'(x) dx
g(x) dx
Back
Proof
Define the numbers
m = inf{ f(x): x [a, b] }
M = sup{ f(x): x [a, b] }
g(x) dx
f(x) g(x) dx M
g(x) dx
g(x) dx =
f(x) g(x) dx
But since f is continuous on [a, b] and d is between m and M, we can apply the
Intermediate Value Theorem to find a number c such that f(c) = d. Then
f(c)
g(x) dx =
f(x) g(x) dx
[a, b] }
We first prove a simpler version of the trapezoid rule using the Mean Value Theorem for
integrals and integration by parts.
Simple Trapezoid Rule: Let f be a function defined on the interval [0, 1] so
that f is twice continuously differentiable. Then there exists a number c [0,
1] so that
f(x) dx = 1/2 (f(0) + f(1)) - 1/12 f''(c)
The trick to prove this statement is to define a function
v(x) = 1/2 x (1 - x)
v''(x) = -1 (green)
Then
f(x) dx = -
v''(x) f(x) dx
v''(x) f(x) dx = v'(1) f(1) - v'(0) f(0) = -1/2 f(1) - 1/2 f(0) -
v'(x) f'(x) dx =
v'(x) f'(x) dx
v(x) f''(x) dx =
v(x) f''(x) dx =
- f''(c)
where we used the Mean Value Theorem for Integration with some number c inside the
interval [0, 1]. Taking everything together (careful with the negative signs) we then have:
= 1/h
g(x) dx =
f(u(x)) dx =
f(u) du
Therefore:
where
sin(x) cos(x) dx
obtained by using (a) a left Riemann sum and (b) the Trapezoid Rule, using
a partition of size 5 and of size 100.
Back
First, let's determine the exact value of the integral by substitution. Let
u = sin(x) so that
du/dx = cos(x) or du = cos(x) dx
Then
sin(x) cos(x) dx = u du =
= 1/2 (sin2(1) - sin2(0)) = 0.5(0.7080734183 - 0) = 0.3540367092
To find the left Riemann sum, we let f(x) = sin(x) cos(x) and compute:
f(0) = 0
f(0.2) = 0.1947091712
f(0.4) = 0.3586780454
f(0.6) = 0.4660195430
f(0.8) = 0.4997868015
and therefore
R(P, f) = 1/5*(f(0) + f(0.2) + f(0.4) + f(0.6) + f(0.8)) =
= 0.2 * 1.519193562 = 0.3038387122
The error between the approximate and exact value is about 0.05, or 14%. To estimate the
error using the trapezoid rule, we compute f''(x) = -4 sin(x) cos(x) so that K = 4. Then
|R| < 4/12 * 1 * 0.22 = 0.04 / 3 = 0.013
so that even the theoretically worst error is a lot better (less than 4%). To find the value
using the trapezoid rule, we need to evaluate f at the same values as before and also
compute f(1) = 0.4546487134. Then the trapezoid rule, combining these numbers a little
different than the left Riemann sum, gives:
[1/2 f(0) + f(0.2) + f(0.4) + f(0.6) + f(0.8) + 1/2 f(1)] * 0.2 =
= [0 + 1.519193562 + 0.5 * 0.4546487134 ] * 0.2 =
= 0.3493035838
which is indeed much closer than our previous approximation (the error is only about 1%).
Incidentally, if we increased the size of the partition to 100, the error committed by the
trapezoid rule would already be less than 1/3 * 0.012 = 0.00003 or 0.01%, whereas the
left Riemann sum still has an error of about 0.003, or about 1% (compare the value of
the applet with the exact value above).
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth
Proof:
The proof is easiest when allowing complex numbers instead of reals and is not given.
Interactive Real Analysis, ver. 1.9.3
(c) 1994-2000, Bert G. Wachsmuth