Sei sulla pagina 1di 34

Solving Systems of

Equations
Rule of Thumb:
More equations than unknowns system is unlikely to have a
solution.
Same number of equations as unknowns system is likely to
have a unique solution.
Fewer equations than unknowns the system is likely to have
infinitely many solutions.
Solving a System with Several
Equations in Several Unknowns
We are interested in this last situation!
Systems of Equations
Linear system.
Matrix methods.
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ + =
+ =
+ =
Note: We expect to have as many free variables
as the difference between the number of unknowns
and the number of equations.
Systems of Equations
4 3 ) cos( 4
1 ) sin( 3
= + +
= +
t t y
t x y
Linear system.
Matrix methods.
Non-linear system.
Can be solved analytically
More difficult!
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ + =
+ =
+ =
cos( ) 3 sin( ) 4 0 x y x y + =
Linear Systems of Equations
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ =
+ =
+ =
We can write this as a matrix/vector equation as follows:
1
2
3
1
2
2 2 1 2 11 0
1 0 4 2 5 0
1 1 1 1 5 0
y
y
y
x
x
(
(

( (
(
( (
( =
( (
(
( (

(
(

We can solve this
system using
elementary row
operations
Solving the System
1
2
3
1
2
1 0 0 2 1 0
0 1 0 3 5 0
0 0 1 0 1 0
y
y
y
x
x
(
(

( (
(
( (
= (
( (
(
( (

(
(

1 0 0 2 1
0 1 0 3 5
0 0 1 0 1

(
(
(
(

Row-reduced coefficient matrix:
Now we can reinterpret this reduced
matrix as a solution to the system of
equations.
1 1 2
2 1 2
3 2
2
3 5
y x x
y x x
y x
= +
=
=
The Solution Function
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ =
+ =
+ =
1 1 2
2 1 2
3 2
2
3 5
y x x
y x x
y x
= +
=
=
We started with three equations in five unknowns.

As expected, our first three unknowns depend on
two free variables, two being the difference
between the number of unknowns and the number
of equations.
Where are we?
Solving a system of m equations in n unknowns
is equivalent to finding the zeros of a vector
valued function from

n
.

When m > n, such a system will typically
have infinitely many solutions. In nice cases,
the solution will be a function from

m-n

n
.


What about finding solutions?
4 3 ) cos( 4
1 ) sin( 3
= + +
= +
t t y
t x y
most are impossible
to solve analytically!
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ + =
+ =
+ =
Though we completely understand
how to solve systems of linear
equations . . .
and we luck out with a small
number of non-linear systems . . .
cos( ) 3 sin( ) 4 0 x y x y + =
The Implicit Function Theorem. . .

Now we Introduce a useful theoretical framework
. . . gives us conditions under which a system of m equations
in n unknowns will have a solution.

. . . and it tells us that (under reasonable hypotheses) the
solution function will be well behaved.
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11
4 2 5
5
y y y x x
y y x x
y y y x x
+
| |
|
= + =
|
|
+
\ .
F 0
1 2
1 2 1 2
2
2
( , ) 3 5
x x
x x x x
x
+
| |
|
=
|
|

\ .
g
1 2 3 1 2
1 3 1 2
1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
y y y x x
y y x x
y y y x x
+ =
+ =
+ =
1 1 2 3 1 2
2 1 3 1 2
3 1 2 3 1 2
2 2 2 11 0
4 2 5 0
5 0
f y y y x x
f y y x x
f y y y x x
= + =
= + =
= + =
1 1 2
2 1 2
3 2
2
3 5
y x x
y x x
y x
= +
=
=

F:
5

3

g :
2

3


So How do we Proceed?
cos( ) 3 sin( ) 4 0 x y x y + =
This equation is
already too hard.
Well start with something a bit simpler.
What would you do with this one?
( )
3 2
1
4 25 150 0
75
x x x + + =
So How do we Proceed?
cos( ) 3 sin( ) 4 0 x y x y + =
This equation is
already too hard.
Well start with something a bit simpler.
What would you do with this one?
( )
3 2
1
4 25 150 0
75
x x x + + =
So How do we Proceed?
cos( ) 3 sin( ) 4 0 x y x y + =
This equation is
already too hard.
Well start with something a bit simpler.
What would you do with this one?
( )
3 2
1
4 25 150 0
75
x x x + + =
Iteration!
Recall:
x
0
x
0
x
0
R
R
we are looking for a fixed point
Solving equations
Suppose we want to find a root for the equation



We can add x to both sides of the equation



and reinterpret the problem as a fixed point problem.
( )
3 2
1
4 25 150 0.
75
x x x + + =
( )
3 2
1
4 25 150
75
x x x x x + + =
Root finding via iteration
When we iterate the function


starting at x
0
=0, we get
( )
3 2
1
( ) 4 25 150
75
R x x x x x = + +
0 1 2 10
20 40 100
0, 2, 4.77, 8.91, ,
8.80, , x 8.78, , x 8.78,
x x x x
x
= ~ ~ =
~ ~ ~
( )
3 2
1
4 25 150 0.
75
x x x + + =
The limit is, indeed, a solution to the equation
Problem. . .
This works nicely when the iterated map converges, but
it may not converge.
For instance, if we try the same trick with


the iterated map goes to infinity.

Whats the difference? Whats going on?
( )
3 3
3 2 4 0 3 2 4 x x x x x x + = + + =
Recall: Attractors and repellors
x
0
The key difference is the
derivative of the function
at the fixed point.
M
The Derivative at the Fixed Point
Theorem: Suppose that R is continuously differentiable on
[a,b] and that p is a fixed point of R lying in [a,b]. If
then there exists some subinterval I of [a,b]
containing p such that if x
0
is in I, then




Conversely, if then there is some interval I about
p such that every iterated map based a point of I (other than
p, of course) will eventually leave I and will thus not
converge to p.
2 3 4
0 0 0 0 0
, ( ), ( ), ( ), ( ) . x R x R x R x R x p
( ) 1 R p
'
<
( ) 1 R p
'
>
Let x be a real number and K a non-zero real number.
Note that x is a solution to the equation


if and only if it is also a solution to the equation



Rethinking the scheme
( ) 0 f x =
( ) 0. K f x =
So our fixed point root-finding scheme can be applied to
any equation of the form
( ) . x K f x x + =
Choose the value
of K so that the
derivative is
favorable
The best of all possible worlds
It is not hard to show that the rate at which the iterated maps
on R converge to the fixed point p is governed by the size of
| R '(p) | . The closer | R '(p) | is to zero, the faster the
convergence rate.

While we are making choices, we might as well be ambitious
and choose K so that we not only get convergence to the
root, but we get really fast convergence!

For we want

( ) ( ) R x x K f x = +
( ) 1 ( ) 0 R p K f p
' '
= + =

1
( )
K
f p
=
'
Works, provided
that f '(p) = 0.
So in the best of all possible worlds, we want to iterate the
function


where p is a root of f.
( )
( )
( )
f x
R x x
f p
=
'
Notice that this looks remarkably like the iteration function
for Newtons method:
( )
( )
( )
f x
N x x
f x
=
'
Constant (given p!),
and ideally chosen for
convergence near p.
Evaluated each step,
but not far from f (p)
when x is close to p.
Quasi-Newton methods: Iterative ways to find p with f (p) = 0
Must evaluate new f (x) each step.
Requires no magic knowledge and
converges in a neighborhood of p .
( )
( )
( )
f x
N x x
f x
=
'
Newtons method
Pretty good method
D
x f
x x Q
) (
) ( =
( )
( )
( )
f x
R x x
f p
=
'
Leibnitz method
If we choose D near enough to f (p) so
that |Q(p)| < 1/2, iterated maps will
converge in a neighborhood of p .
Best of all possible worlds: Since
R(p) = 0, iterated maps converge
very quickly in a neighborhood of p .
One Equation in Two Unknowns
cos( ) 3 sin( ) 4 0? x y x y + =
Can we use this set of ideas to tackle
Yes!
First we recall that the solutions to this equation are the zeros of
the 2-variable function
( , ) cos( ) 3 sin( ) 4 f x y x y x y = +
0-Level Curves
( , ) cos( ) 3 sin( ) 4 0 f x y x y x y = + =
cos( ) 3 sin( ) 4 0 x y x y + =
The pairs (x,y) that satisfy the
equation f(x,y)=0 lie on the
intersection of the graph of f
and the horizontal plane z = 0.

That is, they lie on the 0-level
curves of the function f.

Taking a piece
The 0-level curves of f.
Though the points on
the 0-level curves of f
do not form a function,
portions of them do.
) , ( y x f z =
0 ) , ( = y x f
x
y
Consider the contour line f (x,y) = 0 in the xy-plane.
Idea: At least in small regions, this curve might be described
by a function y = g(x) .
Our goal: find such a function!
x
y
0
) , (
= |
.
|

\
|
c
c
=
b a
y
f
D
(a,b)
(a,b)
y = g(x)
Start with a point (a,b) on the
contour line, where partial
with respect to y is not 0:
In a small box around (a,b),
we can hope to find g(x).
Make sure all of the y-partials
in this box are close to D.
( )
( )
D
y x f
y y
x
,
= |
(a,b)
How to construct g(x):
For this x, construct function
Iterate |
x
(y) to find its fixed
point, where f (x,y) = 0. Let
fixed point be g(x).
x
y
z
f (x,y) (x fixed)
Whats going on here?
For our fixed value of x, we find
g(x) so that f (x,g(x)) = 0. We do
this by iterating |
x
.
g(x)
Fix x.
( )
( )
D
y x f
y y
x
,
= |
(a,b)
How to construct g(x)
Given x, construct function
Iterate |
x
(y) to find its fixed
point, where f (x,y) = 0. Let
fixed point be g(x).
x
How do we know this works?
For x = a, this is just the ideal Leibnitz
method for finding g(a) = b.
For small enough boxes, this is a pretty
good method elsewhere.
Non-trivial issues:
Make sure the iterated
map converges.
Make sure you get the
right fixed point.
(Dont leave the box!)
b
x
y
(a,b)
(a,b)
y = g(x)
Staying in the Box?
Staying in the box?
Not so obvious . . . we
have to work a bit to
make sure we get the
right fixed point. (We
dont leave the box!)
(a,b)
x
y
x
Looking ahead
We are interested in solving systems of
equations.
The Implicit Function Theorem will tell us
when there are solutions to a system and will
describe the nature of those solutions.
We will prove the implicit function theorem by
employing quasi-Newtons methods as a
theoretical tool.

Potrebbero piacerti anche