Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
R. Kress
Introduction
Some forty years ago when I was working on my thesis I fell in love with
integral equations as one of the most beautiful topics in both pure and applied
analysis. This article is intended to stimulate the reader to share this love
with me.
The term integral equation was first used by Paul du Bois-Reymond in
1888. In general, an integral equation is an equation where an unknown
function occurs under an integral. Typical examples are
Z 1
K(x, y)(y) dy = f (x)
(1)
0
and
Z
(x) +
(2)
In these equations the function is the unknown, and the so-called kernel K
and the right hand side f are given functions. Solving these integral equations
amounts to determining a function such that (1) or (2), respectively, are
satisfied for all 0 x 1. The equations (1) and (2) carry the name of
Ivar Fredholm and are called Fredholm integral equations of the first and
second kind, respectively. In the first equation the unknown function only
occurs under the integral whereas in the second equation it also appears
outside the integral. Later on we will show that this is more than just a
formal difference between the two types of equations. A first impression on
the difference can be obtained by considering the special case of a constant
kernel K(x, y) = c 6= 0 for all x, y [0, 1]. On one hand, it is easily seen that
the equation of the second kind (2) has a unique solution given by
Z 1
c
f (y) dy
=f
1+c 0
1
R1
provided c 6= 1. If c = 1 then (2) is solvable if and only if 0 f (y) dy = 0
and the general solution is given by = f + with an arbitrary constant .
One the other hand, the equation of the first kind (1) is solvable if and only
if f is a constant, let say, f (x) = for all x. In this case every function
with mean value is a solution.
Of course, the integration domains in (1) and (2) are not restricted to the
interval [0, 1]. In particular, the integration domain can be multidimensional
and for the integral equation of the first kind the domain where the equation
is required to be satisfied need not to coincide with the integration domain.
The main aim of this article is to first guide the reader through part of the
historical development of the theory and the applications of these equations.
In particular, we exhibit their close connection to partial differential equations and emphasize their fundamental role in initiating the development of
functional analysis as the appropriate abstract framework for studying integral and also differential equations. Then in the second part we will illustrate
how integral equations play an important role in current mathematical research on inverse and ill-posed problems in areas such as medical imaging
and non-destructive evaluation.
Two mathematical problems are called inverse to each other if the formulation of the first problem contains the solution of the second problem
and vice versa. According to this definition at first glance it seems arbitrary
to distinguish one of the two problems as an inverse problem. However,
in general, one of the two problems is easier and more intensively studied
whereas the other is more difficult and less explored. Then the first problem
is denoted as the direct problem and the second as the inverse problem.
A wealth of inverse problems arise in the mathematical modeling of noninvasive evaluation and imaging methods in science, medicine and technology.
For imaging devices such as conventional X-rays or X-ray tomography the
corresponding direct problem consists of determining the images, i.e., projections from the known density distribution of an object. Conversely, the
inverse problem asks to reconstruct the density from the images. More generally, inverse problems answer questions about the cause of a given effect
whereas in the corresponding direct problem the cause is known and the effect is to be determined. A common feature of such inverse problems is their
ill-posedness or instability, i.e., small changes in the measured effect may
result in large changes in the estimated cause.
The equations (1) and (2) are linear equations since the unknown function appears in a linear fashion. Though nonlinear integral equations also
2
d, y > 0,
(3)
f (y) =
y
0
for the total time f (y) required for the object to fall from P = ((y), y) to
P0 = (0, 0) where
s
1 + [ 0 ]2
:=
2g
and g denotes the earth gravity. Equation (3) is known as Abels integral
equation. Given the shape the falling time f is obtained by simply evaluating the integral on the right hand side of (3). However, the solution of the
generalized tautochrone problem requires the solution of the inverse problem,
i.e., given the function f the solution of the integral equation (3) has to
be found which certainly is a more challenging task. This solution can be
shown to be given by
Z y
1 d
f ()
(y) =
d, y > 0.
(4)
dy 0
y
p
For the special case of a constant f = a/2g with a > 0 one obtains from
3
a
1
y
and it can be seen that the solution of this equation is given by the cycloid
with parametric representation
(x(t), y(t)) =
a
(t + sin t, 1 cos t),
2
0 t .
Early years
where denotes the unit normal vector to the boundary curve directed into
the exterior of D. Now the so-called jump relations from potential theory
require that
Z
ln |x y|
1
1
(y)
(x) +
ds(y) = f (x)
(8)
(y)
(9)
and
Z
(x) +
K(y, x)(y) dy = 0
(10)
and
Z
(x) +
(12)
have a unique continuous solution and for each continuous right hand
side f and g, respectively, or the homogeneous equations (9) and (10) have
the same finite number of linearly independent solutions and the inhomogeneous
integral equations are solvable if and only if the right hand sides satisfy
R1
f (x)(x) dx = 0 for all solutions to the homogeneous adjoint equation
0
R1
(10) and 0 (x)g(x) dx = 0 for all solutions to the homogeneous equation
(9).
We explicitly note that this theorem implies that for the first of the two
alternatives each one of the four properties implies the three other ones.
Hence, in particular, uniqueness for the homogeneous equation (9) implies
existence of a solution to the inhomogeneous equation (11) for each right
hand side. This is notable, since it is almost always much simpler to prove
uniqueness for a linear problem than to prove existence.
Fredholms existence results also clarify the existence of a solution to
the boundary integral equation (8) for the Dirichlet problem for the Laplace
6
on the space C[0, 1] of continuous functions maps bounded sets into relatively
compact sets, i.e., in todays terminology A is a compact operator.
What is fascinating about the work of Riesz is that his proofs are still
usable and can almost literally be transferred from the case of an integral
operator in the space of continuous functions to the general case of a compact
operator in a normed space. Riesz knew about the generality of his method
since he explicitly noted that the restriction to continuous functions was not
relevant.
The results of Riesz can be summarized into the following theorem where
I denotes the identity operator.
Theorem 4.1 For a compact linear operator A : X X in a normed space
X either I + A is injective and surjective and has a bounded inverse or the
null space of N (I + A) := { : + A = 0} has nonzero finite dimension
and the image space (I + A)(X) is a proper subspace of X.
The central and most valuable piece of the Riesz theory is again the equivalence of injectivity and surjectivity. Theorem 4.1 does not yet completely
encompass the alternative for Fredholm integral equations of Theorem 3.1
since a link with an adjoint equation and the characterization of the image
space in the second case of the alternative are missing. Due to space limitations we have to omit closing this gap through the results of Schauder from
1929 and more recent developments in the 1960s (see [4]).
The following theorem is just a consequence of the fact that the identity
operator on a normed space is compact if and only if X has finite dimension.
8
It explains why the difference between the two integral equations (1) and (2)
is more than just formal.
Theorem 4.2 Let X and Y be normed spaces and let A : X Y be a
compact linear operator. Then A cannot have a bounded inverse if X is
infinite dimensional.
Numerical solution
The conceptually most straight forward idea for the numerical solution of
integral equations of the second kind goes back to Nystrom in 1930 and
consists in replacing the integral in (2) by numerical integration. Using a
quadrature formula
Z 1
n
X
ak g(xk )
g(x) dx
0
k=1
n
X
ak K(x, xk )(xk )
(14)
k=1
for x [0, 1], i.e., we apply the quadrature formula for g = K(x, ). Then
the solution to the integral equation of the second kind + A = f is
approximated by the solution of n + An n = f, which reduces to solving a
finite-dimensional linear system as follows. If n is a solution of
n (x) +
n
X
(15)
k=1
for x [0, 1] then clearly the values n,j := n (xj ) at the quadrature points
satisfy the linear system
n,j +
n
X
k=1
(16)
n
X
ak K(x, xk )n,k
(17)
k=1
for x [0, 1] can be seen to solve equation (15). Under appropriate assumptions on the kernel K and the right hand side f for a convergent sequence
of quadrature rules it can be shown that the corresponding sequence (n ) of
approximate solutions converges uniformly to the solution of the integral
equation if n . Furthermore it can be established that the error estimates for the quadrature rules carry over to error estimates for the Nystrom
approximations.
We conclude this short discussion on the numerical solution of integral
equations by pointing out that in addition to the Nystrom method a wealth of
other methods is available such as collocation and Galerkin methods (see [4]).
Ill-posed problems
u(0, t) = u(1, t) = 0,
and the initial condition
0 x 1,
u(x, 0) = (x),
2 2 t
an en
sin nx
(18)
n=1
an = 2
(y) sin ny dy
(19)
of the given initial values. This initial value problem is well-posed: The
final temperature f := u( , T ) clearly depends continuously on the initial
temperature because of the exponentially decreasing factors in the series
f (x) =
2 2 T
an en
sin nx.
(20)
n=1
2 2 T
bn en
sin nx
(21)
n=1
X
2 2
K(x, y) = 2
en T sin nx sin ny.
n=1
11
In general, integral equations of the first kind with continuous kernels provide
typical examples for ill-posed problems as consequence of Theorem 4.2.
Of course, the ill-posed nature of an equation has consequences for its numerical solution. The fact that an operator does not have a bounded inverse
means that the condition numbers of its finite-dimensional approximations
grow with the quality of the approximation. Hence, a careless discretization
of ill-posed problems leads to a numerical behavior that at first glance seems
to be paradoxical. Namely, increasing the degree of discretization, i.e., increasing the accuracy of the approximation for the operator, will cause the
approximate solution to the equation to become less and less reliable.
Regularization
(22)
This decomposition shows that the error consists of two parts: the first term
reflects the influence of the incorrect data and the second term is due to
the approximation error between R and A1 . Assuming that X is infinite
dimensional, R cannot be uniformly bounded since otherwise A would have
a bounded inverse. Consequently, the first term will be increasing as 0
whereas the second term will be decreasing as 0 according to (22).
Every regularization scheme requires a strategy for choosing the parameter in dependence on the error level and the data f in order to achieve
an acceptable total error for the regularized solution. On one hand, the accuracy of the approximation asks for a small error kR A k, i.e., for a small
parameter . On the other hand, the stability requires a small kR k, i.e., a
large parameter . A popular strategy is given by the discrepancy principle.
Its motivation is based on the consideration that, in general, for erroneous
data the residual kA f k should not be smaller than the accuracy of
the measurements of f , i.e., the regularization parameter should be chosen
such that kAR f f k .
We now assume that X and Y are Hilbert spaces and denote their inner products by (, ) with the space L2 [0, 1] of Lebesgue square integrable
complex valued functions on [0, 1] as a typical example. Each bounded linear operator A : X Y has a unique adjoint operator A : Y X with
the property (A, g) = (, A g) for all X and g Y . If A is compact then A is also compact. The adjoint of the compact integral operator
A : L2 [0, 1] L2 [0, 1] defined by (13) is given by the integral operator with
the kernel K(y, x) where the bar indicates the complex conjugate.
Extending the singular value decomposition for matrices from linear algebra, for each compact linear operator A : X Y there exists a singular
system consisting of a monotonically decreasing null sequence (n ) of positive numbers and two orthonormal sequences (n ) in X and (gn ) in Y such
that
An = n gn , A gn = n n , n IN.
For each X we have the singular value decomposition
=
(, n )n + P
n=1
13
space of A and
A =
n (, n )gn .
n=1
From the singular value decomposition it can be readily deduced that the
equation of the first kind A = f is solvable if and only if f is orthogonal to
the null space of A and satisfies
X
1
|(f, gn )|2 < .
2
n
n=1
(23)
X
1
(f, gn )n .
n
n=1
(24)
The solution (24) clearly demonstrates the ill-posed nature of the equation
A = f . If we perturb the right hand side f by f = f + gn we obtain the
solution = + 1
n n . Hence, the ratio k k/kf f k = 1/n can
be made arbitrarily large due to the fact that the singular values tend to
zero. This observation suggest to regularize by damping the influence of the
factor 1/n in the solution formula (24). In the Tikhonov regularization this
is achieved by choosing
R f :=
X
n=1
n
(f, gn )n .
+ 2n
(25)
Computing R f does not require the singular system since for injective A it
can be shown that R = (I + A A)1 A . Hence := R f can be obtained
as the unique solution of the well-posed equation of the second kind
+ A A = A f.
Computerized tomography
i.e., the scanning process in principle provides the line integrals over all lines
traversing the scanned cross-section. The transform that maps a function
in IR2 onto its line integrals is called the Radon transform and the inverse
problem of computerized tomography requires its inversion. Already in 1917
Radon gave an explicit inversion formula which however is not immediately
applicable for practical computations.
For the formal description of the Radon transform it is convenient to
parameterize the line L by its unit normal vector and its signed distance s
from the origin in the form L = {s + t : t IR}, where is obtained by
rotating counter-clockwise by 90 degrees. Now the two-dimensional Radon
transform R is defined by
Z
f (s + t ) dt, S 1 , s IR,
(Rf )(, s) :=
and maps L1 (IR2 ) into L1 (S 1 IR). Given the measured line integrals g, the
inverse problem of computerized tomography consists of solving
Rf = g
(26)
for f . Although it is not of the conventional form (1) or (2), clearly (26) can
be viewed as an integral equation. Its solution can obtained using Radons
inversion formula
1
R H
Rf
(27)
f=
4
s
with the Hilbert transform
Z
1
g(t)
(Hg)(s) :=
dt, s IR,
s t
applied with respect to the second variable in Rf . The operator R is the
adjoint of R with respect to the L2 inner products on IR2 und S 1 IR. By
interchanging integrations, it can be seen to be given by
Z
(R g)(x) =
g(, x ) d, x IR2 ,
S1
15
i.e., it can be considered as integrating over all lines through x and therefore
is called the back projection operator. Because of the occurrence of the
Hilbert transform in (27), inverting the Radon transform is not local, i.e.,
the line integrals through a neighborhood of the point x do not suffice for the
reconstruction of f (x). Due to the derivative appearing in (27) the inverse
problem to reconstruct the function f from its line integrals is ill-posed. For
a detailed study of computerized tomography we refer to [5].
In practice the integrals can be measured only for a finite number of lines
and correspondingly a discrete version of the Radon transform has to be
inverted. The most widely used inversion algorithm is the filtered back projection algorithm which may be considered as an implementation of Radons
inversion formula with the middle part H/s replaced by a convolution,
i.e., a filter in the terminology of image processing. However, so-called algebraic reconstruction techniques are also used where the function f is decomposed into pixels, i.e., approximated by piecewise constants on a grid of
little squares. The resulting sparse linear system for the pixel values then is
solved iteratively, for example, by Kaczmarzs method.
For the case of a radially symmetric function f , that is, f (x) = f0 (|x|)
clearly Rf does not depend on , that is, (Rf )(, s) = g0 (s) where
Z
f0 ( s2 + t2 )dt, s 0.
g0 (s) = 2
0
Substituting t =
s 0,
which is an Abel type integral equation again. Its solution can be shown to
be given by
Z
1
1
g00 (s)
ds, r 0.
f0 (r) =
r
s2 r 2
This approach can be extended to a full inversion formula by expanding both
f and g = Rf in a Fourier series with respect to the polar angle. Then the
Fourier coefficients of f and g are related by Abel type integral equations
involving Chebyshev polynomials.
X-ray tomography was first suggested by the physicist Allan Cormack in
1963 and due to the efforts of the electrical engineer Godfrey Hounsfield it
was introduced into medical practice in the 1970s. For their contributions
16
to X-ray tomography Cormack and Hounsfield were awarded with the 1979
Nobel prize for medicine.
Inverse scattering
Scattering theory is concerned with the effects that obstacles and inhomogeneities have on the propagation of waves and in particular time-harmonic
waves. Inverse scattering provides the mathematical tools for such fields as
radar, sonar, medical imaging and nondestructive testing.
For time-harmonic waves the time dependence is factored out in the form
U (x, t) = Re {u(x) eit } with a positive frequency , i.e, the complex valued
space dependent part u represents the real valued amplitude and phase of the
wave and satisfies the Helmholtz equation u + k 2 u = 0 with a positive wave
number k. For a vector d IR3 with |d| = 1, the function eik xd satisfies the
Helmholtz equation for all x IR3 . It is called a plane wave, since ei(k xdt)
is constant on the planes k x d t = const. Assume that an incident field
is given by ui (x) = eik xd . Then the simplest obstacle scattering problem is
to find the scattered field us as a solution to the Helmholtz equation in the
exterior of a bounded scatterer D IR3 such that the total field u = ui + us
satisfies the Dirichlet boundary condition u = 0 on D modeling a soundsoft obstacle or a perfect conductor. In addition, to ensure that the scattered
wave is outgoing it has to satisfy the Sommerfeld radiation condition
s
u
s
lim r
iku = 0,
(28)
r
r
where r = |x| and the limit holds uniformly in all directions x/|x|. This
ensures uniqueness of the solution to this exterior Dirichlet problem for
the Helmholtz equation. Existence of the solution has been established via
boundary integral equations in the spirit of Section 3 in the 1950s by Vekua,
Weyl and M
uller.
The radiation condition (28) can be shown to be equivalent to the asymptotic behavior
1
eik|x|
s
u (
x) + O
, |x| ,
u (x) =
|x|
|x|
uniformly for all directions x = x/|x|, where the function u defined on
the unit sphere S 2 is known as the far field pattern of the scattered wave.
17
18
L2 (S 2 ) with kernel u (
x, d) given by
Z
u (
x, d)g(d) ds(d),
(F g) :=
x S 2 .
S2
1 ik xz
e
4
eik|xz|
,
4|x z|
x 6= z,
of the Helmholtz equation with source point in z IR3 the linear sampling
method is based on the ill-posed equation
F gz = ( , z)
(29)
(30)
An essential tool in the linear sampling method are Herglotz wave functions
with kernel g defined as the superposition of plane waves given by
Z
vg (x) :=
eik xd g(d) ds(d), x IR3 .
S2
It can be shown that if z D then the value of the Herglotz wave function vgz, (z) with the kernel gz, given by the solution of (29) obtained
by Tikhonov regularization with parameter remains bounded as 0
whereas it is unbounded if z 6 D. This criteria can be evaluated numerically on a sufficiently fine grid of points to visualize the scatterer D. The
main feature of the factorization method is the fact that the equation (30)
is solvable in L2 (S 2 ) if and only if z D. This can be utilized to visualize
the scatterer on a grid with the aid of the solubility condition (23) in terms
of a singular system of F . For details we refer to [3].
The references [1, 2, 4] are evidence of my continuing love for integral
equations.
19
References
[1] Colton, D. and Kress, R. 1998 Inverse Acoustic and Electromagnetic Scattering Theory. 2nd. ed. Springer-Verlag, New York.
[2] Ivanyshyn, O., Kress, R. and Serranho, P. 2010 Huygens principle and
iterative methods in inverse obstacle scattering. Advances in Computational Mathematics 33, 413429.
[3] Kirsch, A. and Grinberg, N. 2008 The Factorization Method for Inverse
Problems. Oxford University Press, Oxford.
[4] Kress, R. 1997 Linear Integral Equations. 2nd ed. Springer-Verlag, New
York.
[5] Natterer, N. 2001 The Mathematics of Computerized Tomography. SIAM,
Philadelphia.
20