Sei sulla pagina 1di 8

Illumination Invariant Face Recognition: A Survey

Xuan Zou, Josef Kittler and Kieron Messer

Abstract- The illumination variation problem is one of illumination invariant face recognition can be found in [75]
the well-known problems in face recognition in uncontrolled [33][14][32]. However, these reviews focused either only on
environment. In this paper an extensive and up-to-date survey passive approaches or only on active approaches. The review
of the existing techniques to address this problem is presented.
This survey covers the passive techniques that attempt to solve presented in this paper is more extensive than previous
the illumination problem by studying the visible light images in reviews and covers more recent techniques in both groups,
which face appearance has been altered by varying illumination, such as Generalized Photometric Stereo[76], Total-Variance
as well as the active techniques that aim to obtain images of Quotient Image[13], and Active Differential Imaging[78].
face modalities invariant to environmental illumination. These techniques have been shown to be successful in
I. INTRODUCTION dealing with illumination variations.
For many applications, the performance of face recogni- II. PASSIVE APPROACHES
tion systems in controlled environments has now reached a Passive approaches can be divided into four groups: illu-
satisfactory level; however, there are still many challenges mination variation modelling, illumination invariant features,
posed by uncontrolled environments. Some of these chal- photometric normalisation, and 3D morphable model. The
lenges are posed by the problems caused by variations in performance of typical passive approaches is presented in
illumination, face pose, expression, and etc. The effect of Table VI.
variation in the illumination conditions in particular, which
causes dramatic changes in the face appearance, is one of A. Illumination Variation Modelling
those challenging problems [75] that a practical face recog- The modelling of face images under varying illumination
nition system needs to face. To be more specific, the varying can be based on a statistical model or physical model. For
direction and energy distribution of the ambient illumination, statistical modelling, no assumption concerning the surface
together with the 3D structure of the human face, can lead property is needed. Statistical analysis techniques, such as
to major differences in the shading and shadows on the PCA(Eigenface) and LDA(Fisherface), are applied to the
face. Such variations in the face appearance can be much training set which contains faces under different illumina-
larger than the variation caused by personal identity [37]. tions to achieve a subspace which covers the variation of
The variations of both global face appearance and local possible illumination. In physical modelling, the model of
facial features also cause problems for automatic face detec- the process of image formation is based on the assumption
tion/localisation, which is the prerequisite for the subsequent of certain object surface reflectance properties, such as
face recognition stage. Therefore the situation is even worse Lambertian reflectance.
for a fully automatic face recognition system. Moreover, in a 1) Linear Subspaces: Hallinan[22] showed that five
practical application environment, the illumination variation eigenfaces were sufficient to represent the face images under
is always coupled with other problems such as pose variation a wide range of lighting condition. Shashua proposed Photo-
and expression variation, which increase the complexity of metric Alignment approach to find the algebraic connection
the automatic face recognition problem. between all images of an object taken under varying illumi-
A number of illumination invariant face recognition ap- nation conditions [52]. An order k Linear Reflectance Model
proaches have been proposed in the past years. Existing for any surface point p is defined as the scalar product x a,
approaches addressing the illumination variation problem fall where x is a vector in the k-dimensional Euclidean space of
into two main categories. We call the approaches in the first invariant surface properties( such as surface normal, albedo,
category "passive" approaches, since they attempt to over- and so forth), and a is an arbitrary vector. The image intensity
come this problem by studying the visible spectrum images I(p) of an object with an order k reflection model can be
in which face appearance has been altered by illumination represented by a linear combination of a set of k images of
variations. The other category contains "active" approaches, the object. For lambertian surface under distant point sources
in which the illumination variation problem is overcome by and in the absence of shadows, all the images lie in a 3D
employing active imaging techniques to obtain face images linear subspace of the high dimensional image space, which
captured in consistent illumination condition, or images of means that they can be represented by a set of 3 images,
illumination invariant modalities. Existing reviews related to each from a linearly independent source. Given three images
X.Zou, J.Kittler and K. Messer are with Centre for Vision, Speech and
of this surface under three known and linearly independent
Signal Processing,University of Surrey, United Kingdom light sources, the surface normal and the albedo can be
{x.zou, j.kittler, k.messer}@surrey.ac.uk recovered. This is known as Photometric Stereo. Shashua

978-1-4244-1597-7/07/$25.00 ©2007 IEEE

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
claimed the attached shadows, which are caused by points basis images are built based on a collection of 2D basis
where the angle between surface normal and the direction images. For a given training face image, the basis images for
of light source is obtuse(np s < 0, therefore I(p) = 0), this face can be estimated based on Maximum A Posterior
do not have a significant adverse effect on the photometric estimation. In the second method a 3D morphable model
alignment scheme. However, the cast shadows caused by and the harmonic representation are combined to perform
occlusion cannot be modeled using the above framework. face recognition with both illumination and pose variation.
Belhumeur et al. [5] presented the so-called 3D linear 4) Nine point lights: Lee et al. [31] showed that there
subspace method for illumination invariant face recognition, exists a configuration of nine point source directions such
which is a variant of the photometric alignment method. In that a subspace resulting from nine images of each individual
this linear subspace method, three or more images of the under these nine lighting sources is effective at recognition
same face taken under different lighting are used to construct under a wide range of illumination conditions. The advantage
a 3D basis for the linear subspace. The recognition proceeds of this method is that there is no need to obtain a 3D model
by comparing the distance between the test image and each of surface as in the spherical harmonics approach[46], or to
linear subspace of the faces belonging to each identity. The collect a large number of training images as in the statistical
Fisher Linear Discriminant(also called FisherFace) method modelling approaches.
is also proposed in [5] in order to maximise the ratio of the 5) Generalized Photometric Stereo: Recently, Zhou et al.
between-class scatter and the within-class scatter of the face [76] analyzed images of the face class with both the Lam-
image set to achieve better recognition performance. bertian reflectance model and the linear subspace approach.
Batur and Hayes[3] proposed a segmented linear subspace The human face is claimed to be an example of a so-called
model to generalize the 3D linear subspace model so that Linear Lambertian Object, which is not only an object with
it is robust to shadows. Each image in the training set is Lambertian surface, but also a linear combination of basis
segmented into regions that have similar surface normals by objects with lambertian surfaces. The albedo and surface
k-Mean clustering, then for each region a linear subspace is normal vectors of each basis object for the face class form a
estimated. Each estimation only relies on a specific region, matrix called class-specific albedo/shape matrix, which can
so it is not influenced by the regions in shadow. be recovered by a Generalized Photometric Stereo process
2) Illumination Cone: Belhumeur and Kriegman [6] from the bootstrap set. The model is trained using Vetter's
proved that all images of a convex object with Lambertian 3D face database [7]. Excellent performance was reported.
surface from the same viewpoint but illuminated by an The work was further extended for multiple light sources.
arbitrary number of distant point sources form a convex
Illumination Cone. The dimension of this illumination cone B. Illumination Invariant Features
is the same as the number of distinct surface normals. This Adini et al. [1] presented an empirical study that evaluates
illumination cone can be constructed from as few as three the sensitivity of several illumination insensitive image repre-
images of the surface, each under illumination from an sentations to changes in illumination. These representations
unknown point source. The illumination cone is a convex include edge map, image intensity derivatives, and image
combination of extreme rays given by xij =Tmax(Bsij, 0), convolved with a 2D Gabor-like filter. All of the above
where sij = bi x bj, and bi,bj are two different rows of representations were also followed by a log function to
a matrix B where each row is the product of albedo with generate additional representations. However, the recogni-
surface normal vector. Kriegman and Belhumeur showed in tion experiment on a face database with lighting variation
[30] that for any finite set of point sources illuminating indicated that none of these representations is sufficient by
an object viewed under either orthographic or perspective itself to overcome the image variation due to the change of
projection, there is an equivalence class of object shapes illumination direction.
having the same set of shadows. These observations are 1) Features Derived from Image Derivatives: Line edge
exploited by Georghiades et al. [19] for face recognition map [18] is proposed for face recognition by Gao and Leung.
under variable lighting. The edge pixels are grouped into line segments, and a revised
3) Spherical Harmonics: Spherical Harmonics method is Hausdorff Distance is designed to measure the similarity be-
proposed by Basri and Jacobs [45], and contemporarily by tween two line segments. Chen et al. [12] showed that for any
Ramamoorthi and Hanrahan [44]. Assuming arbitrary light image, there are no discriminative functions that are invariant
sources (point sources or diffuse sources) distant from an to illumination, even for objects with Lambertian surface.
object of Lambertian reflectance property, Basri and Jacobs However, they showed that the probability distribution of
[46] show that ignoring cast shadow the intensity of object the image gradient is a function of the surface geometry and
surface can be approximated by a 9-dimensional linear reflectance, which are the intrinsic properties of the face.
subspace based on a Spherical Harmonic representation. The direction of image gradient is revealed to be insensitive
Zhang and Samaras [71] proposed two methods for face to illumination change. The recognition performance using
recognition under arbitrary unknown lighting by using the gradient direction is close to the illumination cone approach.
spherical harmonics representation, which requires only one Relative Image Gradient feature is applied by Wei and
training image per subject and no 3D shape information. Lai [65] and Yang et al. [69] for robust face recognition
In the first method[70] the statistical model of harmonic under lighting variation. The relative image gradient G(x, y)

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
is defined as G(x, y) = m vw(x,y) where called Generic Intrinsic Illumination Subspace, which can
I(x, y) is the image intensity, V is the gradient operator, be obtained from a bootstrap set. For a given image the
W(x, y) is a local window centered at (x, y), and c is a illumination image can be estimated by L = Bi, where
constant value to avoid dividing by zero. I = argminrBI L* Here B is the basis matrix of the
Zhao and Chellappa [74] presented a method based on intrinsic illumination subspace, and L* is an initial estimation
Symmetric Shape from Shading for illumination insensitive of illumination image based on smoothed input image. Fi-
face recognition. The symmetry of every face and the shape nally p(x, y) can be obtained by p(x, y) = I(x, y)/L(x, y).
similarity among all faces are utilized. A prototype image The method was evaluated on CMU-PIE and Yale B face
with normalized illumination can be obtained from a single databases and showed significantly better results than the
training image under unknown illumination. Their experi- Quotient Image method. It is also shown that enforcing non-
ments showed that using the prototype image significantly negative light constraint will further improve the results.
improved the face recognition based on PCA and LDA. 3) Retinex approach: In Retinex approaches the lumi-
Sim and Kanade [58] developed a statistical shape from nance is estimated by the smoothed image. The image can
shading model to recover face shape from a single image then be divided by the luminance to obtain the reflectance,
and to synthesize the same face under new illumination. which is an invariant feature to illumination. A single Gaus-
The surface radiance i (x) for location x is modeled as sian function is applied to smooth the image in the single
i(x) = n(x)T x s + e, where n(x) is the surface normal scale retinex approach [26], and the sum of several Gaussian
with albedo, s is the light source vector, e is an error term functions with different scales is applied in the multi-scale
which models shadows and specular reflections. A bootstrap retinex approach[27]. Logarithm transform is employed to
set of faces with labeled varying illuminations is needed to compress the dynamic range in [26] and [27].
train the statistical model for n(x) and e. The illumination Wang et al. [64] defined Self-Quotient Image, which is
for an input image can be estimated using kernel regression essentially a multi-scale retinex approach, however instead of
based on the bootstrap set, then n(x) can be obtained by using isotropic smoothing as in [27], anisotropic smoothing
Maximum A Posterior estimation and the input face under a functions with different scales are applied. Each anisotropic
new illumination can be synthesized. smoothing function is a Gaussian weighted by a threshold-
2) Quotient Image: Shashua and Riklin-Raviv [53] treat ing function. Zhang et al. [72] proposed a Morphological
face as an Ideal Class of Object, i.e. the objects that have Quotient Image(MQI) method in which mathematical mor-
the same shape but differ in the surface albedo. The Quotient phology operation is employed to smooth the original image
Image Qy(u, v) of object y against object a is defined to obtain a luminance estimate.
P (un ) where Py (U, V), Pa (U, v) are albedo Gross and Brajovic [21] solve luminance L for the retinex
by Qy (u, v)
approach by minimizing an anisotropic function over the im-
of the two objects. The image Qy depends only on the
age region Q: J(L) =ffA p(x, y) (L I)2dxdy + AffAQ (L2 +
relative surface texture information, and is independent of L2)dxdy, where p(x, y) is space varying permeability weight
illumination. A bootstrap set containing N faces under three which controls the anisotropic nature of the smoothing.
unknown independent illumination directions is employed.
Lx and Ly are the spacial derivatives of L, and I is the
Qy of a probe image Y(u, v) can be calculated as Qy(u, v) intensity image. The isotropic version of function J(L) can
Aj(u,v) where Aj (u, v) is the average of images under be obtained by discarding p(x, y).
illumination j in the bootstrap set, and xj can be determined In the Total-Variation based Quotient Image(TVQI) ap-
from all the images in bootstrap set and Y(u, v). Then the proach [13], the luminance u(x) is obtained by minimizing
recognition is performed based on the quotient image. ffQ Vu(x) + A I(x) -u(x) dx over all points x in image
Based on the assumption that faces are an Ideal Class I(x).
of objects, Shan et al. [51] proposed Quotient Illumination 4) Transformation domain features: Recently methods
Relighting. When the illumination in the probe image and based on the frequency domain representation have received
the target illumination condition are both known and exist attention. Savvides et al. [49] performed PCA in the phase
in the bootstrap set, the rendering can be performed by a domain and achieved impressive results on the CMU-PIE
transformation learnt from the bootstrap set. database [57]. This so-called Eigenphase approach improved
Chen and Chen [11] proposed a Generic Intrinsic Illumi- the performance dramatically compared to Eigenface, Fish-
nation Subspace approach. Given the Ideal Class assump- erface and 3D linear subspace approach. Meanwhile, they
tion, all objects of the same ideal class share the same further showed that even with partial face images the per-
generic intrinsic illumination subspace. Considering attached formance of the Eigenphase approach remains excellent
shadows, the appearance image of object i in this class and the advantages over other approaches are even more
under a combination of k illumination sources { ,}k 1 is significant. Heo et al. [23] showed that applying Support
represented by Ii(x,y) =P(X, y) Z=1 max(n(x, y)lj, 0), Vector Machines directly on phase can lead to even better
where pi(x, y) is the albedo, and n(x, y) is the surface performance than the Eigenphase approach mentioned above.
normal vector of all objects in the class. The illumination In [67] a quaternion correlation method in a wavelet
image is defined as L(x, y) ,k max(n(x, y)lj, 0). The
=
domain is proposed and good performance is achieved on
illumination images of a specific Ideal Class form a subspace the CMU-PIE database with only one training sample per

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
subject. The subband images after discrete wavelet decom- the logarithm of the image I(x, y) to reduce the luminance
position are encoded into a 2-D quaternion image. Quater- part, which is the low frequency component of the image,
nion Fourier Transform is then performed to transfer the and amplify the reflectance part, which corresponds to the
quaternion image to quaternion frequency domain, where high frequency component.
a quaternion correlation filter is applied. Qing et al. [43] Du and Ward [16] performed illumination normalization
showed that the Gabor phase is tolerant to illumination in the wavelet domain. Histogram equalisation is applied to
change and has more discriminative information than phase low-low subband image of the wavelet decomposition, and
in the Fourier spectrum. simple amplification is performed for each element in the
Savvides et al. proposed a series of work based on advance other 3 subband images to accentuate high frequency com-
correlation filters[48][47]. A pre-whitening spectrum stage is ponents. Uneven illumination is removed in the reconstructed
usually adopted to emphasize higher frequency components image obtained by employing inverse wavelet transform on
followed by phase matching. the modified 4 subband images.
Llano et al. [34] examined the sensitivity of several fre- Xie and Lam [68] proposed an illumination normalization
quency domain representations of face image to illumination method which is called Local Normalization. They split the
change. Those representations are the magnitude, phase, real face region into a set of triangular facets, the area of which
part and imaginary part of the Fourier spectrum of original is small enough to be considered as planar patch. The main
face image, and those of gradient image. The gradient image idea of this approach is to normalize the intensity values
is defined as an image where each pixel has a complex value within each facet to be of zero mean and unit variance.
with the horizontal gradient of the original image as the Short et al. [55] compared five photometric normaliza-
real part, and the vertical gradient as imaginary part. The tion methods, namely illumination insensitive eigenspaces,
experimental results on the normal illumination set and the multiscale Retinex method, homomorphic filtering, a method
darken set of the XM2VTS face database showed that the using isotropic smoothing to estimate luminance, and one
real part of the Fourier spectrum of the gradient image is less using anisotropic smoothing [21]. Each method is tested
sensitive to illumination change than other representations. with/without histogram equalisation performed in advance.
5) Local Binary Pattern: Local Binary Pattern (LBP) is Interestingly it was found that histogram equalisation helped
a local feature which characterizes the intensity relationship in every case. It is shown that using anisotropic smoothing
between a pixel and its neighbors. LBP is unaffected by any method as photometric normalisation led to the most consis-
monotonic grayscale transformation in that the pixel intensity tent verification performance for experiments across the Yale
order is not changed after such a transformation. Further- B, BANCA[2] and XM2VTS [35] databases.
more, for a region with a number of pixels, a histogram of Chen et al. [14] employed DCT to compensate for il-
the LBP patterns associated with respective pixels within this lumination variation in the logarithm domain. The uneven
region tends to be a good feature for face recognition. LBP illumination is removed in the image reconstructed by inverse
has been used in [32] [24] as an illumination invariant feature. DCT after a number of DCT coefficients corresponding to
C. Photometric Normalization low frequency are discarded.
Histogram Equalisation [20] is the most commonly used
approach. By performing histogram equalisation, the his- D. 3D Morphable Model
togram of the pixel intensities in the resulting image is Blanz and Vetter [7] proposed face recognition based on
flat. It is interesting that even for images with controlled fitting a 3D morphable model. The 3D morphable model
illumination (such as face images in the XM2VTS database), describes the shape and texture of face separately based
applying histogram equalisation still offers performance gain on the PCA analysis of the shape and texture obtained
in face recognition [54]. from a database of 3D scans. To fit a face image under
Shan et al. [51] proposed Gamma Intensity Correction for unknown pose and illumination to the model, an optimisation
illumination normalisation. The corrected image G(x, y) can process is needed to optimize shape coefficients, texture
be obtained by performing an intensity mapping: G(x, y) coefficients along with 22 rendering parameters to minimise
cI(x, y) where c is a gray stretch parameter, and -y is the the difference of the input image and the rendered image
Gamma coefficient. based on those coefficients. The rendering parameters include
In Homomorphic filtering approach [20] the logarithm of pose angles, 3D translation, ambient light intensities, directed
the equation of the reflectance model is taken to separate light intensities and angles, and other parameters of the
the reflectance and luminance. The reflectance model often camera and color channels. The illumination model of Phong
adopted is described by I(x, y) = R(x, y) * L(x, y), where is adopted in the rendering process to describe the diffuse
I(x, y) is the intensity of the image, R(x, y) is the reflectance and specular reflection of the surface. After fitting both
function, which is the intrinsic property of the face, and the gallery images and the probe images to the model, the
L(x, y) is the luminance function. Based on the assumption recognition can be performed based on the model coefficients
that the illumination varies slowly across different locations for shape and texture. Good recognition performance across
of the image and the local reflectance changes quickly across pose and illumination is achieved in experiments on CMU-
different locations, a high-pass filtering can be performed on PIE and FERET face database.

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
III. ACTIVE APPROACHES 1) Thermal Infrared: A survey on visual and infrared face
In active approaches additional devices (optical filters, recognition is presented in [29]. Wilder at al. [66] showed
active illumination sources or specific sensors) usually need that with minor illumination changes and for subjects without
to be involved to actively obtain different modalities of face eyeglasses, applying thermal image for face recognition does
not lead to significant difference compared to visible images.
images that are insensitive to or independent of illumination
change. Those modalities include 3D face information [8] However, for scenarios with huge illumination changes and
and face images in those spectra other than visible spectra, facial expressions, superior performance was achieved based
on radiometrically calibrated thermal face images than that
such as thermal infrared image [29] and near-infrared hyper-
spectral image [40]. based on visible image[62][59]. The experiments in [15]
show that the face recognition based on thermal images
A. 3D information degrades more significantly than visible images when there
is a substantial passage of time between the acquisition of
3D information is one of the intrinsic properties of a gallery images and probe images. This result was proved to
face, which is invariant to illumination change. The surface be reproducible by [61], however it is shown that with a
normal information is also used in some passive approaches more sophisticated recognition algorithm the difference of
described in the previous section, however, they are recov- recognition performance across time based on thermal face
ered from the intensity images captured by the visible light and visible face is small.
camera. In this section we talk about the 3D information Despite the independence of visible light, the thermal
acquired by active sensing devices like 3D laser scanners or imagery has its own disadvantages. The temperature of the
stereo vision systems. environment, physical conditions and psychological condi-
3D information can be represented in different ways. tions will affect the heat pattern of the face[4]. Meanwhile,
The most commonly used representations are range image, the infrared is opaque to eyeglasses. All the above motivate
profile, surface curvature, Extended Gaussian Image(EGI), the fusion of thermal infrared image with visible images for
Point Signature, and etc. Surveys on 3D face recognition face recognition. Various fusion schemes have been proposed
approaches can be found in [8][9] and [50]. The 3D modality [29] [15] [4] [60] and shown to lead to better performance than
can be fused with 2D modality, i.e. texture, to achieve better the recognition based on either modality alone. The thermal
performance [9] [10]. Nevertheless, it should be noted that the face recognition experiments are usually conducted on the
2D face images which are combined with 3D face info as face database from the University of Notre Dame [15] or the
reported in [9][10] are captured in a controlled environment. Equinox face database [17]. The former contains the visible
It is still not clear how much the fusion will help in the spectrum images and LWIR images of 240 subjects without
case of uncontrolled environment due to the impact of glasses, but with different lighting and facial expressions.
uncontrolled illumination on the 2D face intensity images. The latter was collected by Equinox Corp. and contains the
Kittler et al. [28] reviewed the full spectrum of 3D face visible images and LWIR images of a total of 115 subjects.
processing, from sensing to recognition. The review covers In [60], thermal face recognition is performed in an
the currently available 3D face sensing technologies, various operational scenario, where both indoor and outdoor face
3D face representation models and the different ways to use data of 385 subjects is captured. When the system is trained
3D model for face recognition. In addition to the discussion on indoor sessions and tested on outdoor sessions, the
on separate 2D and 3D based recognition and the fusion of performance degrades no matter whether one is using thermal
different modalities, the approach involving 3D assisted 2D imagery or visible imagery. However, the thermal imagery
recognition is also addressed. substantially outperformed visible imagery. With the fusion
of both modalities, the outdoor performance can be close to
B. Infrared indoor face recognition.
Visible light spectrum ranges from 0.4,um-0.7,um in the 2) Active Near-IR Illumination: The Near-IR band falls
electromagnetic spectrum. The infrared spectrum ranges into the reflective portion of the infrared spectrum, between
from 0.7,um -10mm. It can be divided into 5 bands, namely: the visible light band and the thermal infrared band. It
Near-Infrared(Near-IR) (0.7-0.9,um), the Short-Wave In- has advantages over both visible light and thermal infrared.
frared (SWIR) (0.9-2.4,um), the Mid-Wave Infrared(MWIR) Firstly, since it can be reflected by objects, it can serve as
(3.0-8.0,um), the Long-Wave Infrared(LWIR) (8.0-14.0,um), an active illumination source, in contrast to thermal infrared.
and Far-Infrared(FIR) (14.Oum-10mm). Near-IR and SWIR Secondly, it is invisible, making active Near-IR illumination
belong to reflected infrared (0.7-2.4,um), while MWIR and unobtrusive. Thirdly, unlike thermal infrared, Near-IR can
LWIR belong to thermal infrared (2.4,um-14mm). Similar easily penetrate glasses.
to the visible spectrum, the reflected infrared contains the Pan et al. [40] performed face recognition in hyperspectral
information about the reflected energy from the object sur- images. A CCD camera with a liquid crystal tunable filter
face, which is related to the illumination power and the was used to collect images with 31 bands over near-infrared
surface reflectance property. Thermal Infrared directly relates range. It was shown the hyperspectral signatures of the skin
to the thermal radiation from object, which depends on the from different persons are significantly different, while those
temperature of the object and emissivity of the material[29]. belonging to the same person are stable. Above 91 % rank-

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
one correct identification rate is obtained in the recognition In a practical scenario there can be possible violation of the
experiments on frontal hyperspectral face images. assumptions that active differential imaging relies on. Firstly
Most recently, Li et al. [32] proposed a face recognition the subject may move during the capture. The solution to
system based on active Near-IR lighting provided by Near- this problem would be employing a fast capture system or a
IR Light-Emitting Diodes(LEDs). The Near-IR face image motion compensation technique as in [80]. Secondly, to avoid
captured by this device is subject to a monotonic transform saturation under strong sunshine, the dynamic intensity range
in the gray tone, then LBP feature is extracted to compen- of the image illuminated by active illumination source has
sate for this monotonic transform to obtain an illumination to be sacrificed. A sensor with large dynamic intensity range
invariant face representation. Zhao and Grigat [73] performed (eg. 16 bit) can serve as a solution to this problem.
face recognition in Near-IR images based on Discrete Cosine
Transform(DCT) feature and SVM classifier. IV. DISCUSSION
Although infrared image is invariant to visible illumina- The performance of typical passive approaches is shown
tion, it is not independent of the environmental illumination. in Table VI. The first subset of Yale B or Harvard is used as
This is because environmental illumination contains energy training/gallery set for the tests on the rest subsets. There
in a wide range of spectrum, including infrared. The variation are some passive approaches which are not included in
of the infrared component in the environmental illumination this table, this is because either their performance was less
will impose variation in the captured image. significant, or the experiment was not done on a standard
One solution to maximize the ratio between the active database or using a standard protocol, therefore it is not
source and the environmental source is to apply synchronized fair to compare their performance with those listed in the
flashing imaging by Hizem et al. [25]. A powerful active table. It can be seen that some passive approaches achieved
illumination source is desirable. Illuminants such as LEDs excellent performance. However, it can not be concluded
can provide very powerful flash but only for very short time that the illumination problem is well solved. Although good
to avoid the internal thermal effects which might destroy the performance is usually reported for most techniques, each
LEDs. The idea in Hizem et al. [25] is to synchronise the technique has its own drawbacks. The illumination modeling
sensor exposure time with the powerful flash. The sensor is methods require training samples from controlled illumina-
exposed to the environmental illumination only for the same tion. The physical modeling of the image formation generally
short exposure time as the flash. Since the power of the flash requires the assumption that the surface of the object is
is usually much stronger than the environmental illumination, Lambertian, which is violated for real human faces. The
the contribution of the environmental illumination to the statistic modeling methods require training samples from as
captured image will be minimised. many as possible different illumination conditions to ensure
Nevertheless, the illumination variation problem can only a better performance. The performance of many photometric
be alleviated but not completely solved by the above men- normalisation methods severely depends on the choice of
tioned approach. For indoor environment, the infrared energy parameters[56].
in environmental illumination is low and will not cause much In addition to the individual drawbacks for each approach,
problem, while in outdoor environment, the infrared energy there are some common issues. Firstly, it should be noticed
can be very strong. that all the experiments are conducted on manually marked
3) Active Differential Imaging: Active Differential Imag- data. The sensitivity to localisation error still needs to be
ing can be applied to solve the illumination variation problem investigated for all approaches. Secondly, the databases to
even for the outdoor application. In active differential imag- evaluate techniques are of very small size. For instance, Yale
ing, an image is obtained by the pixel-to-pixel differentiation B contains images of only 10 subjects. The performance of
between two successive frames: one with an active illuminant illumination invariant approaches on large scale face dataset
on and one with the illuminant off. Assuming a still scene, a is still unknown. How much discriminative information has
linear response of the camera to the scene radiation and no been lost after the suppression of the illumination effect?
saturation, the difference image of these two frames contains As for active approaches, it should be investigated how
only the scene under the active illuminant, and is completely much the active sensing process will be influenced by the
independent of the environmental illumination. Due to its environmental illumination. For example, a number of 3D
invisibility, Near-IR illumination usually serves as the active acquisition techniques make use of intensity/infrared image
illumination for the active differential imaging system. Face pairs for 3D reconstruction. The variation in the environmen-
recognition experiments carried out on the face images cap- tal illumination might cause a problem for the accuracy of
tured by active Near-IR differential imaging system achieved 3D reconstruction in those systems.
very low error rates even in the scenario with dramatic Usually it is difficult to compare the performance be-
ambient illumination changes[78][79][38][25]. It is further tween passive approaches and active approaches because
shown in [77] that this active Near-IR differential imaging the datasets to test passive approaches are captured under
technique significantly outperformed some typical passive different illumination conditions with the databases captured
approaches for illumination problem. Specific sensors which by active approaches. There is only one exception: in [78],
can perform differential imaging have been proposed in faces under ambient illumination were collected simultane-
[36][63][39]. ously when collecting a database with the active differential

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
imaging system. Based on experiments on this face database [20] R. Gonzalez and R. Woods. Digital Image Processing. Prentice Hall,
the active differential imaging technique is shown to be second edition, 1992.
[21] R. Gross and V. Brajovic. An image pre-processing algorithm for
significantly superior to the conventional passive approaches. illumination invariant face recognition. In Proc. 4th Int'l Conf on
Moreover, it would be interesting to see whether the active AVBPA, 2003.
approaches and passive approaches can be combined to better [22] P. Hallinan. A low-dimensional representation of human faces for
arbitrary lighting conditions. In Proc. IEEE conf. CVPR, 1994.
solve the illumination problem in a practical scenario. [23] J. Heo, M.Savvides, and B.V.K. Vijaya Kumar. Illumination tolerant
Face Recognition Grand Challege(FRGC) experiment 4 face recognition using phase-only support vector machines in the
was designed for uncontrolled illumination problem. For frequency domain. In Proc. ICAPR. Springer-Verlag Lecture Notes
in Computer Science, 2005.
more information about the performance of those algorithms [24] G. Heusch, Y Rodriguez, and S. Marcel. Local binary patterns as an
submitted for evaluation please refer to[42][41]. image preprocessing for face authentication. In Proc. IEEE AFGR,
2006.
V. CONCLUSION [25] W. Hizem, E. Krichen, Y Ni, B. Dorizzi, and S. Garcia-Salicetti.
An extensive survey of passive and active approaches Specific sensors for face recognition. In Proc. IAPR Int'l Conf on
to address the illumination problem has been presented in Biometrics, 2006.
[26] D. J. Jobson, Z. Rahman, and G. A. Woodel. Properties and
this paper. The performance, advantages and drawbacks of performance of a center/surround retinex. IEEE Trans. on Image
passive and active approaches have been summarised. Future Processing: special issue on color processing, 1996.
directions for the research on illumination invariant face [27] D. J. Jobson, Z. Rahman, and G. A. Woodel. A multiscale retinex for
recognition have been suggested. bridging the gap between color images and the human observation of
scences. IEEE Trans. on Image Processing, 6(7), 1997.
VI. ACKNOWLEDGMENTS [28] J. Kittler, A. Hilton, M. Hamouz, and J. Illingworth. 3d assisted
face recognition: A survey of 3d imaging,modelling and recognition
The support from DTI Project No.TP/6/ICT/6/S/K1533J approaches. In Proc. IEEE conf CVPR, 2005.
is gratefully acknowledged. [29] S.G. Kong, J. Heo, B. Abidi, J. Paik, and M. Abidi. Recent advances
in visual and infrared face recognition - a review. CVIU, 2004.
REFERENCES [30] D. Kriegman and P. Belhumeur. What shadows reveal about object
[1] Y. Adini, Y. Moses, and S. Ullman. Face recognition: The problem structure. In Proc. the European Conf on Computer Vision, 1998.
of compensating for changes in illumination direction. IEEE Trans. [31] K. C. Lee, J. Ho, and D. Kriegman. Nine points of lights: Acquiring
PAMI, 19(7), 1997. subspaces for face recognition under variable lighting. In Proc. IEEE
[2] E. Bailly-Bailliere and etc. The banca database and evaluation conf. CVPR, 2001.
protocal. In Int'l Conf on AVBPA, 2003. [32] Stan Z. Li, R. Chu, S. Liao, and L. Zhang. Illumination invariant face
[3] A.U. Batur and M. H. III. Hayes. Linear subspaces for illumination recogntion using near-infrared images. IEEE Trans. PAMI, 29(4):627-
robust face recognition. In Proc. IEEE conf CVPR, 2001. 639, 2007.
[4] G. Bebis, A. Gyaourouva, S. Singh, and I. Pavlidis. Face recognition [33] D. Liu, K. Lam, and L. Shen. Illumination invariant face recognition.
by fusing thermal infrared and visible imagery. Image and Vision Pattern Recognition, 38:1705-1716, 2005.
Computing, 2006. [34] E. G. Llano, J. Kittler, and K. Messer. An illumination insensitive
[5] P. Belhumeur, J. Hespanha, and D. Kriegman. Eigenfaces vs fish- representation for face verification in the frequency domain. In Proc.
erfaces Recognition using class specific linear projection. IEEE ICPR, 2006.
Trans.PAMI, 19:711-720, 1997. [35] K. Messer, J. Matas, J. Kittler, J. Luettin, and G. Maitre. Xm2vtsdb:
[6] P. Belhumeur and D. Kriegman. What is the set of images of an object The extended m2vts database. In Int'l Conf on AVBPA, pages 72-77,
under all possible illumination conditions. IJCV, 28(3):245-260, 1998. 1999.
[7] V. Blanz and T. Vetter. Face recognition based on fitting a 3d [36] Hiroki Miura and et al. A 100frame/s cmos. active pixel sensor for 3d-
morphable model. IEEE Trans. PAMI, 25(9):1063-1073, 2003. gesture recognition system. In Proc. IEEE Solid-State Circuits Conf.,
[8] K. W. Bowyer, K. Chang, and P. Flynn. A survey of approaches to pages 142-143, 1999.
three-dimensional face recognition. In Proc. ICPR, 2004. [37] Y. Moses, Y. Adani, and S.Ullman. Face recognition: the problem of
[9] K. W. Bowyer, K. Chang, and P. Flynn. A survey of approaches and compensating for the illumination direction. In Proc. ECCV, 1994.
challenges in 3d and multi-modal 3d+2d face recognition. In CVIU, [38] Y. Ni, E. Krichen, S. Salicetti, and B. Dorizzi. Active differential
2006. imaging for human face recognition. IEEE Signal Processing letter,
[10] K. C. Chang, K. W. Bowyer, and P. J. Flynn. An evaluation of 13(4):220-223, 2006.
multimodal 2d+3d face biometrics. IEEE Trans. PAMI, 27(4), 2005. [39] Y. Ni and X. L. Yan. Cmos active differential imaging device with
[11] C.P. Chen and C.S. Chen. Lighting normalization with generic intrinsic single in-pixel analog memory. In Proc. IEEE Eur. Solid-State Circuits
illumination subspace for face recognition. In Proc. IEEE ICCV, 2005. Conf., pages 359-362, 2002.
[12] H. Chen, P. Belhumeur, and D. Jacobs. In search of illumination [40] Z. Pan, G. Healey, M. Prasad, and B. Tromberg. Face recognition in
invariants. In Proc. IEEE conf CVPR, 2000. hyperspectral images. IEEE Trans. PAMI, 25:1552-1560, 2003.
[13] T. Chen and et al. Illumination normalization for face recognition [41] P. J. Phillips and et al. Frvt2006 and ice2006 large-scale results.
and uneven background correction using total variation based image Technical report, NIST, 2007.
models. In Proc. IEEE CVPR, 2005. [42] P.J. Phillips and et al. Preliminary face recognition grand challege
[14] W. Chen, M.J. Er, and S. Wu. Illumination compensation and normal- results. In Proc. IEEE AFGR, 2006.
ization for robust face recognition usinf discrete cosine transform in [43] L. Qing, S. Shan, X. Chen, and W. Gao. Face recognition under
logrithm domain. IEEE Trans. SMC-B, 36(2):458-466, 2006. varying lighting based on the probabilistic model of gabor phase. In
[15] X. Chen, P. J. Flynn, and K. W. Bowyer. Visible-light and infrared Proc. IAPR ICPR, 2006.
face recognition. In Proc. Workshop on Multimodal User Authen., [44] R. Ramamoorthi and P. Hanrahan. On the relationship between
2003. radiance and irradiance: Determining the illumination from images of
[16] Shan Du and R. Ward. Wavelet-based illumination normalization for a convex lambertain object. Journal of Optical Society of American,
face recognition. In Proc. IEEE ICIP, 2005. 18(10):2448-2459, 2001.
[17] Equinox. http://www.equinoxsensors.com/product/hid.html. [45] R.Basri and D. Jacobs. Lambertian reflectance and linear subspaces.
[18] Y. Gao and M. Leung. Face recognition using line edge map. IEEE In Proc. IEEE ICCV, 2001.
Trans. on PAMI, 2002. [46] R.Basri and D. Jacobs. Lambertian reflectance and linear subspaces.
[19] A. S. Georghiades, P. N. Belhumeur, and D. J. Kriegman. From few IEEE Trans. PAMI, 25(2):218-233, 2003.
to many: Illumination cone models for face recognition under variable [47] M. Savvides and et al. Partial and holistic face recognition on frgc-ii
lighting and pose. IEEE Trans. Pattern Anal. Mach. Intelligence, data using support vector machine kernel correlation feature analysis.
23(6):643-660, 2001. In Proc. IEEE CVPRO6, 2006.

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.
TABLE I
IDENTIFICATION ERROR RATES OF PASSIVE ILLUMINATION INVARIANT APPROACHES (%)
Category Method Yale B Harvard PIE
Subset 2 3 4 5 overall subset 2 3 4 5
Illumination Eigenface[5] 0 25.8 75.7 0 4.4 41.5
Variation Eigenface w/o lst3[5] 0 19.2 66.4 0 4.4 27.7
Modelling (statistical Fisherface[5] 0 0 4.6
physical 3D Linear Subspace[5] 0 0 15.0 0 4.4 9.2
modelling Cone-attached[19] 0 0 8.6
Cone-Casted [19] 0 0 0
Segment Linear Subspace[3] 0 0 0
Generic Intrinsic Illum. Sub.[ll] 0 0 11 3.0
Spherical Harmonics[71] 0 0 2.8
9 point lights [31] 0 0 2.8 1.9
Generalized Photo. Stereo[76] 0.9
Illumination Gradient Angle[12] 0 0 1.4
Insensitive Gabor Phase[43] 0 0 2.8 4.7
Features Quotient Image[53] 1.7 38.1 65.9 76.7
Quotient Image Relighting[51] 0 0 9.4 17.6 0 16.2 42.9 71.6
Self Quotient Image[64] 2.0 1.0 3.0
Morphological Quotient Img[72] 0 0 0 0.5
Total-Variation Quotient Img[13] 0 0 0
Anitsotropic smoothing[21] 1.0 1.0 6.0
Photometric Histogram Equalization[51] 0 11.0 44.9 55.6 1.1 47.9 70.0 87.6
Normalisation Gama Intensity Correc.[51] 0 11.9 60.1 72.5 5.6 54.6 77.1 84.6
DCT in Log domain[14] 0 0 0.18 1.71 0.36
Wavelet Reconstruction[16] 0 0 5.24 9.17
Local normalization[68] 0.5 0
3D Morphable Model[7] _ 0.2

[48] M. Savvides, B.V.K. Vijaya Kumar, and P.K. Khosla. Corefaces-robust [65] S. Wei and S. Lai. Robust face recognition under lighting variations.
shift invariant pca based correlation filter for illumination tolerant face In Proc. ICPR, 2004.
recognition. In Proc. IEEE CVPRO4, 2004. [66] J. Wilder, P. J. Phillips, C. Jiang, and S. Wiener. Comparison of visible
[49] M. Savvides, B.V.K. Vijaya Kumar, and P.K. Khosla. Eigenphases vs. and infrared imagery for face recognition. In Proc. IEEE AFGR, 1996.
eigenfaces. In Proc. IEEE ICPR, 2004. [67] Chunyan Xie, M. Savvides, and B.V.K. Vijayakumar. Quaternion
[50] A. Scheenstra, A. Ruifork, and R. C. Veltkamp. A survey of 3d face correlation filters for face recognition in wavelet domain. In Proc.
recognition methods. In Proc. Int'l Conf. on AVBPA, 2005. IEEE Int'l Conf: on Acoustics, Speech, and Signal Processing, 2005.
[51] S. Shan, W. Gao, B. Cao, and D. Zhao. Illumination normalization for [68] X. Xie and K. Lam. An efficient illumintion normalization method
robust face recognition against varying lighting conditions. In Proc. for face recognition. Pattern Recognition Letters, 27, 2005.
IEEE workshop on AMFG, 2003. [69] C. Yang, S. Lai, and L. Chang. Robust face matching under different
[52] A. Shashua. On photometric issue in 3d visual recognition from a lighting conditions. EURASIP Journal on App. Sig. Proc., 2004.
single 2d image. IJCV, 1997. [70] L. Zhang and D. Samaras. Face recognition under variable lighting
[53] A. Shashua and T. Riklin-Raviv. The quotient image: class-based re- using harmonic image exemplars. In Proc. IEEE Computer Society
conf CVPR, 2003.
rendering and recognition with varying illuminations. IEEE Trans. [71] L. Zhang and D. Samaras. Face recognition from a single training
PAMI, 23(2):129-139, 2001. image under arbitrary unknown lighting using spherical harmonics.
[54] J. Short. Illumination Invariance for Face verification. PhD thesis, IEEE Trans. PAMI, 29, 2006.
University of Surrey, 2006. [72] Y Zhang and et al. Mqi baed face recognition under uneven
[55] J. Short, J. Kittler, and K. Messer. A comparison of photometric illumination. In Proc. ICB, 2007.
normalisation algorithm for face verification. In Proc. Int'l conf [73] S. Zhao and R. Grigat. An automatic face recognition system in the
AFGR, pages 254-259, 2004. near infrared spectrum. In Proc. MLDM, 2005.
[56] J. Short, J. Kittler, and K. Messer. Photometric normalisation for face [74] W. Zhao and R. Chellappa. Illumination-insensitive face recognition
verification. In Proc. Int'l Conf on AVBPA, 2005. using symmetric shape from shading. In Proc. IEEE Computer Society
[57] T. Sim, S. Baker, and M. Bsat. The CMU pose, illumination, and Conf on CVPR, volume 4, pages 286-293, 2000.
expression (PIE) database of human faces. Technical Report CMU-RI- [75] W. Zhao, R. Chellappa, and A. Rosenfeld. Face recognition: A
TR-01-02, The Robotics Institute, Carnegie Mellon University, 2001. literature survey. ACM Computing Surveys, 35:399-458, 2003.
[58] T. Sim and T. Kanade. Combining models and exemplars for face [76] S. Zhou, G. Aggarwal, R. Chellappa, and D. Jacobs. Appearance
recogntion: An illumination example. In Proc. Workshop on Models characterization of linear lambertian objects, generalized photometric
versus Exemplars in Computer Vision,CVPR, 2001. stereo and illumination-invariant face recognition. IEEE Trans. PAMI,
[59] D. Socolinsky and A. Selinger. A comparative study of face recogni- 29(2):230-245, 2007.
tion performance with visible and thermal infrared imagery. In Proc. [77] X. Zou. Illumination Invariant Face Recognition based on Active
ICPR, 2002. Near-IR Differential Imaging. PhD thesis, Univerity of Surrey, 2007.
[60] D. Socolinsky and A. Selinger. Thermal face recognition in an [78] X. Zou, J. Kittler, and K. Messer. Face recognition using active Near-
operational scenario. In Proc. IEEE conf. CVPR, 2004. IR illumination. In Proc. BMVC, pages 209-219, 2005.
[61] D. Socolinsky and A. Selinger. Thermal face recognition over time. [79] X. Zou, J. Kittler, and K. Messer. Ambient illumination variation
In Proc. ICPR, 2004. removal by active Near-IR imaging. In Proc. IAPR Int'l Conf on
[62] D. Socolinsky, A. Selinger, and J. Neuheise. Face recognition with Biometrics, pages 19-25, 2006.
visible and thermal infrared imagery. CVIU, 91:72-114, 2003. [80] X. Zou, J. Kittler, and K. Messer. Motion compensation for face
recognition based on active differential imaging. In Proc. ICB, 2007.
[63] A. Teuner and et. al. A survey of surveillance sensor systems using
cmos imagers. In Proc. Int'l Conf on Img. Analy. and Proc., 1999.
[64] H. Wang, S. Li, and Y Wang. Face recognition under varying lighting
condition using self quotient image. In Proc. IEEE AFGR, 2004.

Authorized licensed use limited to: UNIVERSIDADE DE PERNAMBUO. Downloaded on August 05,2010 at 19:15:50 UTC from IEEE Xplore. Restrictions apply.

Potrebbero piacerti anche