Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
net/publication/3906043
CITATIONS READS
28 30
3 authors:
Ales Leonardis
University of Birmingham
288 PUBLICATIONS 8,202 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Horst Wildenauer on 01 June 2014.
233
$10.00 0 2001 IEEE
0-7695-1143-0/01
images of objects that are illumination invariant. However, that the representation obtained by first filtering and then
it was demonstrated in that paper that from the image gra- performing PCA is not efficient (in terms of the number of
dient a measure which is insensitive to illumination can be eigenimages needed). This can be explained by the fact that
developed and used for probabilistic object recognition. In the filtered (edge) images are less correlated, therefore more
a similar spirit, Jacobs et al. [9] developed a measure from eigenimages are needed for an accurate representation.
the ratio of two images for comparing images under variable The representation based on first performing PCA and
illumination. It was stated that this measure can be inter- then filtering has the additional advantage of increased flex-
preted as a simple comparison between image edges. Sim- ibility. Namely, once we have performed the PCA we can
ilar results have been obtained in [14]. From these studies, apply any linear filter without recalculating the eigenimages
we can conclude that edges and gradients are useful mea- (therefore we also do not need to access to the original train-
sures to deal with variations in illumination. However, in ing set). Whereas in the other approach, any time when we
the approaches mentioned above, the authors did not take require to use a new filter, we first have to filter the original
into account multiple poses. In particular, we would like training images and then perform PCA.
to be able to deal with the illumination variations in cases The main aim of this paper is to exploit the filtered
when a set of images is compressed-in an eigenspace and eigenspace representation to obtain illumination insensitiv-
the input objects to be recognized are taken under' various ity in object recognition. In section 2 we show how to ob-
illumination conditions. This would enable us to perform tain the coefficients of the eigenimage expansion from the
recognition in an efficient manner. ' ' filtered images. We demonstrate that by taking a filter bank
In this paper we propose an approach based on eigen- of gradient filters, one can already achieve illumination in-
images filtered by gradient-based operators. Namely, we sensitivity to a large degree. As an extension, we show how
show how to recover eigenspace coefficients in a global the coefficients can be robustly recovered in order to deal
eigenspace representation from responses of local filter with cluttered background and other types of non-Gaussian
banks. Based on that, we discuss how this approach can be noise. In section 3 we extensively evaluate the approach on
used for illumination insensitive object recognition. This object recognition with different illuminations and compare
work extends the approaches mentioned above in the direc- the results to the one obtained with other approaches. In
tion that we can deal also with multiple instances of objects section 4 we discuss how this global-local representation
in multiple poses in a compressed representation. can be used to further enhance the current capabilities of
To be more specific, let y = [ y l , .. . , ymIT E Rm be an recognition methods based on eigenimage representations.
individual image, and Y = { yl, . . . yn} be a set of images.
To simplify the notation we assume Y to be normalized, 2. Eigenimages and Local Filters
having zero mean. The set of eigenvectors obtained from
Y is denoted by E = { e l , .. . e n } ;ei = [ e i l , . . . ,eimIT E
Rm. Let us denote the PCA transform by P C A p ( Y )= &p, In the standard approach, the parameters ai of the eigen-
image expansion are obtained by projecting an image in the
where p indicates that usually only p , p < TI, eigenvectors
form of data vector x onto the eigenspace
(those with the largest eigenvalues) are needed to represent
the yi to a sufficient degree of accuracy as a linear combi- m
nation of eigenvectors ei ai(x) =< x,ei >= C z j e i j i = 1.. . p . (2)
j=1
P
234
This has the additional advantage that the eigenimages need
not be orthogonal.
We take Eq. (3) as the starting point of our derivation of
a new approach which preserves all the advantages of the
eigenspace method, and augments it with the rich local im-
age structure. We can derive the following property which
holds due to the linearity of the equation
P
(f * 4 ( d = C.4f * e2>(.3) , (4)
(a) Original im-
age
(b) Image illumi-
nated from right
(c) Reconstruc-
tion obtained
2=1 from (b)
This equation states that we can calculate the coefficients Figure 3. Demonstration of illumination in-
u2 from the filtered eigenimages and the filtered input im- sensitivity of filtered eigenspaces.
age. We have demonstrated [4] that the recovered coeffi-
cients in filtered and subsampled images can be calculated
and remain stable despite the fact that due to filtering and
subsampling the eigenimages are no longer orthogonal. 2. we can recover the coefficients from a single point us-
ing q filter responses at that single point;
3. we can use a combination of the above two options.
As preliminary experiments have shown, we get the most
reliable (numerically stable) results using the third option.
We have now the freedom to choose a bank of linear filters
which are insensitive to illumination variations. From [5,9]
we can conclude that gradient-based filters achieve the de-
sired effect. In particular, we have selected a set of steerable
filters [ 131. Figure 2 depicts 8 eigenimages filtered with 6
derivative filters in different orientations.
Fig. 3 demonstrates the insensitivity of the filtered
eigenspaces to illumination changes. Though there is a sig-
nificant change in the illumination conditions between the
images acquired in the training phase (see also section 3)
we can still recover the correct coefficients as can be seen
from the reconstruction in Fig. 3(b).
235
one can tolerate a significant amount of (non-Gaussian)
noise, considerable occlusions, and does not need to pre- (a) Lion with (b) Robustly (c) Reconstruc-
segment the images. shadow reconstructed tion of (a) using
image obtained the standard
Fig. 4 shows the result of using the robust procedure for from (a) method
calculating the coefficients of the same object as in Fig. 3(a).
Note that in this case, despite the fact, that we have already Figure 5. Demonstration of robustness
reached saturation in the frontal part of the image, we were . and illumination insensitivity of filtered
able to calculate the coefficients accurately, as can be seen eigenspaces (cast shadow).
from Fig. 4(b). Using the standard method (Eq. ( 2 ) ) to re-
cover the coefficients, we get the reconstruction depicted
in Fig. 4(c). Fig. 5 depicts another example, where more
6. As has been shown in [lo], under certain conditions,
than 60% of the figure is in a shadow. In this case the re-
variations in the number of hypotheses and the number of
construction is not as good as in the previous case but still
points do not change the results significantly. One should
sufficiently good for a reliable recognition.
note that from the point of view of solving Eq. ( 5 ) we can
decrease the number of filters if we increase the number of
points and vice-versa.
236
Figure 9. Illumination conditions tested in the
recognition experiment.
Figure 7. Test-set of “rhino” object under dif-
ferent illuminations.
Figure 8. Comparison between the standard images contain the appropriate discriminative information.
and our robust method.
4. Summary and Conclusions
237
Acknowledgments
Table 1. Comparison of the recognition rates
for different methods. The tables show the H. B. and H. W. were supported by a grant from the Aus-
confusion matrix and the achieved recogni- trian National Fonds zur Forderung der wissenschaftlichen
tion rate (“A) for the individual objects. For Forschung (P13981INF). H. B. acknowledges the support
those objects that were correctly recognized by the K plus Competence Center ADVANCED COM-
we also calculated the mean absolute error of PUTER VISION. A. L. acknowledges the support from the
the pose estimation (ang.). Ministry of Science and Technology of Republic of Slove-
Robust filtered method - all eigenvectors used. nia (Project 52-0414).
obj.1 1 2 3 4 51 % I ang.
1 I 360 0 0 0 0 1 100.0 5.25 I References
1 1 2i 25i 1
Standard method - all eigenvectors used.
obj. 1 2 3 4 5 % ang. I vs. Fisherfaces: Recognition Using Class Specific Linear
Projection. IEEE Trans. on Pattern Analysis and Machine
1 1 141 0 14 26 179 39.2 I 10.50 Intelligence, 19(7):711-720, July 1997.
18.90
3;:
38 249
18;
44
i;::
69.2
3.47
7.1 1
[3] P. Belhumeur and D. Kriegman. What is the set of images
of an object under all possible illumination conditions? Int.
J. Computer Vision, 28(3):245-260, July 1998.
51 0 557 91.0 6.82 [4] H. Bischof and A. Leonardis. Robust recognition of scaled
avg. 70.3 8.53 eigenimages through a hierarchical approach. In Proc. of
CVPR98, pages 664-670, 1998.
Robust filtered method - w/o first three eigenvectors
[5] H. Chen, P. Belhumeur, and D. Jacobs. In search of illumi-
obj. 1 I 2 3 4 5 I
% ang. I nation invariants. In CVPROO, pages I:254-261,2000.
1 I 359 1 0 0 0 I 99.7 ‘1 4.87 [6] H. Farid and E. Adelson. Separating reflections and lighting
using independent components analysis. In CVPR99, pages
1:262-267, 1999.
[71 A. Georghiades, D. Kriegman, and P. Belhumeur. Illumina-
238