Sei sulla pagina 1di 25

INTRODUCTION

FDDB: A Benchmark for Face Detection in


Unconstrained Settings

Vidit Jain and Erik Learned-Miller

Technical Report UM-CS-2010-009, Dept. of Computer Science,


University of Massachusetts, Amherst. 2010.

Referent
http://vis-www.cs.umass.edu/lfw/index.html
http://www.decom.ufop.br/menotti/rp122/sem/
http://faculty.ucmerced.edu/mhyang/course/cse185/lectures/face_detection.ppt

INTRODUCTION
Several algorithms have been developing for face detections,
however remain difficult to compare due to the lack of enough detail
to reproduce the published results.

This paper presents a new data set of face images with more faces
and more accurate annotations for face regions. Also, propose two
rigorous and precise methods for evaluating the performance of face
detection algorithms. Finally, reports results of several standard
algorithms on the new benchmark.

Face detection data set


2845 images with a total of 5171 faces.
The data set includes occlusions, difficult poses, low resolutions
and out-of-focus faces.
Especification of face regions as elliptical regions.

Both grayscale and color images.

Near-duplicate detection

Annotating face regions

Challenges in face labeling. For some image regions,


deciding whether or not it represents a face can be challenging.
Several factors such as low resolution (green, solid), occlusion
(blue, dashed), and pose of the head (red, dotted) may make
this determination ambiguous.

Evaluation
A detection corresponds to a contiguous image region.
Any post-processing required to merge overlapping detections
has already been done.
Each detection corresponds to exactly one entire face.

10

Matching detections and annotations

A matching of detections to face regions in


this graph corresponds to the selection of a
set of edges M E.
Mathematically, the desired matching M
maximizes the cumulative matching score
while satisfying the following constraints:

The determination of the minimum weight matching in a weighted bipartite graph has an
equivalent dual formulation as finding the solution of the minimum weighted (vertex) cover
problem on a related graph.

11

Evaluation metrics

A score of 1 is assigned to the detected region and 0 otherwise.

This ratio is used as the score for the detected region.

12

Experimental Setup
10 fold cross-validation
A 10-fold cross-validation is performed using a fixed partitioning of
the data set into ten folds.

Unrestricted training
Data outside the FDDB data set is permitted to be included in the
training set.

13

Benchmark

14

The
one use primary_up X4 of Dlib library. The
one use primary_up
X2 of Dlib library we use. The
one not use primary_up.
detectContROC

Thursday, August 06, 2015

detectDiscROC

15

The red one use primary_up X4. The black one use primary_up X2
and the blue one not use primary_up. The graph a bit change of primary_up X4.
detectDiscROC

16

continuous ROC

17

discontinued ROC

18

Can see the


Rank of FDDB Face Detection Data Set and
Benchmark.xlsx
Excel file.
To easy to see the Rank.

19

Continuous ROC at 0-140 False positive


The red one is primary_up X4 and
The black one is primary_up X2

Thursday, August 06, 2015

20

Discontinuous ROC at 0-140 False


continuous
positive ROC at 0-140 False positive
The red one is primary_up X4 and
The black one is primary_up X2

Thursday, August 06, 2015

21

I find that FDDB use 10-fold cross-validation in the evaluation and I find
the information.

k-fold cross-validation1
k-fold cross-validation In k-fold cross-validation, the original sample is randomly
partitioned into k equal sized subsamples. Of the k subsamples, a single subsample is
retained as the validation data for testing the model, and the remaining k 1 subsamples
are used as training data. The cross-validation process is then repeated k times (the
folds), with each of the k subsamples used exactly once as the validation data. The k
results from the folds can then be averaged (or otherwise combined) to produce a single
estimation. The advantage of this method over repeated random sub-sampling (see
below) is that all observations are used for both training and validation, and each
observation is used for validation exactly once. 10-fold cross-validation is commonly
used,[7] but in general k remains an unfixed parameter.

1 https://en.wikipedia.org/wiki/Cross-validation_(statistics)

Thursday, August 06, 2015


2

https://wipawanblog.files.wordpress.com/2013/08/lab_datamining.pdf

22

Thursday, August 06, 2015

http://image.slidesharecdn.com/processminingchapter03datamining-110510153206phpapp02/95/process-mining-chapter-3-data-mining-29-728.jpg?cb=1305044621

23

Other Training Sets


Mainly used for
face recognition
Need to crop
and pre process
the face images
in the data set
Cropped and pre processed data set with face and non-face images
provided by MIT CBCL: http://www.ai.mit.edu/projects/cbcl/software-datasets/index.html

Other Standard Test Sets


MIT test set (http://www.cs.cmu.edu/~har): subsumed by CMU
test set
CMU test set (http://www.cs.cmu.edu/~har) (de facto
benchmark): 130 gray scale images with a total of 507 frontal
faces
CMU profile face test set
(http://eyes.ius.cs.cmu.edu/usr20/ftp/testing_face_images.tar.gz
) : 208 images with faces in profile views
Kodak data set (Eastman Kodak Corp): faces of multiple size, pose
and varying lighting conditions in color images

Potrebbero piacerti anche