Академический Документы
Профессиональный Документы
Культура Документы
Clustering
Soheil Shafiee, Farhad Kamangar, Vassilis Athitsos, and Junzhou Huang
Department of Computer Science and Engineering, University of Texas at Arlington, Arlington, TX, USA
1. Introduction
Face recognition is one of the most challenging realworld applications in computer vision and over the past few
decades, several methods and algorithms have been proposed
to solve this problem. The face recognition techniques could
be broadly divided into two main groups, i.e. geometric
feature-based and template matching approaches. The main
idea in the geometric feature-based methods is to extract
and record the relative position and salient properties of
distinctive face features such as eyes, mouth corners, chin,
and nose. The geometrical features are then used to create
labeled graphs and structural models which can be used to
discriminate between different classes (individuals) [1],[2].
The template matching algorithms, in their simplest forms,
present the image of each face as a vector of intensity values
in a multi-dimensional space. An unknown face is then
compared to a set of known classes by using a metric, such
as Euclidean distance. There are, of course, variations and
(1)
(2)
In fact, vector b
x1 should not necessarily be sparse to be
recovered by (2). It may be sparse in some domain (other
than b
x1 s original domain) [12]. For instance, vector b
x1 ,
could be a general non-sparse signal which has a sparse
representation in frequency or Wavelet domain.
In practice, due to the existence of noise in measurements,
the solution to y = Ax is not exact. In other words, the
system of linear equations, y = Ax should be modified as
y = Ax + n, where n is an n dimensional noise vector. In
this case, the optimization problem (2) may be reformulated
as follows:
b
x1 = arg min kxk1 subject to kAx yk2 ,
(3)
(4)
(7)
30
25
Count
20
15
10
5
0
8
10
12
14
16
Number of clusters per class
18
20
22
(8)
where b
y Rd represents the extracted feature vector of the
original unknown image y, and R Rdm with d m,
b Rdn represents the
is a feature extraction matrix. A
training matrix with reduced dimensionality. Using equation
(8), fewer training face images are needed to form an underb as the
determined system of linear equations. Using A
training matrix, equation (3) can be reformulated as:
b
b
x1 = arg min kxk1 subject to
Ax
b
y
. (9)
2
Different dimensionality reduction matrices such as random projection, down-sampling, Eigen, Fisher, and Laplacian, are studied in [7]. In this paper we utilize three of
the proposed methods: random projection, down-sampling,
and Eigen features. Random projection algorithm employs
a normal distribution function to randomly select a normalized subset of the original face images as features. Downsampling method uses a bi-cubic algorithm to down-sample
CL 1 with 27 faces
CL 5 with 10 faces
CL 2 with 13 faces
CL 6 with 10 faces
CL 3 with 13 faces
CL 7 with 9 faces
CL 4 with 12 faces
25
20
15
10
CL 8 with 6 faces
2000
4000
6000
Number of columns in matrix A
8000
10000
CL 10 with 6 faces
u i
Nj
uX
2
u
i,j
i
,
t
MaxDisti = max
(10)
u
C
j
l
l=1
3.1 Results
The relationship between the running time and size of
the training matrix for SRC algorithm is shown in Fig. 4.
The average running time for both SRC and the proposed
method is the same because they are both using the same
process to solve the l1 norm minimization problem, which
is quadratic to the number of columns in the training matrix.
It should be noted that at a same computational complexity,
K-SRC achieves higher recognition rates than the original
SRC, as shown by the experiments in section 4.
100
Recognition Rate (%)
1500
1000
90
80
70
60
50
200
SRCRandom
KSRCRandom
400
600
800
1000
1200
1400
Number of columns in matrix A
(a)
1600
1800
2000
500
100
0.5
1.5
2
2.5
3
MaxDist threshold ()
3.5
4.5
2000
70
60
SRCEigen
KSRCEigen
50
400
600
800
1000
1200
1400
Number of columns in matrix A
(b)
1600
1800
2000
100
Recognition Rate (%)
80
40
200
4. Experiments
4.1 The Face Dataset
90
90
80
70
60
SRCDownsampled
KSRCDownsampled
50
40
200
400
600
800
1000
1200
1400
Number of columns in matrix A
(c)
1600
1800
2000
5. Conclusion
In this paper, we proposed an efficient method based on
sparse representation classification (SRC) for face recognition application. In this approach, instead of using the
original face images, clusters of training face images are
used to form the training matrix. An adaptive K-means
method is implemented to form the training data clusters.
Adjusting the parameters of the adaptive clustering algorithm
allows for a controlled trade off between speed and accuracy
of the recognition process. Three different feature extraction
algorithms were used in the experiments and results were
compared with the original SRC algorithm. We have tested
our method on a face dataset with 14794 training face
images from 100 classes. Results show that the proposed
approach reduces the running time and memory requirements
of the recognition task while preserving the classification
K-SRC
Acc(%)
Time(s)
95.0
95.0
94.7
0.03
0.18
0.08
SRC
A Col.
300
900
500
Acc(%)
94.0
94.2
94.2
Time(s)
A Col.
0.81
1.33
2.64
2000
2500
3500
100
90
80
70
60
50
0
SRCRandom
KSRCRandom
100
200
300
400
500
600
Average time per test (msec)
(a)
700
800
900
100
90
80
70
60
SRCEigen
KSRCEigen
50
40
0
100
200
300
400
500
600
Average time per test (msec)
(b)
700
800
900
100
Recognition Rate (%)
90
80
70
60
SRCDownsampled
KSRCDownsampled
50
40
0
100
200
300
400
500
600
Average time per test (msec)
(c)
700
800
900
References
[1] R. Jafri and H. R. Arabnia, A survey of face recognition techniques,
Journal of Information Processing Systems, vol. 5, no. 2, pp. 4168,
2009.
[2] S. Chitra and G. Balakrishnan, A survey of face recognition on
feature extraction process of dimensionality reduction techniques,
Journal of Theoretical and Applied Information Technology, vol. 36,
no. 1, pp. 92100, 2005.
[3] M. Turk and A. Pentland, Eigenfaces for recognition, IEEE International Conference on Computer Vision and Pattern Recognition,
1991.
[4] P. Belhumeur, J. Hespanda, and D. Kriegman, Eigenfaces versus
fisherfaces: Recognition using class specific linear projection, IEEE
Transaction on Pattern Analysis and Machine Intelligence, vol. 19,
no. 7, pp. 711720, 1997.
[5] X. He, S. Yan, Y. Hu, P. Niyogi, and H. Zhang, Face recognition
using laplacianfaces, IEEE Transaction on Pattern Analysis and
Machine Intelligence, vol. 27, no. 3, pp. 328340, 2005.
[6] M. Kirby and L. Sirovich, Application of the karhunen-loeve procedure for the characterization of human faces, IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 12, no. 1, pp. 103
108, 1990.
[7] J. Wright, A. Yang, A. Ganesh, and S. Sastry, Robust face recognition
via sparse representation, IEEE Transaction on Pattern Analysis and
Machine Intelligence, vol. 31, no. 2, pp. 210227, 2009.
[8] D. Donoho, Compressed sensing, IEEE Transactions on Information
Theory, vol. 52, no. 4, pp. 12891306, 2006.
[9] E. Candes, Compressive sampling, International Congress of Mathematics, 2006.
[10] D. Donoho and Y. Tsaig, Fast solution of l1-norm minimization
problems when the solution may be sparse, Available Online:
http://dsp.rice.edu/sites/dsp.rice.edu/files/cs/FastL1.pdf, 2006.
[11] E. Amaldi and V. Kann, On the approximability of minimizing
nonzero variables or unsatisfied relations in linear systems, Theoretical Computer Science, 1998.
[12] R. Baraniuk, Compressive sensing, IEEE Signal Processing Magazine, vol. 24, no. 4, pp. 118121, 2007.
[13] R. Basri and D. Jacobs, Lambertian reflectance and linear subspaces,
IEEE Transactions on Pattern Analysis and Machine Intelligence,
vol. 25, no. 2, pp. 218233, 2003.
[14] J. B. MacQueen, Some methods for classification and analysis of
multivariate observations, Proceedings of 5-th Berkeley Symposium
on Mathematical Statistics and Probability, 1967.
[15] R. Xu and D. Wunsch, Survey of clustering algorithms, IEEE
Transactions on Neural Networks, vol. 16, no. 3, pp. 645678, 2005.
[16] P. Phillips, P. Flynn, T. Scruggs, and K. Bow, Overview of the face
recognition grand challenge, IEEE Conference on Computer Vision
and Pattern Recognition, 2005.
[17] E. Candes and J. Romberg, L1-magic : Recovery of sparse signals via
convex programming, Available Online: http://users.ece.gatech.edu
/ justin/l1magic/downloads/l1magic.pdf, 2005.