Академический Документы
Профессиональный Документы
Культура Документы
n
i=1
u
i
a
i
(3)
where | | represents the absolute value.
In the second stage, since the number of feature points
inside the ear vary, it is represented as y and z for the two
images being compared and it is assumed that y > z. Let
V
21
= [
1
,
2
,
3
, ...,
y
] and V
22
= [
1
,
2
,
3
, ...,
z
] be the
feature vectors of the two images. Let m
2
be the number of
points that are matched between two images, for a specic
normal line and it is calculated as
m
2
=
y
i=1
z
j=1
b
i
f
j
(4)
The number of matching points, m
2
is rst initialised to zero.
Then, based on the difference of pairs of points on the two
images, m
2
is updated as
m
2
= m
2
+1, if |b
i
f
i
| , threshold
= m
2
, if |b
i
f
i
| . threshold
(5)
This process is repeated for all feature points corresponding
to each normal line in the ear image. So, two points are
said to be matched, if their angles are nearly the same with
respect to each normal line. As the size of the second
feature vector is not xed, the percentage of matched
points, M
2
, is calculated as
M
2
= m
2
/ min v
1
, v
2
_ _
(6)
where v
1
, v
2
are the size of second feature vectors of the two
images. The value of M
2
represents the total number of
matching points between the two images. These scores are
then converted into distance scores, D
2
by subtracting them
from the highest matching score, so that D
2
can be
combined with D
1
scores of the rst stage.
4 Fusion of distance scores
Initially the verication rates are found out separately by
using the matching scores obtained through the two stages
of feature extraction. Then, the nal distance score, s
ear
, is
obtained by adding the distances D
1
and D
2
by using
weighted sum fusion that is given by
s
ear
= w
1
D
1
+w
2
D
2
(7)
where w
1
and w
2
are the weights assigned to scores of the two
stages of feature extraction. Equal weights are applied as the
two stages produce almost equal verication rates.
5 Optimisation of fusion parameters using
PSO
PSO is an evolutionary search algorithm motivated from the
social behaviour of a ock of birds trying to y to a
favourable environment. PSO is employed to nd the
solution for the adaptive selection of combination of
individual points, which are referred to as the particles in
multi-dimensional search space [20]. Each particle
(representing a bird in the ock), characterised by its
position and velocity, represents the possible solution in the
search space. The behaviour of particles in PSO imitates the
way in which birds communicate with each other while
ying. During this communication, each bird reviews its
new position in space with respect to the best position it
has covered so far. Birds in the ock also identify the bird
that has reached the best position. On knowing this
information, other birds update their velocity (that depends
on a birds local best position as well as position of the best
Fig. 4 Illustration of the feature extraction process
a Input image
b Cropped image
c Edge image
d Image with a reference line
e Image with normal lines
Fig. 5 Cropped ear image with reference and normal lines
www.ietdl.org
IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106 101
doi: 10.1049/iet-bmt.2012.0049 & The Institution of Engineering and Technology 2013
bird in the ock) and y towards the best bird. The process of
regular communication and updation of velocity repeats until
the ock nds a favourable position.
PSO algorithm is implemented to nd the optimal
threshold that maximises accuracy. Each particle in PSO
stores the best position visited by it so far, called pbest
(local best position) in its memory. The particle also
interacts with all neighbours and stores the best position
visited by any particle in the search space and experiences a
pull towards this position called gbest (global best
position). The values of pbest and gbest are updated after
each iteration, if a more dominating solution is found by
the particle and by the population, respectively. This
process continues iteratively until either the desired result is
achieved or the computational power is exhausted [18].
Each particle in the k-dimensional space is dened as X
a
=
(x
a1
, x
a2
, , x
ak
), where the subscript a is the particle number
and the second subscript denotes the dimension. The previous
best position is represented as P
a
= ( p
a1
, p
a2
, , p
ak
) and
velocity along each dimension as V
a
= (v
a1
, v
a2
, , v
ak
).
The particles move to a new position in multi-dimensional
solution space depending on the particles best position
[local best position ( p
ak
) and global best position ( p
gk
)].
The positions p
ak
and p
gk
are updated after each iteration,
whenever a suitable lower cost solution is located by the
particle. The velocity vector of each particle determines the
movement details. The velocity and position update (8) and
(9) of a particle of the PSO for the instance (t + 1) can be
represented as follows
n
ak
t +1 ( ) = en
ak
(t) +c
1
r
1
r
ak
(t) x
ak
(t)
_ _
+c
2
r
2
r
gk
(t) x
ak
(t)
_ _
(8)
x
ak
(t +1) = x
ak
(t) +n
ak
(t +1) (9)
where is the inertia weight and ranges between 0 and 1 that
provides a balance between global and local search abilities of
the algorithm and controls the memory of the particle. The
accelerator coefcients c
1
and c
2
are positive constants and
r
1
and r
2
are two random numbers in the range (01),
usually assigned the same value to give equal weights and
they determine the relative inuence of pbest and gbest.
Also the particles are initialised randomly between 0 and 1.
The weights represented by particles are applied to the
training data and errors are found by computing FAR and
FRR, and then calculating the tness of each particle. The
error rate E is given by
E = C
1
FAR +C
2
FRR (10)
where C
1
is the cost of falsely accepting an imposter, C
2
is the
cost of falsely rejecting the genuine individual, FAR is the
global false acceptance rate and FRR is the global false
rejection rate. Equation (10) can be minimised by choosing
the appropriate decision threshold and can be rewritten in
terms of single cost as
C
2
= 2 C
1
(11)
which makes (10) as
E = C
1
FAR + 2 C
1
_ _
FRR (12)
Equation (12) results from the fact that global error is simply a
sum of the two error rates FAR and FRR, if cost is ignored,
that is, the costs are assumed to be 1. In this algorithm, rst
FAR is evolved for each trait and then thresholds are
calculated based on the FAR and updated. The global FAR
and FRR for the optimum threshold are then calculated
from the individual values of FAR and FRR.
6 Experimental results and discussion
The proposed method uses USTB database containing 308
ear images of 77 persons [21]. The images are taken under
Fig. 6 Different stages of ear image verication
www.ietdl.org
102 IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106
& The Institution of Engineering and Technology 2013 doi: 10.1049/iet-bmt.2012.0049
different illuminations and angles. From this set, 280 images
are taken up for the analysis and 1000 intra and inter class
comparisons are made. Different stages of the proposed ear
authentication system are illustrated in Fig. 6.
Distance scores obtained out of the comparison of images
are plotted in Fig. 7, which shows that fusion of scores
produce more separation between intra and inter classes
than that of the individual scores. In addition, most of the
genuine scores obtained from features of the outer ear are
found to occupy a smaller range compared with that of the
inner ear. Similarly imposter scores obtained from the outer
ear show wider variation than inner ear. These observations
indicate the presence of more unique features in the outer
ear shape.
Performance of the two stages of feature extraction and
their fusion is shown in the receiver operating
characteristic (ROC) graphs of Fig. 8, which is plotted by
nding genuine acceptance, false acceptance and false
rejection rates using the distance scores. Here, equal
weights are used for combining the scores from the two
stages.
Table 1 compares the performance of the proposed system,
when it is applied on the complete ear image and on the
middle portion of ear. When the proposed system is
implemented on a system with Intel core 2 Duo processor
T6600 operating at 2.20 GHz and 3 GB RAM, it is found
that the computation time with a feature set from the middle
portion of the ear is reduced to half of the actual time taken
for processing the complete ear image. It is also seen that
verication rate has signicantly improved when the two
stages of feature extraction are combined by weighted sum
fusion.
Fig. 7 Inter and intra-class variations for scores
a First stage of feature extraction
b Second stages of feature extraction
c Fusion of scores of these two stages
www.ietdl.org
IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106 103
doi: 10.1049/iet-bmt.2012.0049 & The Institution of Engineering and Technology 2013
Table 2 shows the verication rates obtained by varying the
weights applied to the scores from the two stages of feature
extraction. It is observed that optimal threshold values
found by PSO enhance the accuracy of the system as shown
in the last column of the table. It is also seen that the
weight combination w1 = 0.6 and w2 = 0.4 gives the highest
recognition accuracy. Thus, the use of PSO gives better
performance results by nding the optimal threshold for the
given weight values.
Role of partial ear images with varying levels of
occlusion is studied by taking an ear image from the
USTB database. A portion of the ear image is removed by
cropping (from 10 to 50% of the area) the top and bottom
locations and is shown in Fig. 9. Each of these partial
images is separately fed into the proposed system and the
authentication accuracy is found out.
The experimental results listed in Table 3 demonstrate that
the verication rate is considerably higher up to 30%
occluded ear images. But when 40% of ear portion is
removed from the top and bottom put together, it is found
that the verication rate drops abruptly. The result shows
that partial ear images can reduce the effect of noise
(because of hair and ear rings) to some extent, especially
for slighter occlusion. But when the occlusion is heavier, it
degrades the performance of the system.
Performance of the proposed method with two stages of
feature extraction is compared with existing methods and is
shown by the ROC graphs in Fig. 10. As seen from the
graph, the proposed fusion method performs well, in spite
of the lesser performance of the individual stages of fusion
compared with other methods.
Fig. 9 Partial ear image with varying levels of occlusion
a Original image
b 10%
c 20%
d 30%
e 40%
f 50%
Table 1 Performance comparison of the proposed ear
authentication system
Method used Verification
rate, %
Error
rate
Time
taken, s
feature extraction using
complete ear image
95.6 0.055 40.24
feature extraction from
middle ear (stage-1)
93.2 0.07 22.82
feature extraction from
middle ear (stage-2)
91.5 0.077 23.34
fusion of scores from
stages 1 and 2
98.3 0.03 24.52
Table 3 Effect of partial ears on authentication accuracy
Portion of ear removed (top and
bottom), %
10 20 30 40 50
Verification rate, % 99.6 98.8 98.3 86.4 71.6
Fig. 8 ROC curves for ear scores plotted between
a FAR and GAR
b FAR and FRR
Table 2 Performance comparison of the proposed system
before and after applying PSO
Fusion rule
applied
Weights
used
(w1, w2)
GAR
(before
PSO), %
Optimal
threshold
GAR
(after
PSO), %
weighted
sum
(0.5, 0.5) 98.3 0.798 98.9
(0.6, 0.4) 98.6 0.891 99.2
(0.4, 0.6) 97.8 0.674 98.4
(0.7, 0.3) 98.1 0.921 98.8
(0.3, 0.7) 96.5 0.562 97.3
www.ietdl.org
104 IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106
& The Institution of Engineering and Technology 2013 doi: 10.1049/iet-bmt.2012.0049
Table 4 compares the performance of the proposed method
with a few existing methods of ear recognition. As it can be
seen, the proposed hierarchical method gives enhanced
performance, which is because of the removal of occlusion
in the ear image and the fusion of scores of the two stages
of feature extraction.
7 Conclusion
A two-stage geometric approach that is both scale and rotation
invariant is adopted in this work for extracting the unique
features of an ear image. As occlusion because of ear rings
and hair signicantly affect the efciency of ear recognition
process, only the middle portion of ear is considered in the
analysis, which is normally free from occlusion. It is found
that the computation time is reduced by 40%, as the size of
the ear image used for authentication is reduced to half of
its original size. The results also show an increase in the
verication rate with partial ear images, which signies the
presence of more unique features in the middle part of ear
than in other parts. Finally, particle swarm optimisation
technique is applied on the matching scores in order to
optimise the fusion parameters such as decision threshold
and weights. Use of this optimal fusion rule results in
further improved verication rates compared with the fusion
of scores without optimisation. Thus, the fusion of distance
scores obtained from two-stage feature extraction is found to
enhance the system accuracy level signicantly.
8 Acknowledgment
The authors acknowledge the University of Science and
Technology Beijing (USTB) for providing the ear image
database for this research work.
9 References
1 Stan, Z.L. (Ed.): Encyclopedia of biometrics (Springer, 2009), Vol. 2
2 Iannarelli, A.: Ear identication, forensic identication series
(Paramount Publishing Company, Freemont, CA, 1989)
3 Pug, A., Busch, C.: Ear biometrics: a survey of detection, feature
extraction and recognition methods, IET Biometrics, 2012, 1, (2),
pp. 114129
4 Choras, M.: Ear biometrics based on geometrical feature extraction,
Electron. Lett. Comput. Vis. Image Anal., 2005, 5, (3), pp. 8495
5 Choras, M.: Ear biometrics in passive human identication systems.
Proc. Sixth Int. Workshop on Pattern Recognition in Information
Systems, 2006, pp. 20042009
6 Choras, M.: Image feature extraction methods for ear biometrics a
survey. Sixth Int. Conf. on Computer Information Systems and
Industrial Management Applications, 2007, pp. 261265
7 Shailaja, D., Gupta, P.: A simple geometric approach for ear recognition.
Proc. Ninth Int. Conf. Information Technology, 2006, pp. 164167
8 Mu, Z., Yuan, L., Xu, Z., Xi, D., Qi, S.: Shape and structural feature
based ear recognition (Springer Berlin, Heidelberg, 2004) (LNCS,
3338), pp. 663670
9 Burge, M., Burger, W.: Ear biometrics in computer vision. Proc. 15th
Int. Conf. on ICPR, Barcelona, Spain, 2000, Vol. 2, pp. 2822
10 Burge, M., Burger, W.: Biometrics: personal identication in
networked society. Chapter Ear Biometrics, Kluwer Academic,
Boston, 1999, pp. 273285
11 Yuan, L., Mu, Z., Xu, Z.: Using ear biometrics for personal
recognition, in Li, S.Z., Sun, Z., Tan, T., Pankanti, S., Chollet, G.,
Zhang, D. (Eds.): IWBRS 2005 (Springer, Heidelberg, 2005)
(LNCS, 3781), pp. 221228
12 Yuan, L., Mu, Z.C.: Ear recognition based on local information fusion,
Pattern Recognit. Lett., 2012, 33, (2), pp. 182190
13 Zhang, H., Mu, Z., Qu, W., Liu, L.-M., Zhang, C.-Y.: A novel approach
for ear recognition based on ICA and RBF network. Proc. Int. Conf. on
Machine Learning and Cybernetics, 2005, Vol. 7, pp. 45114515
14 Nosrati, M., Faez, K., Faradji, F.: Using 2D wavelet and principal
component analysis for personal identication based on 2D ear
structure. Proc. Int. Conf. on Intelligent and Advanced Systems,
Kuala Lumpur, Malaysia, November 2528, 2007, pp. 616620
15 Zhang, H., Mu, Z.: Compound structure classier system for ear
recognition. Int. Conf. on Automation and Logistics, Qingdao,
September 13, 2008, pp. 23062309
Fig. 10 Comparison of ROC curves for existing and the proposed methods
Table 4 Performance comparison of existing ear recognition
systems with the proposed system
Authors Number of
images
taken
Database
used
Verification
rate, %
Prakash et al. [22] 150 IITK 95.2
Kumar and Wu [16] 465 IITD 96.3
Shailaja and Gupta
[7]
160 Self dataset 95
Kisku et al. [23] 102 IITK 93.4
proposed method
(with 30% ear
occlusion)
280 USTB 98.3
proposed method
(with optimised
parameters)
280 USTB 99.2
www.ietdl.org
IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106 105
doi: 10.1049/iet-bmt.2012.0049 & The Institution of Engineering and Technology 2013
16 Kumar, A., Wu, C.: Automated human identication using ear
imaging, Pattern Recognit., 2012, 45, (3), pp. 956968, doi:
pi10.1016/j.patcog.2011.06.005
17 Zhou, J., Cadavid, S., Abdel-Mottaleb, M.: An efcient 3-D ear
recognition system employing local and holistic features, IEEE Trans.
Inf. Forensics Sec., 2012, 7, (3), pp. 978991
18 Kumar, A., Passi, A.: Comparison and combination of iris matchers
for reliable personal authentication, Pattern Recognit., 2010, 23, (3),
pp. 10161026
19 Gonzalez, R.C., Woods, R.E.: Digital image processing (Prentice-Hall,
Upper Saddle River, NJ, 2008, 3rd edn.)
20 Eberhart, R.C., Kennedy: Swarm Intelligence (Morgan Kaufmann, San
Diego, CA, 2001)
21 Ear recognition laboratory at university of science and technology
Beijing, http: //www.ustb.edu.cn/resb/en/index.htm (visited on 20 May
2010)
22 Prakash, S., Jayaraman, U., Gupta, P.: Ear localization from side face
images using distance transform and template matching. Proc. IEEE
Int. Workshop on Image Processing Theory, Tools and Applications,
Sousse, Tunisia, November 2008, pp. 18
23 Kisku, D.R., Gupta, P., Mehrotra, H., Sing, J.K.: Multimodal belief
fusion for face and ear biometrics, Intell. Inf. Manage. Sci. Res.,
2009, 1, (3), pp. 166171
www.ietdl.org
106 IET Biom., 2013, Vol. 2, Iss. 3, pp. 97106
& The Institution of Engineering and Technology 2013 doi: 10.1049/iet-bmt.2012.0049