
Any of a number of rules for adjusting a raw simi-
larity score in a way that takes into account factors
such as the amount of data on which its calculation was
based, or the quality of the data. One purpose of score
normalization in biometrics is to prevent the arising of
false matches simply because only a few elements (e.g.,
biometric features) were available for comparison. So
an accidental match by chance would be more like
tossing a coin only a few times to produce a perfect
run of all head. Another purpose of score normalization
is to make it possible to compare or to fuse different
types of measurements, as in multibiometrics. For ex-
ample, Z-score normalization redefines every observa-
tion in units of standard deviation from the mean,
thereby allowing incommensurable scores (like height
and weight) to become commensurable (e.g., he is 3.2
standard deviations heavier than normal but 2.3 stan-
dard deviations taller than normal). Frequently the goal
of score normalization is to map samples from different
distributions into normalized samples from a universal
distribution. For example, in iris recognition a decision is
made only after the similarity score (fractional Hamming
Distance) has been converted into a normalized score
that compensates for the number of bits that were avail-
able for comparison, thereby preventing accidental False
Matches just because of a paucity of visible iris tissue.
▶ Score Normalization Rules in Iris Recognition
▶ Session Effects on Speaker Modeling
▶ Speaker Matching
Score Normalization Rules in Iris
Recognition
JOHN DAUGMAN
Cambridge University, Cambridge, UK
Synonyms
Commensurability; Decision criterion adjustment;
Error probability non-accumulation; Normalised
Hamming Distance
Definition
All biometric recognition systems are based on similarity
metrics that enable decisions of ‘‘same’’ or ‘‘different’’ to
be made. Such metrics require normalizations in order
to make them commensurable across comparison cases
that may differ greatly in the quantity of data available, or
in the quality of the data. Is a ‘‘perfect match’’ based only
on a small amount of data better or worse than a less
perfect match based on more data? Another need for
score normalization arises when interpreting the best
match found after an exhaustive search, in terms of the
size of the database searched. The likelihood of a good
match arising just by chance between unrelated templates
must increase with the size of the search database, simply
because there are more opportunities. How should a
given ‘‘best match ’’ score be interpr eted? Addressing
these questions on a principled basis requires models
of the underlying probability distributions that describe
the likelihood of a given degree of similarity arising by
chance from unrelated sources. Likewise, if comparisons
are required over an increasing range of image orienta-
tions because of uncertainty about image tilt, the proba-
bility of a good similarity score arising just by chance
from unrelated templates again grows automatically, be-
cause there are more opportunities. In all these respects,
biometric similarity
▶ score normalization is needed,
and it plays a critical role in the avoidance of False
Matches in the publicly deployed algorithms for iris
recognition.
Introduction
Biometric recognition of a person’s identity requires
converting the observed degree of similarity between
presenting and previously enrolled features into a deci-
sion of ‘‘same’’ or ‘‘different.’’ The previously enrolled
features may not be merely a single feature set obtained
from a single asserted identity, but may be a vast number
of such feature sets belonging to an entire national
population, when identification is performed by exhaus-
tively searching a database for a sufficiently good match.
The
▶ similarity metrics used for each comparison
between samples might be simple correlation statistics,
or vector projections, or listings of the features (like
fingerprint minutiae coordinates and directions) that
agreed and of those that disagreed as percentages of the
total set of features extracted. For each pair of feature
sets being compared, vary ing amounts of data may be
available, and the sets mig ht need to be compared
under various transformations such as image rotations
when the orientation is uncertain. An example is seen
Score Normalization Rules in Iris Recognition
S
1135
S