Академический Документы
Профессиональный Документы
Культура Документы
org/2/1/8/ 121
Albert J. Ahumada, Jr. NASA Ames Research Center, Moffett Field, CA, USA
For the linear discrimination of two stimuli in white Gaussian noise in the presence of internal noise, a method is described
for estimating linear classification weights from the sum of noise images segregated by stimulus and response. The
recommended method for combining the two response images for the same stimulus is to difference the average images.
Weights are derived for combining images over stimuli and observers. Methods for estimating the level of internal noise
are described with emphasis on the case of repeated presentations of the same noise sample. Simple tests for particular
hypotheses about the weights are shown based on observer agreement with a noiseless version of the hypothesis.
The first visual classification images were linear 2.2 The Linear Observer Model
combinations of four averaged noise images, one for each
of the four stimulus-response categories. For a given The linear observer classifying the noisy signals by
stimulus, the average of all the added noises has zero responding R = 0 or R = 1 or would use a vector w and
mean, so the sum of noises from one response class has respond R = 1 if and only if
an expectation equal to the negative of the expectation of w g T > β, (2.2.1)
the sum of the noises from the other response class, so we T
knew to combine the two response noise images with where β is a response criterion and indicates the matrix
opposite sign. It appeared in the initial images that the transpose operator, so that
error images were clearer than the correct response m
images, so we took the difference of the averages rather w gT = ∑ w ( i ) g (i )
than the sums, realizing that this was an arbitrary i =1
decision. We also arbitrarily combined the images from
the two stimuli with equal weight to get a single overall
Also, without lack of generality, we will assume that w has
image. By symmetry, this must be the right weighting to
unit length (w and β have already been divided by the
use if the observer is making the same number of errors
length of w) so that
to equal numbers of each kind of stimulus, which was
approximately the case. In the next section, there is an ||w|| = (w w T) 0.5 = 1. (2.2.2)
analysis showing that for a simplified theoretical
situation, it is possible to show that the averaging is nearly The performance of an observer is characterized by
optimal and to find expressions for good weighting the error rates
functions for the cases of unequal stimulus presentation p 0, 1 = Pr{R = 1 | s 0},
rates and unsymmetrical response biases. The beginning
of the section introduces notation for a standard signal the probability of signal s 0 being followed by response
detection experiment as analyzed by Green and Swets R = 1, and
(1966). p 1, 0 = Pr{R = 0 | s 1},
the probability of signal s 1 being followed by response
2 Template Estimation for Linear R = 0.
Classification of Two Signals in For the linear classifier with vector w and criterion β,
w n T is Gaussian with mean zero and unit variance.
Additive White Gaussian Noise Hence
p 0, 1 = Pr{w (s 0 + n) T > β } = 1− Φ( β – w s 0 T )
2.1 The Signals and Noise and
s 0 and s 1 are 1 by m signal vectors, presented with
probabilities p 0 and p 1 = (1 − p 0) for N trials. On each p 1, 0 = Pr{w (s 1 + n) T < β } = Φ( β – w s 1 T ), (2.2.3)
trial, a random noise sample vector n is added to the where Φ(·) is the cumulative standard Gaussian
signal, so the trial stimulus distribution function.
g = s0+ n If we define sensitivity and bias parameters
or d 0' = w (s 1 − s 0) T (2.2.4)
= ∫ z φ ( z ) dz Φ ( β0 )
β0
= − α φ (α β 0) / Φ (α β 0)
−∞
= − α φ (β H) / Φ (β H)
= φ (β 0)/ Φ (β 0) . (2.5.2.1)
where φ(z) is the standard normal density function and = − α φ (Z(p 0, 0))/p 0, 0 . (2.5.3.7)
the integration of z exp( z 2 /2) is enabled by the variable The effect of the criterion noise on v 0, 0 is to reduce it
substitution by the factor α.
t = −z 2/2. Similarly,
Similarly, v 0, 1 = E[n 0, 1] = E[n | n > β 0, H]
E[n 0, 1] = E[n | n > β 0] = E[α (x + γ y) | x > α β 0]
= ∫ z φ ( z ) dz
∞
0β
(1 − Φ (β0 ) ) = α E[ x | x > α β 0]
= α φ (α β 0) / (1 − Φ (α β 0))
= φ (β 0)/ (1−Φ (β 0)). (2.5.2.2)
= α φ (β H) / (1 − Φ (β H))
2.5.3 Single pixel, noisy criterion
The Gaussian criterion case can be reduced to the = α φ (Z(p 0, 0)) / (1−p 0, 0)
fixed criterion case by a change of variables. Let z be the
standard Gaussian used to form the criterion β 0, H, so = α φ (Z(p 0, 1)) / p 0, 1, (2.5.3.8)
that
because p 0, 1 = 1− p 0, 0 and
β 0, H = γ z + β 0 . (2.5.3.1)
Z(p) = −Z(1 − p). (2.5.3.9)
Then
If false alarms are less frequent than correct rejections
v 0, 0 = E[n 0, 0] = E[n | n < β 0, H] (p 0, 1 < p 0, 0 ), then
Again, if misses are less frequent than hits (p 1, 0<p 1,1), = (1−w(i)2) + w(i)2 Var[z 1 | s s , R ] , (2.5.4.6)
then
which is bounded by Var[z 1 | s s , R] and one. Truncation
|v 1, 0| > |v 1, 1|, (2.5.3.13) of a Gaussian can only decrease the variance, so
a larger absolute expected value on a miss than a hit trial. Var[n s, R (i)] < 1. (2.5.4.7)
Regardless of the signal, the expected value depends only
on the response proportion and the criterion variability. Because ||w|| = 1, if there are very many significant
weights in w, they will have to be small, so that
2.5.4 The multiple pixel case
Var[n s, R (i)] ~ 1. (2.5.4.8)
Another independent variable transformation allows
the single pixel case result to solve the multiple pixel case. Let a s, R be the average value of a number N s, R of n s, R.
Let us first examine the means and variances of the pixels Any combination of the form
of n 0, 0. For any vector w of unit length, it is possible to w est = k 0, 1 a 0, 1 – k 0, 0 a 0, 0
construct an orthonormal transformation U whose first
row is w, that is
+ k 1, 1 a 1, 1 – k 1, 0 a 1, 0 , (2.5.4.9)
T T T T
U = (w w 2 … w m ) , (2.5.4.1) with positive weights k s,R will be an estimate of w times a
such that positive constant.
U T U = I, (2.5.4.2) 2.5.5 Combining the categorized noises
where I is the identity transformation (the transpose of U If we have two independent estimates b and c of the
is its inverse). same quantity (having the same expected value,
When this transformation is applied to n T, we get a E(b)=E(c)) with variances σ b2 and σ c2 , the linear
new noise U n T whose distribution is the same as that of combination of the two estimates with the same expected
n T, but whose first pixel is w n T. On an s 0 trial, a noise value and the smallest variance is
vector n will be classified as n 0, 0 if and only if the first (b / σ b2 + c /σ c2 ) / (1 / σ b2 + 1 / σ c2 ). (2.5.5.1)
pixel of U n T,
That is, each estimate should be weighted inversely by its
z 1 = w n T < β 0, H. (2.5.4.3) variance.
The rest of the pixels (z 2, ..., z m) of U n T are To obtain a minimum variance estimate of w(i) from
independent standard Gaussian variables (with mean zero a sample with N s, R approximately independent samples
and variance 1). of each type, the individual estimates of w(i),
n s, R (i)/ v s, R should be weighted inversely by their
E[n 0, 0 T] = E[ n T | w n T < β 0, H] variances, which are approximately
Var[n s, R (i)/v s, R] = 1/v s, R2, (2.5.5.2)
= E[U T U n T | w n T < β 0, H]
so we should weight each n s, R (i) by v s, R . Because the
= E[U T (z 1, z 2, ... , z m )T | z 1 < β 0, H] weights do not depend on i, we can then just weight n s, R
by v s, R . A good un-normalized estimate of the classifier
= U T E[(z 1, z 2, ... , z m )T | z 1 < β 0, H] w is thus given by the v s, R weighted sums N s, R a s, R ,
w G = v 0, 1 N 0, 1 a 0, 1 + v 0, 0 N 0, 0 a 0, 0
= U T (v 0, 0, 0, ... , 0 )T
+ v 1, 1 N 1, 1 a 1, 1 + v 1, 0 N 1, 0 a 1, 0 . (2.5.5.3)
= v 0, 0 w T. (2.5.4.4)
If we replace p s, R in v s, R of Equations (2.5.3.7, 8, 11, 12)
A similar argument for the other cases leads to the by p s, R = N s, R / N s, where N s = N s, 0 + N s, 1 , and take
general result that advantage of the fact that
E[n s, R ] = v s, R w . (2.5.4.5) φ (Z(p)) = φ (Z(1−p)), (2.5.5.4)
The mean of a classified noise is proportional to the we obtain
classifying vector w. The variance of individual elements
of n s, R, w G = α [ N 0 φ(Z( p 0, 1)) (a 0, 1 − a 0, 0 )
) ∑ U ( i, j )2 Var ( z j ) + w ( i )2 Var ( z 1 ss , R )
m
(
Var n s , R ( i ) = + N 1 φ (Z( p 1, 0)) (a 1, 1 − a 1, 0) ]. (2.5.5.5)
j =2
The more frequent stimulus should be given more
weight and, because for p < .5, φ(Z(p)) increases
Ahumada 127
monotonically, the more error-prone stimulus should be To obtain this probability, we will compute it conditional
given more weight. These weights take into account all on the value of e and then average over the possible
the parameters assumed to determine the observer's values of e.
performance. If both stimuli are equally frequent and the Conditional on the value of e, for s = 0, the
error rates are equal, the formula is proportional to probability of a response R = 0 is given by
w ave = a 0, 1 – a 0, 0 + a 1, 1 – a 1, 0 , (2.5.5.6) Pr{R i = 0 | s 0, i , e } = Pr{ e < β 0, H | e}
the combination rule originally used by Ahumada and
Beard (Ahumada, 1996; Ahumada & Beard, 1998, 1999; = Pr{ e < γ z 0 + β 0 | e}
Beard & Ahumada, 1997, 1998, 2000).
In the next section, we add a third subscript to refer = Pr{ (e − β 0) / γ < z 0 | e}
to the observer. The good weighting scheme for
combining average classification images a s, R, O over = Φ ((β 0 − e)/ γ). (3.1.2)
responses, stimuli, and observers can be described as a For another response to the same signal and the same
sequential process. To combine over responses, just take noise, the criterion variability is independent, so the
the difference, probability of two R = 0 responses is given by
w G, s, O = a s, 1, O − a s, 0, O. (2.5.5.7) Pr{R i = R j = 0 | s 0, i , s 0, j , e } = Pr{ e < β 0, H | e}2
To then combine images for different stimuli, weight
by factors involving the relative frequencies of the stimuli = Φ((β 0 − e)/ γ) 2
and the extremeness of the error proportions for the
stimuli = Φ((β H / α −e) / γ) 2
w G, O = k 1, O w G, 1, O + k 0, O w G, 0, O (2.5.5.8)
= Φ((β H (1 + γ2 ) 0.5 − e) / γ) 2. (3.1.3)
where
2
The probability of two correct responses R i = R j = 0 to
k s, O = N s, O exp( − Z( p s, r, O) /2) . (2.5.5.9) the same noise is then
If estimates are to be combined over M observers, they
( )
2
2
need to be weighted by the square root of the observer's ∞ βH 1 + γ − e
proportion of decision variance due to the external noise, p A,0,0 = ∫ Φ φ ( e ) de , (3.1.4)
α2 = 1/(1 + γ2), and the number of trials run by the
−∞
γ
observer (which is already included here in the N s, O).
Estimates of the internal noise (α or γ) are needed in The equations for arbitrary s can be written
order to use the above formula (2.5.5.10) to combine
( )
2
estimates over observers. Internal noise estimates are also
needed to compute the variance of an estimate of w to ∞ ( 2
)
Z p s ,0 1 + γ − e
p A, s ,0 = ∫ Φ φ ( e ) de , (3.1.6)
plan the number of trials that need to be run (see −∞ γ
Equation 2.5.5.2). For this section, we are using the
model of section 2.4, relaxing the linearity assumptions
about the classification function and the external noise to and
( )
2
the assumption that the term
e = w n Tis a standard Gaussian. ∞ ( 2
)
Z p s ,0 1 + γ − e
The subscripts i and j are added to indicate two p A, s ,1 = ∫ 1 − Φ φ ( e ) de . (3.1.7)
−∞
γ
separate trials. The response agreement probability for a
particular signal and response with the same noise is
denoted by Either of these equations can be used to solve for an
p A, s, R = Pr{R i = R j = R | s s, i , s s, j , n i = n j }. (3.1.1) estimate of γ (the same estimate results by using either
one).
Ahumada 128
3.2 Estimating Observer Noise Using (3.2.1). However, responses from any other template of
the same efficiency would correlate worse with the
a Model Classifier observer responses than the correct template, leading to a
Ahumada and Beard (1998) also derived other larger estimate of internal noise. Thus only the correct
estimates for γ based on the assumption that e comes template can lead to agreement among all three noise
from a known model (and is distributed as a Gaussian) estimates.
and that Gaussian internal noise is added by the observer.
This allows falsification of the model if the estimates of γ 3.3 Example of Model Testing Based
are not consistent. The model they tested was a particular on Internal Noise Estimates
parametric linear filter estimated by Barth, Beard, and
This section illustrates model testing based on noise
Ahumada (1999), but any model can be tested using their
estimates using the data of observer C.S. from Ahumada
scheme if the performance of the noiseless model can be
and Beard (1999). The observer was detecting the
computed for the same noises presented to the observer.
presence or absence of a Gabor signal (2 cpd or 16 cpd) in
The noiseless model's performance index, d 0', is needed
noise. Figure 2 shows that the classification images for 2
as is the trial by trial agreement of the observer and the
cpd signal present and signal absent and for 16 cpd signal
model.
present resembled the signal, but no image emerged for
One estimate of γ is based on the ratio of the
the 16 cpd signal absent, a result consistent with position
performance of the model, d 0', to that of the observer, dH'
or phase uncertainty.
From (2.4.6) above we have
d H' = α d 0' ,
so
α = d H' / d 0' . (3.2.1)
and
γ = ((d 0' / d H')) 2 − 1) 0.5 . (3.2.2)
Another γ estimate comes from the trial by trial
agreement between the observer and the model. For a
given value of e, the event M 0 that the model will
respond that the signal was s 0 when it actually was s s will
occur if e is less than the criterion for the model,
β M, s = Z(pM, s, 0), (3.2.3)
which is β 0 if s = 0 or β 0 − d 0' if s = 1. The probability
that the observer and the model agree on this response
will be
P {R = 0, M 0 } = ∫
βm,s
( 2
)
βH 1 + γ − e
Φ φ ( e ) de . (3.2.4)
−∞
γ
Note that the other possibilities can be computed
from this and simpler ones from the relationships
Pr{R 0, M 0}+ Pr{R 0, M 1} = Pr{R 0} , (3.2.5)
Pr{R 0, M 0}+ Pr{R 1, M 0} = Pr{M 0} , (3.2.6)
and
Pr{R 1, M 1}+ Pr{R 0, M 1} = Pr{M 1} . (3.2.7)
The question arises as to which alternative models
could or could not be rejected by this test? One reader Figure 2. Top. Gabor signals with spatial frequencies of 2 cpd
hypothesized that any template with the same sampling (left) and 16 cpd (right). Center. Classification images for the
efficiency (relative to the ideal observer) could not be signal trials with the 2-cpd target (left) and the 16-cpd target
rejected. Within the current framework, this is true for trials (right). Bottom. Classification images for the no-signal
tests comparing the noise estimate based on observer self- trials with the 2-cpd target (left) and the 16-cpd target (right).
agreement of Section 3.1 with the estimate based on the The observer was C.S., who ran 5,900 trials at 2 cpd and
ratio of observer performance to model performance 8,000 trials at 16 cpd (Ahumada & Beard, 1999).
Ahumada 129
Abbey, C. K., & Eckstein, M. P. (2001b). Theory for Beard, B. L., & Ahumada, A. J., Jr. (1997). Relevant
estimating human-observer templates in two- image features for vernier acuity [Abstract].
alternative forced-choice experiments. In M. F. Perception, 26, 38.
Insana & R. Leahy (Eds.), Proceedings of the 17th Beard, B. L., & Ahumada, A. J., Jr. (1998). Technique to
International Conference on Inferential Processes in extract relevant image features for visual tasks.
Medical Imaging. Berlin: Springer-Verlag. Proceedings of SPIE, 3299, 79-85. [Link]
Abbey, C. K. & Eckstein, M. P. (2002). Classification Beard, B. L., & Ahumada, A. J., Jr. (1999). Detection in
image analysis: Estimation and statistical inference fixed and random noise in foveal and parafoveal
for two-alternative forced-choice experiments. Journal vision explained by template learning. Journal of the
of Vision, 2(1), 66-78, Optical Society of America A, 16, 755-763. [PubMed]
http://journalofvision.org/2/1/5/, DOI
10.1167/2.1.5. [Link] Beard, B. L., & Ahumada, A. J. (2000). Response
classification images for parafoveal vernier acuity
Abbey, C. K., Eckstein, M. P., & Bochud, F. O. (1999). [Abstract]. Investigative Ophthalmology and Visual
Estimation of human-observer templates for 2 Science, 41(Suppl. 4), S804. [Link]
alternative forced choice tasks. Proceedings of SPIE,
3663, 284-295. Burgess, A. E., & Colborne, B. (1988). Visual signal
detection. IV. Observer inconsistency. Journal of the
Ahumada, A. J., Jr. (1967). Detection of tones masked by Optical Society of America A, 5, 617-628. [PubMed]
noise: A comparison of human observers with digital-
Findlay, J. M. (1973). Feature detectors and vernier
computer-simulated energy detectors of varying bandwidths.
Unpublished doctoral dissertation, University of acuity. Nature, 241, 135-137. [PubMed]
California, Los Angeles. Available as Technical Foley, J. M. (1994). Human luminance pattern-vision
Report No. 29, Human Communications mechanisms: Masking experiments require a new
Laboratory, Department of Psychology, University of model. Journal of the Optical Society of America A, 11,
California, Los Angeles. 1710-1719. [PubMed]
Ahumada, A. J., Jr. (1996). Perceptual classification Green, D. M., & Swets, J. A. (1966). Signal detection
images from vernier acuity masked by noise theory. New York: Wiley.
[Abstract]. Perception, 26(Suppl. 1), 18. [Link] Murray, R. F., Bennett, P. J., & Sekuler, A. B. (2002).
Ahumada, A. J., Jr., & Beard, B. L. (1998). Response Optimal methods for calculating classification
classification images in vernier acuity [Abstract]. images: Weighted sums. Journal of Vision, 2(1), 79-
Investigative Ophthalmology and Visual Science, 104, http://journalofvision.org/2/1/6/, DOI
39(Suppl. 4), S1109. [Link] 10.1167/2.1.6 [Link]
Ahumada, A. J., Jr., & Beard, B. L. (1999). Classification Richards, V. M., & Zhu, S. (1994). Relative estimates of
images for detection [Abstract]. Investigative combination weights, decision criteria, and internal
Ophthalmology and Visual Science, 40(Suppl. 4), S572. noise based on correlation coefficients. Journal of the
[Link] Acoustical Society of America, 95, 423-434. [PubMed]
Ahumada, A. J., Jr., & Lovell, J. (1971). Stimulus features Solomon, J. A. (2002). Noise reveals visual mechanisms of
in signal detection. Journal of the Acoustical Society of detection and discrimination. Journal of Vision, 2(1),
America, 49, 1751-1756. 105-120, http://journalofvision.org/2/1/7/, DOI
Ahumada, A. J., Jr., Marken, R., & Sandusky, A. (1975). 10.1167/2.1.7. [Link]
Time and frequency analyses of auditory signal Waugh, S. J., Levi, D. M., & Carney, T. (1993).
detection. Journal of the Acoustical Society of America, Orientation, masking, and vernier acuity for line
57, 385-390. [PubMed] targets. Vision Research, 33, 1619-1638. [PubMed]
Barth, E., Beard, B. L., & Ahumada, A. J., Jr. (1999).
Nonlinear features in vernier acuity. Proceedings of
SPIE, 3644, 88-96. [Link]