Attention and Performance
Attention and Performance
5 Attention and
performance
INTRODUCTION
Attention is invaluable in everyday life. We use attention to avoid being
hit by cars when crossing the road, to search for missing objects and to
perform two tasks together. The word “attention” has various meanings
but typically refers to selectivity of processing as emphasised by William
James (1890, pp. 403–404):
There is a final important distinction (the last one, I promise you!) KEY TERMS
between external and internal attention. External attention is “the selec-
Cocktail party problem
tion and modulation of sensory information” (Chun et al., 2011). In con- The difficulties involved
trast, internal attention is “the selection, modulation, and maintenance of in attending to one voice
internally generated information, such as task rules, responses, long-term when two or more people
memory, or working memory” (Chun et al., 2011, p. 73). The connec- are speaking at the same
time.
tion to Baddeley’s working memory model is especially important (e.g.,
Baddeley (2012); see Chapter 6). The central executive component of Dichotic listening task
working memory is involved in attentional control and is crucially involved A different auditory
message is presented to
in internal and external attention.
each ear and attention
Much attentional research has two limitations. First, the emphasis is has to be directed to one
on attention to externally presented task stimuli rather than internally gen- message.
erated stimuli (e.g., worries; self-reflection). One reason is that it is easier to
Shadowing
assess and to control external attention. Second, what participants attend Repeating one auditory
to is determined by the experimenter’s instructions. In contrast, what we message word for word
attend to in the real world is mostly determined by our current goals and as it is presented while a
emotional states. second auditory message
is also presented; it is
Two important topics related to attention are discussed elsewhere.
used on the dichotic
Change blindness (see Glossary), which shows the close links between atten- listening task.
tion and perception, is considered in Chapter 4. Consciousness (including
its relationship to attention) is discussed in Chapter 16.
Treisman and Riley (1969) asked listeners to shadow one of two audi- Figure 5.1
tory messages. They stopped shadowing and tapped when they detected A comparison of
a target in either message. Many more target words were detected on the Broadbent’s theory (top),
Treisman’s theory (middle),
shadowed message.
and Deutsch and Deutsch’s
Aydelott et al. (2015) asked listeners to perform a task on attended theory (bottom).
target words. When unattended words related in meaning were presented
shortly before the target words themselves, performance on the target
words was enhanced when unattended words were presented as loudly as
attended ones. Thus, the meaning of unattended words was processed.
There is often more processing of unattended words that have a special
significance for the listener. For example, Li et al. (2011) obtained evidence
that unattended weight-related words (e.g., fat; chunky) were processed
more thoroughly by women dissatisfied with their weight. Conway et al.
(2001) found listeners often detected their own name on the unattended
message. This was especially the case if they had low working memory
capacity (see Glossary) indicative of poor attentional control.
Coch et al. (2005) asked listeners to attend to one of two auditory
inputs and to detect targets presented on either input. Event-related poten-
tials (ERPs; see Glossary) provided a measure of processing activity. ERPs
100 ms after target presentation were greater when the target was presented
on the attended rather than the unattended message. This suggests there
was more processing of the attended than unattended targets.
Greater brain activation for attended than unattended auditory stimuli
may reflect enhanced processing for attended stimuli and/or suppressed
processing for unattended stimuli. Horton et al. (2013) addressed this
182 Visual perception and attention
issue. Listeners heard separate speech messages presented to each ear with
instructions to attend to the left or right ear. There was greater brain acti-
vation associated with the attended message (especially around 90 ms after
stimulus presentation). This difference depended on enhancement of the
attended message combined with suppression of the unattended message.
Classic theories of selective auditory attention (those of Broadbent,
Treisman, and Deutsch and Deutsch) de-emphasised the importance of
suppression or inhibition of the unattended message shown by Horton
et al. (2013). For example, Schwartz and David (2018) reported suppres-
sion of neuronal responses in the primary auditory cortex to distractor
sounds. More generally, all the classic theories de-emphasise the flexibility
of selective auditory attention and the role of top-down processes in selec-
tion (see below).
pathways from the auditory cortex to brain areas involved in early auditory
processing (Robinson & McAlpine, 2009). Various top-down factors based
on listeners’ knowledge and/or expectations are involved. For example, lis-
teners are more accurate at identifying what one speaker is saying in the
context of several other voices if they have previously heard that speaker’s
voice in isolation (McDermott, 2009).
Woods and McDermott (2018) investigated top-down processes in
selective auditory attention in more detail. They argued, “Sounds pro-
duced by a given source often exhibit consistencies in structure that might
be useful in separating sources” (p. E3313). They used the term “schemas”
to refer to such structural consistencies.
Listeners showed clear evidence of schema learning leading to rapid
improvements in their listening performance. An important aspect of such
learning is temporal coherence – a given source’s sound features are typ-
ically all present when it is active and absent when it is silent. Shamma
et al. (2011) discussed research showing that if listeners can identify one
distinctive feature of the target voice, they can then distinguish its other
sound features via temporal coherence.
Evans et al. (2016) compared patterns of brain activity when attended
speech was presented on its own or together with competing unattended
speech. Brain areas associated with attentional and control processes (e.g.,
frontal and parietal regions) were more activated in the latter condition.
Thus, top-down processes relating to attention and control are important
in selective auditory processing.
Finally, Golumbic et al. (2013) suggested individuals at actual cocktail
parties can potentially use visual information to assist them in understand-
ing what a given speaker is saying. Listeners heard two simultaneous mes-
sages (one in a male voice and the other in a female voice). Processing of
the attended message was enhanced when they saw a video of the speaker
talking.
In sum, listeners generally achieve the complex task of selecting one
speech message from among several such messages. There has been pro-
gress in identifying the top-down processes involved. For example, if lis-
teners can identify at least one consistently distinctive feature of the target
voice, this makes it easier for them to attend only to that voice. Top-down
processes often produce a “winner-takes-all” situation where the processing
of one auditory input (the winner) suppresses the brain activity associated
with all other inputs (Kurt et al., 2008).
Findings
Support for the zoom-lens model was reported by Müller et al. (2003). On
each trial, observers saw four squares in a semi-circle and were cued to
attend to one, two or all four. Four objects were then presented (one in
each square) and observers decided whether a target (e.g., a white circle)
was among them. Brain activation in early visual areas was most wide-
spread when the attended region was large (i.e., attend to all four squares)
and was most limited when it was small (i.e., attend to one square). As pre-
dicted by the zoom-lens theory, performance (reaction times and errors)
was best with the smallest attended region and worst with the largest one.
Chen and Cave (2016, p. 1822) argued the optimal attentional zoom
setting “includes all possible target locations and excludes possible distrac-
tor locations”. Most findings indicated people’s attentional zoom setting
is close to optimal. However, Collegio et al. (2019) obtained contrary
findings. Drawings of large objects (e.g., jukebox) and small objects (e.g.,
watch) were presented so their retinal size was the same. The observer’s
area of focal attention was greater with large objects because they made
top-down inferences concerning their real-world sizes. As a result, the area
of focal attention was larger than optimal for large objects.
Goodhew et al. (2016) pointed out that nearly all research has focused
only on spatial perception (e.g., identification of a specific object). They
focused on temporal perception (was a disc presented continuously or
were there two presentations separated by a brief interval?). Spotlight size
had no effect on temporal acuity, which is inconsistent with the theory.
How can we explain these findings? Spatial resolution is poor in peripheral
vision but temporal resolution is good. As a consequence, a small atten-
tional spotlight is more beneficial for spatial than temporal acuity.
We turn now to split attention. Suppose you had to identify two digits
that would probably be presented to two cued locations a little way apart
Figure 5.2
(a) Shaded areas indicate
the cued locations; the
near and far locations are
not cued. (b) Probability of
target detection at valid
(left or right) and invalid
(near or far) locations.
Based on information in Awh
and Pashler (2000).
(see Figure 5.2a). Suppose also that on some trials a digit was presented
between the two cued locations. According to zoom-lens theory, the area
of maximal attention should include the two cued locations and the space
in between. As a result, the detection of digits presented in the middle
should have been very good. In fact, Awh and Pashler (2000) found it was
poor (see Figure 5.2b). Thus, attention can resemble multiple spotlights, as
predicted by the split-attention approach.
Morawetz et al. (2007) presented letters and digits at five locations
simultaneously (one in each quadrant of the visual field and one in the
centre). In one condition, observers attended to the visual stimuli at the
upper left and bottom right locations and ignored the other stimuli. There
were two peaks of brain activation corresponding to the attended areas but
less activation corresponding to the region in between. Overall, the pattern
of activation strongly suggested split attention.
Niebergall et al. (2011) recorded the neuronal responses of monkeys
attending to two moving stimuli while ignoring a distractor. In the key
condition, there was a distractor between (and close to) the two attended
stimuli. In this condition, neuronal responses to the distractor decreased
compared to other conditions. Thus, split attention involves a mecha-
nism reducing attention to (and processing of) distractors located between
attended stimuli.
186 Visual perception and attention
KEY TERM In most research demonstrating split attention, the two non-adjacent
stimuli being attended simultaneously were each presented to a different
Hemifield
One half of the visual hemifield (one half of the visual field). Note that the right hemisphere
field. Information from receives visual signals from the left hemifield and the left hemisphere
the left hemifield of each receives signals from the right hemifield. Walter et al. (2016) found per-
eye proceeds to the formance was better when non-adjacent stimuli were presented to different
right hemisphere and
hemifields rather than the same hemifield. Of most importance, the assess-
information from the right
hemifield proceeds to the ment of brain activity indicated effective filtering or inhibition of stimuli
left hemisphere. presented between the two attended stimuli only when presented to differ-
ent hemifields.
In sum, we can use visual attention very flexibly. Visual selective atten-
tion can resemble a spotlight, a zoom lens or multiple spotlights, depending
on the current situation and the observer’s goals. However, split attention
may require that two stimuli are presented to different hemifields rather
than the same one. A limitation with all these theories is that metaphors
(e.g., attention is a zoom lens) are used to describe experimental findings but
these metaphors fail to specify the underlying mechanisms (Di Lollo, 2018).
What is selected?
Why might selective attention resemble a spotlight or zoom lens? Perhaps
we selectively attend to an area or region of space: space-based attention.
Alternatively, we may attend to a given object or objects: object-based
attention. Object-based attention is prevalent in everyday life because visual
attention is mainly concerned with objects of interest to us (see Chapters 2
and 3). As expected, observers’ eye movements as they view natural scenes
are directed almost exclusively to objects (Henderson & Hollingworth,
1999). However, even though we typically focus on objects of potential
importance, our attentional system is so flexible we can attend to an area of
space or a given object.
There is also feature-based attention. For example, suppose you are
looking for a friend in a crowd. Since she nearly always wears red clothes,
you might attend to the feature of colour rather than specific objects or
locations. Leonard et al. (2015) asked observers to identify a red letter
within a series of rapidly presented letters. Performance was impaired when
a # symbol also coloured red was presented very shortly before the target.
Thus, there was evidence for feature-based attention (e.g., colour; motion).
Findings
Visual attention is often object-based. For example, O’Craven et al. (1999)
presented observers with two stimuli (a face and a house), transparently
overlapping at the same location, with instructions to attend to one of
them. Brain areas associated with face processing were more activated
when the face was attended to than when the house was. Similarly, brain
areas associated with house processing were activated when the house was
the focus of attention.
Egly et al. (1994) devised a much-used method for comparing object-
based and space-based attention (see Figure 5.3). The task was to select
a target stimulus as rapidly as possible. A cue presented before the target
Attention and performance 187
Figure 5.4
(a) Possible target locations
(same object far, same
object near, valid, different
object far) for a given cue.
(b) Performance accuracy at
the various target locations.
From Hollingworth et al. (2012).
© 2011 American Psychological
Association.
more to locations or objects. The evidence is mixed (see Chen, 2012). List
and Robertson (2007) used Egly et al.’s (1994) task shown in Figure 5.4
and found location- or space-based inhibition of return was much stronger
than object-based inhibition of return.
Theeuwes et al. (2014) found location- and object-based inhibition of
return were both present at the same time. According to Theeuwes et al.
(p. 2254), “If you direct your attention to a location in space, you will
automatically direct attention to any object . . . present at that location,
and vice versa.”
There is considerable evidence of feature-based attention (see Bartsch
et al., 2018, for a review). In their own research, Bartsch et al. addressed
Attention and performance 189
Evaluation
Research on whether visual attention is object- or location-based have
produced variable findings and so few definitive conclusions are pos-
sible. However, the relative importance of object-based and space- or
location-based attention is flexible. For example, individual differences are
important (Pilz et al., 2012). Note that visual attention can be both object-
based and space-based at the same time.
What are the limitations of research in this area? First, most research
apparently demonstrating that object-based attention is more important
than space- or location-based attention has involved the use of spatial
cues. Recent evidence (Donovan et al., 2017) suggests such cues may bias
visual attention and that visual attention is not initially object-based in
their absence.
Second, space-, object- and feature-based forms of attention often inter-
act with each other to enhance object processing (Kravitz & Behrmann,
2011). However, we have as yet limited theoretical understanding of the
mechanisms involved in such interactions.
Third, there is a need for more research assessing patterns of attention
under naturalistic conditions. In recent research where observers view artifi-
cial stimuli while performing a specific task, it is unclear whether attentional
processes resemble those when they engage in free viewing of natural scenes.
Load theory
Lavie’s (2005, 2010) load theory has been an influential approach to
understanding distraction effects. It distinguishes between perceptual
Findings
There is much support for the hypothesis that high perceptual load reduces
distraction effects. Forster and Lavie (2008) presented six letters in a circle
and participants decided which target letter (X or N) was present. The five
non-target letters resembled the target letter more closely in the high-load
condition. On some trials a picture of a cartoon character (e.g., Spongebob
Squarepants) was presented as a distractor outside the circle. Distractors
interfered with task performance only under low-load conditions.
According to the theory, brain activation associated with distractors
should be less when individuals are performing a task involving high per-
ceptual load. This finding has been obtained with visual tasks and distrac-
tors (e.g., Schwartz et al., 2005) and also with auditory tasks and distractors
(e.g., Sabri et al., 2013).
Why is low perceptual load associated with high distractibility? Biggs
and Gibson (2018) argued this happens because observers generally adopt
a broad attentional focus when perceptual load is low. They tested this
hypothesis using three low-load conditions in which participants decided
whether a target X or N was presented and a distractor letter was some-
times presented (see Figure 5.5). They argued that observers would adopt
the smallest attentional focus in the circle condition and the largest atten-
tional focus in the solo condition. As predicted, distractor interference was
greatest in the solo condition and least in the circle condition. Thus, dis-
traction effects depend strongly on size of attentional focus as well as per-
ceptual load.
The hypothesis that distraction effects should be greater when cognitive
or working memory load is high rather than low was tested by Burnham
et al. (2014). As predicted, distraction effects on a visual search task were
Figure 5.5
Sample displays for three
low perceptual load X X X
* * * *
conditions in which the T T T
task required deciding * * * *
whether a target X or N * *
was presented. See text for
further details.
From Biggs and Gibson (2018).
Attention and performance 191
Evaluation
The distinction between perceptual and cognitive load has proved useful
in predicting when distraction effects will be small or large. More specifi-
cally, the prediction that high perceptual load is associated with reduced
distraction effects has received much empirical support. In applied research,
load theory successfully predicts several aspects of drivers’ attention and
behaviour (Murphy & Greene, 2017). For example, drivers exposed to high
perceptual load responded more slowly to hazards and drove less safely.
What are the theory’s limitations? First, the terms “perceptual load”
and “cognitive load” are vague, making it hard to test the theory (Murphy
et al., 2016). Second, the assumption that perceptual and cognitive load
have separate effects on attention is incorrect (Linnell & Caparos, 2011).
Third, perceptual load and attentional breadth are often confounded.
Fourth, the prediction that high cognitive load is associated with high
distractibility has been disproved when task and distracting stimuli are
easily distinguishable. Fifth, the theory de-emphasises several relevant
factors including the salience or conspicuousness of distracting stimuli and
the spatial distance between distracting and task stimuli (Murphy et al.,
2016).
KEY TERM One attention network is goal-directed or endogenous whereas the other is
stimulus-driven or exogenous.
Covert attention
Attention to an object
in the absence of an eye
Posner’s (1980) approach
movement towards it.
Posner (1980) studied covert attention in which attention shifts to a given
spatial location without an accompanying eye movement. In his research,
people responded rapidly to a light. The light was preceded by a central
cue (arrow pointing to the left or right) or a peripheral cue (brief illumina-
tion of a box outline). Most cues were valid (i.e., indicating where the target
light would appear) but some were invalid (i.e., providing inaccurate infor-
mation about the light’s location).
Responses to the light were fastest to valid cues, intermediate to neutral
cues (a central cross) and slowest to invalid cues. The findings were com-
parable for central and peripheral cues. When the cues were valid on only
a small fraction of trials, they were ignored when they were central cues.
However, they influenced performance when they were peripheral cues.
The above findings led Posner (1980) to distinguish between two atten-
tion systems:
Figure 5.6
The brain areas associated
with the dorsal or goal-
directed attention network
and the ventral or stimulus-
driven network. The full
names of the areas involved
are indicated in the text.
From Corbetta and Shulman
(2011). © Annual Reviews. With
permission of Annual Reviews.
(IPS), inferior frontal junction (IFJ), frontal eye field (FEF), middle tem-
poral area (MT) and V3A (a visual area). Key areas within the ventral
attention network are as follows: inferior frontal junction (IFJ), inferior
frontal gyrus (IFG), supramarginal gyrus (SMG), superior temporal gyrus
(STG) and insula (Ins). The temporo-parietal junction also forms part of
the ventral attention network.
The existence of two attention networks makes much sense. The
goal-directed system (dorsal attention network) allows us to attend to
stimuli directly relevant to our current goals. If we only had this system,
however, our attentional processes would be dangerously inflexible. It is
also important to have a stimulus-driven attentional system (ventral atten-
tion network) leading us to switch attention away from goal-relevant stimuli
to unexpected threatening stimuli (e.g., a ferocious animal). More generally,
the two attention networks typically interact effectively with each other.
Findings
Corbetta and Shulman (2002) supported their two-network model by car-
rying out meta-analyses of brain-imaging studies. In essence, they argued,
brain areas most often activated when participants expect a stimulus that
has not yet been presented form the dorsal attention network. In contrast,
brain areas most often activated when individuals detect low-frequency
targets form the ventral attention network.
Hahn et al. (2006) tested Corbetta and Shulman’s (2002) theory by
comparing patterns of brain activation when top-down and bottom-up
processes were required. As predicted, there was little overlap between the
brain areas associated with top-down and bottom-up processing. In addi-
tion, the brain areas involved in each type of processing corresponded rea-
sonably well to those identified by Corbetta and Shulman.
Chica et al. (2013) reviewed research on the two attention systems and
identified 15 differences between them. For example, stimulus-driven atten-
tion is faster than top-down attention and is more object-based. In addi-
tion, it is more resistant to interference from other peripheral cues once
activated. The existence of so many differences strengthens the argument
the two attentional systems are separate.
Considerable research evidence (mostly involving neuroimaging) indi-
cates the dorsal and ventral attention systems are associated with distinct
194 Visual perception and attention
neural circuits even during the resting state (Vossel et al., 2014). However,
neuroimaging studies cannot establish that any given brain area is necessar-
ily involved in stimulus-driven or goal-directed attention processes. Chica
et al. (2011) provided relevant evidence by using transcranial magnetic
stimulation (TMS; see Glossary) to interfere with processing in a given
brain area. TMS applied to the right temporo-parietal junction impaired
the functioning of the stimulus-driven system but not the top-down one.
In the same study, Chica et al. (2011) found TMS applied to the right
intraparietal sulcus impaired the functioning of both attention systems.
This provides evidence of the two attention systems working together.
Evidence from brain-damaged patients (discussed below, see pp. 196–200)
is also relevant to establishing the brain areas necessarily involved in goal-
directed or stimulus-driven attentional processes. Shomstein et al. (2010)
had brain-damaged patients complete two tasks, one requiring stimulus-
driven attentional processes whereas the other required top-down processes.
Patients having greater problems with top-down attentional processing typ-
ically had brain damage to the superior parietal lobule (part of the dorsal
attention network). In contrast, patients having greater problems with
stimulus-driven attentional processing typically had brain damage to the
temporo-parietal junction (part of the ventral attention network).
Wen et al. (2012) investigated interactions between the two visual atten-
tion systems. They assessed brain activation while participants responded to
target stimuli in one visual field while ignoring all stimuli in the unattended
visual field. There were two main findings. First, stronger causal influences
of the top-down system on the stimulus-driven system led to superior per-
formance on the task. This finding suggests the appearance of an object
at the attended location caused the top-down attention system to suppress
activity within the stimulus-driven system.
Second, stronger causal influences of the stimulus-driven system
on the top-down system were associated with impaired task performance.
This finding suggests activation within the stimulus-driven system pro-
duced by stimuli not in attentional focus disrupted the attentional set
maintained by the top-down system.
Recent developments
Corbetta and Shulman’s (2002) theoretical approach has been developed
in recent years. Here we briefly consider three such developments. First, we
now have a greater understanding of interactions between their two atten-
tion networks. Meyer et al. (2018) found stimulus-driven and goal-directed
attention both activated frontal and parietal regions within the dorsal
attention network, suggesting it has a pivotal role in integrating bottom-up
and top-down processing.
Second, previous research reviewed by Corbetta and Shulman (2002)
indicated the dorsal attention network is active immediately prior to the
presentation of an anticipated visual stimulus. However, this research did
not indicate how long this attention network remained active. Meehan
et al. (2017) addressed this issue and discovered that top-down influences
associated within the dorsal attention network persisted over a relatively
long time period.
Third, brain networks relevant to attention additional to those within KEY TERM
Corbetta and Shulman’s (2002) theory have been identified (Sylvester
Default mode network
et al., 2012). One such network is the cingulo-opercular network including A network of brain
the anterior insula/operculum and dorsal anterior cingulate cortex (dACC; regions that is active
see Figure 5.7). This network is associated with non-selective attention or “by default” when an
alertness (Coste & Kleinschmidt, 2016). individual is not involved
in a current task; it is
Another additional network is the default mode network including
associated with internal
the posterior cingulate cortex (PCC), the lateral parietal cortex (LP), the processes including mind-
inferior temporal cortex (IT), the medial prefrontal cortex (MPF) and the wandering, remembering
subgenual anterior cingulate cortex (sgACC). The default mode network the past and imagining
is activated during internally focused cognitive processes (e.g., mind- the future.
wandering; imagining the future). What is the relevance of this network
to attention? In essence, performance on tasks requiring externally focused
attention is often enhanced if the default mode network is deactivated
(Amer et al., 2016a).
Finally, there is the fronto-parietal network (Dosenbach et al., 2008),
which includes the anterior dorsolateral prefrontal cortex (aDLPFC), the
(a) aDLPFC
IPS
TPJ
LP
VLPFC
anterior insula IT
(b) MCC
dACC
PCC
MPF
Figure 5.7
This is part of a theoretical
approach based on several
functional networks of
sgACC
relevance to attention:
the four networks shown
(fronto-parietal; default
Networks
mode; cingulo-opercular;
Key: Fronto-parietal Default mode and ventral attention) are all
discussed fully in the text.
Cingulo-opercular Ventral attention Sylvester et al., 2012, p. 528.
Reprinted with permission of
Elsevier.
196 Visual perception and attention
KEY TERMS middle cingulate cortex (MCC) and the intraparietal sulcus (IPS). It is
associated with top-down attentional and cognitive control.
Neglect
A disorder involving
right-hemisphere damage
Evaluation
(typically) in which the
left side of objects and/ The theoretical approach proposed by Corbetta and Shulman (2002) has
or objects presented to
several successes to its credit. First, there is convincing evidence for some-
the left visual field are
undetected; the condition what separate stimulus-driven and top-down attention systems, each with
resembles extinction but its own brain network. Second, research using transcranial magnetic stimu-
is more severe. lation suggests major brain areas within each attention system play a causal
Pseudo-neglect role in attentional processes. Third, some interactions between the two net-
A slight tendency in works have been identified. Fourth, research on brain-damaged patients
healthy individuals to supports the theoretical approach (see next section, pp. 196–200).
favour the left side of What are the limitations of this theoretical approach? First, the precise
visual space.
brain areas associated with each attentional system have not been clearly
identified. Second, there is more commonality (especially within the pari-
etal lobe) in the brain areas associated with the two attention networks
than assumed theoretically by Corbetta and Shulman (2002). Third, there
are additional attention-related brain networks not included within the
original theory. Fourth, much remains to be discovered about how differ-
ent attention systems interact.
Figure 5.8
On the left is a copying
task in which a patient with
unilateral neglect distorted
or ignored the left side of
the figures to be copied
(shown on the left). On the
right is a clock drawing
task in which the patient
was given a clock face and
told to insert the numbers
into it.
Reprinted from Danckert and
Ferber (2006). Reprinted with
permission from Elsevier.
with allocentric than egocentric neglect, whereas the opposite was the case
with damage to the temporo-parietal junction.
Extinction is often found in neglect patients. Extinction involves a
failure to detect a stimulus presented to the side opposite the brain damage
when a second stimulus is presented to the same side as the brain damage.
Extinction and neglect are closely related but separate deficits (de Haan
et al., 2012). We will focus mostly on neglect because it has attracted much
more research.
Which brain areas are damaged in neglect patients? Neglect is a het-
erogeneous condition and the brain areas damaged vary considerably
across patients. In a meta-analysis, Molenberghs et al. (2012) found the
main areas damaged in neglect patients are in the right hemisphere and
include the superior temporal gyrus, the inferior frontal gyrus, the insula,
the supramarginal gyrus and the angular gyrus (gyrus means ridge). Nearly
all these areas are within the stimulus-driven or ventral attention network Interactive feature:
(see Figure 5.6) suggesting brain networks are damaged rather than simply Primal Pictures’
specific brain areas (Corbetta & Shulman, 2011). 3D atlas of the brain
We also need to consider functional connectivity (correlated brain
activity between brain regions). Baldassarre et al. (2014, 2016) discovered
widespread disruption of functional connectivity between the hemispheres
in neglect patients. This disruption did not involve the bottom-up and top- KEY TERM
down attention networks. Of importance, recovery from attention deficits Extinction
in neglect patients was associated with improvements in functional con- A disorder of visual
nectivity in bottom-up and top-down attention networks (Ramsey et al., attention in which a
2016). stimulus presented to the
side opposite the brain
The right-hemisphere temporo-parietal junction and intraparietal sulcus damage is not detected
are typically damaged in extinction patients (de Haan et al., 2012). When when another stimulus is
transcranial magnetic stimulation is applied to these areas to interfere with presented at the same
processing, extinction-like behaviour results (de Haan et al., 2012). Dugué time to the side of the
et al. (2018) confirmed the importance of the temporo-parietal junction brain damage.
198 Visual perception and attention
Theoretical considerations
Corbetta and Shulman (2011) discussed neglect in the context of their
two-system theory (discussed earlier, see pp. 192–196). In essence, the
bottom-up ventral attention network is damaged. Strong support for this
assumption was reported by Toba et al. (2018a) who found in 25 neglect
patients that impaired performance on tests of neglect was associated with
damage to parts of the ventral attention network (e.g., angular gyrus;
supramarginal gyrus). Since the right hemisphere is dominant in the ventral
attention network, neglect patients typically have damage in that hem-
isphere. Of importance, Corbetta and Shulman (2011) also assumed that
damage to the ventral network impairs the functioning of the goal-directed
dorsal attention network (even though not itself damaged).
How does the damaged ventral attention network impair the dorsal
attention network’s functioning? The two attention networks interact and
so damage to the ventral network inevitably affects the functioning of the
dorsal network. More specifically, damage to the ventral attention network
“impairs non-spatial [across the entire visual field] functions, hypoactivates
[reduces activation in] the right hemisphere, and unbalances the activity of
the dorsal attention network” (Corbetta & Shulman, 2011, p. 592).
de Haan et al. (2012) proposed a theory of extinction based on two
major assumptions:
Findings
According to Corbetta and Shulman (2011), the dorsal attention network in
neglect patients functions poorly because of reduced activation in the right
hemisphere and associated reduced alertness and attentional resources.
Thus, increasing patients’ general alertness should enhance their detection
of left-field visual targets. Robertson et al. (1998) found the slower detec-
tion of left visual field stimuli compared to those in the right visual field was
no longer present when warning sounds were used to increase alertness.
Bonato and Cutini (2016) compared neglect patients’ ability to detect
visual targets with (or without) a second, attentionally demanding task.
Detection rates were high for targets presented to the right visual field in
both conditions. In contrast, patients detected only approximately 50% as
many targets in the left visual field as the right when performing another
task. Thus, neglect patients have limited attentional resources.
Corbetta and Shulman (2011) assumed neglect patients have an essen-
tially intact dorsal attention network. Accordingly, neglect patients might
use that network effectively if steps were taken to facilitate its use. Duncan
et al. (1999) presented arrays of letters and neglect patients recalled only
those in a pre-specified colour (the dorsal attention network could be used
to select the appropriate letters). Neglect patients resembled healthy con-
trols in showing equal recall of letters presented to each side of visual space.
The two attention networks typically work closely together. Bays et al.
(2010) studied neglect patients. They used eye movements during a visual
search to assess patients’ problems with top-down and stimulus-driven
attentional processes. Both types of attentional processes were equally
impaired (as predicted by Corbetta and Shulman, 2011). Of most impor-
tance, there was a remarkably high correlation of +.98 between these two
types of attentional deficit.
Toba et al. (2018b) identified two reasons for the failure of neglect
patients to detect left-field stimuli:
KEY TERM 24% of patients. Accordingly, Toba et al. argued we should develop
multi-component models of visual neglect to account for such individual
Visual search
A task involving the rapid differences.
detection of a specified We turn now to extinction patients. According to de Haan et al.
target stimulus within a (2012), extinction occurs because of biased competition between stimuli.
visual display. If two stimuli could be integrated, that might minimise competition and so
reduce extinction. Riddoch et al. (2006) tested this prediction by present-
ing objects used together often (e.g., wine bottle and wine glass) or never
used together (e.g., wine bottle and ball). Extinction patients identified both
objects more often in the former condition than the latter (65% vs 40%,
respectively).
The biased competition hypothesis has been tested in other ways. We
can impair attentional processes in the intact left hemisphere by applying
transcranial magnetic stimulation to it. This should reduce competition
from the left hemisphere in extinction patients and thus reduce extinction.
Some findings are consistent with this prediction (Oliveri & Caltagirone,
2006).
de Haan et al. (2012) also identified reduced attentional capacity as a
factor causing extinction. Bonato et al. (2010) studied extinction with or
without the addition of a second, attentionally demanding task. As pre-
dicted, extinction patients showed a substantial increase in the extinction
rate (from 18% to over 80%) with this additional task.
Overall evaluation
Research has produced several important findings. First, neglect and
extinction patients can process unattended visual stimuli in the absence of
conscious awareness of those stimuli. Second, most neglect patients have
damage to the ventral attention network leading to impaired functioning of
the undamaged dorsal attention network. Third, extinction occurs because
of biased competition for attention and reduced attentional capacity.
What are the limitations of research in this area? First, it is hard
to produce theoretical accounts applicable to all neglect or extinction
patients because the precise symptoms and regions of brain damage
vary considerably across patients. Second, neglect patients vary in their
precise processing deficits (e.g., Toba et al., 2018b), but this has been
de-emphasised in most theories. Third, the precise relationship between
neglect and extinction remains unclear. Fourth, the dorsal and ventral
networks generally interact but the extent of their interactions remains to
be determined.
VISUAL SEARCH
We spend much time searching for various objects (e.g., a friend in a
crowd). The processes involved have been studied in research on visual
search where a specified target is detected as rapidly as possible. Initially,
we consider an important real-world situation where visual search can be
literally a matter of life-or-death: airport security checks. After that, we
consider an early very influential theory of visual search before discussing
more recent theoretical and empirical developments.
Attention and performance 201
Figure 5.9
Each bag contains one illegal item. From left to right: a large bottle; a dynamite stick; and a gun part.
From Mitroff & Biggs (2014).
There are two major reasons it is often hard for airport security screeners to detect dangerous
items. First, illegal and dangerous items are (thankfully!) present in only a minute fraction of pas-
sengers’ luggage. This rarity of targets makes it hard for airport security screeners to detect them.
Mitroff and Biggs (2014) asked observers to detect illegal items in bags (see Figure 5.9). The
detection rate was only 27% when targets appeared under 0.15% of the time: they termed this
the “ultra rare item effect”. In contrast, the detection rate was 92% when targets appeared more
than 1% of the time.
Peltier and Becker (2016) tested two explanations for the reduced detection rate with rare targets:
(1) a reduced probability that the target is fixated (selection error); and (2) increased caution about
reporting targets because they are so unexpected (identification error). There was evidence for
both explanations. However, most detection failures were selection errors (see Figure 5.10).
Accuracy
1 Target present Target absent
0.9
0.8
0.7
Accuracy
0.6
0.5
0.4
0.3
0.2
0.1
0
10 50 90
Prevalence
Figure 5.10
Frequency of selection and identification errors when targets were present on 10%, 50% or 90% of trials.
From Peltier and Becker (2016).
202 Visual perception and attention
Second, security screeners search for numerous different objects. This increases search difficulty.
Menneer et al. (2009) found target detection was worse when screeners searched for two catego-
ries of objects (metal threats and improvised explosive devices) rather than one.
How can we increase the efficiency of security screening? First, we can exploit individual dif-
ferences in the ability to detect targets. Rusconi et al. (2015) found individuals scoring high on a
questionnaire measure of attention to detail had superior target-detection performance than low
scorers.
Second, airport security screeners can find it hard to distinguish between targets (i.e., danger-
ous items) and similar-looking non-targets. Geng et al. (2017) found that observers whose training
included non-targets resembling targets learned to develop increasingly precise internal target
representations. Such representations can improve the speed and accuracy of security screening.
Third, the low detection rate when targets are very rare can be addressed. Threat image pro-
jection (TIP) can be used to project fictional threat items into x-ray images of luggage to increase
the apparent frequency of targets. When screeners are presented with TIPs plus feedback when
they miss them, screening performance improves considerably (Hofer & Schwaninger, 2005). In
similar fashion, Schwark et al. (2012) found providing false feedback to screeners to indicate they
had missed rare targets reduced their cautiousness about reporting targets and improved their
performance.
(1) Basic visual features are processed rapidly and pre-attentively in par-
allel across the visual scene.
(2) Stage (1) is followed by a slower serial process with focused atten-
tion providing the “glue” to form objects from the available features
(e.g., an object that is round and has an orange colour is perceived
as an orange). In the absence of focused attention, features from
different objects may be combined randomly producing an illusory
conjunction.
It follows from the above assumptions that targets defined by a single feature
(e.g., a blue letter or an S) should be detected rapidly and in parallel. In
contrast, targets defined by a conjunction or combination of features (e.g., a
green letter T) should require focused attention and so should be slower to
detect. Treisman and Gelade (1980) tested these predictions using both types
of targets; the display size was 1–30 items and a target was present or absent.
As predicted, response was rapid and there was very little effect of
KEY TERM
display size when the target was defined by a single feature: these findings
Illusory conjunction
Mistakenly combining
suggest parallel processing (see Figure 5.11). Response was slower and was
features from two strongly influenced by display size when the target was defined by a con-
different stimuli to junction of features: these findings suggest there was serial processing.
perceive an object that is According to the theory, lack of focused attention can produce illusory
not present. conjunctions based on random combinations of features. Friedman-Hill
Attention and performance 203
Figure 5.11
Performance speed on
a detection task as a
function of target definition
(conjunctive vs single
feature) and display size.
Adapted from Treisman and
Gelade (1980).
Limitations
What are the theory’s limitations? First, Duncan and Humphreys (1989,
1992) identified two factors not included within feature integration theory:
(1) When distractors are very similar to each other, visual search is faster
because it is easier to identify them as distractors.
(2) The number of distractors has a strong effect on search time to
detect even targets defined by a single feature when targets resemble
distractors.
Second, Treisman and Gelade (1980) estimated the search time with con-
junctive targets was approximately 60 ms per item and argued this repre-
sented the time taken for focal attention to process each item. However,
research with other paradigms indicates it takes approximately 250 ms for
attention indexed by eye movements to move from one location to another.
Thus, it is improbable focal attention plays the key role assumed within the
theory.
Third, the theory assumes visual search is often item-by-item. However,
the information contained within most visual scenes cannot be divided up
into “items” and so the theory is of limited applicability. Such considera-
tions led Hulleman and Olivers (2017) to produce an article entitled “The
impending demise of the item in visual search”.
Fourth, visual search involves parallel processing much more than
implied by the theory. For example, Thornton and Gilden (2007) used
29 different visual tasks and found 72% apparently involved parallel
204 Visual perception and attention
processing. We can explain such findings by assuming that each eye fixa-
tion permits considerable parallel processing using information available in
peripheral vision (discussed below, see pp. 206–208).
Fifth, the theory assumes that the early stages of visual search are
entirely feature-based. However, recent research using event-related poten-
tials indicates that object-based processing can occur much faster than pre-
dicted by feature integration theory (e.g., Berggren & Eimer, 2018).
Sixth, the theory assumes visual search is essentially random. This
assumption is wrong with respect to the real world – we typically use our
knowledge of where a target object is likely to be located when searching
for it (see below).
Dual-path model
In most of the research discussed so far, the target appeared at a random
location within the visual display. This is radically different from the real
world. Suppose you are outside looking for your missing cat. Your visual
search would be very selective – you would ignore the sky and focus mostly
on the ground (and perhaps the trees). Thus, your search would involve
top-down processes based on your knowledge of where cats are most likely
to be found.
Ehinger et al. (2009) studied top-down processes in visual search by
recording eye fixations of observers searching for a person in 900 real-
world outdoor scenes. Observers typically fixated plausible locations (e.g.,
pavements) and ignored implausible ones (e.g., sky; trees; see Figure 5.12).
Observers also fixated locations differing considerably from neighbouring
locations and areas containing visual features resembling those of a human
figure.
How can we reconcile Ehinger et al.’s (2009) findings with those
discussed earlier? Wolfe et al. (2011) proposed a dual-path model (see
Figure 5.13). There is a selective pathway of limited capacity (indicated
by the bottleneck) with objects being selected individually for recognition.
Figure 5.12
The first three eye fixations made by observers searching for pedestrians. As can be
seen, the great majority of their fixations were on regions in which pedestrians would
most likely be found. Observers’ fixations were much more like each other in the left-
hand photo than in the right-hand one, because there were fewer likely regions in the
left-hand one.
From Ehinger et al. (2009). Reprinted with permission from Taylor & Francis.
Attention and performance 205
Figure 5.13
A two-pathway model of
y
wa visual search. The selective
a th pathway is capacity limited
e p
c tiv and can bind stimulus
ele features and recognise
ns
Early vision No objects. The non-selective
pathway processes the gist
of scenes. Selective and
non-selective processing
occur in parallel to produce
Features effective visual search.
Color
Orientation From Wolfe et al. (2011).
Size Reprinted with permission from
Depth Elsevier.
Select
Motion
Etc.
ive pa
thway
Binding and
recognition
This pathway has been the focus of most research until recently. There is
also a non-selective pathway in which the “gist” of a scene is processed.
Such processing can then guide processing within the selective pathway
(represented by the arrow labelled “guidance”). This pathway allows us to
utilise our stored environmental knowledge and so is of great value in the
real world.
Findings
Wolfe et al. (2011) compared visual searches for objects presented within a
scene setting or at random locations. As predicted, search rate per item was
much faster in the scene setting (10 ms vs 40 ms, respectively). Võ and Wolfe
(2012) explained that finding in terms of “functional set size” – searching in
scenes is efficient because most regions can be ignored. As predicted, Võ
and Wolfe found 80% of each scene was rarely fixated.
Kaiser and Cichy (2018) presented observers with objects typically
located in the upper (e.g., aeroplane; hat) or lower (e.g., carpet; shoe) visual
field. These objects were presented in their typical or atypical location
(e.g., hat in the lower visual field). Observers had to indicate whether an
object presented very briefly was located in the upper or lower visual field.
Observers’ performance was better when objects appeared in their typical
location because of their extensive knowledge of where objects are gener-
ally located.
Chukoskie et al. (2013) found observers can easily learn where
targets are located. An invisible target was presented at random loca-
tions on a blank screen and observers were provided with feedback. There
was a strong learning effect – fixations rapidly shifted from being fairly
206 Visual perception and attention
KEY TERM random to being focused on the area within which the target might be
present.
Fovea
A small area within the Ehinger et al.’s (2009) findings (discussed earlier, see p. 204) suggested
retina in the centre in the that scene gist or context can be used to enhance the efficiency of visual
field of vision where visual search. Katti et al. (2017) presented scenes very briefly (83 ms) followed by
acuity is greatest. a mask. Observers were given the task of detecting a person or a car and
performed very accurately (over 90%) and rapidly. Katti et al. confirmed
that scene gist or context influenced performance. However, performance
was influenced more strongly by features of the target object – the more
key features of an object were visible, the faster it was detected.
What is the take-home message from the above study? The efficiency
of visual search with real-world scenes is more complex than implied by
Ehinger et al. (2009). More specifically, observers may rapidly fixate on the
area close to a target person because they are using scene gist or because
they rapidly process features of the person (e.g., wearing clothes).
Evaluation
Our knowledge of likely (and unlikely) locations for any given object in a
scene influences visual search in the real world. This is fully acknowledged
in the dual-path model. There is also support for the notion that scene
knowledge facilitates visual search by reducing functional set size.
What are the model’s limitations? First, how we use gist knowledge
of a scene very rapidly to reduce the search area remains unclear. Second,
there is insufficient focus on the learning processes that can greatly facil-
itate visual search – the effects of such processes can be seen in the very
rapid and accurate detection of target information by experts in several
domains (see Chapter 11).
Third, it is important not to exaggerate the importance of scene gist or
context in influencing the efficiency of visual search. Features of the target
object can influence visual search more than scene gist (Katti et al., 2017).
Fourth, the assumption that items are processed individually within
the selective pathway is typically mistaken. As we will see shortly, visual
search often depends on parallel processes within peripheral vision and
such processes are not considered within the model.
your thumb and fixating the nail. Foveal vision only covers the nail so the KEY TERM
great majority of what you can see is in peripheral vision.
Visual crowding
We can also compare the value of foveal and peripheral vision by con- The inability to recognise
sidering individuals with impaired eyesight. Those with severely impaired objects in peripheral
peripheral vision (e.g., due to glaucoma) had greater problems with mobil- vision due to the presence
ity (e.g., number of falls; ability to drive) than those who lack foveal vision of neighbouring objects.
(due to macular degeneration) (Rosenholtz, 2016). Individuals with severely
impaired central or foveal vision performed almost as well as healthy con-
trols at detecting target objects in coloured scenes (75% vs 79%, respec-
tively) (Thibaut et al., 2018).
If visual search depends heavily on peripheral vision, what predictions
can we make? First, if each fixation provides observers with a considerable
amount of information about several objects, visual search will typically
involve parallel rather than serial processing. Second, we need to consider
limitations of peripheral vision (e.g., visual acuity is less in peripheral
than foveal vision). However, a more important limitation concerns visual
crowding – a reduced ability to recognise objects or other stimuli because
of irrelevant neighbouring objects or stimuli (clutter). Visual crowding
impairs peripheral vision to a much greater extent than foveal vision.
Rosenholtz et al. (2012) proposed the texture tiling model based on
the assumption peripheral vision is of crucial importance in visual search.
More specifically, processing in peripheral vision can cause adjacent stimuli
to tile (join together) to form an apparent target, thus increasing the diffi-
culty of visual search. Below we consider findings relevant to this model.
Findings
As mentioned earlier (p. 203), Thornton and Gilden (2007) found almost
three-quarters of the visual tasks they studied involved parallel processing.
This is entirely consistent with the emphasis on parallel processing in the
model.
Direct evidence for the importance of peripheral vision to visual search
was reported by Young and Hulleman (2013). They manipulated the visible
area around the fixation point making it small, medium or large. As pre-
dicted by the model, visual search performance was worst when the visible
area was small (so only one item could be processed per fixation). Overall,
visual search was almost parallel when the visible area was large but serial
when it was small.
Chang and Rosenholtz (2016) used various search tasks. According to
feature integration theory, both tasks shown in Figure 5.14 should be com-
parably hard because the target and distractors share features. In contrast,
the texture tiling model predicts the task on the right should be harder
because adjacent distractors seen in peripheral vision can more easily tile
(join together) to form an apparent T. The findings from these tasks (and
several others) supported the texture tiling model but were inconsistent
with feature integration theory.
Finally, Hulleman and Olivers (2017) produced a model of visual
search consistent with the texture tiling model. According to this model,
each eye fixation lasts 250 ms, during which information from foveal and
peripheral vision is extracted in parallel. They also assumed that the area
208 Visual perception and attention
Figure 5.14
The target (T) is easier to (a) Easier search (b) Harder search
find in the display on the
left than the one on the
right.
From Chang and Rosenholtz
(2016). Find the T
around the fixation point within which a target can generally be detected
is smaller when the visual search task is difficult (e.g., because target dis-
criminability is low).
A key prediction from Hulleman and Olivers’ (2017) model is that the
main reason why search times are longer with more difficult search tasks is
because more eye fixations are required than with easier tasks. A computer
simulation based on these assumptions produced search times very similar
to those obtained in experimental studies.
Evaluation
What are the strengths of the texture tiling model? First, the information
available in peripheral vision is much more important in visual search than
assumed previously. The model explains how observers make use of the
information available in peripheral vision.
Second, the model explains why parallel processing is so prevalent
in visual search – it reflects directly parallel processing within peripheral
vision. Third, there is accumulating evidence that search times are gener-
ally directly related to the number of eye fixations.
Fourth, an approach based on eye fixations and peripheral vision can
potentially explain findings from all visual search paradigms, including
complex visual scenes and item displays. Such an approach thus has more
general applicability than feature integration theory.
What are the model’s limitations? First, as Chang and Rosenholtz
(2016) admitted, it needs further development to account fully for visual
search performance. For example, it does not predict search times with
precision. In addition, it does not specify the criteria used by observers to
decide no target is present.
Second, visual search is typically much faster for experts than non-
experts in their domain of expertise (e.g., medical experts examining mam-
mograms) (see Chapter 11). The texture tiling model does not identify
clearly the processes allowing experts to make very efficient use of periph-
eral information.
CROSS-MODAL EFFECTS
Nearly all the research discussed so far is limited in that the visual (or audi-
tory) modality was studied on its own. We might try to justify this approach
by assuming attentional processes in each sensory modality operate
Attention and performance 209
independently from those in other modalities. However, that assumption is KEY TERMS
incorrect. In the real world, we often coordinate information from two or
Cross-modal attention
more sense modalities at the same time (cross-modal attention). An example The coordination of
is lip reading, where we use visual information about a speaker’s lip move- attention across two or
ments to facilitate our understanding of what they are saying (see Chapter 9). more modalities (e.g.,
Suppose we present participants with two streams of lights (as was vision and audition).
done by Eimer and Schröger, 1998), with one stream being presented to the Ventriloquism effect
left and the other to the right. At the same time, we present participants The mistaken perception
with two streams of sounds (one to each side). In one condition, partici- that sounds are
coming from their
pants detect deviant visual events (e.g., longer than usual stimuli) presented
apparent source (as in
to one side only. In the other condition, participants detect deviant audi- ventriloquism).
tory events in only one stream.
Event-related potentials were recorded to assess the allocation of atten-
tion. Unsurprisingly, Eimer and Schröger (1998) found ERPs to deviant
stimuli in the relevant modality were greater to stimuli presented on the
to-be-attended side than the to-be-ignored side. Thus, participants allo-
cated attention as instructed.
Of more interest is what happened to the allocation of attention in
the irrelevant modality. Suppose participants detected visual targets on the
left side. In that case, ERPs to deviant auditory stimuli were greater on
the left side than the right. This is a cross-modal effect: the voluntary or
endogenous allocation of visual attention also affected the allocation of
auditory attention. Similarly, when participants detected auditory targets
on one side, ERPs to deviant visual stimuli on the same side were greater
than ERPs to those on the opposite side. Thus, the allocation of auditory
attention also influenced the allocation of visual attention.
Ventriloquism effect
What happens when there is a conflict between simultaneous visual and
auditory stimuli? We will focus on the ventriloquism effect in which
sounds are misperceived as coming from their apparent visual source.
Ventriloquists (at least good ones!) speak without moving their lips while
manipulating a dummy’s mouth movements. It seems as if the dummy is
speaking. Something similar happens at the movies. The actors’ lips move
on the screen but their voices come from loudspeakers beside the screen.
Nevertheless, we hear those voices coming from their mouths.
Certain conditions must be satisfied for the ventriloquism effect to
occur (Recanzone & Sutter, 2008). First, the visual and auditory stimuli
must occur close together in time. Second, the sound must match expecta-
tions created by the visual stimulus (e.g., high-pitched sound coming from
a small object). Third, the sources of the visual and auditory stimuli should
be close together spatially. More generally, the ventriloquism effect reflects
the unity assumption (the assumption that two or more sensory cues come
from the same object: Chen & Spence, 2017).
The ventriloquism effect exemplifies visual dominance (visual informa-
tion dominating perception). Further evidence comes from the Colavita
effect (Colavita, 1974): participants instructed to respond to all stimuli
respond more often to visual than simultaneous auditory stimuli (Spence
et al., 2011).
210 Visual perception and attention
KEY TERM When during processing is visual spatial information integrated with
auditory information? Shrem et al. (2017) found that misleading visual
Temporal ventriloquism
effect information about the location of an auditory stimulus influenced the pro-
Misperception of the cessing of the auditory stimulus approximately 200 ms after stimulus onset.
timing of a visual stimulus The finding that this effect is still present even when participants are aware
when an auditory stimulus of the spatial discrepancy between the visual and auditory input suggests it
is presented close to it in
occurs relatively “automatically”.
time.
However, the ventriloquism effect is smaller when participants had
previously heard syllables spoken in a fearful voice (Maiworm et al., 2012).
This suggests the effect is not entirely “automatic” but is reduced when the
relevance of the auditory channel is increased.
Why does vision capture sound in the ventriloquism effect? The visual
modality typically provides more precise information about spatial loca-
tion. However, when visual stimuli are severely blurred and poorly local-
ised, sound captures vision (Alais & Burr, 2004). Thus, we combine visual
and auditory information effectively by attaching more weight to the more
informative sense modality.
Temporal ventriloquism
The above explanation for the ventriloquist illusion is a development of
the modality appropriateness and precision hypothesis (Welch & Warren,
1980). According to this hypothesis, when conflicting information is pre-
sented in two or more modalities, the modality having the greatest acuity
generally dominates. This hypothesis predicts the existence of another illu-
sion. The auditory modality is typically more precise than the visual modal-
ity at discriminating temporal relations. As a result, judgements about
the temporal onset of visual stimuli might be biased by auditory stimuli
presented very shortly beforehand or afterwards. This is the temporal
ventriloquism effect.
Research on temporal ventriloquism
was reviewed by Chen and Spence (2017).
A simple example is when the apparent
onset of a flash is shifted towards an abrupt
sound presented slightly asynchronously (see
Figure 5.15). Other research has found that
the apparent duration of visual stimuli can be
distorted by asynchronous auditory stimuli.
We need to consider the temporal ven-
triloquism effect in the context of the unity
assumption. This is the assumption that “two
or more uni-sensory cues belong together
(i.e., that they come from the same object or
event)” (Chen & Spence, 2017, p. 1). Chen
and Spence discussed findings showing that
Figure 5.15 the unity assumption generally (but not
An example of temporal ventriloquism in which the apparent
always) enhances the temporal ventriloquism
time of onset of a flash is shifted towards that of a sound
presented at a slightly different timing from the flash. effect.
From Chen and Vroomen (2013). Reprinted with permission from
Orchard-Mills et al. (2016) extended
Springer. research by using two visual stimuli (one
Attention and performance 211
above and the other below fixation) and two auditory stimuli (low- and
high-pitch). When the visual and auditory stimuli were congruent (e.g.,
visual stimulus above fixation and auditory stimulus high-pitch), the tem-
poral ventriloquism effect was found. However, this effect was eliminated
when the visual and auditory stimuli were incongruent, which prevented
binding of information across the two senses.
Theoretical considerations
Why do so many drivers endanger people’s lives by using mobile phones? Most believe they
can drive safely while using a mobile phone whereas other drivers cannot (Sanbonmatsu et al.,
2016b). Their misplaced confidence depends on limited monitoring of their driving performance:
drivers using a mobile phone make more driving errors but do not remember making more errors
(Sanbonmatsu et al., 2016a).
Why does mobile-phone use impair driving performance? Strayer and Fisher (2016) in their
SPIDER model identified five cognitive processes that are adversely affected when drivers’ attention
is diverted from driving (e.g., by mobile-phone use):
(1) There is less effective visual scanning of the environment for potential threats. Distracted drivers
are more inclined to focus attention on the centre of the road and less inclined to scan objects
in the periphery and their side mirrors (Strayer & Fisher, 2016).
(2) The ability to predict where threats might occur is impaired. Distracted drivers are much less
likely to make anticipatory glances towards the location of a potential hazard (e.g., obstructed
view of a pedestrian crossing) (Taylor et al., 2015).
(3) There is reduced ability to identify visible threats, a phenomenon known as inattentional blind-
ness (see Glossary; and Chapter 4). In a study by Strayer and Drews (2007), 30 objects (e.g.,
pedestrians; advertising hoardings) were clearly visible to drivers. However, those using a
mobile phone subsequently recognised far fewer objects they had fixated than those not using
a mobile phone (under 25% vs 50%, respectively).
(4) It is harder to decide what action is necessary in a threatening situation. Cooper et al. (2009)
found drivers were 11% more likely to make unsafe lane changes when using a mobile
phone.
(5) It becomes harder to execute the appropriate action. Reaction times are slowed (Caird et al.,
2008, discussed above, p. 214).
The SPIDER model is oversimplified in several ways. First, various different activities are associated
with mobile-phone use. Simmons et al. (2016) found in a meta-analytic review that the risk of safety-
critical events was increased by activities requiring drivers to take their eyes off the road (e.g., locat-
ing a phone; dialling; texting). However, talking on a mobile phone did not increase risk.
Second, driving-irrelevant cognitive activities do not always impair all aspects of driving per-
formance. Engstrom et al. (2017, p. 734) proposed their cognitive control hypothesis: “Cognitive
load selectively impairs driving sub-tasks that rely on cognitive control but leaves automatic perfor-
mance unaffected.” For example, driving-irrelevant activities involving cognitive load (e.g., mobile-
phone use) typically have no adverse effect on well-practised driving skills, such as lane keeping
and braking when getting close to the vehicle in front (Engstrom et al., 2017).
Third, individuals using mobile phones while driving are unrepresentative of drivers in general
(e.g., they tend to be relatively young and to engage in more risk-taking activities: Precht et al.,
2017). Thus, we must consider individual differences in personality and risk taking when interpret-
ing accidents associated with mobile-phone use.
Fourth, the SPIDER model implies that performance cannot be improved by adding a secondary
task. However, driving performance in monotonous conditions is sometimes better when drivers
listen to the radio at the same time (see Engstrom et al., 2017). Listening to the radio can reduce
the mind-wandering that occurs when someone drives in monotonous conditions. Drivers indicat-
ing their immediate thoughts during their daily commute reported mind-wandering 63% of the
time and active focus on driving only 15%–20% of the time (Burdett et al., 2018).
Figure 5.16
Wickens’s four-dimensional
multiple resource model.
The details are described in
the text.
From Wickens (2008). © 2008.
Reprinted by permission of
SAGE Publications.
Here is the model’s crucial prediction: “To the extent that two tasks use dif-
ferent levels along each of the three dimensions [excluding (4) above], time-
sharing [dual-task performance] will be better” (Wickens, 2008, p. 450). Thus,
tasks requiring different resources can be performed together more success-
fully than those requiring the same resources. Wickens’s approach bears
some resemblance to Baddeley’s (e.g., 2012) working memory model (see
Chapter 6). According to that model, two tasks can be performed together
successfully provided they use different components or processing resources.
Findings
Research discussed earlier (Treisman & Davies, 1973; McLeod, 1977)
showing the negative effects of stimulus and response similarity on per-
formance are entirely consistent with the theory. Lu et al. (2013) reviewed
research where an ongoing visual-motor task (e.g., car driving) was
performed together with an interrupting task in the visual, auditory or
tactile (touch) modality. As predicted, non-visual interrupting tasks (espe-
cially those in the tactile modality) were processed more effectively than
visual ones and there were no adverse effects on the visual-motor task.
According to the model, there should be only limited dual-task inter-
ference between two visual tasks if one requires focal or foveal vision,
whereas the other requires ambient or peripheral vision. Tsang and Chan
(2018) obtained support for this prediction in a study in which participants
tracked a moving target in focal vision while responding to a spatial task
in ambient or peripheral vision.
Dual-task performance is often more impaired than predicted by
the theory. For example, consider a study by Robbins et al. (1996; see
Attention and performance 217
Chapter 6). The main task was selecting chess moves and we will focus on
the condition where the task performed at the same time was generating
random letters. These two tasks involve different processing codes (spatial
vs verbal, respectively) and they also involve different response types
(manual vs vocal, respectively). Nevertheless, generating random letters
caused substantial interference on the chess task.
Evaluation
The main assumptions of the theory have largely been supported by the
experimental evidence. In other words, dual-task performance is generally
less impaired when two tasks differ with respect to modalities, processing
codes or visual channels than when they do not.
What are the model’s limitations?
Threaded cognition
Salvucci and Taatgen (2008, 2011) proposed a model of threaded cogni-
tion in which streams of thought are represented as threads of processing.
For example, processing two tasks might involve two separate threads. The
central theoretical assumptions are as follows:
Multiple threads or goals can be active at the same time, and as long as
there is no overlap in the cognitive resources needed by these threads,
there is no multi-tasking interference. When threads require the same
resource at the same time, one thread must wait and its performance
will be adversely affected.
(Salvucci & Taatgen, 2011, p. 228)
The model predicts focal working memory is required only in the hard con-
dition. Subtraction was combined with a secondary task: a tracking task
involving visual and manual resources or a tone-counting task involving
working memory.
Nijboer et al. (2013) predicted performance on the easy subtraction
task should be worse when combined with the tracking task because both
compete for visual and manual resources. In contrast, performance on
the hard subtraction task should be worse when combined with the tone-
counting task because there are large disruptive effects when two tasks
compete for working memory resources. The findings were as predicted.
Borst et al. (2013) found there was far less impairment of hard sub-
traction performance by a secondary task requiring working memory when
Attention and performance 219
Evaluation
The model has proved successful in various ways. First, several important
cognitive resources have been identified. Second, the model identifies brain
areas associated with various cognitive resources. This has led to compu-
tational modelling testing the model’s predictions using neuroimaging and
behavioural findings. Thus, the model accounts for dual-task performance
without assuming the existence of a central executive or other executive
process (often vaguely defined in other theories). Fourth, the theory predicts
factors determining switching between two tasks being performed together.
Fifth, individuals often have fewer problems performing two simultaneous
tasks than generally assumed.
What are the model’s limitations? First, it predicts that “Practising two
tasks concurrently [together] results in the same performance as perform-
ing the two tasks independently” (Salvucci & Taatgen, 2008, p. 127). This
de-emphasises the importance of processes coordinating and managing two
tasks performed together (see next section). Second, excluding processes
resembling Baddeley’s central executive is controversial and may well
prove inadvisable. Third, most tests of the model have involved the simul-
taneous performance of two relatively simple tasks and its applicability to
more complex tasks remains unclear. Fourth, the theory does not provide
a full explanation for individual differences in the extent of task switching
(e.g., Katidioti & Taatgen, 2014).
Cognitive neuroscience
The cognitive neuroscience approach is increasingly used to test theoretical
models and enhance our understanding of processes underlying dual-task
performance. Its value is that neuroimaging provides “an additional data
source for contrasting between alternative models” (Palmeri et al., 2017,
p. 61). More generally, behavioural findings indicate the extent to which
dual-task conditions impair task performance but are often relatively unin-
formative about the precise reasons for such impairment.
Suppose we compare patterns of brain activation while participants
perform tasks x and y singly or together. Three basic patterns are shown
in Figure 5.18:
Figure 5.18 (a) Underadditive activation (b) Additive activation (c) Overadditive activation
(a) Underadditive activation; Component task 1 Component task 1 Component task 1
(b) additive activation;
(c) overadditive activation.
White indicates task 1 Time Time Time
Component task 2 Component task 2 Component task 2
activation; grey indicates
task 2 activation; and
black indicates activation Time Time Time
only present in dual-task Dual-task Dual-task Dual-task
conditions.
From Nijboer et al., 2014.
Time Time Time
Reprinted with permission of
Elsevier.
Attention and performance 221
(2) Additive activation: brain activation in the dual-task condition is KEY TERM
simply the sum of the two single-task activations because access to
Underadditivity
resources is integrated efficiently between the two tasks. The finding that brain
(3) Overadditive activation: brain activation in one or more brain areas is activation when tasks
present in the dual-task condition but not the single-task conditions. A and B are performed
This occurs when dual-task conditions require executive processes that at the same time is less
than the sum of the brain
are absent (or less important) with single tasks. These executive pro-
activation when tasks
cesses include the coordination of task demands, attentional control A and B are performed
and dual-task management generally. We would expect such executive separately.
processes to be associated mostly with activation in prefrontal cortex.
Findings
We start with an example of underadditive activation. Just et al. (2001)
used two very different tasks (auditory sentence comprehension and mental
rotation of 3-D figures) performed together or singly. Performance on both
tasks was impaired under dual-task conditions compared to single-task
conditions. Under dual-task conditions, brain activation in language pro-
cessing areas decreased by 53% and reduced by 29% in areas associated
with mental rotation. These findings suggest fewer task-relevant processing
resources were available when both tasks were performed together.
Schweizer et al. (2013) also reported underadditivity. Participants
performed a driving task on its own or with a distracting secondary task
(answering spoken questions). Driving performance was unaffected by
the secondary task. However, driving with distraction reduced activa-
tion in posterior brain areas associated with spatial and visual processing
(underadditivity). It also produced increased activation in the prefrontal
cortex (overadditivity; see Figure 5.19) probably because driving with dis-
traction requires increased attentional or cognitive control within the pre-
frontal cortex.
Figure 5.19
Dual-task performance is often associated with overadditivity due to Effects of an audio
increased activity within the prefrontal cortex (especially the lateral pre- distraction task on brain
frontal cortex) during dual-task performance (see Strobach et al., 2018, for activity associated with a
a review). However, most such findings do not show that this increased straight driving task. There
prefrontal activation is actually required for dual-task performance. were significant increases
More direct evidence that prefrontal areas associated with attentional in activation within the
ventrolateral prefrontal
or cognitive control are causally involved in enhancing dual-task perfor-
cortex and the auditory
mance was reported by Filmer et al. (2017) and Strobach et al. (2018). cortex (in orange). There
Filmer et al. (2017) studied the effects of transcranial direct current stim- was decreased activation
ulation (tDCS; see Glossary) applied to areas of the prefrontal cortex in occipital-visual areas (in
associated with cognitive control. Anodal tDCS during training enhanced blue).
cognitive control and subsequent dual-task performance. From Schweizer et al. (2013).
222 Visual perception and attention
Evaluation
Brain activity in dual-task conditions often differs from the sum of brain
activity of the same two tasks performed singly. Dual-task activity can
exhibit underadditivity or overadditivity. The findings are theoretically
important because they indicate performance of dual tasks can involve
much more cognitive control and other processes than single tasks. Garner
and Dux’s (2015) findings demonstrate that enhanced dual-task perfor-
mance with practice can depend on increased differentiation between the
two tasks with respect to processing and brain activation.
What are the limitations of the cognitive neuroscience approach?
First, increased (or decreased) activity in a given brain area in dual-task
conditions is not necessarily very informative. For example, Dux et al.
(2009) found dual-task performance improved over time because prac-
tice increased the speed of information processing in the prefrontal cortex
rather than because it changed activation within that region. Second, it is
often unclear whether patterns of brain activation are directly relevant to
task processing rather than reflecting non-task processing.
Third, findings in this area are rather inconsistent (Strobach et al., 2018)
and we lack a comprehensive theory to account for these inconsistencies.
Plausible reasons for these apparent inconsistencies are the great variety of
task combinations used in dual-task studies and individual differences in
task proficiency among participants (Watanabe & Funahashi, 2018).
the frequency with which participants switched their attentional focus from KEY TERMS
one task to the other. This led researchers to use much simpler tasks so that
Psychological refractory
they had “better experimental control over the timing of the component period (PRP) effect
task processes” (Koch et al., 2018, p. 575). The slowing of the
The dominant paradigm in recent research is as follows. There are two response to the second
stimuli (e.g., two lights) and two responses (e.g., button presses), one asso- of two stimuli when
presented close together
ciated with each stimulus. Participants respond to each stimulus as rapidly
in time.
as possible. When the two stimuli are presented at the same time (dual-task
condition), performance is typically worse on both tasks than when each Stimulus onset
asynchrony (SOA)
task is presented separately (single-task conditions).
Time interval between the
When the second stimulus is presented shortly after the first, there start of two stimuli.
is typically a marked slowing of the response to the second stimulus:
the psychological refractory period (PRP) effect. This effect is robust –
Ruthruff et al. (2009) obtained a large PRP effect even when participants
were given strong incentives to eliminate it.
The PRP effect has direct real-world relevance. Hibberd et al. (2013)
studied the effects of a simple in-vehicle task on braking performance when
the vehicle in front braked and slowed down. There was a classic PRP
effect – braking time was slowest when the in-vehicle task was presented
immediately before the vehicle in front braked.
How can we explain the PRP effect? It is often argued task perfor-
mance involves three successive stages: (1) perceptual; (2) central response
selection; and (3) response execution. According to the bottleneck model
(e.g., Pashler, 1994),
The response selection stage of the second task cannot begin until the
response selection stage of the first task has finished, although the
other stages . . . can proceed in parallel . . . according to this model,
the PRP effect is a consequence of the waiting time of the second task
because of a bottleneck at the response selection stage.
(Mittelstädt & Miller, 2017, p. 89)
The bottleneck model explains several findings. For example, consider the
effects of varying the time between the start of the first and second stimuli
(stimulus onset asynchrony (SOA)). According to the model, processing on
the first task should slow down second-task processing much more when
the SOA is small than when it is larger. The predicted finding is generally
obtained (Mittelstädt & Miller, 2017).
The bottleneck model remains the most influential explanation of the
PRP effect (and other dual-task costs). However, resource models (e.g.,
Navon & Miller, 2002) are also influential. According to these models,
limited processing capacity can be shared between two tasks so both are
processed simultaneously. Of crucial importance, sharing is possible even
during the response selection process. A consequence of sharing processing
capacity across task is that each task is processed more slowly than if per-
formed on its own.
Many findings can be explained by both models. However, resource
models are more flexible than bottleneck models. Why is this? Resource
models assume the division of processing resources between two tasks
varies freely to promote efficient performance.
KEY TERM Another factor influencing the PRP effect is crosstalk (the two tasks
interfere directly with each other). This mostly occurs when the stimuli
Crosstalk
In dual-task conditions, and/or responses on the two tasks are similar. A classic example of cross-
the direct interference talk is when you try to rub your stomach in circles with one hand while
between the tasks that is patting your head with the other hand (try it!).
sometimes found. Finally, note that participants in most studies receive only modest
amounts of practice in performing two tasks at the same time. As a con-
sequence, the PRP effect may occur at least in part because participants
receive insufficient practice to eliminate it.
Findings
According to the bottleneck model, we would expect to find a PRP effect
even when easy tasks are used and/or participants receive prolonged prac-
tice. Contrary evidence was reported by Schumacher et al. (2001). They
used two tasks: (1) say “one”, “two” or “three” to low-, medium- and
high-pitched tones, respectively; (2) press response keys corresponding to
the position of a disc on a computer screen. These tasks were performed
together for over 2,000 trials, by which time some participants performed
them as well together as singly.
Strobach et al. (2013) conducted a study very similar to that of
Schumacher et al. (2001). Participants took part in over 5,000 trials involv-
ing single-task or dual-task conditions. However, dual-task costs were not
eliminated after extensive practice: dual-task costs on the auditory task
reduced from 185 to 60 ms and those on the visual task from 83 to 20 ms
(see Figure 5.20). How did dual-task practice benefit performance? Practice
speeded up the central response selection stage in both tasks.
Why did the findings differ in the two studies discussed above? In both
studies, participants were rewarded for fast responding on single-task and
dual-task trials. However, the way the reward system was set up in the
Schumacher et al. study may have led participants to exert more effort in
dual-task than single-task trials. This potential bias was absent from the
Figure 5.20
Reaction times for correct
responses only over eight
experimental sessions
under dual-task (auditory
and visual tasks) and single-
task (auditory or visual task)
conditions.
From Strobach et al. (2013).
Reprinted with permission of
Springer.
Attention and performance 225
Strobach et al. study. This difference in reward structure could explain the KEY TERM
greater dual-task costs in the Strobach et al. study.
Backward crosstalk
Hesselmann et al. (2011) studied the PRP effect using event-related effect
potentials. The slowing of responses on the second task was closely Aspects of Task 2
matched by slowing in the onset of the P300 (an ERP component reflecting influence response
response selection). However, there was no slowing of earlier ERP compo- selection and
performance speed on
nents reflecting perceptual processing. Thus, as predicted by the bottleneck
Task 1 in studies on the
model, the PRP effect depended on response selection rather than percep- psychological refractory
tual processes. period (PRP) effect.
According to the resource model approach, individuals choose whether
to use serial or parallel processing on PRP tasks. Miller et al. (2009) argued
that serial processing generally leads to superior performance compared
with parallel processing. However, parallel processing should theoretically
be superior when the stimuli associated with the two tasks are mostly pre-
sented close in time. As predicted, there was a shift from predominantly
serial processing towards parallel processing when that was the case.
Miller et al. (2009) used very simple tasks and it is likely parallel pro-
cessing is most likely to be used with such tasks. Han and Marois (2013)
used two tasks, one of which was relatively difficult. Participants used
serial processing even when parallel processing was encouraged by finan-
cial rewards.
Finally, we consider the theoretically important backward crosstalk
effect: “characteristics of Task 2 of 2 subsequently performed tasks influ-
ence Task 1 performance” (Janczyk et al., 2018, p. 261). Hommel (1998)
obtained this effect. Participants responded to Task 1 by making a left or
right key-press and to Task 2 by saying “left” or “right”. Task 1 responses
were faster when the two responses were compatible (e.g., press right key
+ say “right”) than when they were incompatible (e.g., press right key +
say “left”). Evidence for the backward crosstalk effect was also reported by
Janczyk et al. (2018).
Why is the backward crosstalk effect theoretically important? It indi-
cates that aspects of response selection processing on Task 2 occur before
response selection processing on Task 1 has finished. This effect is incom-
patible with the bottleneck model, which assumes response selection on
Task 1 is completed prior to any response selection on Task 2. In other
words, this model assumes there is serial processing at the response selec-
tion stage. In contrast, the backward crosstalk effect is compatible with the
resource model approach.
model approach. Fischer et al. (2018) also found evidence for much flex-
ibility. There was less interference between the two tasks when finan-
cial rewards were offered because participants devoted more processing
resources to protecting the first task from interference. However, the
resource model approach has the disadvantage compared to the bottleneck
model that its predictions are less precise, making it harder to submit to
detailed empirical testing.
Finally, as Koch et al. (2017, p. 575) pointed out, the bottleneck
model “can be applied (with huge success) mainly for conditions in which
two tasks are performed strictly sequentially”. This is often the case with
research on the psychological refractory period effect but is much less
applicable to more complex dual-task situations.
“AUTOMATIC” PROCESSING
We have seen in studies of divided attention that practice often causes a dra-
matic improvement in performance. This improvement has been explained
by assuming some processes become automatic through prolonged prac-
tice. For example, the huge amount of practice we have had with reading
words has led to the assumption that familiar words are read “automati-
cally”. Below we consider various definitions of “automaticity”. We also
consider different approaches to explaining the development of automatic
processing.
Figure 5.21
Response times on a
decision task as a function
of memory-set size, display-
set size and consistent vs
varied mapping.
Response times on a
decision task as a function
of memory-set size, display-
set size and consistent
vs varied mapping. Data
from Shiffrin and Schneider
(1977).
American Psychological
Association.
Limitations
What are the limitations with this approach? First, the distinction between
automatic and controlled processes is oversimplified (discussed below).
Second, Shiffrin and Schneider (1977) argued automatic processes operate
in parallel and place no demands on attentional capacity and so decision
speed should be unrelated to the number of items. However, decision
228 Visual perception and attention
speed was slower when the memory set and visual display both contained
several items (see Figure 5.21). Third, the theory is descriptive rather than
explanatory – it does not explain how serial controlled processing turns into
parallel automatic processing.
Definitions of automaticity
Shiffrin and Schneider (1977) assumed there is a clear-cut distinction
between automatic and controlled processes. More specifically, automatic
processes possess several features (e.g., inflexibility; very efficient because
they have no capacity limitations; occurring in the absence of attention). In
essence, it is assumed there is perfect coherence or consistency among the
features (i.e., they are all found together).
Moors and De Houwer (2006) and Moors (2016) identified four key
features associated with automaticity:
Why might these four features (or the similar ones identified by Shiffrin
and Schneider (1977)) often be found together? Instance theory (Logan,
1988; Logan et al., 1999) provides an influential answer. It is assumed
task practice leads to storage of information in long-term memory which
facilitates subsequent performance on that task. In essence, “Automaticity
is memory retrieval: performance is automatic when it is based on a
single-step direct-access retrieval of past solutions from memory” (Logan,
1988, p. 493). For example, if you were given the problem “24 × 7 = ???”
numerous times, you would retrieve the answer (168) “automatically”
without performing any mathematical calculations.
Instance theory makes coherent sense of several characteristics of
automaticity. Automatic processes are fast because they require only the
retrieval of past solutions from long-term memory. They make few demands
on attentional resources because the retrieval of heavily o ver-learned
information is relatively effortless. Finally, there is no conscious awareness
of automatic processes because no significant processes intervene between
stimulus presentation and retrieval of the correct response.
In spite of its strengths, instance theory is limited (see Moors, 2016).
First, the theory implies the key features of automaticity will typically all
be found together. However, this is not the case (see below). Second, it
is assumed practice leads to automatic retrieval of solutions with learners
having no control over such retrieval. However, Wilkins and Rawson (2011)
found evidence learners can exercise top-down control over retrieval: when
the instructions emphasised accuracy, there was less evidence of retrieval
Attention and performance 229
than when they emphasised speed. Thus, the use of retrieval after practice
is not fully automatic.
Melnikoff and Bargh (2018) argued that the central problem with the
traditional approach is that no research has shown the four features asso-
ciated with “automaticity” occurring together. As they pointed out, “No
attempt has been made to estimate the probability of a process being inten-
tional given that is conscious versus unconscious, or the probability of a
process being controllable given that it is efficient versus inefficient, and so
forth” (p. 282).
2nd threshold
Current stimulus representation factors Unconscious
Current stimulus factors Attention
• Quality of stimulus representation: duration, processing
• Stimulus quality: duration, intensity
• Un/expectedness intensity, distinctiveness
• Goal in/congruence ~ Accessibility of stimulus representation for 1st threshold
• Novelty/familiarity processing
Figure 5.22
Factors that are hypothesised to influence representational quality within Moors’ (2016) theoretical approach.
From Moors (2016).
230 Visual perception and attention
Findings
According to Moors’ (2016) theoretical framework, there is a flexible rela-
tionship between controlled and conscious processing. This contrasts with
Schneider and Shiffrin’s (1977) assumption that executive control is always
associated with conscious processing.
Diao et al. (2016) reported findings consistent with Moors’ prediction.
They used a Go/No-Go task where participants made a simple response
(Go trials) or withheld it (No-Go trials). High-value or low-value financial
rewards were available for successful performance. Task stimuli were pre-
sented above or below the level of conscious awareness.
What did Diao et al. (2016) find? Performance was better on high-
reward than low-reward trials even when task processing was uncon-
scious. In addition, participants showed superior unconscious inhibitory
control (assessed by event-related potentials) on high-reward trials. Thus,
one feature of automaticity (unconscious processing) was present whereas
another feature (goal-uncontrolled) was not.
Huber-Huber and Ansorge (2018) also reported problems for the tra-
ditional approach. Participants received target words indicating an upward
or downward direction (e.g., above; below). Prior to the target word, a
prime word also indicating an upward or downward direction was pre-
sented below the level of conscious awareness. Response times to the target
words were slower when there was a conflict between the meanings of the
prime and target words than when they were congruent in meaning. As in
the study by Diao et al. (2016), unconscious processing was combined with
control, a combination that is inconsistent with the traditional approach.
Evaluation
The theoretical approach to automaticity proposed by Moors (2016) has
several strengths. First, the assumption that various features associated with
automaticity often correlate poorly with each other is clearly superior to
the earlier notion that these features exhibit perfect coherence. Second, her
assumption that processes vary in the extent to which they are “automatic”
is much more realistic than the simplistic division of processes into auto-
matic and non-automatic. Third, the approach is more comprehensive than
previous ones because it considers more factors relevant to “automaticity”.
What are the limitations with Moors’ (2016) approach? First, numer-
ous factors are assumed to influence representational quality (and thus the
extent to which processes are automatic) (see Figure 5.22). It would thus
require large-scale experimental research to assess the ways all these factors
interact. Second, the approach provides only a partial explanation of the Interactive exercise:
underlying mechanisms causing the various factors to influence representa- Definitions of attention
tional quality.
CHAPTER SUMMARY
• Focused auditory attention. When two auditory messages
are presented at the same time, there is less processing of the
unattended than the attended message. Nevertheless, unattended
messages often receive some semantic processing. The restricted
processing of unattended messages may reflect a bottleneck at
various stages of processing. However, theories assuming the
existence of a bottleneck de-emphasise the flexibility of selective
auditory attention. Attending to one voice among several (the
cocktail party problem) is a challenging task. Human listeners use
top-down and bottom-up processes to select one voice. Top-
down processes include the use of various control processes
(e.g., focused attention; inhibitory processes) and learning about
structural consistencies present in the to-be-attended voice.