2022wirescogsci Noyce
2022wirescogsci Noyce
DOI: 10.1002/wcs.1610
PERSPECTIVE
1 | INTRODUCTION
Attention is a set of processes that modulate what information gets represented in the brain. These processes act
similarly—and are even shared—across auditory and other sensory modalities.
In 1953, Colin Cherry ignited research on the cocktail party problem, describing how attention alters perception in a
crowded setting with multiple sound sources (Cherry, 1953). However, by the 1990s, many hearing researchers, driven
to optimize information preserved in telephone communication and to understand the impact of hearing loss, had
turned to developing quantitative, bottom-up models that assumed an ideal observer (Egan, 1971; Fletcher &
Galt, 1950; Henning, 1967; Siebert, 1970). This approach specifically ignores any role of central cognitive processes like
memory and attention. Ideal observer models account well for performance on simple psychoacoustic tasks, but not for
how we cope with the cacophony of sounds in daily life.
Because vision scientists dominated attention studies in the ensuing decade (Corbetta & Shulman, 2002; Hopfinger
et al., 2000; Posner & Petersen, 1990; Treisman, 2006; Yantis, 2008), many discussions of attention and its underlying
brain mechanisms focused on results, paradigms, and stimuli from visual research. For instance, many visual para-
digms present a static scene of competing objects that, to make tasks demanding, is flashed on only briefly. Auditory
information, however, is intrinsically temporal: information is primarily conveyed by changes in time-varying acoustic
signals, such as amplitude and frequency modulation. There is no meaning in a “snapshot” of auditory stimuli. The
dominance of visual studies led to an emphasis on how visual features such as spatial adjacency influence the deploy-
ment of attention. Conversely, there was little work examining how attention operates through time to track dynamic
This is an open access article under the terms of the Creative Commons Attribution-NonCommercial License, which permits use, distribution and reproduction in any
medium, provided the original work is properly cited and is not used for commercial purposes.
© 2022 The Authors. WIREs Cognitive Science published by Wiley Periodicals LLC.
objects, an issue critical to how attention works to track an auditory source that extends through time (known as a
“stream”). Here, as auditory researchers, we argue that most attention effects identified by visual studies operate analo-
gously in auditory attention; in addition, however, there are key temporal phenomena critical to how attention governs
auditory attention that likely impact visual attention as well.
Another key issue is that attentional abilities differ across individuals. For instance, people with sensorineural hear-
ing deficits or who use cochlear implants often have trouble focusing auditory attention (Dai et al., 2018; Shinn-
Cunningham & Best, 2008). Certain neurological disorders can compromise top-down cognitive control important for
focusing attention, including those with post-traumatic stress disorder (Bressler et al., 2017; Leskin & White, 2007;
Vasterling et al., 1998), autism spectrum disorder (Schwartz et al., 2020), schizophrenia (Mathalon et al., 2004), and
attention deficit and hyperactivity disorder (Hasler et al., 2016). Understanding the mechanisms of attention, including
how attention operates through time, are thus important for developing treatments and interventions to help individ-
uals who have trouble deploying attention, whether in vision or audition.
2 | D E S C R I B I N G A T T EN T I O N
Attention prioritizes certain information at the expense of other information. It determines whether we notice the broc-
coli on our plate or the serving dish it came from, and whether we register the gravel under our feet as we walk in the
park. Attention depends not only on how well we focus volitionally (top-down attention), but also on automatic
responses to salient events (bottom-up attention). It can be used to prepare for an event, to follow an information source
over time, and to reorient after a distraction.
Attention allows us to cope with the fact that human cognition is capacity limited. We cannot fully process the bar-
rage of information reaching us; instead, our brains favor processing information that is “important.” In top-down
attention, the observer consciously decides what to process based on their goals, while in bottom-up attention, the brain
blithely ignores predictable, expected information while automatically prioritizing new and unexpected events (e.g., a
crash of thunder or flash of lightning). The schematic in Figure 1 shows examples of top-down, object-based, and
bottom-up attention in an auditory scene with several sources.
Mechanisms of attention are distributed across every stage of cognitive processing. Throughout the cortex,
feedforward and feedback connections increase the representation of attended stimuli and decrease the representation
of ignored stimuli (Yantis, 2008). Attention can be directed either externally, to sensory input, or internally, to a mem-
ory representation (Panichello & Buschman, 2021), and even modify what gets encoded into memory (Payne
et al., 2013; Payne & Sekuler, 2014).
3 | TOP-DOWN ATTENTION
Top-down attention allows us to consciously focus on upcoming events, select a specific feature in a scene, or search for
a target amongst distractors. It is top-down attention that, in a setting with competing for sensory inputs, allows us to
deliberately bias what is enhanced and what is suppressed. For instance, Figure 1a illustrates the process of a listener
preparing to attend to a source from a known (desired) location; once that stream begins, attention selects that stream
by focusing an “attentional spotlight.”
When people successfully focus attention on a target, the neural responses to that target are magnified, and those to
competing distractors are suppressed (Clark & Hillyard, 1996; Foster et al., 2020; Woldorff et al., 1993). For example,
when subjects are instructed to listen to one source amongst overlapping sounds, attention enhances the event-related
potential (ERP) evoked by the onsets of events in the attended source and reduces ERPs to events in the other sources
(Choi et al., 2014; Hillyard et al., 1973).
Figure 2a shows a typical paradigm for studying top-down attention in audition. A listener is cued to attend to one
spatial location. Then, three melodies of complex tones are presented from the listener's left, center, and right.
Figure 2b shows grand average ERPs recorded during such a task. Onset times of tones in the left and right melodies
are shown by the blue and red vertical lines, respectively. The traces show the average voltage over the course of a trial.
The negative peaks approximately 0.15 s after each tone's onset reflect sensory processing. When subjects attend to the
left melody (blue traces), the peaks elicited by the left tones (blue circles) are larger than when subjects attend to the
right stream (red traces); the converse is true for peaks elicited by the right (red) stream. Many studies with similar
NOYCE ET AL. 3 of 11
Non-spatial features
Spa res (a) (b) (c) (d)
fea
tu
tial
Time
F I G U R E 1 Auditory stimuli can be considered in a multidimensional space, with axes including time, spatial features (such as
interaural differences and spectral location cues), and nonspatial features, such as pitch and timbre, here collapsed into a single dimension
for visualization. This figure shows examples of how attention might behave in an auditory scene with several sources, depicted as acoustic
waveforms. (a) A prestimulus cue, such as “Listen to the left stream”, can orient a listener to engage top-down attention (visualized as the
highlighted area on the plots) towards an upcoming stream with the desired spatial features. Once that stream (illustrated in red) is selected,
attention focuses down to also emphasize other, nonspatial features of that stream. (b) Object-based attention allows the listener to follow
the object across momentary silent gaps. (c) When two sound sources (illustrated in red and yellow) are near each other in feature space,
they can be confused: Attention may begin tracking the distractor instead of the target, especially at moments of ambiguity, such as an
instantaneous silence in the target. (d) The sudden appearance of a new stream (illustrated in purple) can capture attention involuntarily
(via bottom-up processes), regardless of its similarity to the previous attentional target.
−3
−6 “Listen Right”
F I G U R E 2 (a) Schematic of a typical top-down auditory attention task. A subject is cued to attend to one of three spatially separated
streams of sound and report its contents. (b) ERP traces (electrode Cz) while subjects attend to either a left-spatialized melody (onset times
shown by blue vertical lines) or a right-spatialized melody (red lines). The auditory cue telling subjects to listen to either the left or right
stream occurs at the time denoted by the vertical black line. When subjects attend to the left stream (blue traces), negative-going ERPs
elicited by notes in the left melody are enhanced (blue circles) and those elicited by notes from the right are suppressed, and vice versa. Data
from Choi et al. (2014)
results demonstrate that attention modulates neural responses, explicitly altering information's representation in the
brain. Note that the very first tone onset elicits equally strong responses regardless of the task instructions, an example
of bottom-up attentional capture (see below).
In vision studies, researchers distinguish between focusing attention to a particular location (e.g., 15 degrees left of
fixation) versus to some nonspatial attribute (e.g., a red object). In part, this reflects the fact that visual inputs at the
4 of 11 NOYCE ET AL.
retina are encoded according to where objects are relative to an observer's eyes (an organization that is topographically
preserved throughout the visual system), while other features must be centrally computed. In contrast, the auditory sys-
tem is not organized by spatial location. Sound location is itself a feature that must be computed from the signals
reaching the two ears. Still, listeners can focus attention to sounds based on either location or other features, such as
pitch, timbre, or talker identity. And, as we discuss below, even though both spatial and nonspatial acoustic features
must be centrally computed, spatial auditory attention recruits different brain networks than does nonspatial auditory
attention, paralleling functional differences between spatial and feature-based attention in vision.
Most headphone-based studies of auditory spatial attention use simplified spatialization cues (such as a pure time
delay between the signals at the two ears), ignoring the frequency-dependent differences in the levels and timing of the
signals that would reach the left and right ears in the real world. Many even use dichotic presentations, in which
entirely separate streams of information are presented to each ear (Hashimoto et al., 2000; Jäncke et al., 2003;
Kimura, 1967; Treisman, 1960). Listeners can successfully direct spatial attention even with such unnatural, degraded
spatialization (Baumgartner et al., 2017; Ross et al., 2010). Some dichotic studies find that people can follow the stream
in the right ear better than the left-ear stream, an effect that has been ascribed to left-lateralized language specialization
in the brain (Hiscock & Kinsbourne, 2011; Kimura, 1961). However, recent work suggests that the right-ear bias may
instead arise from increased powers of attention to the right hemifield (Payne et al., 2017; Tanaka et al., 2021). In addi-
tion, it is not clear whether such asymmetries arise with natural binaural spatial cues; at least one study has shown that
unrealistic spatial simulations do not fully engage the brain networks devoted to spatial auditory attention (Deng, Choi,
et al., 2019).
4 | OBJECT-BASED ATTENTION
When a person focuses on an object because it has a desired feature (such as location, pitch, etc.), top-down attention
tends to focus on the entire object. In other words, the natural unit of attention seems to be an object: a collection of
features that the brain believes came from the same external source (Duncan, 1984; Shinn-Cunningham, 2008).
Figure 1a illustrates this idea: attention focuses initially on one location; however, once a stream at that location
becomes the focus of attention, attentional focus narrows onto the nonspatial features of the attended source.
Relatedly, auditory objects extend across time, often even across silences. For instance, speech includes momentary
gaps, yet is perceived as one perceptual stream. Once we attend an element of a stream (e.g., a syllable in ongoing
speech), that stream tends to remain the focus of attention over time, automatically (Best et al., 2008, 2010; Billig &
Carlyon, 2016; Bressler et al., 2014; Woods & McDermott, 2015), and attention becomes more tightly focused to that
stream (Choi et al., 2014; Golumbic et al., 2013). Figure 1b illustrates this idea: attention remains focused to pick up the
next syllable from the ongoing stream even after a momentary gap.
When two streams are similar in spatial and other features, they can be confused across time. Figure 1c illustrates
this: a competing stream that is similar to the attended stream in high-dimensional feature space can be confused with
the attended stream, especially after a momentary silence in the attended stream. When a listener is asked to focus
attention on the speech from one location (e.g., “always report the words spoken from the left”) and the left talker sud-
denly switches locations with a competing talker, subjects often track the original talker rather than the original loca-
tion (Mehraei et al., 2018). That is, successful attention relies on object formation and segregation through time, and
object-based attention can even override volitional attention.
This critical role of attention as a mechanism for following an object through time is most evident in the auditory
system, where the temporal dynamics of scenes and stimuli cannot be ignored. However, this also affects vision and
other senses. One of the most challenging visual attention paradigms—multiple object tracking—requires subjects to
use attention to track moving objects among identical distractors (Alvarez & Cavanagh, 2004; Pylyshyn & Storm, 1988).
Similarly, everyday interactions with our surroundings require us to attend to objects over time: to catch a ball, to judge
whether it is safe to cross a street, or to fill a water glass without it overflowing.
5 | BOTTOM-UP ATTENTION
Attention can also be drawn by unusual or salient sensory inputs. For example, a plate smashing on the floor involun-
tarily captures attention. Automatic attentional capture occurs even for less dramatic events, such as the onset of a tone
NOYCE ET AL. 5 of 11
or appearance of an image, disrupting top-down and object-based attention (Bulger et al., 2020; Maryott et al., 2011;
Noyce & Sekuler, 2014a). Figure 1d shows an example of a new sound source's onset grabbing attention involuntarily,
away from the previously attended object.
In the audition, involuntary attentional capture interferes with top-down auditory attention regardless of the spatial
location of the interrupting event. Figure 3 illustrates this effect when a listener is attending a stream comprising three
speech syllables from one direction and ignoring a similar, competing stream from the opposite hemifield (taken from
Liang et al., 2022). Top-down attention allows the listener to select one stream, and object-based attention lets them fol-
low it over time. But when an unexpected sound (a cat “MEOW”) occurs just before the second target syllable, it cap-
tures attention, disrupting the ongoing processing and requiring the listener to reorient back to the intended target.
This interference is equally strong whether the MEOW comes from the same hemifield as the target or the opposite
hemifield.
Another interesting point demonstrated in Figure 3 is that salient interruptions interfere not only with attention per
se, but also with storing attended information in memory. In particular, the interrupting MEOW harms performance
on the first target syllable, which finishes playing before the MEOW even begins (Liang et al., 2022). Indeed, roughly
two-third of the 45 subjects tested in this study showed a significant decrease in their ability to recall the first syllable
due to the interrupting MEOW. Thus, attentional capture not only interferes with ongoing attention to a stream, but
with the memory encoding and storage of items that already occurred (see also Section 7).
Attention can also be disrupted, requiring subsequent reorienting, when the attended stream contains task-
irrelevant discontinuities. For instance, when subjects are instructed to attend to a source at a particular location,
changes in pitch impair performance; when subjects are attending to sources with a particular pitch, changes in location
impair performance (Maddox & Shinn-Cunningham, 2012). In both cases, the degree of disruption increases with the
magnitude of the task-irrelevant discontinuity. In some cases, this may be due to attention following a perceived object
in opposition to the task instructions (Mehraei et al., 2018), but disruptions of attention from stimulus discontinuity
can happen even if there is no competition for attention from other sources. For instance, if a listener is attending to
75%
Contralateral interrupter
50%
Ipsilateral interrupter
25%
1st syllable 2nd syllable 3rd syllable
F I G U R E 3 (Left) Schematic of the spatial locations of two competing speech streams from the same talker, and, when present, an
interrupting cat “MEOW.” Subjects direct top-down spatial attention to follow either the left or right speech stream. The interrupting
MEOW, which happens randomly on one-fourth of trials, appears randomly from either the same or opposite hemifield as the target stream.
(Right) Syllable identification accuracy. The MEOW significantly affects the recall of not only the second and third, but also the first syllable.
Critically, contralateral and ipsilateral interrupters are equally disruptive
6 of 11 NOYCE ET AL.
an isolated stream of speech, but the talker suddenly changes, attention to the stream is disrupted and must be
re-established (Lim et al., 2019; Mullennix & Pisoni, 1990; Nusbaum & Morin, 1992). Both electroencephalography
(EEG) and pupillometry show strong responses to task-irrelevant stimulus discontinuities, supporting the idea that such
discontinuities trigger bottom-up attention that requires re-orienting within the scene (Lim et al., 2021; Mehraei
et al., 2018).
6 | B R A I N NE T W O R K S C O N T R O L L I N G A T T E N T I O N
The most common framework for understanding the brain regions that participate in attention posits two anti-
correlated networks, one for top-down attentional control, and one for bottom-up attentional capture (Corbetta &
Shulman, 2002; Fox et al., 2005; Power et al., 2011; Yeo et al., 2011). However, evidence for these two networks comes
primarily from visual attention studies, in part due to the challenges of auditory studies in fMRI (Peelle, 2014). (See Lee
et al., 2014 for a review of auditory exceptions.)
We find that there are at least two similar, but differently specialized networks recruited for top-down attention
(Figure 4). One, the well-established frontoparietal network, is specialized for visual and spatial attention. However, a
second auditory-biased attention network includes distinct regions in lateral frontal cortex that interleave anatomically
with the frontoparietal network (Braga et al., 2013; Michalka et al., 2015; Noyce et al., 2017; Tobyne et al., 2018). These
complementary networks extend broadly throughout frontal cortex (Noyce et al., 2021; Tobyne et al., 2017).
The complementary affinities of vision for spatial information and audition for timing information play out in how
tasks recruit sensory-biased cortical networks. A rhythm memory task with purely visual stimuli recruits not only the
visual-biased frontal network, but also the auditory-biased network (Michalka et al., 2015). A spatial memory task with
purely auditory stimuli recruits the auditory-biased network, as well as visual-biased regions of both lateral frontal cor-
tex (Michalka et al., 2015) and anterior parietal cortex (Michalka et al., 2016). Magnetoencephalography (MEG) and
EEG reveal further evidence of parietal recruitment for spatial—but not nonspatial—auditory attention (Lee
et al., 2012). Contralateral alpha-band (8–14 Hz) oscillations over parietal cortex, long associated with visual–spatial
attention (Klimesch et al., 1999; Worden et al., 2000), also occur during auditory spatial attention, even tracking syllabic
events over time (Bonacci et al., 2019, 2020; Deng et al., 2020; Wöstmann et al., 2021). Contralateral transcranial alter-
nating current stimulation (tACS) in the alpha frequency disrupts spatial but not nonspatial auditory attention (Deng,
Reinhart, et al., 2019; Wöstmann et al., 2018).
Regardless of whether spatial or nonspatial attention guides auditory selective attention, the net result is similar:
information about the attended objects is represented more robustly in the brain. This is seen as a relative enhancement
of ERPs elicited by attended objects compared to those elicited by unattended objects (as illustrated in Figure 2),
stronger neural entrainment to the envelope of an attended vs. an unattended stream (Golumbic et al., 2013;
F I G U R E 4 Maps of bilateral visual-biased and auditory-biased regions in frontal cortex. Three visual-biased regions and five auditory-
biased regions control attention and working memory tasks. Adapted from Noyce et al. (2021)
NOYCE ET AL. 7 of 11
F I G U R E 5 Memory performance (d0 ) for short sequences, when cued to track either the locations (left) or timing (right, orange) of
events in a sequence. Subjects are better at encoding locations from a visual than an auditory sequence, but better at encoding timing in an
auditory than visual sequence. Adapted from Noyce et al. (2016)
Mesgarani & Chang, 2012), and increased success in decoding neural activity to reconstruct an attended versus an
unattended stimulus (Bednar & Lalor, 2020; Mesgarani & Chang, 2012).
Bottom-up attention has been ascribed to a ventral attention network, which includes nodes in cingulate,
temporoparietal, and opercular cortex, but this work has almost entirely been done in visual paradigms (Corbetta &
Shulman, 2002; Devaney et al., 2019; Dosenbach et al., 2007; Seeley et al., 2007). While few studies have investigated
brain networks controlling attentional reorienting to unexpected auditory events, at least one MEG study hints that the
“visual” ventral network is multisensory: auditory attention switching engages a region within this canonical bottom-
up attention network (Larson & Lee, 2013).
Attention, especially top-down attention, and working memory (the ability to hold and manipulate information) are
intricately linked (Bettencourt et al., 2011; Panichello & Buschman, 2021). Working memory requires attention to be
directed to an internal representation that is maintained after sensory inputs are gone (Lim et al., 2015), and if attention
is disrupted the memory representation degrades (Payne et al., 2013; Unsworth & Robison, 2016). Attentional limits
likely restrict working memory capacity.
Attention also affects how objects are encoded into memory. Attended objects tend to be remembered (Maryott
et al., 2011; Noyce & Sekuler, 2014b); relatedly, events that disrupt attention also impair recall of events surrounding
the disruption (Lim et al., 2019). Even when an object is attended, not all of its features will get into memory: attended
features are encoded more robustly (e.g., Noyce et al., 2016).
Figure 5 shows discrimination performance when subjects are asked to attend to either the locations or time inter-
vals of a short sequence, which is composed of either visual or auditory events; they then must compare their stored
representation of the sequence to a subsequent, similar sequence of events in the other sensory modality. Performance
shows a significant interaction between sensory modality and the feature to be recalled. As we discussed above, atten-
tion co-opts the specialized neural circuitry for visual cognition when performing spatial tasks, and for auditory cogni-
tion when performing temporal tasks (Michalka et al., 2015, 2016). Consistent with this, it is easier to attend to and
stores spatial information from visual inputs, and easier to attend to and store timing information from auditory signals
(Noyce et al., 2016), as supported by better performance in conditions relying on these “natural” memory
representations.
8 | C ON C L U S I ON
Attention performs essential pre-processing for the rest of human cognition. It is a tug of war between our desires and
our distractions, in a finely tuned and complex interplay between top-down, bottom-up, and object-based attention.
8 of 11 NOYCE ET AL.
This allows us to select, to focus, to remember… while not getting eaten by a bear or hit by a taxi. After attention filters
important information from the barrage of neural events, the brain can analyze those important events in detail, letting
us understand speech, appreciate a sunset, or merely cross a busy intersection unscathed.
A U T H O R C ON T R I B U T I O NS
Abigail Noyce: Conceptualization (equal); investigation (equal); supervision (supporting); visualization (equal); writ-
ing – original draft (lead); writing – review and editing (equal). Jasmine Kwasa: Conceptualization (supporting);
investigation (equal); visualization (equal); writing – review and editing (supporting). Barbara Shinn-Cunningham:
Conceptualization (supporting); funding acquisition (lead); investigation (equal); project administration (lead); supervi-
sion (lead); visualization (equal); writing – original draft (supporting); writing – review and editing (lead).
FUNDING INFORMATION
This work was supported by NIDCD R01 DC013825 (to Barbara G. Shinn-Cunningham), ONR project
N00014-20-1-2709 (to Barbara G. Shinn-Cunningham), and NINDS F99 NS115331 (to Jasmine A. C. Kwasa).
CONFLICT OF INTEREST
The authors have declared no conflicts of interest for this article.
ORCID
Abigail L. Noyce https://orcid.org/0000-0002-4339-4113
Jasmine A. C. Kwasa https://orcid.org/0000-0001-5537-6054
Barbara G. Shinn-Cunningham https://orcid.org/0000-0002-5096-5914
R EL ATE D WIR Es AR TI CL E
Visual attention
R EF E RE N C E S
Alvarez, G. A., & Cavanagh, P. (2004). The capacity of visual short-term memory is set both by visual information load and by number of
objects. Psychological Science, 15(2), 106–111.
Baumgartner, R., Reed, D. K., Toth, B., Best, V., Majdak, P., Colburn, H. S., & Shinn-Cunningham, B. (2017). Asymmetries in behavioral and
neural responses to spectral cues demonstrate the generality of auditory looming bias. Proceedings of the National Academy of Sciences of
the United States of America, 114(36), 9743–9748.
Bednar, A., & Lalor, E. C. (2020). Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using
EEG. NeuroImage, 205, 116283.
Best, V., Ozmeral, E. J., Kopco, N., & Shinn-Cunningham, B. G. (2008). Object continuity enhances selective auditory attention. Proceedings
of the National Academy of Sciences of the United States of America, 105(35), 13174–13178.
Best, V., Shinn-Cunningham, B. G., Ozmeral, E. J., & Kopco, N. (2010). Exploring the benefit of auditory spatial continuity. The Journal of
the Acoustical Society of America, 127(6), EL258-64.
Bettencourt, K. C., Michalka, S. W., & Somers, D. C. (2011). Shared filtering processes link attentional and visual short-term memory capac-
ity limits. Journal of Vision, 11(10), 1–9. https://doi.org/10.1167/11.10.22
Billig, A. J., & Carlyon, R. P. (2016). Automaticity and primacy of auditory streaming: Concurrent subjective and objective measures. Journal
of Experimental Psychology. Human Perception and Performance, 42(3), 339–353.
Bonacci, L. M., Bressler, S., & Shinn-Cunningham, B. G. (2020). Nonspatial features reduce the reliance on sustained spatial auditory atten-
tion. Ear and Hearing, 41, 1635–1647. https://doi.org/10.1097/AUD.0000000000000879
Bonacci, L. M., Dai, L., & Shinn-Cunningham, B. G. (2019). Weak neural signatures of spatial selective auditory attention in hearing-
impaired listeners. The Journal of the Acoustical Society of America, 146(4), 2577–2589.
Braga, R. M., Wilson, L. R., Sharp, D. J., Wise, R. J. S., & Leech, R. (2013). Separable networks for top-down attention to auditory non-spatial
and visuospatial modalities. NeuroImage, 74, 77–86.
Bressler, S., Goldberg, H., & Shinn-Cunningham, B. (2017). Sensory coding and cognitive processing of sound in veterans with blast expo-
sure. Hearing Research, 349, 98–110.
Bressler, S., Masud, S., Bharadwaj, H., & Shinn-Cunningham, B. (2014). Bottom-up influences of voice continuity in focusing selective audi-
tory attention. Psychological Research, 78(3), 349–360.
NOYCE ET AL. 9 of 11
Bulger, E., Shinn-Cunningham, B. G., & Noyce, A. L. (2020). Distractor probabilities modulate flanker task performance. Attention, Percep-
tion, & Psychophysics, 83, 866–881. https://doi.org/10.3758/s13414-020-02151-7
Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and with two ears. The Journal of the Acoustical Society of
America, 25(5), 975–979.
Choi, I., Wang, L., Bharadwaj, H., & Shinn-Cunningham, B. (2014). Individual differences in attentional modulation of cortical responses
correlate with selective attention performance. Hearing Research, 314, 10–19.
Clark, V. P., & Hillyard, S. A. (1996). Spatial selective attention affects early extrastriate but not striate components of the visual evoked
potential. Journal of Cognitive Neuroscience, 8(5), 387–402.
Corbetta, M., & Shulman, G. L. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews. Neuroscience,
3(3), 201–215.
Dai, L., Best, V., & Shinn-Cunningham, B. G. (2018). Sensorineural hearing loss degrades behavioral and physiological measures of human
spatial selective auditory attention. Proceedings of the National Academy of Sciences of the United States of America, 115(14), E3286–
E3295.
Deng, Y., Choi, I., & Shinn-Cunningham, B. (2020). Topographic specificity of alpha power during auditory spatial attention. NeuroImage,
207, 116360.
Deng, Y., Choi, I., Shinn-Cunningham, B., & Baumgartner, R. (2019). Impoverished auditory cues limit engagement of brain networks con-
trolling spatial selective attention. NeuroImage, 202, 116151.
Deng, Y., Reinhart, R. M., Choi, I., & Shinn-Cunningham, B. G. (2019). Causal links between parietal alpha activity and spatial auditory
attention. eLife, 8, e51184. https://doi.org/10.7554/eLife.51184
Devaney, K. J., Rosen, M. L., Levin, E. J., & Somers, D. C. (2019). Identification of visual attentional regions of the Temporoparietal junction
in individual subjects using a vivid, novel oddball paradigm. Frontiers in Human Neuroscience, 13, 424.
Dosenbach, N. U. F., Fair, D. A., Miezin, F. M., Cohen, A. L., Wenger, K. K., Dosenbach, R. A. T., Fox, M. D., Snyder, A. Z., Vincent, J. L.,
Raichle, M. E., Schlaggar, B. L., & Petersen, S. E. (2007). Distinct brain networks for adaptive and stable task control in humans. Proceed-
ings of the National Academy of Sciences of the United States of America, 104(26), 11073–11078.
Duncan, J. (1984). Selective attention and the organization of visual information. Journal of Experimental Psychology. General, 113(4),
501–517.
Egan, J. P. (1971). Auditory masking and signal detection theory. Audiology, 10(1), 41–47.
Fletcher, H., & Galt, R. H. (1950). The perception of speech and its relation to telephony. The Journal of the Acoustical Society of America,
22(2), 89–151.
Foster, J. J., Bsales, E. M., & Awh, E. (2020). Covert spatial attention speeds target individuation. The Journal of Neuroscience, 40, 2717–2726.
https://doi.org/10.1523/JNEUROSCI.2962-19.2020
Fox, M. D., Snyder, A. Z., & Vincent, J. L. (2005). The human brain is intrinsically organized into dynamic, anticorrelated functional net-
works. Proceedings of the National Academy of Sciences of the United States of America, 102(27), 9673–9678.
Golumbic, E. M. Z., Ding, N., Bickel, S., Lakatos, P., Schevon, C. A., McKhann, G. M., Goodman, R. R., Emerson, R., Mehta, A. D.,
Simon, J. Z., Poeppel, D., & Schroeder, C. E. (2013). Mechanisms underlying selective neuronal tracking of attended speech at a “cocktail
party”. Neuron, 77(5), 980–991.
Hashimoto, R., Homae, F., Nakajima, K., Miyashita, Y., & Sakai, K. L. (2000). Functional differentiation in the human auditory and language
areas revealed by a dichotic listening task. NeuroImage, 12(2), 147–158.
Hasler, R., Perroud, N., Meziane, H. B., Herrmann, F., Prada, P., Giannakopoulos, P., & Deiber, M.-P. (2016). Attention-related EEG markers
in adult ADHD. Neuropsychologia, 87, 120–133.
Henning, G. B. (1967). A model for auditory discrimination and detection. The Journal of the Acoustical Society of America, 42(6), 1325–1334.
Hillyard, S. A., Hink, R. F., Schwent, V. L., & Picton, T. W. (1973). Electrical signs of selective attention in the human brain. Science,
182(4108), 177–180.
Hiscock, M., & Kinsbourne, M. (2011). Attention and the right-ear advantage: What is the connection? Brain and Cognition, 76(2), 263–275.
Hopfinger, J. B., Buonocore, M. H., & Mangun, G. R. (2000). The neural mechanisms of top-down attentional control. Nature Neuroscience,
3(3), 284–291.
Jäncke, L., Specht, K., Shah, J. N., & Hugdahl, K. (2003). Focused attention in a simple dichotic listening task: An fMRI experiment. Brain
Research. Cognitive Brain Research, 16(2), 257–266.
Kimura, D. (1961). Cerebral dominance and the perception of verbal stimuli. Canadian Journal of Psychology, 15(3), 166–171.
Kimura, D. (1967). Functional asymmetry of the brain in dichotic listening. Cortex; a Journal Devoted to the Study of the Nervous System and
Behavior, 3(2), 163–178.
Klimesch, W., Doppelmayr, M., Schwaiger, J., Auinger, P., & Winkler, T. (1999). ‘Paradoxical’ alpha synchronization in a memory task.
Brain Research. Cognitive Brain Research, 7(4), 493–501.
Larson, E., & Lee, A. K. C. (2013). The cortical dynamics underlying effective switching of auditory spatial attention. NeuroImage, 64,
365–370.
Lee, A. K. C., Larson, E., Maddox, R. K., & Shinn-Cunningham, B. G. (2014). Using neuroimaging to understand the cortical mechanisms of
auditory selective attention. Hearing Research, 307, 111–120.
10 of 11 NOYCE ET AL.
Lee, A. K. C., Rajaram, S., Xia, J., Bharadwaj, H., Larson, E., Hämäläinen, M. S., & Shinn-Cunningham, B. G. (2012). Auditory selective
attention reveals preparatory activity in different cortical regions for selection based on source location and source pitch. Frontiers in
Neuroscience, 6, 190.
Leskin, L. P., & White, P. M. (2007). Attentional networks reveal executive function deficits in posttraumatic stress disorder. Neuropsychology,
21(3), 275–284.
Liang, W., Brown, C. A., & Shinn-Cunningham, B. G. (2022). Cat-astrophic effects of sudden interruptions on spatial auditory attention. Jour-
nal of the Acoustical Society of America, 151, 3219–3233.
Lim, S.-J., Carter, Y. D., Michelle Njoroge, J., Shinn-Cunningham, B. G., & Perrachione, T. K. (2021). Talker discontinuity disrupts attention
to speech: Evidence from EEG and pupillometry. Brain and Language, 221, 104996. https://doi.org/10.1101/2021.01.28.428718
Lim, S.-J., Shinn-Cunningham, B. G., & Perrachione, T. K. (2019). Effects of talker continuity and speech rate on auditory working memory.
Attention, Perception & Psychophysics, 81(4), 1167–1177.
Lim, S.-J., Wöstmann, M., & Obleser, J. (2015). Selective attention to auditory memory Neurally enhances perceptual precision. The Journal
of Neuroscience, 35(49), 16094–16104.
Maddox, R. K., & Shinn-Cunningham, B. G. (2012). Influence of task-relevant and task-irrelevant feature continuity on selective auditory
attention. Journal of the Association for Research in Otolaryngology, 13(1), 119–129.
Maryott, J., Noyce, A., & Sekuler, R. (2011). Eye movements and imitation learning: Intentional disruption of expectation. Journal of Vision,
11(1), 7.
Mathalon, D. H., Heinks, T., & Ford, J. M. (2004). Selective attention in schizophrenia: Sparing and loss of executive control. The American
Journal of Psychiatry, 161(5), 872–881.
Mehraei, G., Shinn-Cunningham, B., & Dau, T. (2018). Influence of talker discontinuity on cortical dynamics of auditory spatial attention.
NeuroImage, 179, 548–556.
Mesgarani, N., & Chang, E. F. (2012). Selective cortical representation of attended speaker in multi-talker speech perception. Nature,
485(7397), 233–236.
Michalka, S. W., Kong, L., Rosen, M. L., Shinn-Cunningham, B. G., & Somers, D. C. (2015). Short-term memory for space and time flexibly
recruit complementary sensory-biased frontal lobe attention networks. Neuron, 87(4), 882–892.
Michalka, S. W., Rosen, M. L., Kong, L., Shinn-Cunningham, B. G., & Somers, D. C. (2016). Auditory spatial coding flexibly recruits anterior,
but not posterior, visuotopic parietal cortex. Cerebral Cortex, 26(3), 1302–1308.
Mullennix, J. W., & Pisoni, D. B. (1990). Stimulus variability and processing dependencies in speech perception. Perception & Psychophysics,
47(4), 379–390.
Noyce, A., & Sekuler, R. (2014a). Oddball distractors demand attention: Neural and behavioral responses to predictability in the flanker task.
Neuropsychologia, 65, 18–24.
Noyce, A., & Sekuler, R. (2014b). Violations of newly-learned predictions elicit two distinct P3 components. Frontiers in Human Neuroscience,
8, 374.
Noyce, A. L., Cestero, N., Michalka, S. W., Shinn-Cunningham, B. G., & Somers, D. C. (2017). Sensory-biased and multiple-demand
processing in human lateral frontal cortex. The Journal of Neuroscience, 37(36), 8755–8766.
Noyce, A. L., Cestero, N., Shinn-Cunningham, B. G., & Somers, D. C. (2016). Short-term memory stores organized by information domain.
Attention, Perception & Psychophysics, 78(3), 960–970.
Noyce, A. L., Lefco, R. W., Brissenden, J. A., Tobyne, S. M., Shinn-Cunningham, B. G., & Somers, D. C. (2021). Extended frontal networks
for visual and auditory working memory. Cerebral Cortex, 32, 855–869. https://doi.org/10.1093/cercor/bhab249
Nusbaum, H. C., & Morin, T. M. (1992). Paying attention to differences among talkers. In Speech Perception, Production and Linguistic Struc-
ture (pp. 113–134). Sage.
Panichello, M. F., & Buschman, T. J. (2021). Shared mechanisms underlie the control of working memory and attention. Nature, 592, 601–
605. https://doi.org/10.1038/s41586-021-03390-w
Payne, L., Guillory, S., & Sekuler, R. (2013). Attention-modulated alpha-band oscillations protect against intrusion of irrelevant information.
Journal of Cognitive Neuroscience, 25(9), 1463–1476.
Payne, L., Rogers, C. S., Wingfield, A., & Sekuler, R. (2017). A right-ear bias of auditory selective attention is evident in alpha oscillations.
Psychophysiology, 54(4), 528–535.
Payne, L., & Sekuler, R. (2014). The importance of ignoring: Alpha oscillations protect selectivity. Current Directions in Psychological Science,
23(3), 171–177.
Peelle, J. E. (2014). Methodological challenges and solutions in auditory functional magnetic resonance imaging. Frontiers in Neuroscience,
8, 253.
Posner, M. I., & Petersen, S. E. (1990). The attention system of the human brain. Annual Review of Neuroscience, 13, 25–42.
Power, J. D., Cohen, A. L., Nelson, S. M., Wig, G. S., Barnes, K. A., Church, J. A., Vogel, A. C., Laumann, T. O., Miezin, F. M.,
Schlaggar, B. L., & Petersen, S. E. (2011). Functional network organization of the human brain. Neuron, 72(4), 665–678.
Pylyshyn, Z. W., & Storm, R. W. (1988). Tracking multiple independent targets: Evidence for a parallel tracking mechanism. Spatial Vision,
3(3), 179–197.
Ross, B., Hillyard, S. A., & Picton, T. W. (2010). Temporal dynamics of selective attention during dichotic listening. Cerebral Cortex, 20(6),
1360–1371.
NOYCE ET AL. 11 of 11
Schwartz, S., Wang, L., Shinn-Cunningham, B. G., & Tager-Flusberg, H. (2020). Neural evidence for speech processing deficits during a cock-
tail party scenario in minimally and low verbal adolescents and young adults with autism. Autism Research, 13(11), 1828–1842. https://
doi.org/10.1002/aur.2356
Seeley, W. W., Menon, V., Schatzberg, A. F., Keller, J., Glover, G. H., Kenna, H., Reiss, A. L., & Greicius, M. D. (2007). Dissociable intrinsic
connectivity networks for salience processing and executive control. The Journal of Neuroscience: The Official Journal of the Society for
Neuroscience, 27(9), 2349–2356.
Shinn-Cunningham, B. G. (2008). Object-based auditory and visual attention. Trends in Cognitive Sciences, 12(5), 182–186.
Shinn-Cunningham, B. G., & Best, V. (2008). Selective attention in normal and impaired hearing. Trends in Amplification, 12(4), 283–299.
Siebert, W. M. (1970). Frequency discrimination in the auditory system: Place or periodicity mechanisms? Proceedings of the IEEE, 58(5),
723–730.
Tanaka, K., Ross, B., Kuriki, S., Harashima, T., Obuchi, C., & Okamoto, H. (2021). Neurophysiological evaluation of right-ear advantage dur-
ing dichotic listening. Frontiers in Psychology, 12, 696263.
Tobyne, S. M., Osher, D. E., Michalka, S. W., & Somers, D. C. (2017). Sensory-biased attention networks in human lateral frontal cortex rev-
ealed by intrinsic functional connectivity. NeuroImage, 162, 362–372.
Tobyne, S. M., Somers, D. C., Brissenden, J. A., Michalka, S. W., Noyce, A. L., & Osher, D. E. (2018). Prediction of individualized task activa-
tion in sensory modality-selective frontal cortex with “connectome fingerprinting.”. NeuroImage, 183, 173–185.
Treisman, A. M. (1960). Contextual cues in selective listening. The Quarterly Journal of Experimental Psychology, 12(4), 242–248.
Treisman, A. M. (2006). How the deployment of attention determines what we see. Visual Cognition, 14(4–8), 411–443.
Unsworth, N., & Robison, M. K. (2016). The influence of lapses of attention on working memory capacity. Memory & Cognition, 44(2),
188–196.
Vasterling, J. J., Brailey, K., Constans, J. I., & Sutker, P. B. (1998). Attention and memory dysfunction in posttraumatic stress disorder. Neuro-
psychology, 12(1), 125–133.
Woldorff, M. G., Gallen, C. C., Hampson, S. A., Hillyard, S. A., Pantev, C., Sobel, D., & Bloom, F. E. (1993). Modulation of early sensory
processing in human auditory cortex during auditory selective attention. Proceedings of the National Academy of Sciences of the
United States of America, 90(18), 8722–8726.
Woods, K. J. P., & McDermott, J. H. (2015). Attentive tracking of sound sources. Current Biology, 25(17), 2238–2246.
Worden, M. S., Foxe, J. J., Wang, N., & Simpson, G. V. (2000). Anticipatory biasing of visuospatial attention indexed by retinotopically spe-
cific α-bank electroencephalography increases over occipital cortex. Journal of Neuroscience, 20(6), RC63.
Wöstmann, M., Maess, B., & Obleser, J. (2021). Orienting auditory attention in time: Lateralized alpha power reflects spatio-temporal filter-
ing. NeuroImage, 228, 117711. https://doi.org/10.1016/j.neuroimage.2020.117711
Wöstmann, M., Vosskuhl, J., Obleser, J., & Herrmann, C. S. (2018). Opposite effects of lateralised transcranial alpha versus gamma stimula-
tion on auditory spatial attention. Brain Stimulation, 11(4), 752–758.
Yantis, S. (2008). The neural basis of selective attention: Cortical sources and targets of attentional modulation. Current Directions in Psycho-
logical Science, 17(2), 86–90.
Yeo, B. T. T., Krienen, F. M., Sepulcre, J., Sabuncu, M. R., Lashkari, D., Hollinshead, M., Roffman, J. L., Smoller, J. W., Zöllei, L.,
Polimeni, J. R., Fischl, B., Liu, H., & Buckner, R. L. (2011). The organization of the human cerebral cortex estimated by intrinsic func-
tional connectivity. Journal of Neurophysiology, 106(3), 1125–1165.
How to cite this article: Noyce, A. L., Kwasa, J. A. C., & Shinn-Cunningham, B. G. (2022). Defining attention
from an auditory perspective. WIREs Cognitive Science, e1610. https://doi.org/10.1002/wcs.1610