Attenuation of auditory evoked potentials for hand and eye-initiated sounds
Introduction
Our brains must regulate a continuous stream of sensory input in order to flexibly generate behaviour and allow interaction with the world. A well-established example of such regulation is sensory attenuation, where the sensory input evoked by self-initiated actions is marked by reduced phenomenological (e.g., Blakemore, Frith, & Wolpert, 1999; Cardoso-Leite, Mamassian, Schütz-Bosbach, & Waszak, 2010; Sato, 2008) and neurophysiological representations (e.g., Baess, Jacobsen, & Schröger, 2008; Houde, Nagarajan, Sekihara, & Merzenich, 2002; Schafer & Marcus, 1973) compared to identical, externally initiated sensory input—a phenomenon typified by the difficulty of tickling ourselves (Weiskrantz, Elliott, & Darlington, 1971). Functionally, sensory attenuation serves to conserve attentional resources and to enable sensory processing in situations where volitional actions would otherwise desensitize sensory receptors, such as during speech production (Bendixen, SanMiguel, & Schröger, 2012). It has also been proposed as fundamental for self-identity, such that dysfunctional attenuation could lead to psychotic symptomology (Feinberg, 1978, Ford et al., 2001).
In the auditory domain, the N1 or N1 m component (an evoked potential or magnetic field which appears approximately 100 ms after the onset of an auditory stimulus) is used as a cortical index of sensory attenuation, because its amplitude, compared to externally initiated stimuli, is consistently reduced for both self-initiated vocalizations (Curio, Neuloh, Numminen, Jousmaki, & Hari, 2000; Heinks-Maldonado, Mathalon, Gray, & Ford, 2005; Houde et al., 2002) and button-press-initiated stimuli (Aliu, Houde, & Nagarajan, 2009; Baess et al., 2008; Martikainen, Kaneko, & Hari, 2005; McCarthy & Donchin, 1976; Schafer & Marcus, 1973; Sowman, Kuusik, & Johnson, 2012). The predominant explanation for these findings invokes a theory of motor control in which a forward model predicts future behavioural states and their sensory consequences (Wolpert, Ghahramani, & Jordan, 1995). According to this theory, the sensory consequences of volitional action can be predicted based on an efference copy (i.e., a copy of the motor command), and sensory attenuation reflects the subtraction of this prediction from actual sensory input (Bays & Wolpert, 2007). Conversely, externally initiated sensory input, for which there is no motor information to form an accurate prediction, will remain unmodulated by the efference copy mechanism (see Timm, SanMiguel, Keil, Schroger, & Schonwiesner, 2014).
Forward prediction is strongly implicated in speech production (Hickok, 2012), which involves a well-defined range of motor output (e.g., shape formed by lips) producing specific, habitual, sensory consequences (i.e., uttered syllables). As candidate language fibre pathways have been identified (Dick & Tremblay, 2012), upon which the efference copies of speech motor output conceivably travel, there is also a plausible neural basis for forward prediction in speech production. However, the N1 attenuation literature largely comprises experiments based on arbitrary action–sensation contingencies, usually hand movements to press a button and elicit a tone. This is problematic, because it is not clear that motor-to-sensory mapping can be generalized from speech production to auditory input evoked by motor actions unrelated to speech (Horváth, 2015). Unlike speech-induced auditory attenuation, for which there are distinct neural networks proposed to be involved (Behroozmand et al., 2016; Chang, Niziolek, Knight, Nagarajan, & Houde, 2013; Greenlee et al., 2013) which likely encode specific acoustic properties of the upcoming sound, internal predictions arising from non-speech motor actions might be comparatively more crude. This presents the possibility that sensory attenuation of speech and non-speech stimuli are driven by different mechanisms.
An alternate (or perhaps complementary) explanation of sensory attenuation relates to learned associations between actions and sensations. Operationally, learned associations have been proposed to “pre-activate” potential sensory input, increasing baseline neural activity at a greater rate than the signal increase resulting from an incoming stimulus (Roussel, Hughes, & Waszak, 2013). This account differs from forward prediction in that sensory attenuation is attributed to poorer stimulus discrimination (and consequent reduction of the corresponding sensory representations) rather than an efference copy directly reducing the internal response to a predicted stimulus. An associative explanation can be reasonably applied to the classic paradigm of pressing a button to hear an auditory stimulus, given that such a contingency conceivably draws upon a wealth of pre-existing action–sensation associations that have been experienced and learnt over the course of a lifetime. Indeed, there is evidence to suggest that prior experience can affect sound perception (Repp & Knoblich, 2007), and some suggestion that contingency strength alters neurophysiological response in the auditory (Baess et al., 2008; SanMiguel, Widmann, Bendixen, Trujillo-Barreto, & Schröger, 2013) and visual (Roussel, Hughes, & Waszak, 2014) domains (see Horváth, 2015 for a discussion). Hence, efference copy modulation may not be wholly responsible for all cases of sensory attenuation, and an account based on learned associations may explain observed effects in contingencies that lack the highly specific motor-to-sensory mapping that exists during speech production.
One possible test of the contribution of learned associations would be to employ a novel contingency (i.e., one never experienced before) between motor output and sensory input. For example, volitional eye movements cannot directly cause sounds in our natural environment, and therefore it seems implausible that an efference copy of the eye motor command would be sent to the auditory cortex. Hence, for a contingency between an eye movement and an auditory stimulus, it is almost certain that no learned association exists, which provides a strong test of whether learned associations are an essential component of sensory attenuation. Moreover, if neurophysiological auditory attenuation were still observed for this novel contingency, it implies that motor-to-sensory mapping is indeed generalizable to indirect action–sensation contingencies. This result would support the forward prediction model of sensory attenuation. However, if sensory attenuation was not observed, a limit for generalizability will have been identified, which might suggest that learned associations are driving button-press-elicited auditory N1 attenuation. Alternatively, sensory attenuation may be driven by a combination of these processes; for example, it seems possible that forward prediction could be established for novel contingencies, such that levels of N1 attenuation are mediated by associative strength.
Accordingly, the primary aim of the present study was to determine whether auditory event-related potential (ERP) attenuation, particularly with respect to the N1 component, would occur following an action–sensation contingency for which no prior learning exists, and thus provide a valuable contribution to the discussion about the processes underlying sensory attenuation. To do so, our experimental design contained a new condition which associated eye movements (the motor output) with tones (the auditory sensory input), a pairing which cannot occur outside of artificial contexts, and for which it is difficult to imagine an analogous, naturally occurring pairing of events. Specifically, participants were presented a pure tone (as is common in auditory ERP studies of this nature, e.g., Baess et al., 2008) following a singular, volitional saccadic movement. To determine whether N1 attenuation occurred, their resultant electrophysiological response, following subtraction of a motor condition (i.e., the same eye movement without stimulus presentation), was compared to an externally initiated condition in which tones were presented without any participant input. Given the possibility raised above that a combination of efference copy and associative processes drive sensory attenuation, we expected to observe N1 attenuation for saccade-initiated stimuli compared to the externally initiated condition. Even so, it seemed unlikely that a novel contingency could produce the same level of attenuation as previously seen in button-press-initiated experiments, as we have substantial prior experience with auditory sensory input following hand motor output.
In view of this hypothesis, our secondary aim was to quantitatively compare self-initiation effects associated with different regions of motor output (i.e., hand and eye). To achieve this aim, our within-subjects design also included an established button-press-initiated condition, for which convincing N1 and P2 attenuation has been demonstrated (Mifsud et al., 2016, Oestreich et al., 2016, Whitford et al., 2011). We chose to additionally assess the P2 component (a large voltage positivity which peaks approximately 200 ms after stimulus onset), which reflects the processing of specific auditory features (Shahin, Roberts, Pantev, Trainor, & Ross, 2005), for two reasons. First, SanMiguel, Todd, and Schröger (2013) have suggested that it may provide a more direct measure of sensory-specific prediction effects than the N1, because, unlike N1 effects, P2 attenuation was uniform over different stimulus onset asynchronies. Hence, it seems prudent to report P2 effects so as to enable comparison between different paradigms. Second, the P2 component has previously been shown to discriminate between self-initiated conditions which differ by motor output region (i.e., hand and foot; van Elk, Salomon, Kannape, & Blanke, 2014), if not necessarily contingency strength (that is, we have substantive experience with both hand and feet producing auditory stimulation), and is therefore relevant given the disparate eye and hand regions targeted in the present study.
Section snippets
Participants
Forty participants were recruited using an online experiment management system at UNSW Australia. There were 18 females, 36 were right-handed, and mean age was 23 years (SD = 7). Participants provided written, informed consent and received either course credit (n = 24) or financial imbursement (n = 16, A$30) in exchange for their time. This study was approved by the UNSW Human Research Ethics Advisory Panel (Psychology).
Procedure
Participants sat in a quiet, artificially lit room, 60 cm in front of a computer
Results
Fig. 2 shows the grand-averaged ERPs at electrode Cz and its neighbouring electrodes (i.e., FCz, C1, C2, and CPz), which represent locations at which the N1 and P2 components are typically maximal, especially subsequent to bilateral auditory stimulation (Ford, Gray, Faustman, Roach, & Mathalon, 2007; Luck, 2012, Näätänen and Picton, 1987). In addition, Fig. 3 shows a complete view of scalp activity during the component windows for each stimulus condition. We conducted separate one-way
Discussion
The present study investigated sensory attenuation of self-initiated stimuli in terms of the effect of motor output region (i.e., hand or eye actions) on neurophysiological response to identical contingent sensory input (i.e., auditory tones). We found that button-press-initiated stimuli evoked significantly reduced N1 and P2 component amplitude compared to both saccade- and externally initiated stimuli. In the saccade-initiated condition, we observed an intermediary level of N1 attenuation,
Acknowledgements
Nathan G. Mifsud is supported by an Australian Postgraduate Award. Thomas J. Whitford is supported by a Career Development Fellowship from the National Health and Medical Research Council of Australia (APP1090507), a Discovery Project from the Australian Research Council (DP140104394), and a Young Investigator Award (Barbara and John Streicker Investigator) from the NARSAD Brain and Behavior Research Foundation (17537).
References (46)
- et al.
Motor-induced suppression of the auditory cortex
Journal of Cognitive Neuroscience
(2009) - et al.
Suppression of the auditory N1 event-related potential component with unpredictable self-initiated tones: Evidence for internal forward models with dynamic stimulation
International Journal of Psychophysiology
(2008) - et al.
Computational principles of sensorimotor control that minimize uncertainty and variability
Journal of Physiology
(2007) - et al.
Neural correlates of vocal production and motor control in human heschl's gyrus
Journal of Neuroscience
(2016) - et al.
Early electrophysiological indicators for predictive processing in audition: A review
International Journal of Psychophysiology
(2012) - et al.
Fusion of visual and auditory stimuli during saccades: A Bayesian explanation for perisaccadic distortions
Journal of Neuroscience
(2007) - et al.
Spatio-temporal prediction modulates the perception of self-produced stimuli
Journal of Cognitive Neuroscience
(1999) - et al.
A new look at sensory attenuation. Action-effect anticipation affects sensitivity, not response bias
Psychological Science
(2010) - et al.
Human cortical sensorimotor network underlying feedback control of vocal pitch
Proceedings of the National Academy of Sciences of the United States of America
(2013) - et al.
Speaking modifies voice-evoked activity in the human auditory cortex
Human Brain Mapping
(2000)
Extraretinal control of saccadic suppression
Journal of Neuroscience
Beyond the arcuate fasciculus: Consensus and controversy in the connectional anatomy of language
Brain
Efference copy and corollary discharge: Implications for thinking and its disorders
Schizophrenia Bulletin
Neurophysiological evidence of corollary discharge dysfunction in schizophrenia
American Journal of Psychiatry
Dissecting corollary discharge dysfunction in schizophrenia
Psychophysiology
Sensory-motor interactions for vocal pitch monitoring in non-primary human auditory cortex
Public Library Of Science
Fine-tuning of auditory cortex during speech production
Psychophysiology
Computational neuroanatomy of speech production
Nature Reviews Neuroscience
Action-related auditory ERP attenuation: Paradigms and hypotheses
Brain Research
The cortical computations underlying feedback control in vocal production
Current Opinion in Neurobiology
Modulation of the auditory cortex during speech: An MEG study
Journal of Cognitive Neuroscience
Independent component analysis: Algorithms and applications
Neural Networks
The cerebellum generates motor-to-auditory predictions: ERP lesion evidence
Journal of Cognitive Neuroscience
Cited by (13)
Attenuation of auditory N2 for self-modulated tones during continuous actions
2021, Biological PsychologyCitation Excerpt :In the present study, for example, consequential features of tones were different between when participants turned the steering wheel from a rotation angle of 20° to 30° and when they turned the steering wheel from 60° to 70°. Furthermore, previous studies have shown that the degree of N1 attenuation decreased when the association between motor actions and the presentation of tones was weak (Mifsud et al., 2016; Mifsud & Whitford, 2017). Therefore, the ambiguity in continuous actions could be responsible for the absence of precise prediction based on an efference copy of motor commands, which may be available for lower-order auditory processing reflected in N1.
Processing of self-initiated sound motion in the human brain
2021, Brain ResearchAction–effect contingency modulates the readiness potential
2018, NeuroImageAttenuation of visual evoked responses to hand and saccade-initiated flashes
2018, CognitionCitation Excerpt :In other words, there is insufficient support for assuming that findings from button-press studies can be generalized to other action–sensation contingencies. In the auditory domain, the limits of this assumption have been tested by van Elk, Salomon, Kannape, and Blanke (2014) and Mifsud, Beesley, Watson, and Whitford (2016), who employed paradigms using foot and saccade initiation respectively to demonstrate that differences in auditory-evoked potentials were dependent on the region of motor output used to produce the incoming stimulus. In the study by Mifsud, Beesley, et al. (2016), for example, a greater degree of auditory N1 attenuation was observed for button press-initiated tones than for saccade-initiated tones, consistent with the fact that while hand movements are strongly associated with auditory sensations (e.g., the sound of one’s fingers typing on a keyboard), eye movements are rarely, if ever, accompanied by auditory feedback.
Role of the Cerebellum in Adaptation to Delayed Action Effects
2017, Current Biology