Elsevier

Biological Psychology

Volume 120, October 2016, Pages 61-68
Biological Psychology

Attenuation of auditory evoked potentials for hand and eye-initiated sounds

https://doi.org/10.1016/j.biopsycho.2016.08.011Get rights and content

Highlights

  • Saccade initiation of tones evoked less sensory attenuation than button-presses.

  • Motor-to-sensory mapping can be partly generalized to multiple contingencies.

  • Predictive and associative processes may both contribute to sensory attenuation.

Abstract

Reduction of auditory event-related potentials (ERPs) to self-initiated sounds has been considered evidence for a predictive model in which copies of motor commands suppress sensory representations of incoming stimuli. However, in studies which involve arbitrary auditory stimuli evoked by sensory-unspecific motor actions, learned associations may underlie ERP differences. Here, in a new paradigm, eye motor output generated auditory sensory input, a naïve action–sensation contingency. We measured the electroencephalogram (EEG) of 40 participants exposed to pure tones, which they produced with either a button-press or volitional saccade. We found that button-press-initiated stimuli evoked reduced amplitude compared to externally initiated stimuli for both the N1 and P2 ERP components, whereas saccade-initiated stimuli evoked intermediate attenuation at N1 and no reduction at P2. These results indicate that the motor-to-sensory mapping involved in speech production may be partly generalized to other contingencies, and that learned associations also contribute to the N1 attenuation effect.

Introduction

Our brains must regulate a continuous stream of sensory input in order to flexibly generate behaviour and allow interaction with the world. A well-established example of such regulation is sensory attenuation, where the sensory input evoked by self-initiated actions is marked by reduced phenomenological (e.g., Blakemore, Frith, & Wolpert, 1999; Cardoso-Leite, Mamassian, Schütz-Bosbach, & Waszak, 2010; Sato, 2008) and neurophysiological representations (e.g., Baess, Jacobsen, & Schröger, 2008; Houde, Nagarajan, Sekihara, & Merzenich, 2002; Schafer & Marcus, 1973) compared to identical, externally initiated sensory input—a phenomenon typified by the difficulty of tickling ourselves (Weiskrantz, Elliott, & Darlington, 1971). Functionally, sensory attenuation serves to conserve attentional resources and to enable sensory processing in situations where volitional actions would otherwise desensitize sensory receptors, such as during speech production (Bendixen, SanMiguel, & Schröger, 2012). It has also been proposed as fundamental for self-identity, such that dysfunctional attenuation could lead to psychotic symptomology (Feinberg, 1978, Ford et al., 2001).

In the auditory domain, the N1 or N1 m component (an evoked potential or magnetic field which appears approximately 100 ms after the onset of an auditory stimulus) is used as a cortical index of sensory attenuation, because its amplitude, compared to externally initiated stimuli, is consistently reduced for both self-initiated vocalizations (Curio, Neuloh, Numminen, Jousmaki, & Hari, 2000; Heinks-Maldonado, Mathalon, Gray, & Ford, 2005; Houde et al., 2002) and button-press-initiated stimuli (Aliu, Houde, & Nagarajan, 2009; Baess et al., 2008; Martikainen, Kaneko, & Hari, 2005; McCarthy & Donchin, 1976; Schafer & Marcus, 1973; Sowman, Kuusik, & Johnson, 2012). The predominant explanation for these findings invokes a theory of motor control in which a forward model predicts future behavioural states and their sensory consequences (Wolpert, Ghahramani, & Jordan, 1995). According to this theory, the sensory consequences of volitional action can be predicted based on an efference copy (i.e., a copy of the motor command), and sensory attenuation reflects the subtraction of this prediction from actual sensory input (Bays & Wolpert, 2007). Conversely, externally initiated sensory input, for which there is no motor information to form an accurate prediction, will remain unmodulated by the efference copy mechanism (see Timm, SanMiguel, Keil, Schroger, & Schonwiesner, 2014).

Forward prediction is strongly implicated in speech production (Hickok, 2012), which involves a well-defined range of motor output (e.g., shape formed by lips) producing specific, habitual, sensory consequences (i.e., uttered syllables). As candidate language fibre pathways have been identified (Dick & Tremblay, 2012), upon which the efference copies of speech motor output conceivably travel, there is also a plausible neural basis for forward prediction in speech production. However, the N1 attenuation literature largely comprises experiments based on arbitrary action–sensation contingencies, usually hand movements to press a button and elicit a tone. This is problematic, because it is not clear that motor-to-sensory mapping can be generalized from speech production to auditory input evoked by motor actions unrelated to speech (Horváth, 2015). Unlike speech-induced auditory attenuation, for which there are distinct neural networks proposed to be involved (Behroozmand et al., 2016; Chang, Niziolek, Knight, Nagarajan, & Houde, 2013; Greenlee et al., 2013) which likely encode specific acoustic properties of the upcoming sound, internal predictions arising from non-speech motor actions might be comparatively more crude. This presents the possibility that sensory attenuation of speech and non-speech stimuli are driven by different mechanisms.

An alternate (or perhaps complementary) explanation of sensory attenuation relates to learned associations between actions and sensations. Operationally, learned associations have been proposed to “pre-activate” potential sensory input, increasing baseline neural activity at a greater rate than the signal increase resulting from an incoming stimulus (Roussel, Hughes, & Waszak, 2013). This account differs from forward prediction in that sensory attenuation is attributed to poorer stimulus discrimination (and consequent reduction of the corresponding sensory representations) rather than an efference copy directly reducing the internal response to a predicted stimulus. An associative explanation can be reasonably applied to the classic paradigm of pressing a button to hear an auditory stimulus, given that such a contingency conceivably draws upon a wealth of pre-existing action–sensation associations that have been experienced and learnt over the course of a lifetime. Indeed, there is evidence to suggest that prior experience can affect sound perception (Repp & Knoblich, 2007), and some suggestion that contingency strength alters neurophysiological response in the auditory (Baess et al., 2008; SanMiguel, Widmann, Bendixen, Trujillo-Barreto, & Schröger, 2013) and visual (Roussel, Hughes, & Waszak, 2014) domains (see Horváth, 2015 for a discussion). Hence, efference copy modulation may not be wholly responsible for all cases of sensory attenuation, and an account based on learned associations may explain observed effects in contingencies that lack the highly specific motor-to-sensory mapping that exists during speech production.

One possible test of the contribution of learned associations would be to employ a novel contingency (i.e., one never experienced before) between motor output and sensory input. For example, volitional eye movements cannot directly cause sounds in our natural environment, and therefore it seems implausible that an efference copy of the eye motor command would be sent to the auditory cortex. Hence, for a contingency between an eye movement and an auditory stimulus, it is almost certain that no learned association exists, which provides a strong test of whether learned associations are an essential component of sensory attenuation. Moreover, if neurophysiological auditory attenuation were still observed for this novel contingency, it implies that motor-to-sensory mapping is indeed generalizable to indirect action–sensation contingencies. This result would support the forward prediction model of sensory attenuation. However, if sensory attenuation was not observed, a limit for generalizability will have been identified, which might suggest that learned associations are driving button-press-elicited auditory N1 attenuation. Alternatively, sensory attenuation may be driven by a combination of these processes; for example, it seems possible that forward prediction could be established for novel contingencies, such that levels of N1 attenuation are mediated by associative strength.

Accordingly, the primary aim of the present study was to determine whether auditory event-related potential (ERP) attenuation, particularly with respect to the N1 component, would occur following an action–sensation contingency for which no prior learning exists, and thus provide a valuable contribution to the discussion about the processes underlying sensory attenuation. To do so, our experimental design contained a new condition which associated eye movements (the motor output) with tones (the auditory sensory input), a pairing which cannot occur outside of artificial contexts, and for which it is difficult to imagine an analogous, naturally occurring pairing of events. Specifically, participants were presented a pure tone (as is common in auditory ERP studies of this nature, e.g., Baess et al., 2008) following a singular, volitional saccadic movement. To determine whether N1 attenuation occurred, their resultant electrophysiological response, following subtraction of a motor condition (i.e., the same eye movement without stimulus presentation), was compared to an externally initiated condition in which tones were presented without any participant input. Given the possibility raised above that a combination of efference copy and associative processes drive sensory attenuation, we expected to observe N1 attenuation for saccade-initiated stimuli compared to the externally initiated condition. Even so, it seemed unlikely that a novel contingency could produce the same level of attenuation as previously seen in button-press-initiated experiments, as we have substantial prior experience with auditory sensory input following hand motor output.

In view of this hypothesis, our secondary aim was to quantitatively compare self-initiation effects associated with different regions of motor output (i.e., hand and eye). To achieve this aim, our within-subjects design also included an established button-press-initiated condition, for which convincing N1 and P2 attenuation has been demonstrated (Mifsud et al., 2016, Oestreich et al., 2016, Whitford et al., 2011). We chose to additionally assess the P2 component (a large voltage positivity which peaks approximately 200 ms after stimulus onset), which reflects the processing of specific auditory features (Shahin, Roberts, Pantev, Trainor, & Ross, 2005), for two reasons. First, SanMiguel, Todd, and Schröger (2013) have suggested that it may provide a more direct measure of sensory-specific prediction effects than the N1, because, unlike N1 effects, P2 attenuation was uniform over different stimulus onset asynchronies. Hence, it seems prudent to report P2 effects so as to enable comparison between different paradigms. Second, the P2 component has previously been shown to discriminate between self-initiated conditions which differ by motor output region (i.e., hand and foot; van Elk, Salomon, Kannape, & Blanke, 2014), if not necessarily contingency strength (that is, we have substantive experience with both hand and feet producing auditory stimulation), and is therefore relevant given the disparate eye and hand regions targeted in the present study.

Section snippets

Participants

Forty participants were recruited using an online experiment management system at UNSW Australia. There were 18 females, 36 were right-handed, and mean age was 23 years (SD = 7). Participants provided written, informed consent and received either course credit (n = 24) or financial imbursement (n = 16, A$30) in exchange for their time. This study was approved by the UNSW Human Research Ethics Advisory Panel (Psychology).

Procedure

Participants sat in a quiet, artificially lit room, 60 cm in front of a computer

Results

Fig. 2 shows the grand-averaged ERPs at electrode Cz and its neighbouring electrodes (i.e., FCz, C1, C2, and CPz), which represent locations at which the N1 and P2 components are typically maximal, especially subsequent to bilateral auditory stimulation (Ford, Gray, Faustman, Roach, & Mathalon, 2007; Luck, 2012, Näätänen and Picton, 1987). In addition, Fig. 3 shows a complete view of scalp activity during the component windows for each stimulus condition. We conducted separate one-way

Discussion

The present study investigated sensory attenuation of self-initiated stimuli in terms of the effect of motor output region (i.e., hand or eye actions) on neurophysiological response to identical contingent sensory input (i.e., auditory tones). We found that button-press-initiated stimuli evoked significantly reduced N1 and P2 component amplitude compared to both saccade- and externally initiated stimuli. In the saccade-initiated condition, we observed an intermediary level of N1 attenuation,

Acknowledgements

Nathan G. Mifsud is supported by an Australian Postgraduate Award. Thomas J. Whitford is supported by a Career Development Fellowship from the National Health and Medical Research Council of Australia (APP1090507), a Discovery Project from the Australian Research Council (DP140104394), and a Young Investigator Award (Barbara and John Streicker Investigator) from the NARSAD Brain and Behavior Research Foundation (17537).

References (46)

  • S.O. Aliu et al.

    Motor-induced suppression of the auditory cortex

    Journal of Cognitive Neuroscience

    (2009)
  • P. Baess et al.

    Suppression of the auditory N1 event-related potential component with unpredictable self-initiated tones: Evidence for internal forward models with dynamic stimulation

    International Journal of Psychophysiology

    (2008)
  • P.M. Bays et al.

    Computational principles of sensorimotor control that minimize uncertainty and variability

    Journal of Physiology

    (2007)
  • R. Behroozmand et al.

    Neural correlates of vocal production and motor control in human heschl's gyrus

    Journal of Neuroscience

    (2016)
  • A. Bendixen et al.

    Early electrophysiological indicators for predictive processing in audition: A review

    International Journal of Psychophysiology

    (2012)
  • P. Binda et al.

    Fusion of visual and auditory stimuli during saccades: A Bayesian explanation for perisaccadic distortions

    Journal of Neuroscience

    (2007)
  • S.J. Blakemore et al.

    Spatio-temporal prediction modulates the perception of self-produced stimuli

    Journal of Cognitive Neuroscience

    (1999)
  • P. Cardoso-Leite et al.

    A new look at sensory attenuation. Action-effect anticipation affects sensitivity, not response bias

    Psychological Science

    (2010)
  • E.F. Chang et al.

    Human cortical sensorimotor network underlying feedback control of vocal pitch

    Proceedings of the National Academy of Sciences of the United States of America

    (2013)
  • G. Curio et al.

    Speaking modifies voice-evoked activity in the human auditory cortex

    Human Brain Mapping

    (2000)
  • M.R. Diamond et al.

    Extraretinal control of saccadic suppression

    Journal of Neuroscience

    (2000)
  • A.S. Dick et al.

    Beyond the arcuate fasciculus: Consensus and controversy in the connectional anatomy of language

    Brain

    (2012)
  • I. Feinberg

    Efference copy and corollary discharge: Implications for thinking and its disorders

    Schizophrenia Bulletin

    (1978)
  • J.M. Ford et al.

    Neurophysiological evidence of corollary discharge dysfunction in schizophrenia

    American Journal of Psychiatry

    (2001)
  • J.M. Ford et al.

    Dissecting corollary discharge dysfunction in schizophrenia

    Psychophysiology

    (2007)
  • J.D. Greenlee et al.

    Sensory-motor interactions for vocal pitch monitoring in non-primary human auditory cortex

    Public Library Of Science

    (2013)
  • T.H. Heinks-Maldonado et al.

    Fine-tuning of auditory cortex during speech production

    Psychophysiology

    (2005)
  • G. Hickok

    Computational neuroanatomy of speech production

    Nature Reviews Neuroscience

    (2012)
  • J. Horváth

    Action-related auditory ERP attenuation: Paradigms and hypotheses

    Brain Research

    (2015)
  • J.F. Houde et al.

    The cortical computations underlying feedback control in vocal production

    Current Opinion in Neurobiology

    (2015)
  • J.F. Houde et al.

    Modulation of the auditory cortex during speech: An MEG study

    Journal of Cognitive Neuroscience

    (2002)
  • A. Hyvarinen et al.

    Independent component analysis: Algorithms and applications

    Neural Networks

    (2000)
  • F. Knolle et al.

    The cerebellum generates motor-to-auditory predictions: ERP lesion evidence

    Journal of Cognitive Neuroscience

    (2012)
  • Cited by (13)

    • Attenuation of auditory N2 for self-modulated tones during continuous actions

      2021, Biological Psychology
      Citation Excerpt :

      In the present study, for example, consequential features of tones were different between when participants turned the steering wheel from a rotation angle of 20° to 30° and when they turned the steering wheel from 60° to 70°. Furthermore, previous studies have shown that the degree of N1 attenuation decreased when the association between motor actions and the presentation of tones was weak (Mifsud et al., 2016; Mifsud & Whitford, 2017). Therefore, the ambiguity in continuous actions could be responsible for the absence of precise prediction based on an efference copy of motor commands, which may be available for lower-order auditory processing reflected in N1.

    • Attenuation of visual evoked responses to hand and saccade-initiated flashes

      2018, Cognition
      Citation Excerpt :

      In other words, there is insufficient support for assuming that findings from button-press studies can be generalized to other action–sensation contingencies. In the auditory domain, the limits of this assumption have been tested by van Elk, Salomon, Kannape, and Blanke (2014) and Mifsud, Beesley, Watson, and Whitford (2016), who employed paradigms using foot and saccade initiation respectively to demonstrate that differences in auditory-evoked potentials were dependent on the region of motor output used to produce the incoming stimulus. In the study by Mifsud, Beesley, et al. (2016), for example, a greater degree of auditory N1 attenuation was observed for button press-initiated tones than for saccade-initiated tones, consistent with the fact that while hand movements are strongly associated with auditory sensations (e.g., the sound of one’s fingers typing on a keyboard), eye movements are rarely, if ever, accompanied by auditory feedback.

    View all citing articles on Scopus
    View full text