Elsevier

Hearing Research

Volume 269, Issues 1–2, 1 October 2010, Pages 34-41
Hearing Research

Research paper
What breaks a melody: Perceiving F0 and intensity sequences with a cochlear implant

https://doi.org/10.1016/j.heares.2010.07.007Get rights and content

Abstract

Pitch perception has been extensively studied using discrimination tasks on pairs of single sounds. When comparing pitch discrimination performance for normal-hearing (NH) and cochlear implant (CI) listeners, it usually appears that CI users have relatively poor pitch discrimination. Tasks involving pitch sequences, such as melody perception or auditory scene analysis, are also usually difficult for CI users. However, it is unclear whether the issue with pitch sequences is a consequence of sound discriminability, or if an impairment exists for sequence processing per se. Here, we compared sequence processing abilities across stimulus dimensions (fundamental frequency and intensity) and listener groups (NH, CI, and NH listeners presented with noise-vocoded sequences). The sequence elements were firstly matched in discriminability, for each listener and dimension. Participants were then presented with pairs of sequences, constituted by up to four elements varying on a single dimension, and they performed a same/different task. In agreement with a previous study (Cousineau et al., 2009) fundamental frequency sequences were processed more accurately than intensity sequences by NH listeners. However, this was not the case for CI listeners, nor for NH listeners presented with noise-vocoded sequences. Intensity sequence processing was, nonetheless, equally accurate in the three groups. These results show that the reduced pitch cues received by CI listeners do not only elevate thresholds, as previously documented, but also affect pitch sequence processing above threshold. We suggest that efficient sequence processing for pitch requires the resolution of individual harmonics in the auditory periphery, which is not achieved with the current generation of implants.

Research Highlights

► Sequence processing was studied for normal-hearing and cochlear implant listeners.► Differences across populations for discrimination thresholds were factored out. ► Cochlear implant users displayed a specific deficit for pitch sequence processing. ► Contour cues for efficient sequence processing seem degraded by current implants.

Introduction

The cochlear implant, a surgically-implanted device that bypasses cochlear processing to directly stimulate the auditory nerve, has been used to restore auditory function in many individuals with profound deafness. The original aim of the implant design was, understandably, to enable speech intelligibility. But whereas speech intelligibility in quiet can be achieved with a coarse representation of acoustic information (Shannon et al., 1995), other auditory abilities may require acoustic cues that are not currently available to cochlear implant (CI) users. In particular, pitch perception seems to be impaired when using a cochlear implant (for reviews, see McDermott, 2004, Moore and Carlyon, 2005, Drennan and Rubinstein, 2008). This is problematic, as accurate processing of pitch patterns is essential for speech perception (intonation, tonal languages), music perception (melodies), or auditory scene analysis (streaming, speech in noise). A better understanding of which aspects of pitch patterns’ perception are impaired in CI users is therefore of fundamental importance.

Pitch perception in CI users has often been assessed using pitch-ranking tasks with acoustically presented stimuli (Geurts and Wouters, 2001, Laneau et al., 2004, McDermott, 2004, Pressnitzer et al., 2005, Vandali et al., 2005, Sucher and McDermott, 2007, Looi et al., 2008a, Looi et al., 2008b). A similar procedure was used in all of these studies: two complex sounds (often sung vowels) that differed in fundamental frequency (F0) were presented. Listeners had to rank them on the pitch dimension by indicating which sound was higher in pitch, and threshold was estimated as the smallest F0 difference producing a consistent ranking. Results showed average thresholds much poorer than those observed for normal-hearing (NH) listeners (e.g. around 10% of F0 for the best performers in the CI groups, compared to less than 0.5% for NH, Geurts and Wouters, 2001, Pressnitzer et al., 2005). Moreover, a prominent feature of all results is a large inter-subject variability: for a 50% difference in F0 (7 semitones), performance of CI users may range from chance to near perfect (McDermott, 2004).

Another type of measure has focused on the processing of pitch sequences that extend over time, because of their immediate relevance to music perception (Cooper et al., 2008, Fujita and Ito, 1999, Galvin et al., 2007, Gfeller et al., 2002, Kong et al., 2004, Looi et al., 2008a, Looi et al., 2008b, Pressnitzer et al., 2005, Singh et al., 2009). In some of these studies, familiar melodies were to be recognized from a closed set (e.g. Fujita and Ito, 1999) whereas in others, simple melodic contours were to be discriminated (e.g. Galvin et al., 2007). Results generally indicate that melody perception is poor in CI users, with, again, a large variability between subjects. Galvin et al. (2007), for instance, obtained results ranging from chance to near-NH performance in a 5-note melodic contour identification task.

There are many potential sources for the variability reported across these studies, from different processing strategies to various nerve survival rates in individual CI listeners (Moore and Carlyon, 2005). It is unclear, moreover, if the variability observed in the pitch sequence tasks is simply a reflection of the diverse pitch-ranking abilities of individual CI users. Obviously, a large pitch-ranking threshold should induce poor pitch sequence representation. Looi et al. (2004) found that subjects’ ability to rank pitches was correlated with their ability to recognize melodies. However, it is also possible that, in addition, pitch sequence processing per se is impaired in CI users. For NH listeners, McFarland and Cacace (1992) suggested that sequences of pitch were more accurately processed than sequences of loudness, even though the discriminability between single elements of the sequences was approximately equated on each dimension. Cousineau et al. (2009), using a method that took into account the exact discriminability thresholds of individual listeners on each dimension, confirmed the advantage for pitch sequence processing for NH listeners. Pitch sequence discrimination performance was found to be superior to loudness sequence discrimination performance, presumably because of contour-encoding mechanisms available only for pitch (Demany and Ramos, 2005, Demany et al., 2009). In addition, it was found that the pitch sequence advantage was restricted to sounds made of resolved harmonics; pitch sequences made up of complex tones without any resolved harmonic were processed no more accurately than loudness sequences.

The latter finding leads to the prediction that CI users, being generally unable to resolve the individual harmonics of complex tones (Laneau et al., 2004), may suffer from a specific impairment in pitch sequence processing, independent of their pitch-ranking abilities. The following experiments were designed to test this hypothesis. We used the psychophysical method of Cousineau et al. (2009) to test the perception of sequences varying in either F0 or intensity in three groups of listeners: CI users, NH listeners, and NH listeners presented with noise-vocoded sequences (NH-voc). Importantly, the method aimed to uncouple sequence processing performance from any difference in terms of pitch discriminability that was expected between (and within) the different groups of listeners.

Section snippets

CI group

The CI group consisted of five post-lingually deafened adult CI users (M = 65.4 years, SD = 9.8). These listeners used a variety of implanted devices and processing strategies, which were set to their recommended settings during the experiment. Some relevant details about individual listeners can be found in Table 1. All listeners had already participated in psychophysical experiments and were relatively good performers on consonant and vowel closed-set identification in silence (Poncet-Wallet

Adjustment step

The mean Δ values obtained during the adjustment step are shown in Fig. 3, with the corresponding individual Δ values listed in Table 2. For the NH group, the mean Δ value was 0.39 semitones (ST) for the F0 condition, and 2.95 dB for the I condition. For the CI group, the mean Δ value was 7.80 ST for the F0 condition and 5.66 dB for the I condition. For the NH-voc group, the mean Δ value was 1.98 ST for the F0 condition and 3.18 dB for the I condition.

For statistical testing, we applied a

NH group

In NH listeners, we found that F0 sequences are processed very accurately: performance for discriminating two sequences of four elements was just as good as performance for discriminating two single sounds. These results replicate, with different listeners and stimuli, the findings of Cousineau et al. (2009). Cousineau et al. (2009) pointed out that such a pattern of results was inconsistent with the hypothesis that listeners break down each sequence into its individual elements and perform

Acknowledgments

We would like to thank Brett Swanson for his helpful review and for his contribution in producing the electrodogram of Fig. 2.

References (44)

  • V. Looi et al.

    Pitch discrimination and melody recognition by cochlear implant users

  • C.M. Sucher et al.

    Pitch ranking of complex tones by normally hearing subjects and cochlear implant users

    Hear. Res.

    (2007)
  • F.G. Zeng

    Temporal pitch in electric hearing

    Hear. Res.

    (2002)
  • J.G. Bernstein et al.

    Pitch discrimination of diotic and dichotic tone complexes: harmonic resolvability or harmonic number?

    J. Acoust. Soc. Am.

    (2003)
  • W.B. Cooper et al.

    Music perception by cochlear implant and normal hearing listeners as measured by the Montreal battery for evaluation of amusia

    Ear. Hear.

    (2008)
  • M. Cousineau et al.

    What makes a melody: the perceptual singularity of pitch sequences

    J. Acoust. Soc. Am.

    (2009)
  • A. de Cheveigné et al.

    The case of the missing delay lines: synthetic delays obtained by cross-channel phase interaction

    J. Acoust. Soc. Am.

    (2006)
  • L. Demany et al.

    On the binding of successive sounds: perceiving shifts in nonperceived pitches

    J. Acoust. Soc. Am.

    (2005)
  • L. Demany et al.

    Tuning properties of the auditory frequency-shift detectors

    J. Acoust. Soc. Am.

    (2009)
  • L. Demany et al.

    Auditory change detection: simple sounds are not memorized better than complex sounds

    Psychol. Sci.

    (2008)
  • W.R. Drennan et al.

    Music perception in cochlear implant users and its relationship with psychophysical capabilities

    J. Rehabil. Res. Dev.

    (2008)
  • P.J. Fitzgibbons et al.

    Auditory temporal processing in elderly listeners

    J. Am. Acad. Audiol.

    (1996)
  • L.M. Friesen et al.

    Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants

    J. Acoust. Soc. Am.

    (2001)
  • S. Fujita et al.

    Ability of nucleus cochlear implantees to recognize music

    Ann. Otol. Rhinol. Laryngol.

    (1999)
  • J.J. Galvin et al.

    Melodic contour identification by cochlear implant listeners

    Ear Hear.

    (2007)
  • L. Geurts et al.

    Coding of the fundamental frequency in continuous interleaved sampling processors for cochlear implants

    J. Acoust. Soc. Am.

    (2001)
  • K. Gfeller et al.

    Recognition of familiar melodies by adult cochlear implant recipients and normal-hearing adults

    Cochlear Implants Int.

    (2002)
  • D.M. Green et al.

    Signal Detection Theory and Psychophysics

    (1966)
  • D.D. Greenwood

    A cochlear frequency-position function for several species–29 years later

    J. Acoust. Soc. Am.

    (1990)
  • N. He et al.

    Frequency and intensity discrimination measured in a maximum-likelihood procedure from young and aged normal-hearing subjects

    J. Acoust. Soc. Am.

    (1998)
  • B.A. Henry et al.

    Spectral peak resolution and speech recognition in quiet: normal hearing, hearing impaired, and cochlear implant listeners

    J. Acoust. Soc. Am.

    (2005)
  • Y.Y. Kong et al.

    Music perception with temporal cues in acoustic and electric hearing

    Ear Hear.

    (2004)
  • Cited by (16)

    • Musical and vocal emotion perception for cochlear implants users

      2018, Hearing Research
      Citation Excerpt :

      In the second study, in order to evaluate the effect of CI processing in the same individuals, we use a validated cochlear implant simulation approach. Vocoding algorithms have been successfully employed to simulate the auditory signal degradation through a CI in normal hearing individuals; their use for perception research has been validated for both speech (Fu and Nogaki, 2005; Nogaki et al., 2007; Poissant et al., 2006; Qin and Oxenham, 2003) and music (Cousineau et al., 2010; Limb, 2006; Moore and Tan, 2003). Furthermore, this approach provides easier access to a larger number of homogeneous participants (Driscoll et al., 2009; Poissant et al., 2006).

    • Categorisation of natural sounds at different stages of auditory recovery in cochlear implant adult deaf patients

      2018, Hearing Research
      Citation Excerpt :

      This may be the consequence of the distinction between voice and musical instruments, which was absent at the stage of Int CI. It is evident that spectral degradation by the implant dramatically affects the perception of any melodic sound by CI users such as telephone tunes or any instrumental passages, which rely on fine spectral details that cannot be efficiently transmitted through the implant (Cousineau et al., 2010; Gfeller et al., 2002; Looi et al., 2008). Though cochlear implants can improve the awareness of environmental sounds (Looi and Arnephy, 2010), the identification of such sounds in CI varies from about 45% (Shafiro et al., 2011) to about 80% in experienced CI users (Reed and Delhorne, 2005).

    • Categorization of common sounds by cochlear implanted and normal hearing adults

      2016, Hearing Research
      Citation Excerpt :

      Other aspects of auditory perception remain problematic for many CI users especially concerning the recognition of environmental sounds and music. The reduction in fine spectral information negatively affects the musical abilities of CI users, for example the perception of melody and timbre within instrumental passages relies on fine spectral details that cannot be efficiently transmitted through the implant (Gfeller et al., 2002; Looi et al., 2008; Cousineau et al., 2010). Similarly, whilst a CI can improve the awareness of environmental sounds (Looi and Arnephy, 2010) environmental sound identification in CI users is commonly described as poor with correct identification varying from only 45% (Shafiro et al., 2011) to nearly 80% in experienced CI users (Reed and Delhorne, 2005).

    View all citing articles on Scopus
    View full text