A parietal–temporal sensory–motor integration area for the human vocal tract: Evidence from an fMRI study of skilled musicians

https://doi.org/10.1016/j.neuropsychologia.2007.06.024Get rights and content

Abstract

Several sensory–motor integration regions have been identified in parietal cortex, which appear to be organized around motor-effectors (e.g., eyes, hands). We investigated whether a sensory–motor integration area might exist for the human vocal tract. Speech requires extensive sensory–motor integration, as does other abilities such as vocal musical skills. Recent work found that a posterior superior temporal-parietal region, area Spt, has both sensory (auditory) and motor response properties (for both speech and tonal stimuli). Brain activity of skilled pianists was measured with fMRI while they listened to a novel melody and either covertly hummed the melody (vocal tract effectors) or covertly played the melody on a piano (manual effectors). Activity in area Spt was significantly higher for the covert hum versus covert play condition. A region in the anterior IPS (aIPS) showed the reverse pattern, suggesting its involvement in sensory-manual transformations. This finding suggests that area Spt is a sensory–motor integration area for vocal tract gestures.

Introduction

There has been a great deal of work over the last decade aimed at understanding the neural circuits supporting sensory–motor interaction. Most of this research has centered on sensory–motor integration within the context of the visual system, and much has been learned. For example, regions in the posterior parietal cortex (PPC) in both human and non-human primates have been identified as critical components of visuomotor integration circuits (Andersen, 1997; Milner & Goodale, 1995). These regions appear to be organized primarily around motor-effector systems, such as ocular versus hand/limb action systems (Colby & Goldberg, 1999; Culham & Kanwisher, 2001; Grefkes, Ritzl, Zilles, & Fink, 2004; Kertzman, Schwarz, Zeffiro, & Hallett, 1997) and they may be computing coordinate transformations, mapping between sensory representations and motor commands during movement planning, online updating, and control of movement (Andersen, 1997, Castiello, 2005; Colby & Goldberg, 1999).

Much less research has been conducted on sensory–motor integration within the context of the auditory system. This is likely because most investigators who are interested in sensory–motor interactions have focused on eye or limb movements where visual information plays a dominant role in movement planning and online guidance. However, there is a domain of action where the auditory system plays a critical role, namely in the movements of the vocal tract for speech. For example, the developmental task of learning to articulate the sound patterns of one's language is cued externally by auditory input primarily (somatosensory feedback plays an additional important role, see Tremblay, Shiller, & Ostry, 2003). As adults, we can, simply by listening, learn to pronounce new words, or pick up regional accents, sometimes unconsciously. Experimental work has also shown that delayed or otherwise altered auditory speech feedback affects speech articulation (Houde & Jordan, 1998; Yates, 1963), and it is well known that late onset deafness results in articulatory decline (Waldstein, 1989). Finally, neuropsychological research has shown that damage to left hemisphere auditory regions leads to deficits in speech production (Damasio & Damasio, 1980). All of this shows clearly that the auditory system plays an important role in speech production and therefore, there must be a neural mechanism for interfacing auditory and motor representations of speech (Doupe & Kuhl, 1999). See Hickok and Poeppel (2007) for a recent review.

Until recently, little was known about the neural circuit(s) supporting auditory–motor integration. But recent fMRI experiments have made some progress in this respect. The design of these studies relied on the observation from the visual domain, that many neurons in PPC sensory–motor integration areas have both sensory and motor response properties (Murata, Gallese, Kaseda, & Sakata, 1996). Thus, an area supporting auditory–motor integration for speech should respond both during perception and production of speech (covert production is used in prior studies, including the present study, to avoid overt auditory feedback (Buchsbaum, Hickok, & Humphries, 2001; Hickok, Buchsbaum, Humphries, & Muftuler, 2003). Using this approach, a fronto-parietal–temporal network of auditory + motor responsive regions was identified in human cortex. Included in this network is a region in the left posterior Sylvian fissure at the parietal–temporal boundary, area Spt.1 Area Spt appears to be functionally and anatomically connected with a frontal area known to be important for speech (area 44) (Buchsbaum et al., 2001, Buchsbaum et al., 2005a; Catani, Jones, & Ffytche, 2005; Galaburda & Sanides, 1980; Hickok & Poeppel, 2004) and the area Spt region, when disrupted via lesion or electrical stimulation, results in speech production deficits (Anderson et al., 1999; Damasio & Damasio, 1980). For these reasons, and also because of its anatomical location, area Spt has been hypothesized to be a sensory–motor integration area (Hickok et al., 2003) analogous in function to the sensory–motor integration areas previously identified in the PPC (Andersen, 1997; Colby & Goldberg, 1999).

Additional work showed that area Spt is not speech-specific, but responds equally well to the perception and production (covert humming) of tonal melodies (Hickok et al., 2003). Area Spt also responds during the temporary maintenance of a list of words in short-term memory independently of whether the items to be maintained were presented auditorily or visually (Buchsbaum et al., 2005b). This latter finding parallels claims that visual–motor integration areas have working memory-related properties (Murata et al., 1996), as well as the observation that sensory input from multiple modalities can drive neurons in PPC sensory–motor fields (Cohen, Batista, & Andersen, 2002; Mullette-Gillman, Cohen, & Groh, 2005). In sum, the response properties of area Spt – particularly that it shows both sensory and motor responses – are consistent with the hypothesis that it is a sensory–motor integration region similar to those found in the primate intraparietal sulcus (Buchsbaum et al., 2005b).

While area Spt has been shown to respond quite well in tasks involving a variety of sensory inputs (speech, tones, written words), its response to varying motor output conditions has not yet been investigated. While Baumann et al. (2007) report that the planum temporale is active in musicians generating hand/arm movements associated with playing a piece of music (the piece was “played” on a board), suggesting that Spt is active not only during orofacial-associated output tasks, but also manual tasks, without an orofacial task to use as a contrast it is impossible to gauge whether that region may be relatively selective for one or another output modality. If area Spt is organized in a manner analogous to PPC sensory–motor regions – which are organized around motor-effector systems – it should be fairly tightly linked to the vocal tract-related motor system. This, in turn, predicts that manipulation of the output modality should affect activity levels in area Spt even if the sensory stimulation were held constant. We set out to test this prediction in the present experiment.

One population in which auditory inputs can be mapped efficiently to two different motor-effector systems is skilled musicians. In at least some musicians, an aurally presented novel melody can be reproduced either by covertly humming the melody (vocal track effectors) or by covertly playing the melody (manual effectors). Based on our hypothesis that area Spt is a sensory–motor integration area involving vocal track-related motor systems, we predicted that area Spt would activate more strongly when musicians (pianists) were asked to covertly hum a heard melody than when they were asked to covertly play a heard melody. As a secondary hypothesis, we postulated that the use of manual effectors in covertly playing the melody will activate PPC, as this region has been implicated in several manual tasks including visually guided reaching, apraxic assessments, and piano-playing (Burnod et al., 1999, Haslinger et al., 2005; Makuuchi, Kaminaga, & Sugishita, 2005; Meister et al., 2004).

Section snippets

Subjects

Seven (2 male) skilled pianists (age range 18–35 years) participated in the experiment after giving written informed consent. The approval of the University of California, Irvine Institutional Review Board (IRB) was received prior to the start of the study. All participants were screened for prerequisite skill level prior to inclusion in the study. Potential subjects were seated in front of a piano keyboard and presented with novel 3-s piano melodies one at a time. After each melody,

Results

Our main objective was to identify regions that have both perceptual and motor response properties. Our assumption is that rehearsal-related activity is tapping motor-articulatory networks (Meister et al., 2004). The theoretical motivation for focusing on regions that respond both to sensory and motor phases of the task is that these regions are likely to be good candidates for performing sensory–motor integration. A left Sylvian parietal–temporal ROI (area Spt), a region previously identified

Discussion

A region in the posterior Sylvian fissure at the parietal–temporal boundary (area Spt), known to have auditory–motor response properties (Buchsbaum et al., 2001, Buchsbaum et al., 2005b, Hickok et al., 2003; Hickok & Poeppel, 2004; Okada & Hickok, 2006), was found to be substantially more active during covert humming of novel melodies than during covert playing on a keyboard of those same melodies. This finding is in support of the hypothesis that area Spt is a sensory–motor integration area

Acknowledgements

Thanks to Susan Anderson for creating the melodic stimuli.

This work was supported by NIHDC03681.

References (44)

  • G. Hickok et al.

    Dorsal and ventral streams: A framework for understanding aspects of the functional anatomy of language

    Cognition

    (2004)
  • I.G. Meister et al.

    Playing piano in the mind—An fMRI study on music imagery and performance in pianists

    Brain Research Cognitive

    (2004)
  • K. Okada et al.

    Left posterior auditory-related cortices participate both in speech perception and speech production: Neural overlap revealed by fMRI

    Brain and Language

    (2006)
  • O. Simon et al.

    Topographical layout of hand, eye, calculation, and language-related areas in the human parietal lobe

    Neuron

    (2002)
  • R. Andersen

    Multimodal integration for the representation of space in the posterior parietal cortex

    Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences

    (1997)
  • B. Bangert et al.

    Shared networks for auditory and motor processing in professional pianists: Evidence from fMRI conjunction

    NeuroImage

    (2006)
  • Y. Burnod et al.

    Parieto-frontal coding of reaching: an integrated framework

    Experimental Brain Research

    (1999)
  • G.A. Calvert et al.

    Activation of auditory cortex during silent lipreading

    Science

    (1997)
  • U. Castiello

    The neuroscience of grasping

    Nature Reviews Neuroscience

    (2005)
  • M. Catani et al.

    Perisylvian language networks of the human brain

    Annals of Neurology

    (2005)
  • J. Chiou et al.

    A simple simultaneous geometric and intensity correction method in echo planar imaging by EPI based phase modulation

    IEEE TMI

    (2003)
  • Y. Cohen et al.

    Comparison of neural activity preceding reaches to auditory and visual stimuli in the parietal reach region

    Neuroreport

    (2002)
  • Cited by (75)

    • Interaction of the effects associated with auditory-motor integration and attention-engaging listening tasks

      2019, Neuropsychologia
      Citation Excerpt :

      Thus, speech-level information, although clearly present at the level of AC, does not seem to strongly affect auditory-motor integration in posterior STG. Previous studies have shown that, in addition to listening to speech and covert speech production, left posterior STG regions are activated also during non-speech audiomotor tasks such as humming of melodies or playing a musical instrument (Buchsbaum et al., 2001; Hickok et al., 2003a, 2003b, 2009; Pa and Hickok, 2008). Based on the results of these studies, we expected that, similar to vowel repetition, also pitch repetition (humming the pitch of a target) would show enhanced activation relative to vowel-production responding in posterior STG.

    View all citing articles on Scopus
    View full text