Skip to main content Accessibility help
×
Hostname: page-component-848d4c4894-jbqgn Total loading time: 0 Render date: 2024-07-04T21:59:10.403Z Has data issue: false hasContentIssue false

6 - The functional architecture of speech perception

Published online by Cambridge University Press:  08 August 2009

David Poeppel
Affiliation:
Professor Department of Linguistics Cognitive and Neuroscience Language Lab University of Maryland 1401 Marie Mount Hall College Park, MD 20742
Martin Hackl
Affiliation:
Department of Linguistics and Cognitive Science Pomona College Mason Hall 211B 550 Harvard Ave – 110B Mason Hall Claremont, CA 91711
James R. Pomerantz
Affiliation:
Rice University, Houston
Get access

Summary

Introduction

The language system is that aspect of mind/brain function that forms the basis for phonological, morphological, syntactic, and semantic computation. The “currencies” (or the ontology) of this central and abstract computational system are representations that are amodal, for example the concepts “feature” (phonology) or “affix” (morphology) or “phrase” (syntax) or “generalized quantifier” (semantics). Representation and computation with such concepts is typically considered independent of sensory modalities. Of course, the linguistic computational system is not isolated but interacts with other cognitive systems and with sensory–motor interface systems.

With regard to the input and output, the system has at least three modality-specific interfaces: an acoustic-articulatory system (speech perception and production), a visuo-motor system (reading/writing and sign), and a somatosensory interface (Braille). Speech and sign are the canonical interfaces and develop naturally; written language and Braille are explicitly taught: barring gross pathology, every child learns to speak or sign (rapidly, early, without explicit instruction, to a high level of proficiency), whereas learning to read/write Braille requires explicit instruction, is not universal, and occurs later in development.

In this chapter we focus on speech perception, specifically with regard to linguistic constraints and cortical organization. We first outline the key linguistic assumptions, including the concept of “distinctive feature,” and then discuss a functional-anatomic model that captures a range of empirical findings.

The linguistic basis of speech perception

The central importance of words for language use and understanding

An essential part of the cognitive ability underlying the linguistic behavior of a competent speaker of a language consists of knowing the words of the language or their constituents (roots).

Type
Chapter
Information
Topics in Integrative Neuroscience
From Cells to Cognition
, pp. 154 - 180
Publisher: Cambridge University Press
Print publication year: 2008

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

Belin, P., Zatorre, R. J., Lafaille, P., Ahad, P., and Pike, B. (2000). Voice-selective areas in human auditory cortex. Nature, 403(6767), 309–12.CrossRefGoogle ScholarPubMed
Binder, J. R., Frost, J. A., Hammeke, T. A., et al. (2000). Human temporal lobe activation by speech and nonspeech sounds. Cerebral Cortex, 10, 512–28.CrossRefGoogle ScholarPubMed
Binder, J. R., Frost, J. A., Hammeke, T. A., Rao, S. M., and Cox, R. W. (1996). Function of the left planum temporale in auditory and linguistic processing. Brain, 119, 1239–47.CrossRefGoogle ScholarPubMed
Blumstein, S. (1995). The neurobiology of the sound structure of language. In Gazzaniga, M., ed., The Cognitive Neurosciences. Cambridge, MA: MIT Press.Google Scholar
Boatman, D., Hart, J., Lesser, R. P.et al. (1998). Right hemisphere speech perception revealed by amobarbital injection and electrical interference. Neurology, 51(2), 458–64.CrossRefGoogle ScholarPubMed
Boemio, A., Fromm, S., Braun, A., and Poeppel, D. (2005). Hierarchical and asymmetric temporal sensitivity in human auditory cortices. Nature Neuroscience, 8, 389–95.CrossRefGoogle ScholarPubMed
Buchman, A., Garron, D., Trost-Cardamone, J. E., Wichter, M. D., and Schwartz, M. (1986). Word deafness: one hundred years later. Journal of Neurol Neurosurgery Psychiatry, 49(5), 489–99.CrossRefGoogle ScholarPubMed
Buchsbaum, B. R., Hickok, G., and Humphries, C. (2001). Role of left posterior superior temporal gyrus in phonological processing for speech perception and production. Cognitive Science, 25, 663–78.CrossRefGoogle Scholar
Burton, M. W. (2001). The role of inferior frontal cortex in phonological processing. Cognitive Science, 25, 695–709.CrossRefGoogle Scholar
Burton, M. W., Small, S. L., and Blumstein, S. E. (2000). The role of segmentation in phonological processing: an fMRI investigation. Journal of Cognitive Neuroscience, 12, 679–90.CrossRefGoogle Scholar
Chomsky, N. (1995). The Minimalist Program. Cambridge, MA: MIT Press.Google Scholar
Damasio, A. R. (1992). Aphasia. The New England Journal of Medicine, 326(8), 531–9.CrossRefGoogle ScholarPubMed
Dehaene-Lambertz, G., Dupoux, E., and Gout, A. (2000). Electrophysiological correlates of phonological processing: a cross-linguistic study. Journal of Cognitive Neuroscience, 12, 635–47.CrossRefGoogle ScholarPubMed
Dupoux, E., Kakehi, K., Hirose, Y., Pallier, C., and Mehler, J. (1999). Epenthetic vowels in Japanese: a perceptual illusion? Journal of Experimental Psychology: Human Perception and Performance, 25, 1568 –78.Google Scholar
Gandour, J., Wong, D., Hsieh, L., et al. (2000). A crosslinguistic PET study of tone perception. Journal of Cognitive Neuroscience, 12(1), 207–22.CrossRefGoogle ScholarPubMed
Grant, K. W., Wassenhove, V., and Poeppel, D. (2004). Detection of auditory (cross-spectral) and auditory-visual (cross-modal) synchrony. Speech Communication, 44, 43–53.CrossRefGoogle Scholar
Greenberg, S. (1998). A syllable-centric framework for the evolution of spoken language. Brain and Behavioral Sciences, 21(4), 518.CrossRefGoogle Scholar
Griffiths, T. D., Rees, A., and Green, G. G. R. (1999). Disorders of human complex sound processing. Neurocase, 5, 365–78.CrossRefGoogle Scholar
Halle, M. (1990). Phonology. In Osherson, D. and Lasnik, H., eds., Language. An Invitation to Cognitive Science. Cambridge, MA: MIT Press.Google Scholar
Helenius, P., Salmelin, R., Service, E., and Connolly, J. E. (1998). Distinct time courses of word and context comprehension in the left temporal cortex. Brain, 121, 1133–42.CrossRefGoogle ScholarPubMed
Hesling, I., Dilharreguy, B., Clement, S., Bordessoules, M., and Allard, M. (2005). Cerebral mechanisms of prosodic sensory integration using low-frequency bands of connected speech. Human Brain Mapping 26(3), 157–69.CrossRefGoogle ScholarPubMed
Hickok, G. and Poeppel, D. (2000). Towards a functional neuroanatomy of speech perception. Trends Cognitive Sciences, 4, 131–8.CrossRefGoogle ScholarPubMed
Hickok, G. and Poeppel, D. (2004). Dorsal and ventral streams: a framework for understanding aspects of the functional anatomy of language. Cognition, 92, 67–99.CrossRefGoogle ScholarPubMed
Hirsh, I. and Watson, C. S. (1996). Auditory psychophysics and perception. Annual Review of Psychology, 47, 461–84.CrossRefGoogle ScholarPubMed
Indefrey, P. and Levelt, W. J. M. (2004). The spatial and temporal signatures of word production components. Cognition, 92, 101–44.CrossRefGoogle ScholarPubMed
Ivry, R. and Lebby, P. (1998). The neurology of consonant perception: specialized module or distributed processors? In Beeman, M. and Chiarello, C., eds., Right Hemisphere Language Comprehension: Perspectives from Cognitive Neuroscience. Mahwah, NJ: Erlbaum, pp. 3–25.Google Scholar
Ivry, R. B. and Robertson, L. C. (1998). The Two Sides of Perception. Cambridge, MA: MIT Press.Google Scholar
Joliot, M., Ribary, U., and Llinas, R. (1994). Human oscillatory brain activity near 40 Hz coexists with cognitive temporal binding. Proceedings of the National Academy of Sciences USA, 91(24), 11748–51.CrossRefGoogle ScholarPubMed
Kaye, J. (1989). Phonology: A Cognitive View. Hillsdale, NJ: Lawrence Erlbaum Associates.Google Scholar
Kenstowicz, M. (1994). Phonology in Generative Grammar. Cambridge, MA: Blackwell.Google Scholar
Levelt, W. J. M. (1989). Speaking. Cambridge, MA: MIT Press.Google Scholar
Liberman, A. M., Cooper, F. S., Shankweiler, D. P., and Kennedy, Studdert M. (1967). Perception of the speech code. Psychological Review, 74, 431–61.CrossRefGoogle ScholarPubMed
Liberman, A. M. and Mattingly, I. G. (1985). The motor theory of speech perception revised. Cognition, 21, 1–36.CrossRefGoogle ScholarPubMed
Mazoyer, B. M., Dehaene, S., Tzourio, N., et al. (1993). The cortical representation of speech. Journal of Cognitive Neuroscience, 5(4), 467–79.CrossRefGoogle Scholar
Mehler, J. (1981). The role of syllables in speech processing – infant and adult data. Philosophical Transactions of the Royal Society B 295, 333–52.CrossRefGoogle Scholar
Meinschaefer, J., Hausmann, M., and Güntürkün, O. (1999). Laterality effects in the processing of syllable structure. Brain and Language, 70, 287–93.CrossRefGoogle ScholarPubMed
Näätänen, R. (1992). Attention and Brain function. Hillsdale, NJ: Lawrence Erlbaum Associates, Publishers.Google Scholar
Nicholls, M. E. (1996). Temporal processing asymmetries between the cerebral hemispheres: evidence and implications. Laterality, 1 (2), 97–137.CrossRefGoogle ScholarPubMed
Norris, D. and Wise, R. (2000). The study of prelexical and lexical processes in comprehension: psycholinguistics and functional neuroimaging. The New Cognitive Neurosciences. G. M. Cambridge: MIT Press.Google Scholar
Phillips, C. (2001). Levels of representation in the electrophysiology of speech perception. Cognitive Science, 25, 711–31.CrossRefGoogle Scholar
Phillips, C., Pellathy, T., Marantz, A., et al. (2000). Auditory cortex accesses phonological categories: an MEG mismatch study. Journal of Cognitive Neuroscience, 12, 1038–55.CrossRefGoogle ScholarPubMed
Pisoni, D. (1973). Auditory and phonetic memory codes in the discrimination of consonants and vowels. Perception and Psychophysics, 13, 253–60.CrossRefGoogle ScholarPubMed
Poeppel, D. (2001). Pure word deafness and the bilateral processing of the speech code. Cognitive Science, 25, 679–93.CrossRefGoogle Scholar
Poeppel, D., Boemio, A., Simon, J., et al. (2000). High-frequency Response Asymmetry to Auditory Stimuli of Varying Spectral Complexity. New Orleans: Society for Neuroscience.Google Scholar
Poeppel, D., Wharton, C., Fritz, J., et al. (2004). FM sweeps, syllables, and word stimuli differentially modulate left and right non-primary auditory areas. Neuropsychologia, 42, 183–200.CrossRefGoogle Scholar
Poeppel, D., Yellin, E., Phillips, E., et al. (1996). Task-induced asymmetry of the auditory evoked M100 neuromagnetic field elicited by speech sounds. Cognitive Brain Research, 4, 231–42.CrossRefGoogle ScholarPubMed
Pöppel, E. (1997). A hierarchical model of temporal perception. Trends Cognitive Sciences, 1(2), 56–61.CrossRefGoogle ScholarPubMed
Price, C. (2000). The anatomy of language: contributions from functional neuroimaging. Journal of Anatomy, 197, 335–59.CrossRefGoogle ScholarPubMed
Pylkkanen, L., Stringfellow, A., and Marantz, A. (2002). Neuromagnetic evidence for the timing of lexical activation: an MEG component sensitive to phonotactic probability but not to neighborhood density. Brain and Language, 81, 666–78.CrossRefGoogle ScholarPubMed
Rosen, S. (1992). Temporal information is speech: acoustic, auditory, and linguistic aspects. Philosophical Transactions of the Royal Society London B, 336, 367–73.CrossRefGoogle ScholarPubMed
Ross, E. D., Thompson, R. D., and Yenkosky, J. (1997). Lateralization of affective prosody in brain and the callosal integration of hemispheric language functions. Brain and Language, 56(1), 27–54.CrossRefGoogle ScholarPubMed
Scott, S. K., Blank, C. C., Rosen, S., and Wise, R. J. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain, 123, 2400–6.CrossRefGoogle ScholarPubMed
Shamma, S. (2001). On the role of space and time in auditory processing. Trends Cognitive Sciences, 5, 340–8.CrossRefGoogle ScholarPubMed
Singer, W. (1993). Synchronization of cortical activity and its putative role in information processing and learning. Annual Review Physiology, 55, 349–74.CrossRefGoogle Scholar
Tallal, P., Miller, S., and Fitch, R. (1993). Neurobiological basis of speech: a case for the preeminence of temporal processing. Annals New York Academy of Science, 682, 27–47.CrossRefGoogle ScholarPubMed
Theunissen, F. and Miller, J. P. (1995). Temporal encoding in nervous systems: a rigorous definition. Journal of Computational Neuroscience, 2, 149–62.CrossRefGoogle ScholarPubMed
Wassenhove, V., Grant, K., and Poeppel, D. (2005). Visual speech speeds up the neural processing of auditory speech. Proceedings of the National Academy of Sciences USA, 102, 1181–6.CrossRefGoogle ScholarPubMed
Viemeister, N. F. and Plack, C. J. (1993). Time analysis. In Yost, W. A., Popper, A. N., and Fay, R. R., eds., Human Psychophysics. New York: Springer.CrossRefGoogle Scholar
Warren, R. M. (1999). Auditory Perception. Cambridge, UK: Cambridge University Press.Google Scholar
Yabe, H., Tervaniemi, M., Reinikainen, K., and Näätänen, R. (1997). Temporal window of integration revealed by MMN to sound omission. Neuroreport, 8, 1971–4.CrossRefGoogle ScholarPubMed
Zatorre, R. J. (1997). Cerebral correlates of human auditory processing: perception of speech and musical sounds. In Syka, J., ed., Acoustical Signal Processing in the Central Auditory System. New York: Plenum Press, pp. 453–468.CrossRefGoogle Scholar
Zatorre, R. and Belin, P. (2001). Spectral and temporal processing in human auditory cortex. Cerebral Cortex, 11, 946–53.CrossRefGoogle ScholarPubMed
Zatorre, R., Belin, P., and Penhune, V. B. (2002). Structure and function of auditory cortex: music and speech. Trends in Cognitive Sciences, 6(1), 37–46.CrossRefGoogle ScholarPubMed
Zatorre, R. J., Evans, A. C., and Meyer, E. (1994). Neural mechanisms underlying melodic perception and memory for pitch. Journal of Neuroscience, 14, 1908–19.CrossRefGoogle ScholarPubMed
Zatorre, R. J., Evans, A. C., Meyer, E., and Gjedde, A. (1992). Lateralization of phonetic and pitch discrimination in speech processing. Science, 256, 846–9.CrossRefGoogle ScholarPubMed
Zatorre, R. J., Meyer, E., Gjedde, A., and Evans, A. C. (1996). PET studies of phonetic processing of speech: review, replication, and reanalysis. Cerebral Cortex, 6(1), 21–30.CrossRefGoogle ScholarPubMed

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

  • The functional architecture of speech perception
    • By David Poeppel, Professor Department of Linguistics Cognitive and Neuroscience Language Lab University of Maryland 1401 Marie Mount Hall College Park, MD 20742, Martin Hackl, Department of Linguistics and Cognitive Science Pomona College Mason Hall 211B 550 Harvard Ave – 110B Mason Hall Claremont, CA 91711
  • Edited by James R. Pomerantz, Rice University, Houston
  • Book: Topics in Integrative Neuroscience
  • Online publication: 08 August 2009
  • Chapter DOI: https://doi.org/10.1017/CBO9780511541681.010
Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

  • The functional architecture of speech perception
    • By David Poeppel, Professor Department of Linguistics Cognitive and Neuroscience Language Lab University of Maryland 1401 Marie Mount Hall College Park, MD 20742, Martin Hackl, Department of Linguistics and Cognitive Science Pomona College Mason Hall 211B 550 Harvard Ave – 110B Mason Hall Claremont, CA 91711
  • Edited by James R. Pomerantz, Rice University, Houston
  • Book: Topics in Integrative Neuroscience
  • Online publication: 08 August 2009
  • Chapter DOI: https://doi.org/10.1017/CBO9780511541681.010
Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

  • The functional architecture of speech perception
    • By David Poeppel, Professor Department of Linguistics Cognitive and Neuroscience Language Lab University of Maryland 1401 Marie Mount Hall College Park, MD 20742, Martin Hackl, Department of Linguistics and Cognitive Science Pomona College Mason Hall 211B 550 Harvard Ave – 110B Mason Hall Claremont, CA 91711
  • Edited by James R. Pomerantz, Rice University, Houston
  • Book: Topics in Integrative Neuroscience
  • Online publication: 08 August 2009
  • Chapter DOI: https://doi.org/10.1017/CBO9780511541681.010
Available formats
×