Skip to main content Accessibility help
×
Hostname: page-component-89b8bd64d-r6c6k Total loading time: 0 Render date: 2026-05-08T13:31:03.127Z Has data issue: false hasContentIssue false

Phonetics in the Brain

Published online by Cambridge University Press:  02 March 2024

Pelle Söderström
Affiliation:
Lunds Universitet, Sweden

Summary

Spoken language is a rapidly unfolding signal: a complex code that the listener must crack to understand what is being said. From the structures of the inner ear through to higher-order areas of the brain, a hierarchy of interlinked processes transforms the acoustic signal into a linguistic message within fractions of a second. This Element outlines how we perceive speech and explores what the auditory system needs to achieve to make this possible. It traces a path through the system and discusses the mechanisms that enable us to perceive speech as a coherent sequence of words. This is combined with a brief history of research into language and the brain beginning in the nineteenth century, as well as an overview of the state-of-the-art neuroimaging and analysis techniques that are used to investigate phonetics in the brain today. This title is also available as Open Access on Cambridge Core.

Information

Figure 0

Figure 1 Simplified diagram showing the main connections for speech perception and production in the Wernicke-Lichtheim model (adapted from Lichtheim (1885)). The connection m shows the articulatory ‘muscle’ (speech production) pathway between the brainstem and area M (the motor centre of speech, corresponding to Broca’s area), while a signifies the pathway between the auditory brainstem and A (the acoustic word-centre, corresponding to Wernicke’s area). Node B signifies semantic concepts distributed across the brain.

Figure 1

Figure 2 Waveform (top) and spectrogram (bottom) derived using Praat (Boersma & Weenink, 2023) for the sentence dunk the stale biscuits into strong drink (IEEE, 1969) spoken by a young adult female speaker of Australian English. The waveform represents envelope amplitude modulations over time, while the spectrogram shows the frequency content of speech over time. Silences are represented as near-zero amplitudes in the waveform and whiter parts, or less energy, in the spectrogram. More energy is shown as darker parts of the spectrogram. Vowels generally have the most energy, as shown by larger amplitudes in the waveform and darker parts of the spectrogram, and vowels occur over longer timescales than the more transient consonants. Note the high frequency content in the sibilants /s/ in stale, biscuits and strong, exceeding the scale of 10,000 Hz on the y-axis. As can be seen in the word biscuits, silences – in this case, the plosive occlusion phase before the burst in the stop consonant /k/ – are not necessarily reliable cues to word onsets.

Figure 2

Figure 3 Waveforms and spectrograms for the words job (left) and jog (right), produced by an adult speaker of Northern British English. Vertical dashed lines indicate approximate segment boundaries. Coarticulation in the vowel is brought about by the articulators moving to produce either a labial or velar stop. In this example, there were no differences in mean vowel F1 frequency, but F2 and F3 were significantly lower on average in the vowel leading up to the release of jog as compared to job. Listeners take advantage of formant transitions and other subphonemic information in spoken-word recognition.

(Marslen-Wilson & Warren, 1994; McQueen et al., 1999)
Figure 3

Figure 4 Illustration of the signal path from the cochlea to the primary auditory cortex. This heavily simplified figure does not show any differences in lateralisation between the two brain hemispheres, and it does not show any descending (efferent) pathways.

(only ascending/afferent)

Save element to Kindle

To save this element to your Kindle, first ensure no-reply@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Phonetics in the Brain
Available formats
×

Save element to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Phonetics in the Brain
Available formats
×

Save element to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Phonetics in the Brain
Available formats
×