To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This chapter opens with the pivotal scene in Goethe’s bestselling novel, The Sorrows of Young Werther, when Werther reads Ossian to Charlotte. In describing this moment, Goethe reproduces Ossian’s patterns of rhythm and syntax in his own prose. The effect suggests that Werther and Charlotte share an embodied responsiveness to their reading. Goethe here seems to be drawing upon contemporary theories of universal rhythm and debates about prosody. The idea that poetic rhythm is a sensuous experience that can be shared between readers is then pushed to the extreme in the Roman Elegies, in which he playfully compares prosody to sex. The final section of this chapter focuses on Elective Affinities and shows how the novel’s comparison between chemical bonds and bonds of human affection extends also to a comparison between human relationships and the relational structures of language and metaphor.
This study examines the utterance-initial prosodic marking of sarcasm in English and its perception in listeners who did and listeners who did not self-identify as being on the autism spectrum. We ask (i) whether speakers use prosody to mark sarcasm in the early, ‘pre-target’ portion of an utterance (that is, in the portion before a ‘target’ word most closely associated with the sarcastic intent occurs), (ii) whether individuals vary in how they mark sarcasm, (iii) whether listeners reliably recognize sarcasm from pre-target prosody alone, and (iv) whether recognition accuracy varies by speaker or self-identified autistic traits. Eight American English speakers were recorded producing utterances presented in contexts conducive to either sarcasm or sincerity. Pre-target parts were presented in a two-alternative forced-choice experiment to individuals who either did (n=51) or did not (n=44) self-identify as being on the autism spectrum, and were examined for syllable duration and f0-related properties (maximum, minimum, range, and wiggliness). Results show that speakers distinguish sarcasm and sincerity in the pre-target region with duration being the most salient marker. Most listeners recognize sarcasm from pre-target fragments, but there is variation in how well each speaker is perceived. Whether the listener self-identified as being on the autism spectrum or not does not predict sarcasm and sincerity recognition accuracy. The results provide evidence that utterance-initial prosody contributes to sarcasm recognition, with the proviso that speaker and listener variation be taken into account.
Learning a second language (L2) is challenging partly due to perceptual strategies inherited from learners’ first language. For example, speakers of tone languages like Mandarin over-use pitch in English prosody perception and production. We developed a novel training paradigm to help Mandarin learners adopt more native-like strategies by enhancing their use of duration relative to pitch cues during prosody categorization. After prosody training, participants used duration more during phrase boundary categorization but showed no clear change for contrastive focus and lexical stress, suggesting that cue weighting training is most effective when targeting a feature’s primary cue. The control group, who practiced English vocabulary, relied more on pitch in lexical stress categorization and phrase boundary production after training, suggesting that without targeted instruction, listeners default to existing strategies. Our findings demonstrate that although default strategies in L2 speech perception are difficult to resist, lifelong perceptual habits can be adjusted with training.
It is well known that floating tones can associate both within a word and across a word boundary. In relation to floating quantity, however, there is extensive evidence for association within a word, but not across a word boundary. This research report presents evidence for the latter configuration in Shilluk, a West Nilotic language. Shilluk noun forms may end in floating quantity, and this quantity is realized only on following vocalic prefixes, that is, across a word boundary. The investigation includes a descriptive analysis of the phenomenon and a production study based on data from ten Shilluk speakers.
Stress in Gujarati (Indo-Aryan, India and Pakistan) has been alternately claimed to be strictly positional or sensitive to vowel sonority. The latter analyses figure prominently in arguments for scalar markedness constraints (de Lacy 2002, 2006). This study presents acoustic measures and speaker intuitions to evaluate both the positional and sonority-driven stress hypotheses. The acoustic results support weakly cued positional stress, though speaker intuitions for primary stress placement were inconsistent. This replicates Shih's (2018) negative findings, and indicates that Gujarati stress should not figure in discussions of sonority-driven stress or associated theoretical proposals.
Ticuna (ISO: tca; Peru, Colombia, Brazil) displays a larger tone inventory - five level tones - than any other Indigenous American language outside Oto-Manguean. Based on recent fieldwork, this article argues that, in addition to these tone properties, the Cushillococha variety of Ticuna also displays stress. Stress corresponds to morphological structure, licenses additional tonal and segmental contrasts, conditions many phonological processes, and plays a central role in grammatical tone processes marking clause type. Empirically, these findings expand our understanding of word prosody in tone languages in general and Amazonian languages in particular. Theoretically, they challenge current models of stress-conditioned phonology and grammatical tone.
This paper presents rhythmic syncope in Mojeño Trinitario, an Arawak language spoken in lowland Bolivia. In this language, every vowel that is in a weak prosodic position can syncopate. The syncope pattern of Mojeño Trinitario is remarkable for several reasons. First, it involves a regular, categorical and complete deletion rather than a statistical reduction of vowels. Second, it applies similarly to words with either of two stress patterns: iambic words, which make up the great majority of words, and trochaic ones, much less numerous. Third, a great variety of consonant sequences are the result of syncope, and syllabification applies again after syncope. Fourth, rhythmic syncope actually underapplies: almost half of the vowels that are in a position to syncopate are maintained, and vowel quality plays a statistical role in immunity to syncope. Fifth, due to a rich morphology and a set of complex phonotactic rules applying sequentially, syncope leads to extreme opacity. The data presented in tins paper in a theory-neutral way contribute to the typology of rhythmic syncope. It will also be of interest to phonologists considering constraint-based vs. derivational models of phonology.
This paper establishes the lexical tone contrasts in the Nigerian language Izon, focusing on evidence for floating tone. Many tonal languages show effects of floating tone, though typically in a restricted way, such as occurring with only a minority of morphemes, or restricted to certain grammatical enviromnents. For Izon, the claim here is that all lexical items sponsor floating tone, making it ubiquitous across the lexicon and as common as pre-associated tone. The motivation for floating tone comes from the tonal patterns of morphemes in isolation and within tone groups. Based on these patterns, all lexical morphemes are placed into one of four tone classes defined according to winch floating tones they end in. Class A morphemes end in a floating ( ‘wife’), class B in ( ‘salt’), class C in ( ‘sand’), and class D in ( ‘him’). This paper provides extensive empirical support for tins analysis and discusses several issues which emerge under ubiquitous floating tone. Issues include the principled allowance of OCP(T) violations, and the propensity for word-initial vowels and low tone to coincide.
Sound symbolism is a phenomenon with broad relevance to the study of language and mind, but there has been a disconnect between its investigations in linguistics and psychology. This study tests the sound-symbolic potential of ideophones—words described as iconic—in an experimental task that improves over prior work in terms of ecological validity and experimental control. We presented 203 ideophones from five languages to eighty-two Dutch listeners in a binary-choice task, in four versions: original recording, full diphone resynthesis, segments-only resynthesis, and prosody-only resynthesis. Listeners guessed the meaning of all four versions above chance, confirming the iconicity of ideophones and showing the viability of speech synthesis as a way of controlling for segmental and suprasegmental properties in experimental studies of sound symbolism. The success rate was more modest than prior studies using pseudowords like bouba/kiki, implying that assumptions based on such words cannot simply be transferred to natural languages. Prosody and segments together drive the effect: neither alone is sufficient, showing that segments and prosody work together as cues supporting iconic interpretations. The findings cast doubt on attempts to ascribe iconic meanings to segments alone and support a view of ideophones as words that combine arbitrariness and iconicity. We discuss the implications for theory and methods in the empirical study of sound symbolism and iconicity.
The field of linguistic typology has made great strides in mapping the structures to be found across languages. We can now ask whether speakers of all languages distribute their ideas over such structures in the same ways. This question is explored here by examining complement constructions in three genealogically and areally unrelated languages. Each offers glimpses into some factors that shape grammar over time. Crosslinguistic differences in grammar arise from what speakers have chosen to say over millennia, but even languages spoken today can provide snapshots of moments in such processes, if we care to listen to their speakers.
The received wisdom is that word-order alternations in Slavic languages arise as a direct consequence of word-order-related information-structure constraints such as ‘Place given expressions before new ones’. In this article, we compare the word-order hypothesis with a competing one, according to which word-order alternations arise as a consequence of a prosodic constraint: ‘Avoid stress on given expressions’. Based on novel experimental and modeling data, we conclude that the prosodic hypothesis is more adequate than the word-order hypothesis. Yet we also show that combining the strengths of both hypotheses provides the best fit for the data. Methodologically, our article is based on gradient acceptability judgments and multiple regression, which allows us to evaluate whether violations of generalizations like ‘Given precedes new’ or ‘Given lacks stress’ lead to a consistent decrease in acceptability and to quantify the size of their respective effects. Focusing on the empirical adequacy of such generalizations rather than on specific theoretical implementations also makes it possible to bridge the gap between different linguistic traditions and to directly compare predictions emerging from formal and functional approaches.
This is the first comparative analysis of prosody in deaf, native-signing children (ages 5;0-8;5) and adults whose first language is American Sign Language (ASL). The goals of this study are to describe the distribution of prosodic cues during acquisition, to determine which cues across age groups are most predictive in determining clausal and prosodic boundaries, and to ascertain how much isomorphy there is in ASL between syntactic and prosodic units. The results show that all cues are acquired compositionally, and that the prosodic patterns in child and adult ASL signers exhibit important differences regarding specific cues; however, in all groups the manual cues are more predictive of prosodic boundaries than nonmanual markers. This is evidence for a division of labor between the cues that are produced to mark constituents and those that contribute to semantic and pragmatic meaning. There is also more isomorphy in adults than in children, so these results add to the debates about isomorphy, suggesting that while there is clear autonomy between prosody and syntax, productions exhibiting nonisomorphy are relatively rare overall.
This article studies the relationship between prosody and desemanticization in grammaticalization processes by means of a well-described phenomenon, the grammaticalization of ‘type’ nouns (type, kind, sort) in present-day English. To this end, 1,155 tokens of the three nouns, retrieved from the ICE-GB corpus, were semantically classified and prosodically analyzed. Our main result is that different synchronically coexisting prosodic patterns correspond to different degrees of grammaticalization. This result provides evidence that desemanticization and erosion proceed hand in hand. Their parallel development is attributed to the demands of iconicity rather than to frequency effects.
Previous studies on a variety of languages have demonstrated that manual gesture is temporally aligned with prosodic prominence. However, the majority of these studies have been conducted on languages with word-level stress. In this paper, we investigate the alignment of manual beat gestures to speech in local varieties of Standard Indonesian, a language whose word prosodic system has been the subject of conflicting claims. We focus on the varieties of Indonesian spoken in the eastern part of the archipelago and Java. Our findings reveal that there is a strong tendency to align gesture to penultimate syllables in the eastern variety and a tendency to align gesture to final syllables in the Javanese variety. Additionally, while the eastern patterns appear to be word based, the Javanese pattern shows evidence of being phrase based. Surprisingly, the penultimate syllable emerges as a gestural anchor in the eastern variety even for two of the three speakers who showed little to no regular prosodic prominence on this syllable. This suggests that gestural alignment may serve to uncover prosodic anchors even when they are not employed by the phonology proper.
We investigate the synchronization of speech and co-speech actions (i.e., manual game moves) in a dyadic game interaction across different levels of information structure and mutual visibility. We analyze cross-modal synchronization as the temporal distance between co-speech actions and corresponding (1) pitch accent peaks and (2) word onsets. For (1), we find no effect of mutual visibility on cross-modal synchronization. However, pitch accent peaks and co-speech actions are more tightly aligned when game moves are prosodically prominent due to information structural needs. This result is in line with a view of a tightly coupled processing of modalities, where prominence-lending modifications in the prosodic structure attract corresponding manual actions to be realized in a way similar to ‘beat’ gestures. For (2), we do find an effect of mutual visibility, under which co-speech actions are produced earlier and more tightly aligned with word onsets. This result is in line with a view in which co-speech actions act as communicative affordances, which may help an early disambiguation of a message, similar to ‘representational’ co-speech gestures.
Utterance-final weakening refers to a prosodic feature found at the right periphery of some clauses in Pite Saami. This paper provides the most thorough general description of this prosodic phenomenon to date. The dataset used comes from an annotated corpus of spontaneous speech collected during the last 60 years. The phonetic-acoustic correlates are a complete devoicing of all segments in the final syllables of the affected clause, although creaky or breathy voice may also be present. Typically only one syllable is affected, but sometimes multiple syllables are affected. No syntactic units appear to correlate with this, and the weakening phase can even cross word boundaries. The phenomenon marginally correlates with gender, dialect, and age, with the speech of older speakers tending to feature it more frequently and with a longer prosodic scope. Similar utterance-final weakening phenomena are likely found in other languages, especially those in surrounding areas.
In Debussy’s mélodies, the placement of poetic stresses within a bar frequently contradicts the natural rhythm of the French language. In the case of his early songs, scholars have rationalised these occurrences through his ‘casual regard for the text’, his youthfulness, or his ‘budding fascination with the poetry’. However, the irregularities present in his later songs have not been explained. While confirming that Debussy’s translation of the prosody into the musical metre and rhythm is anything but strict, my study of meter and rhythm in his mélodies suggests that an exact rendering of versification was not even intended in many of the songs. To demonstrate this aspect of Debussy’s technique, I compare a few poetic lines from the 1882 and 1892 settings of Verlaine’s ‘En sourdine’ and focus on a single rhythmic alteration, whose location, in the light of the song’s other rhythmic and metrical events, reveals its purpose.
Music & spoken language share many features by combining smaller units (e.g., words, notes) into larger structures (e.g., sentences, musical phrases). This hierarchical organization of sound is culturally contingent & communicates meaning to listeners. Comparisons of music & language from a cognitive neuroscience perspective provide several insights into commonalities & differences between these systems, how they are represented in the brain. The cognitive neuroscience research of music & language, emphasizes the pitfalls & promises identified, including (1) the apparent acoustic & structural similarities between these systems, (2) how both systems convey meaning to listeners, (3) how these systems are learned over the course of development, & (4) the ways in which experience in one domain influences processing in the other domain. We conclude that searching for similarities in how these complex systems are structured (e.g., comparing musical syntax to linguistic syntax) represents a pitfall that researchers should approach with caution. A promising approach in this area of research is to examine how general cognitive mechanisms underlie the learning & maintenance of both systems
Prosody not only signals the speaker’s cognitive states but can also imitate various concepts. However, previous studies on the latter, the iconic function of prosody, have mostly analyzed novel words and nonlinguistic vocalizations. To fill this gap in the literature, the current study has examined the iconic potential of the prosodic features of existing Japanese imitative words known as ideophones. In Experiment 1, female Japanese speakers pronounced 20 sentences containing ideophones in infant-directed speech. They used a higher f0 to express faster and more pleasant movements. Similar iconic associations were observed in Experiment 2, in which Japanese speakers chose the best-matching pitch–intensity–duration combination for each of the ideophones. In Experiment 3, Japanese speakers chose the best-matching voice quality – creaky voice, falsetto, harsh voice or whisper – for the ideophones. Falsetto was preferred for a light object’s fast motion, harsh voice for violent motion and whisper for quiet motion. Based on these results, we entertain the possibility that the iconic prosody of ideophones provides a missing link in the evolutionary theory of language that began with iconic vocalizations. Ideophones with varying degrees of iconic prosody can be considered to be located between nonlinguistic vocalizations and arbitrary words in this evolutionary path.