Hostname: page-component-6b88cc9666-zbq8r Total loading time: 0 Render date: 2026-02-14T06:28:29.571Z Has data issue: false hasContentIssue false

Effects of stuttering and sound avoidance on reference production and memory

Published online by Cambridge University Press:  12 February 2026

Si On Yoon*
Affiliation:
Department of Communicative Sciences and Disorders, New York University , New York, NY, USA
Morgan Schuchard
Affiliation:
Center for Disabilities and Development, University of Iowa Stead Family Children’s Hospital, Iowa City, USA
Anu Subramanian
Affiliation:
Department of Communication Sciences and Disorders, University of Iowa, Iowa City, USA
Naomi H. Rodgers
Affiliation:
Department of Communication Sciences and Disorders, University of Iowa, Iowa City, USA
*
Corresponding author: Si On Yoon; Email: sy4195@nyu.edu
Rights & Permissions [Opens in a new window]

Abstract

Adults who stutter (AWS) frequently engage in language monitoring to anticipate and manage stuttering. This linguistic monitoring may reallocate cognitive resources, with potential consequences for language production and memory. We investigated whether AWS’ increased monitoring during production imposes dual-task costs that limit encoding benefits, or whether it enhances memory through deeper conceptual engagement. Thirty-two AWS and sixty-four adults who do not stutter (AWNS) completed a referential communication task in which they described or identified pictures with an experimenter. To simulate AWS’ linguistic monitoring, half of the AWNS performed a simultaneous sound avoidance task (AWNS-SA), prohibiting certain word-initial phonemes. After the communication task, participants completed a recognition memory test for past referents. Results showed that AWS performed more similarly to AWNS than to AWNS-SA in both language production and memory, although AWS’ memory declined on a trial-by-trial basis when stuttering occurred. These findings suggest that linguistic monitoring in AWS does not impose substantial dual-task costs overall, but that stuttering moments can transiently disrupt memory encoding. Together, these results highlight the adaptive nature of linguistic monitoring in AWS and contribute to a broader understanding of how it supports language production and memory across AWS and AWNS.

Information

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press

Introduction

Producing fluent speech requires the coordination of multiple cognitive processes that operate dynamically and in parallel (Levelt, Reference Levelt1989, Reference Levelt1992; Dell, Reference Dell1986, Reference Dell1988; Dell & Jacobs, Reference Dell, Jacobs, Hickok and Small2016). Speakers must attend to and perceive referents in their discourse context, retrieve lexical and phonological information, and flexibly update memory representations to ensure appropriate reference across discourse contexts (e.g., “the coffee,” “the hot coffee,” “the Cappuccino,” “it”). At the same time, they must monitor their speech output to detect and correct phonological or semantic errors before or during articulation (Levelt, Reference Levelt1983, Reference Levelt1989; Postma, Reference Postma2000). The present study investigates how the demands of this monitoring process affect cognitive resource allocation during language production in adults who stutter (AWS). Specifically, we tested whether AWS allocate cognitive resources differently from adults who do not stutter (AWNS) and how such allocation influences subsequent memory encoding and retrieval.

Stuttering and language production

Stuttering

Stuttering is a multifaceted neurodevelopmental condition that involves behavioral, cognitive, and social-emotional components that interact dynamically as they navigate their daily lives—from speech production to life decisions. Behaviorally, stuttering results from disruptions in speech motor planning that lead to breaks in the forward flow of linguistic output (Chang et al., Reference Chang, Garnett, Etchell and Chow2019). Stuttering-like disfluencies include sound/syllable repetitions (e.g., “muh-muh-muh-my”), sound prolongations (e.g., “mmmmy”), and blocks in which there is a stoppage of airflow and movement before releasing the sound (e.g., “---[tension]---my”). Moments of stuttering can also involve associated behaviors like closing one’s eyes, craning one’s head, moving one’s limbs, or exerting excess physical tension—behaviors that people who stutter may consciously or subconsciously do to try to escape moments of stuttering.

Cognitively, the experience of stuttering is not limited to these surface-level behaviors; rather, it involves complex internal processes of monitoring and anticipation that influence how speakers plan and produce language. It is common for AWS to anticipate moments of upcoming stuttering—an interoceptive awareness that develops with age (Jackson et al., Reference Jackson, Gerlach, Rodgers and Zebrowski2018). This anticipation requires linguistic monitoring and can place cognitive demands on the speaker as they make momentary decisions about how to navigate potential trigger words. Such heightened linguistic monitoring may impose additional demands on attentional cognitive resources, potentially disrupting language production processes compared to AWNS.

Socially and emotionally, repeated experiences of stuttering and ensuing listener reactions can evoke feelings of anxiety and embarrassment, which may further heighten self-focused attention and exacerbate speech monitoring demands (Iverach & Rapee, Reference Iverach and Rapee2014). These emotional reactions can become more pronounced in social contexts where fear of negative evaluation or social anxiety is heightened (Bauerly, Reference Bauerly2024).

Over time, the interplay among these behavioral, cognitive, and social-emotional factors can give rise to a range of strategic avoidance behaviors to minimize overt stuttering-like disfluency, such as by word substitutions and circumlocutions (Jackson et al., Reference Jackson, Rodgers and Rodgers2019). While these avoidances can reduce stuttering that listeners hear, they may also increase the speaker’s cognitive load and change the allocation of cognitive resources during language planning. Understanding how such monitoring and avoidance processes shape language production and downstream memory encoding among AWS is central to the present study.

Language production in AWS

Language production is a multistage process involving conceptualization, formulation, and articulation of messages (Levelt, Reference Levelt1989, Reference Levelt1992; Dell, Reference Dell1986, Reference Dell1988; Roelofs, Reference Roelofs1992, Reference Roelofs1997). Speakers first conceptualize their intended message, then select lexical items and organize them syntactically, and finally articulate the utterance while monitoring for errors via both inner and outer speech loops (Levelt, Reference Levelt1983, Reference Levelt1989; Postma, Reference Postma2000). These processes rely on domain-general cognitive resources, including cognitive control to maintain focus (Strijkers et al., Reference Strijkers, Yum, Grainger and Holcomb2011), monitoring to detect and prevent errors (Nozari & Novick, Reference Nozari and Novick2017), working memory to activate lexical and phonological representations (Schwering & MacDonald, Reference Schwering and MacDonald2020), and attentional selection of appropriate linguistic units (Ferreira & Pashler, Reference Ferreira and Pashler2002).

Research suggests that language production mechanisms are disrupted in AWS at both the lexical-semantic and phonological levels (Dell & O’Seaghdha, Reference Dell and O’Seaghdha1992; Maxfield, Reference Maxfield2015; Maxfield et al., Reference Maxfield, Huffman, Frisch and Hinckley2010, Reference Maxfield, Pizon-Moore, Frisch and Constantine2012, Reference Maxfield, Morris, Frisch, Morphew and Constantine2015). While stuttering is typically associated with disruptions in motor speech execution (Smith & Weber, Reference Smith and Weber2017), AWS also exhibit differences in lexical access and production efficiency. For example, in picture-naming tasks, AWS produce naming errors more frequently and demonstrate slower retrieval of low-frequency words than AWNS (Newman & Ratner, Reference Newman and Ratner2007). One prominent hypothesis is that AWS allocate cognitive resources differently than AWNS during language production, engaging in increased linguistic monitoring that diverts attention from primary production processes. Support for this hypothesis comes from electrophysiological studies showing an attenuated P3 component in AWS during language tasks involving concurrent cognitive demands (Maxfield et al., Reference Maxfield, Olsen, Kleinman, Frisch, Ferreira and Lister2016). The P3 component reflects the availability of cognitive resources for target processing (Luck, Reference Luck1998), and its attenuation in AWS suggests that linguistic monitoring imposes a cognitive load. Importantly, AWS and AWNS exhibit comparable P3 magnitudes in non-linguistic tasks, further supporting the notion that AWS experience increased cognitive demands specifically during linguistic-related activities.

The consequences of heightened linguistic monitoring in AWS—which may function as a dual task, requiring speakers to produce speech while simultaneously monitoring for stuttering-like disfluencies—extend beyond fluency to broader cognitive functions. If AWS engage in such dual-task activities, their ability to encode linguistic information for later retrieval may be impaired. This raises the possibility that AWS’ linguistic monitoring affects not only their real-time speech fluency but also their memory representations of discourse. Alternatively, monitoring may not always be detrimental; increased self-monitoring could also deepen engagement with the linguistic material, fostering more distinctive or elaborated representations that enhance memory (McCurdy et al., Reference McCurdy, Viechtbauer, Sklenar, Frankenstein and Leshikar2020; Yoon et al., Reference Yoon, Benjamin and Brown-Schmidt2016). The present study tests these possibilities by examining how language production and monitoring demands influence memory encoding in AWS.

The interface between language and memory in AWNS

Language production and modifier use

When cognitive resources are taxed, such as under high attentional load or divided focus, speakers often exhibit increased error rates, slower lexical access, and reduced fluency (Ferreira & Pashler, Reference Ferreira and Pashler2002; Schnur et al., Reference Schnur, Schwartz, Kimberg, Hirshorn, Coslett and Thompson-Schill2009). These demands are evidenced during referential communication, in which speakers must select expressions that effectively identify intended referents for listeners. In such contexts, speakers use modifiers (e.g., “the hot coffee” in the presence of hot and iced coffee; see Figure 1, left panel) to disambiguate referents. While modifiers serve an essential communicative function, speakers sometimes overproduce them even when no contrasting alternative exists (e.g., “the hot coffee” when no contrast exists) (Pechmann, Reference Pechmann1989; Tarenskeen, Broersma, & Geurts, Reference Tarenskeen, Broersma and Geurts2015).

Figure 1. Example stimuli in the contrast (left) and non-contrast condition (right) for speakers.

Overmodification has been linked to multiple cognitive and communicative factors. It may arise from limited attentional control or increased processing load, leading speakers to rely on redundant features (Bannon, Reference Bannon2019). Alternatively, speakers may deliberately overspecify to ensure listeners’ comprehension, particularly when they anticipate potential misunderstandings (Rubio-Fernandez, Reference Rubio-Fernandez2016; see also Yoon, Reference Yoon2018). Since language production relies on domain-general control mechanisms to manage competition among lexical alternatives (Nozari, Reference Nozari2018), disruptions in attentional allocation—such as those possibly induced by increased linguistic monitoring among AWS—may influence modifier use and, in turn, potentially affect referential specificity.

The interaction between memory and language

Language production fundamentally shapes how individuals form and retain memory representations. In conversation, speakers typically remember referents they describe better than listeners who hear them (Yoon, Benjamin, & Brown-Schmidt, Reference Yoon, Benjamin and Brown-Schmidt2016, Reference Yoon, Benjamin and Brown-Schmidt2021; McKinley, Benjamin, & Brown-Schmidt, Reference McKinley, Brown-Schmidt and Benjamin2017; Zormpa et al., Reference Zormpa, Brehm, Hoedemaker and Meyer2019). This “speaking benefit” aligns with the well-established generation effect in memory research (Slamecka & Graf, Reference Slamecka and Graf1978), wherein self-generated information is encoded and retrieved better than passively received input. Producing language requires speakers to select lexical items, plan syntactic structures, and integrate conceptual and linguistic information, all of which promote deeper, more elaborative encoding.

Moreover, when speakers describe referents with greater linguistic specificity—such as by adding modifiers—they show enhanced memory for those referents (Yoon et al., Reference Yoon, Benjamin and Brown-Schmidt2016; Reference Yoon, Benjamin and Brown-Schmidt2021; Yoon & Brown-Schmidt, Reference Yoon, Brown-Schmidt, Hartsuiker and Strijkers2023). For example, a speaker who says the hot coffee instead of simply the coffee must retrieve and integrate visual, conceptual, and linguistic information to distinguish that object from others in the same context. Similarly, producing a phrase like the hot coffee in the gray mug requires selecting features that uniquely identify the intended referent and engaging additional cognitive resources during message formulation. The extra information—typically expressed through modifiers—also shapes which aspects of the referent are encoded in memory representations. As a result, increased linguistic and conceptual elaboration strengthens the memory trace for both the referent and its associated features (Broadshaw & Anderson, Reference Broadshaw and Anderson1982; Yoon et al., Reference Yoon, Benjamin and Brown-Schmidt2016). Thus, generating more specific or contrastive descriptions promotes deeper encoding, which may explain why speakers typically remember the items they describe more accurately than listeners who hear them.

Among AWS, however, this language-memory interaction may differ. Heightened linguistic monitoring during language production may divert cognitive resources from encoding referential content, potentially reducing the typical memory advantage observed in AWNS. Conversely, such monitoring may deepen processing by increasing engagement with linguistic planning. The present study examines how monitoring during language production influences memory in AWS relative to AWNS, providing new insight into how cognitive control and fluency interact in communication.

The current research

Although AWS are known to often anticipate and attempt to avoid stuttering, little is known about how these behaviors influence cognitive resource allocation during language production and subsequent memory encoding. In the present study, we applied a psycholinguistic paradigm to test how increased linguistic monitoring in AWS affects both language production and memory for discourse referents.

We hypothesized that heightened monitoring may serve as a dual task, requiring AWS to divide attention between linguistic formulation and monitoring for potential stuttering. This divided attention could disrupt efficient linguistic formulation and reduce the depth of encoding for referential information, leading to poorer memory performance. Alternatively, increased monitoring may promote deeper engagement with linguistic and conceptual content, counteracting potential dual-task costs and resulting in memory performance comparable or superior to AWNS.

To test these possibilities, we compared AWS with AWNS in a referential communication task designed to elicit modifier use. A subgroup of AWNS engaged in a secondary sound-avoidance condition (AWNS-SA), simulating the attentional load AWS may experience during speech monitoring. We predicted that if monitoring primarily taxes cognitive resources, AWS performance would resemble that of the AWNS-SA group, showing differences in modifier use (e.g., omitting necessary or producing unnecessary modifiers) and reduced memory for referents. However, if monitoring instead enhances engagement and processing, AWS may demonstrate preserved memory representations relative to AWNS. By testing these competing hypotheses, this study aims to provide insight into how linguistic monitoring in AWS affects both real-time language production and memory formation and retrieval, shedding light on the cognitive mechanisms underlying stuttering and its broader cognitive consequences.

Experiment method

The Institutional Review Board at the University of Iowa approved this study. All data are publicly available on the Open Science Framework (OSF) and can be accessed at https://osf.io/xa76n/overview.

Participants

A total of 96 native English speakers (ages 18–45) participated in the study, receiving $10 per hour for their participation. Inclusion criteria required that participants (1) report English as their native language and (2) self-report normal hearing and visual acuity (corrected or uncorrected). Sixty-four AWNS participants with no history of communication disorders were recruited from the community. Additionally, 32 AWS participants were recruited from the last author’s research registry, which includes individuals who stutter and have previously participated in her research, consenting to be contacted for future studies. All AWS participants self-identified as people who stutter, reported childhood onset of stuttering, and indicated no co-occurring speech, language, or literacy issues.

Each of the three groups—AWS, AWNS, and AWNS-SA—had an equal number of participants. The AWS group (n = 32; M age = 29.31 years, SD = 7.25) included 14 women, 17 men, and one non-binary person. The AWNS group (n = 32; M age = 28.38 years, SD = 8.72) included 23 women and 9 men. The AWNS-SA group (n = 32; M age = 28.59 years, SD = 8.22) included 13 women, 18 men, and one non-binary person. The AWNS-SA group was designed to simulate the divided attentional demands that AWS may experience in language production due to sound avoidance.

The groups were matched on age and education level, with no significant between-group differences in age (F(2,93) = 0.12, p = .89) or education level (F(2,92) = 1.31, p = .28). Participant demographics are provided in Table 1.

Table 1. Demographic information (mean and standard deviation) for each group

Note: One participant in the AWS group did not report their education level.

Procedure

Following the consent process using REDCap electronic data capture tools hosted at the University of Iowa (Harris et al., Reference Harris, Taylor, Thielke, Payne, Gonzalez and Conde2009; Harris et al., Reference Harris, Taylor, Minor, Elliott, Fernandez, O’Neal, McLeod, Delacqua, Delacqua, Kirby and Duda2019), participants completed the referential communication task, a series of cognitive measures, and an unexpected recognition memory test. To prevent ceiling effects in memory performance (Shepard, Reference Shepard1967), a filler taskFootnote 1 was introduced before the recognition memory test (see also Yoon et al., Reference Yoon, Benjamin and Brown-Schmidt2016, Reference Yoon, Benjamin and Brown-Schmidt2021; Saryazdi, Nuque, & Chambers, Reference Saryazdi, Nuque and Chambers2022). The study was conducted via Zoom and lasted approximately one hour. Audio and video recordings were collected for data coding.

Referential communication task

At the start of the scheduled session, the experimenter emailed participants PowerPoint slides containing the stimuli for the referential communication task. Participants and the experimenter were assigned to begin as either the speaker or the listener. To simulate a face-to-face conversation, Zoom was set to speaker view (self-view off), ensuring that participants could see only the experimenter’s face and vice versa.

For each trial, both participants viewed a slide containing four images (Figures 1 and 2). The target image was marked with a red circle on the speaker’s slide only (Figure 1). The speaker described the target image to the listener (e.g., “The target is coffee”). While the carrier phrase (“The target is …”) was not required, participants often naturally used full sentences (e.g., “I see …” or “The picture is …”). The listener then used their mouse to move a blue square onto the target image (Figure 2). This interactive task mimicked natural communication in that speakers could describe the image freely, but they were not allowed to use location-based descriptions (e.g., “The bottom right one”). Listeners could ask for clarification if a description was insufficiently specific (e.g., “coffee” when both hot and iced coffee were present).

Figure 2. Example stimuli in the contrast (left) and non-contrast condition (right) for listeners. Listeners used their mouse to move the blue square to the target item once the speaker provided the verbal description.

Once the listener selected the target, they instructed the speaker to proceed to the next trial. Each participant completed 48 trials in their initial role before taking a short break and switching roles. The order of speaking and listening was counterbalanced across participants.

In total, participants completed 96 trials, consisting of 48 critical trials (24 production + 24 comprehension) and 48 filler trials (24 production + 24 comprehension). In critical trials, the discourse context was manipulated to include either a contrast item or no contrast item (Figures 1 and 2). In the contrast condition (12 production + 12 comprehension trials), a semantically related contrast image (e.g., target = hot coffee, contrast = iced coffee) was present, requiring a modified expression (e.g., “the hot coffee”) to identify the target. In the non-contrast condition (12 production + 12 comprehension trials), the target image had no related items (e.g., “the coffee” was sufficient). The target and contrast items were counterbalanced across eight lists (e.g., the hot coffee was the target in one list and a competitor in another). Additionally, 48 filler trials were included to increase stimulus variability. Half of these trials required a modifier, while the other half did not. The same filler trials were used across lists.

Sound avoidance condition (AWNS-SA group)

Participants in the attention-divided group (AWNS-SA) simultaneously performed a sound avoidance task during the referential communication task. This task was adapted from the word taboo paradigm used to elicit aphasia-like behaviors in neurotypical adults (Meffert et al., Reference Meffert, Tillmanns, Heim, Jung, Huber and Grande2011). Participants were instructed to avoid using words that began with four common phonemes: /f/, /k/, /p/, and /s/. These phonemes were selected as they were the second, third, fourth, and fifth most common word-initial sounds in the target items. For example, the word “coffee” (initial /k/) was prohibited, requiring participants to use an alternative word (e.g., “espresso” or “drink”). On 38% of trials, the target item’s name began with a prohibited phoneme. Participants were not provided specific alternative strategies beyond selecting different nouns.

Unexpected recognition memory test

After completing the referential communication task and the cognitive measures, participants took an unexpected recognition memory test. In each trial, they were shown a single image and asked whether they had seen it during the previous communication task (e.g., “Have you seen this picture before?”). Participants clicked “yes” if they recognized the image from the communication task, which could include target images (e.g., the hot coffee in Figure 1), contrast images (e.g., the iced coffee in Figure 1), and filler images (e.g., the key from a filler trial). They clicked “no” if they did not recall seeing the image during the task.

Half of the pictures were previously seen (“old” items), and the other half were new, previously unseen images. The old items included 48 critical target images (e.g., the hot coffee) that had been described by either the participant or the experimenter, 48 contrast images (e.g., the iced coffee), which were present but not described during the communication task, and 48 filler images (e.g., the key from a filler trial), half of which had been described, while the other half had been presented but not described. Additionally, there were 144 new images from the same categories as the old items (e.g., a different image of hot coffee). In total, the memory test consisted of 288 trials assessing participants’ recognition of past referents.

Stuttering anticipation scale (AWS only)

At the end of the study, AWS participants rated their anticipated stuttering and avoidance behaviors during the referential communication task on a 5-point Likert scaleFootnote 2 (1 = Never, 5 = Always). On average, AWS rated their anticipation at 3.07 (SD = 1.27, range = 1–5), indicating they sometimes expected to stutter. Their avoidance behavior ratings averaged 2.41 (SD = 1.37, range = 2–5), indicating they sometimes avoided certain words. These patterns align with prior findings on individual differences in stuttering anticipation and avoidance (Jackson et al., Reference Jackson, Gerlach, Rodgers and Zebrowski2018; Rodgers & Jackson, Reference Rodgers and Jackson2021).

Thirty AWS also completed the Stuttering Anticipation Scale (SAS; Jackson et al., Reference Jackson, Gerlach, Rodgers and Zebrowski2018). On this measure, AWS first rated how often they anticipated stuttering in their daily lives on a 5-point Likert scale (M = 3.80, SD = 0.81, range = 2–5). Using the same Likert scale, they then rated how frequently they engaged in 25 common behavioral responses to anticipation. The average sum score was 66.03 (SD = 10.41, range = 43–91) out of a possible range of 25–125.

Coding, planned analyses, and predictions

Communication task

We transcribed the audio recordings from the communication task and coded whether participants produced a modifier in each trial when they spoke. The analysis focused on the initial referring expression produced by the speaker before receiving any feedback from the experimenter, minimizing the potential influence of partner feedback on participants’ descriptions (see Yoon et al., Reference Yoon, Duff and Brown-Schmidt2017). Specifically, we examined the use of pre-noun modifiers (e.g., “the hot coffee”), which reflect pre-planned language prior to production (Brown-Schmidt & Konopka, Reference Brown-Schmidt and Konopka2008). Post-noun modifiers (e.g., “the coffee that is hot”) were excluded because it is less clear whether they result solely from the speaker’s internal language planning processes or are influenced by partner feedback, including both verbal and backchannel cues. For instance, a partner’s hedging may encourage speakers to produce longer utterances, potentially leading to the addition of post-noun modifiers.

We also coded whether AWS stuttered in each trial using a binary coding system: fluent (coded as 0) or stuttered (coded as 1). If a participant produced any stuttering-like disfluencies (i.e., sound/syllable repetitions, prolongations, blocks) at any point while describing the target image—regardless of the type or frequency of stuttering, or whether the stutter occurred on the target noun (e.g., “coffee”) or another word in the utterance—it was coded as 1. Fillers (e.g., “um,” “uh”) were not coded as stuttering, as it is often unreliable to determine whether fillers reflect stuttering avoidance or typical non-stuttered disfluencies due to language processing stallsFootnote 3 .

The final dataset comprised 2,264 picture descriptions (24 critical production trials per participant), excluding 40 trials in which participants did not produce a noun (e.g., producing only a modifier). All of these excluded trials came from the AWNS-SA group.

To analyze the modifier data, we used a binomial (logit-link) mixed-effects model with contrast type (contrast vs. non-contrast) and Group (dummy-coded, with AWS as the reference group: (1) AWS vs. AWNS and (2) AWS vs. AWNS-SA) as fixed effects, with participants and items as random intercepts. Random slopes were included whenever possible. The dependent variable was binary—whether the target noun was modified or not.

All models were fit using the lme4 package (Bates et al., Reference Bates, Mächler, Bolker and Walker2015) in R (version 4.1.0; R Core Team, 2021), using the maximal random effects structure for participants and items. The R package “buildmer” was used to identify a convergent maximal random effects structure (see Voeten, Reference Voeten2023).

Unexpected recognition memory test

Following the analysis process in Yoon et al. (Reference Yoon, Benjamin and Brown-Schmidt2016, Reference Yoon, Benjamin and Brown-Schmidt2021), we calculated discriminability (d′) in the unexpected recognition memory test to assess participants’ ability to distinguish between old and new items. Discriminability was determined by subtracting the standardized false alarm rate (e.g., a “yes” response to a new item they had not seen before, such as a different hot coffee) from the standardized hit rate (e.g., a “yes” response to an old item they had seen before, such as the hot coffee in Figure 1). A higher d’ value indicates better memory of the target item.

Inferential statistics for the recognition memory data were conducted for target items using a logistic mixed-effects model. The model included contrast type (contrast vs. non-contrast), group (dummy-coded with AWS as the reference; (1) AWS vs. AWNS and (2) AWS vs. AWNS-SA), role (speaker vs. listener), and actual item type (old vs. new) as fixed effects, with participants and items as random intercepts. Random slopes were also included. The dependent measure was binary, indicating whether the participant responded “yes” (old image = 1) or “no” (new image = 0).

The intercept term in the model measured response bias, specifically whether participants showed a general inclination to respond “yes” or “no.” Actual item type (old vs. new) was examined to determine whether participants were more likely to respond “yes” when the item was old, representing the primary measure of memory (see Fraundorf, Watson, & Benjamin, Reference Fraundorf, Watson and Benjamin2010; Wright, Horry, & Skagerberg, Reference Wright, Horry and Skagerberg2009). The beta weights associated with the fixed effects indicated the influence of these variables on response bias. Finally, the interaction between contrast type, role, group, and actual item type was analyzed to assess how these factors affected memory for past items.

The final dataset for the unexpected recognition memory test included 9,024 target items (48 old and 48 new items per participant), excluding two AWS participants due to experimenter errors.

We also examined the interaction between stuttering during language production and memory. As part of these exploratory analyses, we investigated whether stuttering during a trial in the communication task or SAS scores affected memory for past referents.

Hypotheses

Our primary goals were to examine (1) modifier use in the communication task, (2) memory performance, and (3) the relationship between referential expressions produced during the communication task and the accuracy of subsequent memory performance.

In the referential communication task, we hypothesized that attentional disruptions (e.g., those associated with anticipating or attempting to avoid moments of stuttering) would affect the production of appropriate referring expressions. Specifically, if increased monitoring acts as a dual-task burden that draws resources away from linguistic formulation, speakers in the AWS and AWNS-SA groups would produce fewer modified expressions in the contrast condition compared to speakers in the AWNS group. Alternatively, if linguistic monitoring promotes deeper engagement with conceptual and linguistic processing, AWS may show preserved or even enhanced modifier use, similar to that of the AWNS group.

With respect to memory, we hypothesized that AWS and AWNS-SA would exhibit comparable memory performance, both showing reduced memory accuracy relative to AWNS, reflecting the shared attentional demands of monitoring or dual-task performance. However, if monitoring enhances rather than depletes cognitive resources, facilitating more elaborative encoding, AWS may demonstrate equivalent or better memory performance than AWNS.

Finally, among AWS, if overt stuttering or behavioral reactions to stuttering anticipation impose additional cognitive costs, we hypothesized that stuttering moments would be associated with reduced memory accuracy within the AWS group. Together, these hypotheses allowed us to test whether monitoring during language production is primarily a resource-dividing cost or an engagement-driven mechanism that may support memory encoding.

Results

Results in the referential communication task

The results of the communication task (Figure 3, Table 2) indicate that speakers across all three groups used more pronoun modifiers to describe target images in the contrast condition (e.g., hot coffee) than in the non-contrast condition (e.g., coffee).

Figure 3. Proportion of pre-noun modified expressions in the referential communication task.

Table 2. Proportion of modified expressions in the referential communication task

We analyzed the use of modified expressions using a binomial (logit-link) mixed-effects model (Table S2). The model revealed significant main effects of contrast (contrast vs. non-contrast; b = −0.58, z = −3.14, p < .01) and Group2 (AWS vs. AWNS-SA; b = −1.05, z = −4.66, p < .0001). Participants in all three groups produced more modifiers in the contrast condition than in the non-contrast condition. Additionally, AWS participants produced more modifiers overall than AWNS-SA participants. However, the overall modifier rate did not significantly differ between AWS and AWNS participants (b = −0.19, z = −0.85, p = .39).

The significant interaction between Group 1 and contrast (b = −0.56, z = −2.30, p = .02) was driven by a larger contrast effect (i.e., the difference in pre-noun modifier rates between the contrast and non-contrast conditions) in the AWNS group (b = −1.15, z = −5.84, p < .0001) compared to the AWS group (b = −0.56, z = −3.34, p = .001).

Similarly, the significant interaction between Group 2 and contrast (b = −0.86, z = −3.53, p < .001) was driven by a larger contrast effect in the AWNS-SA group (b = −1.38, z = −7.50, p < .0001) than in the AWS group (b = −0.56, z = −3.34, p = .001). Interestingly, the AWS group exhibited a smaller contrast effect than the other two groups. This difference was due to a higher rate of modifiers in the non-contrast condition; AWS participants produced more unnecessary modifiers even when they were not required in the local context, such as when only one coffee was present on the right panel in Figure 1.

A separate mixed-effects model comparing only the AWNS and AWNS-SA groups revealed significant main effects of Group (AWNS vs. AWNS-SA; b = −0.88, z = −3.68, p < .001) and contrast (b = −1.29, z = −7.90, p < .001). However, the interaction between Group and contrast was not significant (b = −0.38, z = −1.28, p = .20).

Results in the unexpected recognition memory test

We analyzed recognition memory data for the target (Figure 4), following the methods in Yoon et al. (Reference Yoon, Benjamin and Brown-Schmidt2016, Reference Yoon, Benjamin and Brown-Schmidt2021). Recognition memory for target items was assessed using a binomial (logit-link) mixed-effects model (Table S3 in the appendix). The model included group (dummy-coded with AWS as the reference; (1) AWS vs. AWNS and (2) AWS vs. AWNS-SA), role (speaker vs. listener), and actual item type (old vs. new) as fixed effectsFootnote 4 . The interactions between actual item type and fixed effects are examined, controlling response bias for “yes.” Across all groups, participants exhibited better memory for past referents when they described the item themselves rather than when they listened to the picture description (b = 1.10, z = 4.88, p < .001), consistent with the generation effect (Slamecka & Graf, Reference Slamecka and Graf1978). Interestingly, the AWS group showed better memory for past referents than the AWNS-SA group (b = −0.53, z = −3.49, p < .01) across both speaker and listener roles. However, memory performance did not differ significantly between the AWS and AWNS groups (b = −0.13, z = −0.82, p = .41). These findings suggest that AWS formed memory representations as robust as those of AWNS participants, with speaker versus listener roles influencing memory representations across all groups.

Figure 4. Discriminability (d′) for the target items in the unexpected recognition memory test.

A separate mixed-effects model comparing only the AWNS and AWNS-SA groups revealed significant effects of Group (AWNS vs. AWNS-SA; b = −0.63, z = −2.01, p = .04), Role (speaking vs. listening; b = 1.21, z = 4.74, p < .001), and contrast (b = 0.45, z = 2.02, p = .04).

Effect of stuttering on memory

We further examined how stuttered trials influenced AWS participants’ memory for target images. Specifically, we investigated whether their memory performance differed between fluent and stuttered trials. During the communication task, AWS stuttered in 23% of trials (25.6% in the contrast condition, 20.8% in the non-contrast condition)Footnote 5 . As shown in Table 3, in the non-contrast condition, AWS’ memory for target items declined when they stuttered compared to when they did not. However, in the contrast condition, stuttering did not significantly impact memory performance.

Table 3. Accuracy of the memory test as a function of stuttering in the corresponding trial in the referential communication in AWS

AWS’ memory for target items and the occurrence of stuttering were analyzed using a binomial (logit-link) mixed-effects model, with contrast (contrast vs. non-contrast) and stuttering (whether the participant stuttered in that trial) as fixed effects. The dependent measure was whether the participant correctly recognized the target item. The model revealed significant main effects of contrast (b = −7.18, z = −3.73, p < .001) and Stuttering (b = −1.40, z = −2.18, p = .03). Trial-by-trial analysis indicated that AWS exhibited better memory for target items in the contrast condition than in the non-contrast condition, likely due to the higher modification rate in the contrast condition. Additionally, memory performance for target items declined when AWS stuttered while describing the item. However, the interaction between Condition and Stuttering was not significant (b = 0.44, z = 0.36, p = .72).

We also examined whether SAS responses (i.e., AWS participants’ general stuttering anticipation and behavioral reactions to anticipation) were associated with performance in the communication task and memory test. Neither the overall degree of stuttering anticipation nor the degree of specific behavioral reactions to anticipation was significantly associated with language production (b = 0.03, z = 0.15, p = .88; b = 0.02, z = 1.40, p = .16) or subsequent memory performance (b = −0.45, z = −1.67, p = .10; b = −0.02, z = −1.20, p = .23).

Discussion

In this study, we investigated how linguistic monitoring in AWS influences both language production and memory for discourse referents. Specifically, we were interested in how stuttering anticipation and potential stuttering avoidance shape the allocation of cognitive resources during language production and how these processes influence subsequent memory for previously mentioned referents. During the referential communication task, participants either described target images to a partner or listened to their partner’s descriptions. Afterward, their memory for past referents was tested. To examine the effect of attentional demands associated with monitoring and avoidance behaviors, we included a comparison group of AWNS and a subgroup of AWNS who performed a simultaneous sound avoidance task (AWNS-SA). Below, we discuss three primary findings:

  1. 1. Sensitivity to discourse context and increased modifier use in AWS: participants across all groups adjusted their language production according to discourse context, producing more modifiers when a semantically related contrast item was present compared to when it was absent. Compared to the other two groups, AWS produced more unnecessary modifiers when the discourse context did not require them.

  2. 2. Memory performance across groups: AWNS-SA exhibited reduced memory performance compared to AWS and AWNS, suggesting that explicit divided attention required for the sound avoidance task impaired memory encoding in AWNS-SA, whereas AWS and AWNS showed comparable memory performance.

  3. 3. Effect of stuttering on memory: trial-level analyses revealed that stuttering events disrupted memory for target items in AWS.

Sensitivity to discourse context in language production

Across all three groups, speakers produced more modifiers in contrast trials than in non-contrast trials, demonstrating sensitivity to discourse context even when performing a secondary task (AWNS-SA and possibly AWS). However, AWNS-SA produced fewer modifiers overall, consistent with the idea that divided attention limits resources available for message formulation. Their efforts to avoid taboo sounds in the simultaneous sound avoidance task may have made it harder to retrieve appropriate words quickly. Alternatively, their reduced modifier use may reflect both lexical retrieval difficulties and cognitive interference resulting from maintaining the sound-avoidance constraints.

Interestingly, AWS and AWNS-SA differed in their use of modifiers. AWS produced more modifiers than AWNS-SA, suggesting that the cost of monitoring differs from that of performing an unfamiliar secondary task. Unlike the AWNS-SA group, AWS may be more habituated to monitoring and redistributing attentional resources during language production, minimizing its disruptive effects on language production. Rather than reflecting a simple resource deficit, these findings suggest that monitoring in AWS may be better conceptualized as a dynamic control process—one that enhances engagement with language formulation.

AWS produced more modifiers than both AWNS and AWNS-SA, particularly in the non-contrast condition where modifiers were not necessary. One possible explanation is that AWS uses modifiers as a compensatory strategy to ensure their partner fully understands their intended message. Another possibility is that this reflects overcompensation, in which they provide more words than needed even when those extra details do not improve communication. This overproduction may stem from self-blame or perceived guilt about stuttering, as suggested by Boyle (Reference Boyle2020), with speakers attempting to show speakers their fluency by overelaborating.

Memory performance across groups

The memory findings contribute to understanding of the speaking benefit in memory—the advantage of producing over hearing words—operates under varying cognitive demand. Specifically, they highlight the differences between speakers engaged in typical conversation (AWNS) and those performing under a divided attention condition (AWS and AWNS-SA). Consistent with the generation effect (Slamecka & Graf, Reference Slamecka and Graf1978), results showed that producing words rather than passively hearing them enhanced memory—even in contexts that taxed attentional resources.

However, AWNS-SA exhibited reduced memory accuracy compared to both AWS and AWNS, suggesting that explicit divided attention during language production hinders referential encoding. In contrast, AWS’ memory performance was comparable to that of the AWNS, despite their ongoing monitoring demands. This finding indicates that linguistic monitoring in AWS does not uniformly tax cognitive resources and instead supports the hypothesis that increased monitoring may, under certain conditions, deepen engagement with linguistic processing, thereby offsetting potential dual-task costs.

Although prior work suggests that modified expressions enhance memory for referents (Yoon et al., Reference Yoon, Benjamin and Brown-Schmidt2016), AWS’ higher modifier rate in the referential communication task did not yield better memory performance. This finding appears to suggest that increased linguistic monitoring can reallocate cognitive resources in ways that limit encoding benefits. However, it may instead reflect that monitoring promotes more detailed—but not necessarily durable—encoding. Thus, monitoring does not always impair memory; its effects likely depend on which control resources are engaged and the balance between effortful monitoring and meaningful engagement (Nozari & Novick, Reference Nozari and Novick2017).

Why, then, did AWNS-SA demonstrate lower memory performance than AWS? If we assume that both groups engaged in sound avoidance, one might have expected comparable performance between them. However, our findings suggest that linguistic monitoring in AWS does not uniformly tax cognitive resources. For AWS, linguistic monitoring is a habitual process deeply integrated into speech planning, whereas for AWNS-SA, the sound-avoidance task introduced a novel, cognitively demanding dual-task condition.

AWS may have developed efficient strategies to allocate attention between formulation and monitoring, minimizing interference with memory encoding. In contrast, AWNS-SA participants, for whom the sound-avoidance task represented a novel and cognitively taxing dual-task condition, likely devoted additional working memory resources to maintain the arbitrary constraint, increasing cognitive load and disrupting referential encoding. Despite limitations in fully simulating the experience of stuttering, this comparison provides one of the first explorations of sound avoidance in AWNS, offering a potential model for understanding avoidance strategies in stuttering.

Taken together, these findings suggest that many AWS develop deeply engrained, adaptive skills in verbal avoidance. Over years of managing attentional and fluency demands, AWS may acquire efficient strategies to circumvent words or sound sequences that are likely to trigger disfluency (Newman & Ratner, Reference Newman and Ratner2007). Such practiced control could allow AWS to maintain perceptually fluent speech with little measurable impact on language or subsequent memory for past referents, even under substantial internal monitoring and reformulation. However, the “mental gymnastics” of such linguistic rearranging can feel cognitively taxing. From this perspective, the absence of memory differences between the AWS and AWNS groups may not indicate the absence of cognitive cost per se, but rather one aspect of the experience that can be unintentionally beneficial. These adaptive strategies may also explain why monitoring does not always impair memory: habitual control could reduce the attentional toll of monitoring, thereby preserving encoding efficiency.

Effect of stuttering on language production and memory in AWS

At the trial level, overt stuttering was associated with reduced memory accuracy, suggesting that momentary disruptions in fluency disrupt attention to referential content. While AWS’ overall memory was comparable to AWNS, stuttered trials among AWS showed local impairments in encoding, consistent with transient shifts in cognitive focus from message formulation to motor control.

Measures of anticipation and behavioral reactions to stuttering—as measured by the Stuttering Anticipation Scale (SAS)—were not significantly associated with overall language or memory performance, underscoring the variability in how individuals experience and respond to stuttering anticipation (Rodgers & Jackson, Reference Rodgers and Jackson2021; Tichenor & Yaruss, Reference Tichenor and Yaruss2021; Tichenor, Herring, & Yaruss, Reference Tichenor, Herring and Yaruss2022). This variability highlights that the cognitive consequences of monitoring and avoidance depend on individual adaptation. Future research should further explore how individual differences in monitoring (automatic vs. strategic) differentially affect both real-time language production and long-term memory outcomes.

Limitations

This study investigated the potential impact of linguistic monitoring on language production and subsequent memory in AWS. While AWS are assumed to frequently monitor their speech, it remains unclear how they engaged in the communication task and managed anticipated stuttering on a trial-by-trial basis. Linguistic monitoring is often covert, and AWS may attempt to circumvent stuttering moments through strategies such as word switching or circumlocution, which can prevent overt stuttering from being perceived by listeners. In our exploratory analysis, we examined how overt stuttering in a trial influenced language production and subsequent memory. However, future research is needed to fully understand how stuttering avoidance (resulting in perceptually fluent utterances) affects cognitive resources during language production and memory in AWS. This may be achieved by tailoring stimuli to AWS’ unique list of words that they anticipate stuttering on, which has been done in other studies (e.g., Jackson et al., Reference Jackson, Dravida, Zhang, Noah, Gracco and Hirsch2022; Orpella et al., Reference Orpella, Flick, Assaneo, Shroff, Pylkkänen, Poeppel and Jackson2024).

Another potential limitation is the use of a sound avoidance task in the AWNS-SA group. While this task was designed to simulate the linguistic avoidance strategies used by AWS, it is important to acknowledge that sound avoidance was novel to AWNS-SA participants, whereas AWS has had years of experience managing stuttering-related avoidance strategies. Additionally, AWS may have engaged in an implicit dual task—simultaneously monitoring and producing language—whereas the AWNS-SA group performed an explicit dual task that required them to consciously adhere to the taboo sound rule. These differences may have imposed distinct cognitive demands across the groups. Furthermore, AWS likely employed various strategies beyond avoiding specific sounds or words to prevent overt stuttering. Individual differences in stuttering anticipation and avoidance strategies are clinically meaningful and cannot be fully captured by a single experimental task like the one imposed on AWNS-SA participants in this study.

Conclusion

Our findings suggest that AWS can effectively coordinate monitoring, planning, and production processes without substantial disruption to attentional or cognitive resources. Although AWS tended to produce more words in the communication task, even when additional modifiers were unnecessary, suggesting that AWS may prioritize communicative clarity and social appropriateness, providing additional detail to facilitate their conversation partners’ comprehension or save face rather than to support their own memory. Additionally, we found that moments of overt stuttering while describing an object were associated with reduced memory for referents, indicating that transient disruptions in fluency can momentarily redirect attention away from conceptual encoding. However, the overall similarity in memory performance between AWS and AWNS highlights that long-term adaptation to monitoring demands may mitigate such costs. Taken together, these findings contribute to a more nuanced understanding of linguistic monitoring as a flexible control mechanism, one that can both tax and enhance processing, depending on task demands and individual adaptation. Beyond advancing psycholinguistic models of language and memory, this work underscores the importance of studying not only the nature of stuttering but also how it serves as a window into how speakers balance cognitive control, communicative goals, and social engagement in everyday conversation. Future research should further examine individual differences in how AWS linguistically navigate stuttering anticipation and the impact of these strategies on both speakers’ and listeners’ communication experiences.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/S0142716425100428

Replication package

All data and analytic code have been made publicly available at the Open Science Framework (OSF) and can be accessed at https://osf.io/xa76n.

Acknowledgements

This research was supported by the National Center for Advancing Translational Sciences of the National Institutes of Health under Award Number UM1TR004403. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Competing interests

The author(s) declare none.

Footnotes

1 Participants completed a 20-minute battery of cognitive assessments measuring processing speed, working memory, and inhibitory control. The measures and results are reported in Supplemental Material 1. The cognitive battery, included as a filler task, was used to characterize participant groups and did not influence the results of the referential communication task (e.g., modification rates). Therefore, we do not discuss it further.

2 Responses from five AWS participants, including the two participants who did not complete the stuttering anticipation scale, were not recorded due to technical difficulties or experimenter errors.

3 We coded the use of filler words (e.g., um, uh) separately because filler words were frequently observed across all three groups. The results showed that the speakers in the AWNS (33.6%) and AWNS-SA (34.8%) groups produced more fillers compared to the speakers in the AWS group (23.2%).

4 We also examined a model that included Contrast Type as an additional fixed effect, but its effect was not significant. To maintain a simpler model, we report the model with three fixed effects.

5 To examine whether disfluency (e.g., filler words such as “um” or “uh”) during language production affected memory performance across groups in AWS and AWNS, we conducted an additional analysis, which revealed no significant effect of disfluency on target memory.

References

Bannon, J. (2019). Working memory and referential communication: an investigation of the cognitive factors affecting the production of overspecified referring expressions (Master’s thesis, McMaster University). MacSphere. https://hdl.handle.net/11375/24024 Google Scholar
Bates, D., Mächler, M., Bolker, B., & Walker, S. (2015). Fitting linear mixed-effects models using lme4. Journal of Statistical Software, 67(1), 148. https://doi.org/10.18637/jss.v067.i01 CrossRefGoogle Scholar
Bauerly, K. R. (2024). Characteristics associated with social anxiety in adults with developmental stuttering: A review. Medical Research Archives, 12(10), 5876. https://doi.org/10.18103/mra.v12i10.5876 CrossRefGoogle ScholarPubMed
Boyle, M. P. (2020). Psychological correlates of biological and non-biological explanations for stuttering. International Journal of Speech-Language Pathology, 22, 435443. https://doi.org/10.1080/17549507.2019.1663931 CrossRefGoogle ScholarPubMed
Broadshaw, G. L., & Anderson, J. R. (1982). Elaborative encoding as an explanation of levels of processing. Journal of Verbal Learning and Verbal Behavior, 21(2), 165174. https://doi.org/10.1016/S0022-5371(82)90531-X CrossRefGoogle Scholar
Brown-Schmidt, S., & Konopka, A. E. (2008). Little houses and casas pequeñas: Message formulation and syntactic form in unscripted speech with speakers of English and Spanish. Cognition, 109(2), 274280. https://doi.org/10.1016/j.cognition.2008.07.011 CrossRefGoogle ScholarPubMed
Chang, S. E., Garnett, E. O., Etchell, A., & Chow, H. M. (2019). Functional and neuroanatomical bases of developmental stuttering: Current insights. Neuroscientist, 25(6), 566582. https://doi.org/10.1177/1073858418803594 CrossRefGoogle ScholarPubMed
Dell, G. S. (1986). A spreading-activation theory of retrieval in sentence production. Psychological Review, 93, 283321. https://doi.org/10.1037/0033-295X.93.3.283 CrossRefGoogle ScholarPubMed
Dell, G. S. (1988). The retrieval of phonological forms in production: Tests of predictions from a connectionist model. Journal of Memory and Language, 27, 124142. https://doi.org/10.1016/0749-596X(88)90070-8 CrossRefGoogle Scholar
Dell, G. S., & Jacobs, C. L. (2016). Successful Speaking: Cognitive Mechanisms of Adaption in Language Production. In Hickok, G. & Small, S. L. (Eds), Neurobiology of Language. Academic Press. https://doi.org/10.1016/B978-0-12-407794-2.00018-3 CrossRefGoogle Scholar
Dell, G. S., & O’Seaghdha, P. G. (1992). Stages of lexical access in language production. Cognition, 42(1–3), 287314. https://doi.org/10.1016/0010-0277(92)90046-K CrossRefGoogle Scholar
Ferreira, V. S., & Pashler, H. (2002). Central bottleneck influences on the processing stages of word production. Journal of Experimental Psychology: Learning, Memory, and Cognition, 28(6), 11871199. https://doi.org/10.1037/0278-7393.28.6.1187 Google ScholarPubMed
Fraundorf, S. H., Watson, D. G., & Benjamin, A. S. (2010). Recognition memory reveals just how contrastive accenting really is. Journal of Memory and Language, 63(3), 367386. https://doi.org/10.1016/j.jml.2010.06.004 CrossRefGoogle Scholar
Harris, P. A., Taylor, R., Minor, B. L., Elliott, V., Fernandez, M., O’Neal, L., McLeod, L., Delacqua, G., Delacqua, F., Kirby, J., Duda, S. N., & REDCap Consortium (2019). The REDCap consortium: Building an international community of software platform partners. Journal of Biomedical Informatics, 95, 103208. https://doi.org/10.1016/j.jbi.2019.103208 CrossRefGoogle ScholarPubMed
Harris, P. A., Taylor, R., Thielke, R., Payne, J., Gonzalez, N., & Conde, J. G. (2009). Research electronic data capture (REDCap) – a metadata-driven methodology and workflow process for providing translational research informatics support. Journal of Biomedical Informatics, 42(2), 377381. https://doi.org/10.1016/j.jbi.2008.08.010 CrossRefGoogle ScholarPubMed
Iverach, L., & Rapee, R. M. (2014). Social anxiety disorder and stuttering: Current status and future directions. Journal of Fluency Disorders, 40, 6982. https://doi.org/10.1016/j.jfludis.2013.08.003 CrossRefGoogle ScholarPubMed
Jackson, E. S., Dravida, S., Zhang, X, Noah, J. A., Gracco, V. & Hirsch, J. (2022). Activation in right dorsolateral prefrontal cortex underlies stuttering anticipation. Neurobiology of Language, 3(3), 469494. https://doi.org/10.1162/nol_a_00073 CrossRefGoogle ScholarPubMed
Jackson, E. S., Gerlach, H., Rodgers, N. H., & Zebrowski, P. (2018). My client knows that he is about to stutter: How can we address stuttering anticipation during therapy with young people who stutter? Seminars in Speech and Language, 39, 356370. https://doi.org/10.1055/s-0038-1667164 Google ScholarPubMed
Jackson, E. S., Rodgers, N. H., & Rodgers, D. (2019). An exploratory factor analysis of action responses to stuttering anticipation. Journal of Fluency Disorders, 60, 110. https://doi.org/10.1016/j.jfludis.2019.03.001 CrossRefGoogle ScholarPubMed
Levelt, W. J. M. (1983). Monitoring and self-repair in speech. Cognition, 14, 41104. https://doi.org/10.1016/0010-0277(83)90026-4 CrossRefGoogle ScholarPubMed
Levelt, W. J. M. (1989). Speaking: from intention to articulation. The MIT Press. https://doi.org/10.7551/mitpress/6393.001.0001 CrossRefGoogle Scholar
Levelt, W. J. M. (1992). Accessing words in speech production: stages, processes and representations. Cognition, 42(1–3), 122. https://doi.org/10.1016/0010-0277(92)90038-J CrossRefGoogle Scholar
Luck, S. J. (1998). Sources of dual-task interference: evidence from human electrophysiology. Psychological Science, 9(3), 223227. https://doi.org/10.1111/1467-9280.00043 CrossRefGoogle Scholar
Maxfield, N. D. (2015). Language and the Seaking Mind/Brain in Stuttering, Routledge Handbook of Communication Disorders. New York: Routledge (p. 258267).Google Scholar
Maxfield, N. D., Huffman, J. L., Frisch, S. A., & Hinckley, J. J. (2010). Neural correlates of semantic activation spreading on the path to picture naming in adults who stutter. Clinical Neurophysiology, 121, 14471463. https://doi.org/10.1016/j.clinph.2010.03.026 CrossRefGoogle ScholarPubMed
Maxfield, N. D., Morris, K., Frisch, S. A., Morphew, K., & Constantine, J. L. (2015). Real-time processing in picture naming in adults who stutter: ERP evidence. Clinical Neurophysiology, 126, 284296. https://doi.org/10.1016/j.clinph.2014.05.009 CrossRefGoogle ScholarPubMed
Maxfield, N. D., Pizon-Moore, A. A., Frisch, S. A., & Constantine, J. L. (2012). Exploring semantic and phonological picture-word priming in adults who stutter using event-related potentials. Clinical Neurophysiology, 123, 11311146. https://doi.org/10.1016/j.clinph.2011.10.003 CrossRefGoogle ScholarPubMed
Maxfield, N. D., Olsen, W. L., Kleinman, D., Frisch, S. A., Ferreira, V. S., & Lister, J. J. (2016). Attention demands of language production in adults who stutter. Clinical Neurophysiology, 127(4), 19421960. https://doi.org/10.1016/j.clinph.2016.01.016 CrossRefGoogle ScholarPubMed
McCurdy, M. P., Viechtbauer, W., Sklenar, A. M., Frankenstein, A. N., & Leshikar, E. D. (2020). Theories of the generation effect and the impact of generation constraint: A meta-analytic review. Psychonomic Bulletin & Review, 27(6), 11391165. https://doi.org/10.3758/s13423-020-01762-3 CrossRefGoogle ScholarPubMed
McKinley, G. L., Brown-Schmidt, S., & Benjamin, A. S. (2017). Memory for conversation and the development of common ground. Memory and Cognition, 45, 12811294. https://doi.org/10.3758/s13421-017-0730-3 CrossRefGoogle ScholarPubMed
Meffert, E., Tillmanns, E., Heim, S., Jung, S., Huber, W., & Grande, M. (2011). Taboo: A novel paradigm to elicit aphasia-like trouble-indicating behaviour in normally speaking individuals. Journal of Psycholinguist Research, 40, 307. https://doi.org/10.1007/s10936-011-9170-6 CrossRefGoogle ScholarPubMed
Newman, R. S., & Ratner, N. B. (2007). The role of selected lexical factors on confrontation naming accuracy, speed, and fluency in adults who do and do not stutter. Journal of Speech, Language, and Hearing Research, 50, 196213. https://doi.org/10.1044/1092-4388(2007/016) CrossRefGoogle Scholar
Nozari, N. (2018). How special is language production? Perspectives from monitoring and control. Psychology of Learning and Motivation, 68, 179213. https://doi.org/10.1016/bs.plm.2018.08.006 Google Scholar
Nozari, N., & Novick, J. (2017). Monitoring and control in language production. Current Directions in Psychological Science, 26, 403410. https://doi.org/10.1177/0963721417702419 CrossRefGoogle Scholar
Orpella, J., Flick, G., Assaneo, M. F., Shroff, R., Pylkkänen, L., Poeppel, D., & Jackson, E. S. (2024). Reactive inhibitory control precedes overt stuttering events. Neurobiology of Language, 5(2), 432453. https://doi.org/10.1162/nol_a_00138 CrossRefGoogle ScholarPubMed
Pechmann, T. (1989). Incremental speech production and referential overspecification. Linguistics, 27(1), 89110.10.1515/ling.1989.27.1.89CrossRefGoogle Scholar
Postma, A. (2000). Detection of errors during speech production: A review of speech monitoring models. Cognition, 77, 97131. https://doi.org/10.1016/s0010-0277(00)00090-1 CrossRefGoogle ScholarPubMed
R Core Team (2021). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Vienna, Austria. https://www.R-project.org/.Google Scholar
Rodgers, N. H., & Jackson, E. S. (2021). Temperament is linked to avoidant responses to stuttering anticipation. Journal of Communication Disorders, 93, 106139. https://doi.org/10.1016/j.jcomdis.2021.106139 CrossRefGoogle ScholarPubMed
Roelofs, A. (1992). A spreading-activation theory of lemma retrieval in speaking. Cognition, 42, 107142. https://doi.org/10.1016/0010-0277(92)90041-F CrossRefGoogle ScholarPubMed
Roelofs, A. (1997). The WEAVER model of word-form encoding in speech production. Cognition, 64(3), 249284. https://doi.org/10.1016/S0010-0277(97)00027-9 CrossRefGoogle ScholarPubMed
Rubio-Fernandez, P. (2016). How redundant are redundant color adjectives? An efficiency based analysis of color overspecification. Frontiers in Psychology, 7, 153. https://doi.org/10.3389/fpsyg.2016.00153 CrossRefGoogle ScholarPubMed
Saryazdi, R., Nuque, J., & Chambers, C. G. (2022). Linguistic redundancy and its effects on younger and older adults’ real-time comprehension and memory. Cognitive Science, 46(4), e13123. https://doi.org/10.1111/cogs.13123 CrossRefGoogle ScholarPubMed
Schnur, T. T., Schwartz, M. F., Kimberg, D. Y., Hirshorn, E., Coslett, H. B., & Thompson-Schill, S. L. (2009). Localizing interference during naming: Convergent neuroimaging and neuropsychological evidence for the function of Broca’s area. Proceedings of the National Academy of Sciences, 106(1), 322327. https://doi.org/10.1073/pnas.0805874106 CrossRefGoogle Scholar
Schwering, S. C., & MacDonald, M. C. (2020). Verbal working memory as emergent from language comprehension and production. Frontiers in Human Neuroscience, 14, 68. https://doi.org/10.3389/fnhum.2020.00068 CrossRefGoogle ScholarPubMed
Shepard, R. N. (1967). Recognition memory for words, sentences, and pictures. Journal of Verbal Learning and Verbal Behavior, 6(1), 156163. https://doi.org/10.1016/S0022-5371(67)80067-7 CrossRefGoogle Scholar
Slamecka, N. J., & Graf, P. (1978). The generation effect: delineation of a phenomenon. Journal of Experimental Psychology: Human Learning and Memory, 4, 592604. https://doi.org/10.1037/0278-7393.4.6.592 Google Scholar
Smith, A., & Weber, C. (2017). How stuttering develops: the multifactorial dynamic pathways theory. Journal of Speech, Language, and Hearing Research, 60(9), 24832505. https://doi.org/10.1044/2017_JSLHR-S-16-0343 CrossRefGoogle ScholarPubMed
Strijkers, K., Yum, Y. N., Grainger, J., & Holcomb, P. J. (2011). Early goal-directed top-down influences in the production of speech. Frontiers in Psychology, 2, 371. https://doi.org/10.3389/fpsyg.2011.00371 CrossRefGoogle ScholarPubMed
Tarenskeen, S., Broersma, M., & Geurts, B. (2015). Overspecification of color, pattern, and size: Salience, absoluteness, and consistency. Frontiers in Psychology, 6, 1703. https://doi.org/10.3389/fpsyg.2015.01703 CrossRefGoogle ScholarPubMed
Tichenor, S. E., Herring, C., & Yaruss, J. J. (2022). Understanding the speaker’s experience of stuttering can improve stuttering therapy. Topics in Language Disorders, 42, 5775. https://doi.org/10.1097/tld.0000000000000272 CrossRefGoogle ScholarPubMed
Tichenor, S. E. & Yaruss, J. S. (2021). Variability of stuttering: Behavior and impact. American Journal of Speech Language Pathology, 30, 7588. https://doi.org/10.1044/2020_AJSLP-20-00112 CrossRefGoogle ScholarPubMed
Voeten, C. C. (2023). buildmer: Stepwise elimination and term reordering for mixed-effects regression. https://cran.r-project.org/web/packages/buildmer/buildmer.pdf Google Scholar
Wright, D. B., Horry, R., & Skagerberg, E. M. (2009). Functions for traditional and multilevel approaches to signal detection theory. Behavior Research Methods, 41, 257267. https://doi.org/10.3758/BRM.41.2.257 CrossRefGoogle ScholarPubMed
Yoon, S. O. (2018). The role of discourse context in reference production and comprehension: Insights from the lexical differentiation effect. Psychology of Learning and Motivation, 68, 361392. https://doi.org/10.1016/bs.plm.2018.08.009 Google Scholar
Yoon, S. O., Benjamin, A. S., & Brown-Schmidt, S. (2016). The historical context in conversation: Lexical differentiation and memory for the discourse history. Cognition, 154, 102117. https://doi.org/10.1016/j.cognition.2016.05.011 CrossRefGoogle ScholarPubMed
Yoon, S. O., Benjamin, A. S., & Brown-Schmidt, S. (2021). Referential form and memory for the discourse history. Cognitive Science, 45(4), e12964. https://doi.org/10.1111/cogs.12964 CrossRefGoogle ScholarPubMed
Yoon, S. O., & Brown-Schmidt, S. (2023). Understanding language use in social contexts: the role of past and present discourse contexts. In Hartsuiker, R. & Strijkers, K. (Eds), Current issues in the Psychology of Language (pp. 284303). UK: Routledge Press (Taylor & Francis).Google Scholar
Yoon, S. O., Duff, M. C., & Brown-Schmidt, S. (2017). Learning and using knowledge about what other people do and don’t know despite amnesia. Cortex, 94, 164175. https://doi.org/10.1016/j.cortex.2017.06.020 CrossRefGoogle Scholar
Zormpa, E., Brehm, L.E., Hoedemaker, R.S., & Meyer, A.S. (2019) The production effect and the generation effect improve memory in picture naming. Memory 27(3), 340352. https://doi.org/10.1080/09658211.2018.1510966 CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. Example stimuli in the contrast (left) and non-contrast condition (right) for speakers.

Figure 1

Table 1. Demographic information (mean and standard deviation) for each group

Figure 2

Figure 2. Example stimuli in the contrast (left) and non-contrast condition (right) for listeners. Listeners used their mouse to move the blue square to the target item once the speaker provided the verbal description.

Figure 3

Figure 3. Proportion of pre-noun modified expressions in the referential communication task.

Figure 4

Table 2. Proportion of modified expressions in the referential communication task

Figure 5

Figure 4. Discriminability (d′) for the target items in the unexpected recognition memory test.

Figure 6

Table 3. Accuracy of the memory test as a function of stuttering in the corresponding trial in the referential communication in AWS

Supplementary material: File

Yoon et al. supplementary material

Yoon et al. supplementary material
Download Yoon et al. supplementary material(File)
File 24.7 KB