Highlights
-
• Infants learning ATR harmony languages use ATR cues for speech segmentation.
-
• Multilingual infants in Ghana are sensitive to native language segmentation cues.
-
• Little exposure to harmony language may suffice for multilinguals to track harmony cues.
1. Introduction
Word segmentation poses a fundamental challenge for language-learning infants. This task involves extracting individual words from continuous speech, which is intricate due to the absence of clear pauses or boundaries between words in infants’ input (Cutler & Otake, Reference Cutler and Otake1994; Gonzelez-Gomez & Nazzi, Reference Gonzalez-Gomez and Nazzi2013). Segmentation, therefore, presents a challenge for infants who are yet to acquire a lexicon: they must find out where words start and end in the speech stream (e.g., Jusczyk, Reference Jusczyk1997), as achieving this feat is critical for infants’ lexical acquisition (Cristia et al., Reference Cristia, Seidl, Junge, Soderstrom and Hagoort2014; Junge et al., Reference Junge, Kooijman, Hagoort and Cutler2012; Singh et al., Reference Singh, Reznick and Xuehua2012).
Previous research has revealed that infants rely on various cues from their ambient language to facilitate the detection of words from continuous speech (e.g., Höhle & Weissenborn, Reference Höhle and Weissenborn2003). Some of the cues explored in infants so far include prosodic regularities, which involve rhythmic units such as the strong/weak stress pattern of words in languages like English (e.g., Echols et al., Reference Echols, Crowhurst and Childers1997; Jusczyk et al., Reference Jusczyk, Houston and Newsome1999; Orena & Polka, Reference Orena and Polka2019), and phonotactic regularities, which refers to the rules or constraints governing the possible sound combinations, phoneme positions and syllable structures in a language (e.g., Friederici & Wessels, Reference Friederici and Wessels1993; Gonzalez-Gomez & Nazzi, Reference Gonzalez-Gomez and Nazzi2013; Mattys & Jusczyk, Reference Mattys and Jusczyk2001). The latter (i.e., phonotactic regularities) is the focus of the present study.
Our understanding of infant word segmentation primarily comes from studies conducted with infants growing up in Western contexts learning Indo-European languages, particularly English (Nazzi et al., Reference Nazzi, Iakimova, Bertoncini, Mottet, Serres, de Schonen, Friederici and Thierry2008). This lack of diversity could, however, cause biases in theory building (Aravena-Bravo et al., Reference Aravena-Bravo, Cristia, Garcia, Nicolas, Laranjo, Arokoyo, Benavides-Varela, Benders, Boll-Avetisyan, Ben, Diop, Durán-Urzúa, Manalili, Narasimhan, Omane, Kolberg, Ssemata, Styles, Cristia and Early2023; Cristia et al., Reference Cristia, Foushee, Aravena-Bravo, Cychosz, Scaff and Casillas2023; Kidd & Garcia, Reference Kidd and Garcia2022; Singh et al., Reference Singh, Cristia, Karasik, Rajendra and Oakes2023). The current study contributes to diversifying infant language acquisition research and cross-linguistic understanding of phonological acquisition by focusing on infants growing up in Ghana, Sub-Saharan Africa (an understudied population). Language acquisition in this population is important in particular because infants typically grow up multilingually with several understudied African languages (e.g., Ga, Akan and Ewe) in multi-ethnic and multi-generational homes (see also Omane et al., Reference Omane, Benders, Duah and Boll-Avetisyan2023, Reference Omane, Benders and Boll-Avetisyan2024, Reference Omane, Benders and Boll-Avetisyan2025). To this point, very little is known about how multilingual exposure affects infants’ speech perception.
The present study investigated infants’ use of vowel harmony (VH), a language-specific, phonotactic constraint on the co-occurrence of vowels in a word or syllable as a cue for early speech segmentation. The VH constraint posits that vowels in a di- or multi-syllabic word should share some specific vocalic features. To take the language and the vowel harmony rule under test in the present study as an example: in Akan (a Kwa Niger-Congo language), an Advanced Tongue Root (ATR) harmony language, vowels are grouped into two different sets: +ATR (/i,e,u,o/) and -ATR (/ɪ,ʊ,ɛ,ɔ/), based on the position of the tongue root during the articulation of these vowels. The ATR VH constraint in Akan requires vowels from the same set to occur within a word (with few exceptions), for example, +ATR: kube ‘coconut’ and -ATR: ɛkɔm ‘hunger’ (Dolphyne, Reference Dolphyne1988).
Previous research has established that monolingual infants initially seem to be biased for preferring VH patterns: four-month-old English-learning infants show a listening preference for harmonic over disharmonic syllable sequences (Solá-Llonch & Sundara, Reference Solá-Llonch and Sundara2025). However, from the age of 6 months onwards, only infants learning a VH language exhibit this listening preference (Altan et al., Reference Altan, Kaya, Hohenberger, Haznedar and Ketrez2016; Hohenberger et al., Reference Hohenberger, Altan, Kaya, Tuncer, Avcu, Haznedar and Ketrez2016, Reference Hohenberger, Kaya and Altan2017; van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). In contrast, infants learning a language without VH show no such preference (Gonzalez-Gomez et al., Reference Gonzalez-Gomez, Schmandt, Fazekas, Nazzi and Gervain2019; van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008; Solá-Llonch & Sundara, Reference Solá-Llonch and Sundara2025), suggesting that the preference at later ages is a result of attunement to the input language. Infants growing up multilingually in Ghana typically acquire languages both with and without VH. Our recent work showed that these infants also exhibit a listening preference for VH patterns at the age of 6 months, and this preference did not depend on their degree of regular exposure to VH languages (Omane et al., Reference Omane, Benders and Boll-Avetisyan2024), suggesting that minimal exposure to a VH might be sufficient for maintaining the VH bias. This pattern of results leads to the question of whether VH has a function in language acquisition and processing.
Research on both adults and infants has found that VH affects speech segmentation, facilitating listeners’ detection of potential word boundaries in continuous speech. Specifically, it was found that speakers of languages with VH, but not those of languages without, use the absence of harmony between adjacent syllables to infer the location of word boundaries (e.g., Kabak et al., Reference Kabak, Maniwa and Kazanina2010; Suomi et al., Reference Suomi, McQueen and Cutler1997; Vroomen et al., Reference Vroomen, Tuomainen and De Gelder1998), suggesting that the use of VH is a language-specific cue for speech segmentation. Moreover, two previous studies show that young infants already exploit VH as a segmentation cue (Mintz et al., Reference Mintz, Walker, Welday and Kidd2018; van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). Van Kampen et al. (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008) investigated backness harmony in 9-month-old infants learning Turkish, a VH language. In the head-turn preference paradigm, infants were familiarized with two text passages containing three-syllable nonwords consisting of a bisyllabic stem and a monosyllabic nonsense prefix. In one text passage, the vowel of the prefix harmonized with the vowels of the bisyllabic stem (e.g., nu-namoll), and in the other, it did not (e.g., lo-netiss). At test, each infant heard isolated bisyllabic stems that were familiarized with a harmonizing prefix (e.g., namoll), familiarized with a non-harmonizing prefix (e.g., netiss) or novel (e.g., batull). Infants showed longer looking times for the familiarized stems that had been presented with a non-harmonizing prefix than those that were presented with a harmonizing prefix. This led the authors to conclude that infants inferred a word boundary within the three-syllabic string after the non-harmonizing prefix, indicating that infants use the absence of VH between adjacent syllables as a speech segmentation cue. Furthermore, Mintz et al. (Reference Mintz, Walker, Welday and Kidd2018; experiment 2c) explored whether infants learning English, a language without VH, would rely on VH as a universal cue for speech segmentation using an artificial language learning paradigm. Seven-month-old infants were briefly exposed (for less than a minute) to a continuous speech stream (e.g., ditepubobidetupo). After this, at test, when presented with bisyllabic sequences that had been present in the speech stream, infants preferred listening to sequences containing exclusively front or back vowels (e.g., dite, pubo) over sequences that combined front and back vowels (e.g., detu, bodi). Results from these prior studies suggest that infants are sensitive to VH cues and can use them for segmenting words even after minimal exposure to speech containing vowel harmony patterns. These results give rise to the hypothesis that multilingual infants with some exposure to an ATR harmony language should also be able to use ATR harmony cues for segmentation.
Prior research on bilingual infants’ speech segmentation with cues other than phonotactic cues (e.g., stress) suggests that bilingual infants might segment speech in each of their two languages similarly to monolingual infants. For example, while monolingual French- and English-learning infants can only segment words in their native language (i.e., French-learning infants relying on iambic stress, and English-learning infants on trochaic stress as segmentation cues) but not from a non-native language that has a non-native stress pattern, French-English bilingual infants can segment bisyllabic words from both their languages (Polka & Sundara, Reference Polka and Sundara2003; Polka et al., Reference Polka, Orena, Sundara and Worrall2017). Likewise, in another study with French-English bilingual infants, Orena and Polka (Reference Orena and Polka2019) tested infants in an inter-mixed dual language task. These infants heard text passages that mixed French and English, with the language changing after every two sentences within a passage and with French target words embedded in French sentences and English target words embedded in English sentences. The result was that bilingual infants segmented bisyllabic words in both languages. Segmentation of monosyllabic words has also been found in Spanish–Catalan (Bosch et al., Reference Bosch, Figueras, Teixidó and Ramon-Casas2013) and English–Mandarin (Singh, Reference Singh2018; Singh & Foong, Reference Singh and Foong2012) bilingual infants. Unlike segmentation, in other domains of perception, e.g., in perceptual preferences and speech discrimination, bilingual infants are often found to perform better in the language they are exposed to more (e.g., Liu & Kager, Reference Liu and Kager2015; Sebastián-Gallés & Bosch, Reference Sebastián-Gallés and Bosch2002).
While these findings suggest robust segmentation in bilinguals’ two languages using language-specific cues, it remains an open question whether multilingual infants rely on phonotactic cues for speech segmentation. Moreover, it is not yet known whether segmentation ability is or is not impacted by the degree of exposure to the language(s) with this phonotactic cue. While adults speaking VH, but not adults speaking non-VH languages, use VH cues for segmenting artificial languages (e.g., Kabak et al., Reference Kabak, Maniwa and Kazanina2010; Suomi et al., Reference Suomi, McQueen and Cutler1997), infant research suggests that infants may be able to use VH for segmentation even when exposed to a language without this cue (Mintz et al., Reference Mintz, Walker, Welday and Kidd2018; Van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). However, although Mintz et al.’s (Reference Mintz, Walker, Welday and Kidd2018) findings suggest that even infants without any exposure to VH can use this cue for segmentation, their study used a very simple artificial language, possibly overestimating the use of vowel harmony in children with no exposure to it. The study by Van Kampen et al. (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008) provides some further indication that infants’ use of VH for segmentation from natural speech is not dependent on exclusive exposure to VH languages: the Turkish-learning infants in that study were living in Germany with some exposure to German (a non-VH language). Yet, it is still not clear how infants with different levels of simultaneous exposure to languages with VH alongside those without VH will use this cue for speech segmentation. Answering this question will help us understand how multilingual exposure impacts phonotactic processing of the ambient language, and, more precisely, how multilingual exposure may affect their segmentation ability with cues that are only relevant to a subset of their languages.
2. The present study
For the first time, the current study investigated whether multilingually raised infants learning minimally one language with and one without VH can use VH cues when dealing with the task of detecting word boundaries in fluent speech. We focused on a specific type of VH, the Advanced Tongue Root (ATR) harmony in Akan, which statistically is the most common type of VH in the world given its occurrence in genetically diverse languages (Rose, Reference Rose2018). The acquisition of ATR harmony has only been investigated in our recent study on 6-month-old infants in Ghana, finding sensitivity to ATR harmony irrespective of the degree of exposure to ATR harmony languages (Omane et al., Reference Omane, Benders and Boll-Avetisyan2024). Studying multilingual infants exposed to an ATR harmony language and non-vowel harmony languages will contribute to our understanding of how the degree of exposure to multiple diverse languages affects infants’ ability to use VH, specifically ATR harmony, for segmentation. Research questions, hypotheses and methods, including the analysis plan, were pre-registered (see https://osf.io/pqk2n).
The current study aimed to address two questions. The first question was whether multilingual infants exposed to minimally one ATR harmony language and one non-VH language would rely on ATR harmony cues for speech segmentation. To address this, multilingually raised infants performed a word segmentation task with naturally recorded passages in Akan (familiarization) followed by isolated bisyllables (test), which had been embedded in the passages as target words. Within the passages, target words occurred in contexts manipulated for having either a harmony context or a disharmony context, where vowels in the target word either harmonize or disharmonize in ATR with vowels of an attached modifier. We hypothesized that multilingually raised infants exposed to an ATR harmony language (e.g., Akan) will rely on ATR harmony cues for segmentation. Two predictions are possible: (1) If infants segment words based on VH cues, it will be most likely that they display a familiarity preference, i.e., that they listen longer to familiarized target words that appeared in the disharmony context passages than to both novel words and words that appeared in the harmony context passages. However, infants may also show a novelty preference by listening longer to novel words and to familiarized target words that appeared in the harmony context passage than to words that occurred in a disharmony context passage. Both outcomes would suggest infants’ use of VH for segmentation.Footnote 1
The second question was whether the relative amount of exposure to (an) ATR harmony language(s) would modulate the use of harmony cues in speech segmentation. There are reasons to hypothesize that (1) exposure to an ATR harmony language might modulate the use of harmony cues for segmentation but also reasons to hypothesize that (2) exposure to an ATR harmony language might not modulate the use of harmony cues for segmentation. If exposure modulates the use of harmony cues in segmentation, the difference in looking times to target words familiarized in a disharmony context versus those familiarized in a harmony context will be more enhanced the more infants are exposed to (an) ATR harmony language(s).
3. Methods
3.1. Participants
The final sample consisted of 40 full-term (i.e., 37 weeks or more gestation) infants aged between 9 and 11 months (22 boys; Mean age: 306 days; range = 274–363) being raised multilingually in Accra, Ghana. Infants were recruited from two hospitals (the University of Ghana Hospital, Legon, and a hospital that prefers to remain anonymous) and through personal networks (snowball sampling) in Accra. Previous speech segmentation studies have commonly tested between 14 to 32 infants per group (e.g., Gonzalez-Gomez & Nazzi, Reference Gonzalez-Gomez and Nazzi2013; Jusczyk et al., Reference Jusczyk, Houston and Newsome1999; Orena & Polka, Reference Orena and Polka2019; van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). Since we expected heterogeneity regarding language experience in our sample, we aimed for a larger sample. Consistent with our a-priori estimation, we successfully tested 40 infants during the 6 months of fieldwork available for data collection. An additional six infants were tested, but their data were excluded from the analyses presented here because of parent interference during the experiment (n = 2), the child crying (n = 1), experimenter errors (n = 1) and technical failure – the experiment ending abruptly after a few test trials (n = 2). No infant was reported to have any hearing or vision problems. As an inclusion criterion, all infants were learning Akan (an ATR harmony language) and minimally one non-VH language (e.g., Ewe, Ga, Ghanaian English). Infants’ language exposure was assessed using the Caregiver Interview about Multilingual Exposure (CIME) tool, an interview assessment protocol, and a logbook (see Omane et al., Reference Omane, Benders and Boll-Avetisyan2025; the same tools were used in Omane et al., Reference Omane, Benders and Boll-Avetisyan2024). Infants were exposed to between two and four languages based on the interview assessment (two languages: 30%, three languages: 55% and four languages: 15%), and two to five languages according to the logbook (two languages: 32.5%, three languages: 50%, four languages: 12.5% and five languages: 5%). An overview of the language combinations infants experienced is provided in the Supplementary Materials, Table S1. Participants were compensated with a fee. The authors assert that all procedures contributing to this work comply with the ethical standards of the relevant national and institutional committees on human experimentation and with the Helsinki Declaration of 1975, as revised in 2008. The study was approved by the Ethics Committee for Humanities at the University of Ghana (Ghana) under the project name ‘Multilingual infants’ use of vowel harmony for speech perception in Akan’ (approval number: ECH 226/ 21–22).
3.2. Auditory stimuli
3.2.1. Nonce target words
We created four harmonic nonce target words (henceforth: target words) that infants needed to segment from spoken text passages. These target words were bisyllabic C1V1C2V2 (C = Consonant, V = Vowel) made up of Akan phonemes (see Dolphyne, Reference Dolphyne1988). Four vowels were selected from the vowel inventory of Akan, which comprised two advanced tongue root vowels (+ATR, one front and one back) and two unadvanced tongue root vowels (−ATR, one front and one back) as given in Table 1.
Table 1. The Set of +ATR and −ATR vowels used in creating the stimuli with their standard phonetic description

Note: The phonetic descriptions are based on the International Phonetic Alphabet (IPA).
From the selected vowels, four two-vowel templates were created. For each template, vowels shared the same ATR feature: either both vowels are +ATR (/_i_u, _u_i /) or -ATR (/_ɛ_ɔ, _ɔ_ɛ/). Vowels in each template also shared vowel height (e.g., /_i_u/ are both high vowels; /_ɔ_ɛ/ are both mid-vowels). Four stop consonants /p, b, t, d/ from the consonant inventory of Akan were selected to create the stimuli. The consonants were combined with the vowel templates to create the bisyllabic C1V1C2V2 target words: tupi, dibu, dɔbɛ and tɛpɔ. Two constraints, based on voicing and place of articulation, determined where consonants were placed: (1) The initial consonant was always a coronal (/t/ or /d/), and the second was a bilabial stop (/p/ or /b/). (2) Both consonants in each template shared the same voicing feature (/t_p_/ or /d_b_/). These restrictions on the consonants in the stimuli ensured a constant degree of complexity across the conditions.
3.3. Nonce modifiers
To create the harmonic and disharmonic sequences in which the target words would occur, we followed Van Kampen et al. (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008) in using four monosyllabic nonce modifiers (henceforth: modifiers) that do not exist in the target language, Akan: ke, kɛ, ge and gɛ, which either have a + ATR /e/ or -ATR /ɛ/ mid vowel. We chose to attach the monosyllables after the target words instead of before based on a grammatical property of Akan: we wanted to assign target words to noun positions and be treated as a noun, and the monosyllables to a post-nominal modifier (e.g., adjectives) position, where they must agree in ATR with the preceding vowel.
3.3.1. Target word and modifier combinations
With the bisyllabic target words and the monosyllabic modifiers, we formed three-syllable combinations (i.e., CVCV+CV). For each target word, two different modifiers were attached to manipulate harmonicity: one modifier shares ATR with the vowels in the target word (e.g., tupi-ke) while the other does not (e.g., tupi-kɛ). Harmonic and disharmonic combinations of target word plus modifier were controlled for changes in vowel height and consonant voicing: The -ATR target words always shared their vowel height with the modifiers (all mid-vowels) but not consonant voicing (e.g., tɛpɔ-gɛ or dɔbɛ-kɛ), irrespective of harmony context. Conversely, the +ATR target words share consonant voicing (e.g., tupi-ke) but not vowel height with the modifiers, regardless of harmony context (see Table 2 for the complete list).
Table 2. Bisyllabic target words plus modifiers (italicized) showing the resulting harmonic and disharmonic three-syllable combinations

With this, infants were first familiarized with trisyllabic affixed forms (stem + modifier, CVCV+CV), inserted into text passage templates described in the paragraph below and tested on their recognition of the bisyllabic stem, similar to van Kampen et al.’s (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008) design (see also Kim & Sundara, Reference Kim and Sundara2015). The noun-modifier structure requires that infants use backward cues to segment speech streams into word-sized units, which previous studies suggest they are able to do (Thiessen et al., Reference Thiessen, Onnis, Hong and Lee2019). Consider below the disharmonic and harmonic context examples in (1) and (2), respectively.
-
(1) tupi-kɛ tiee maame no ‘tupi kɛ listened to the woman’
-
(2) tupi-ke tiee maame no ‘tupi ke listened to the woman’
In (1), the modifier ‘kɛ’ is not harmonic with the preceding target word tupi. So, if infants rely on backward cues from vowel harmony for segmentation, they should assume a word boundary before the modifier (i.e., after the target word) since the ATR feature of the modifier vowel (−ATR) does not spread regressively, indicating that the syllables cannot belong to a single word (e.g., ‘tupi kɛ’ would be parsed as ‘kɛ’ modifying ‘tupi’). In (2) by contrast, the modifier harmonizes with the preceding target word, suggesting that the three syllables could belong to a single trisyllabic word (e.g., ‘tupike’).
3.3.2. Text passages
Two different six-sentence text passage templates, ‘A’ and ‘B,’ were created (see Supplementary Materials 2, Table S2). Each sentence had one slot for a target word plus modifier combination: three in the sentence-initial and three in the non-initial position. The structure of the sentences in the passage templates was such that no target word plus modifier combination appeared at the end of a sentence (i.e., within the verb phrase) to avoid regressive spreading of the +ATR harmony feature to nearby -ATR syllables or words, as regressive +ATR vowel harmony across word boundaries within verb phrases is a phonological rule in Akan (Dolphyne, Reference Dolphyne1988; Kügler, Reference Kügler2015). The ATR features of syllables preceding and following the slots were balanced across the two passage templates. Firstly, the slots in both passage templates were followed by +ATR (di, wei, tie) and -ATR (maa, fʊrʊʊ, pɪra) words three times each. Secondly, the slots in non-initial position were preceded once by +ATR (wei) and twice by -ATR (biaa, ɛnʊra) in passage template A and twice by +ATR (afei, nti) and once by -ATR (ɔkyɪna) in passage template B. Each passage template was assigned one +ATR and one -ATR nonword. Specifically, tupi and tɛpɔ were assigned to passage template A, and dibu and dɔbɛ to passage template B. Both the harmonic and disharmonic trisyllabic sequences created from each nonword were assigned to the same passage template. This ensured that every passage template was assigned an equal number of harmonic and disharmonic combinations (see Table 3 below). Consequently, each infant encountered harmonic and disharmonic contexts in different passage templates. The assignments of harmonic and disharmonic context passages were counterbalanced across infants. For example, in the illustration in Table 3, infant X receives one harmony context passage from template A and one disharmony context passage from template B, while infant Y, in contrast, receives a disharmony context passage from template A and a harmony context passage from template B. From here on, we use the term ‘passage(s)’ to refer to passage templates with embedded target word plus modifier combinations.
Table 3. Summary of how target words plus their modifiers were assigned to the passage templates showing harmonic and disharmonic combinations for each passage, and (between parentheses) illustration of counterbalanced assignment to infants

A female native speaker of Akan recorded all stimuli. All target words were produced with a consistent High–Low (H-L) tone pattern. The isolated CVCV test items were elicited in different sentence types: (1) interrogative: wonim tupi? ‘do you know tupi?’ (2) declarative: yɛfrɛ no tupi! ‘s/he is called tupi!’ (3) imperative: Fa tupi! ‘take tupi!’. This was meant to facilitate the realization of acoustic variability across the tokens of each target word. The speaker was instructed to read all sentence types and passages in an infant-directed manner, with no pauses between words or at phrase boundaries. Thirty different recordings (tokens) of each target word (10 in each sentence type) and four renditions of each passage were recorded. Further processing of the stimuli, such as annotation and extraction of sentences for the passages and test items, and checking the pitch height of extracted target test items, was done in Praat (version 6.2.06, [Boersma & Weenink, Reference Boersma and Weenink2022]).
To create the familiarization stimuli (passages), we selected one recorded rendition of each sentence of the eight six-sentence familiarization passages. For each passage, the sentences were re-concatenated with 500 milliseconds of silence between them. The average duration of a passage was about 13.2 seconds (range: 13.05–13.30 s).
For the test stimuli (isolated words), 15 tokens of each target word were selected and concatenated into a single stream trial with 650 milliseconds of silence between the tokens. In each experimental trial, the infant heard 15 different tokens of a single target word sequence (e.g., tupi). The average duration for a test trial was 15.6 seconds (range: 15.33–15.88 s).
3.4. Visual stimuli
There were two visual stimuli: a rotating colorful wheel and a static checkerboard. The rotating wheel was used as an attention-getter before each trial was presented, while the checkerboard served as an unrelated visual attractor when the auditory stimulus was being played.
4. Procedure
4.1. Experiment
Infants were tested individually in a mobile lab setup in one of the three locations: a room at the Department of Linguistics at the University of Ghana or one of two hospitals. We used the single-screen central fixation paradigm (Cooper & Aslin, Reference Cooper and Aslin1990). To minimize the influence of the experimenter’s presence in the room and activities such as coding and switching trials during testing, the rooms were partitioned into two with a curtain. The infant and the caregiver sat on one side of the partitioned room. The infant sat on the caregiver’s lap in front of a 17.3-inch Asus ROG Strix XG17 screen about 40–50 cm away. Two Logitech z120 pc speakers through which the auditory stimuli were presented were placed at the two bottom edges of the screen. A webcam was placed on top of the screen to monitor the infant during the experiment. The experimenter sat on the other side of the partitioned room to observe and record the infant’s looking behavior via the webcam. The experimenter manually coded the infants’ looking times during the experiment. The experiment was run by the Habit2 program, version 2.2.1 (Oakes et al., Reference Oakes, Sperka, DeBolt and Cantrell2019).
At the onset of each trial, the visual attention-getter appeared at the center of the screen to attract the infant’s attention. Once the infant’s attention was captured, the experimenter initiated the auditory trial, which was accompanied by the visually presented checkerboard at the center of the screen.
4.1.1. Familiarization phase
Half of the infants were familiarized with text passages containing the target words tɛpɔ and dibu, while the other half were familiarized with passages containing tupi and dɔbɛ (see text passages with embedded target word plus modifier combinations in the Supplementary Materials, Table S2). Each infant was familiarized with one harmonic and one disharmonic combination of the target word plus modifier, embedded in the passage templates. Each infant heard either template A passage with a harmonic target word plus a modifier combination or template B passage with a disharmonic combination or vice versa, as illustrated in Table 3. The order of presentation of passages was pseudorandomized across infants. Each passage was repeated two times; hence, every infant heard four passages in total.
4.1.2. Test phase
The auditory trials were presented simultaneously with an unrelated visual stimulus (static checkerboard). Each infant listened to the two target words they heard during familiarization and one novel target word. Novel target words were randomly assigned to infants. Note that a novel target word for one infant was a familiarized word for another. Each target word was presented four times, making a total of 12 trials. A trial consisted of 15 repeated isolated tokens of the same target word. Each infant heard a maximum of 12 test trials, and fewer if the infant became fussy and could not complete all trials. The average duration for a test trial was about 15.6 seconds. The trial ended after the infant looked away for more than 2 consecutive seconds. The infants’ listening time was measured by the length of time that they looked at the screen when an auditory stimulus was being played. Thus, infants’ looking time was used as a proxy for how long they listened to the auditory stimulus.
4.2. Language exposure assessment
4.2.1. Caregiver Interview about Multilingual Exposure (CIME)
Caregivers completed the CIME before the experiment to determine whether infants met the inclusion criteria of gestational age (37 or more weeks of gestation), were learning Akan (an ATR harmony language) and minimally one non-VH language (e.g., Ewe, Ga, Ghanaian English; the same intake criteria as in Omane et al., Reference Omane, Benders and Boll-Avetisyan2024). The CIME asked about the languages directed to the infant by a range of speakers (e.g., father, mother, grandparents, older sibling[s] and other caregivers) as well as those that the infant generally overhears (e.g., at home and religious gatherings). Exposure to each language was calculated as a proportion of time an infant heard a given language over the total time of exposure to all languages (collapsing over both speech directed to the infant and overheard speech). Similar calculations were performed to arrive at ATR-harmony exposure (combining all languages in each infant’s input with ATR harmony, such as Akan and Dagbani; see Supplementary Table S1).
4.2.2. Logbook
In addition to the CIME, a logbook method was used to estimate language exposure throughout an infant’s entire day, as in Omane et al. (Reference Omane, Benders and Boll-Avetisyan2024, Reference Omane, Benders and Boll-Avetisyan2025). Caregivers were given a logbook in the form of a small card placed in the pocket of a vest worn by the baby at home after the experiment to complete at home. They were instructed to indicate, by ticking different options, the different languages their infant heard every half an hour for an entire day between 7 am and 7 pm while the baby was awake. A demonstration accompanied the instruction to ensure that caregivers understood the task. Caregivers typically filled out the logbook a day or two after the experiment. The experimenter picked up the logbook from the majority of the caregivers’ homes after completion (which usually took around 2 hours due to traffic), with only a few caregivers agreeing to return the logbook to the hospital where the experiment was conducted. Language and ATR-harmony exposure were similarly calculated from the logbook data.
5. Data preprocessing and analyses
5.1. Data exclusion
Following our pre-registration and previous studies (e.g., Bosch et al., Reference Bosch, Figueras, Teixidó and Ramon-Casas2013; Junge et al., Reference Junge, Everaert, Porto, Fikkert, de Klerk, Keij and Benders2020), we excluded 11 test trials (2.3% of total trials) across eight infants from the analyses, as total looking time at these trials was less than 1 second. Six of those trials were with familiarized words (disharmonic context: n = 2, harmonic context: n = 4), and five with novel words.
5.2. Statistical analysis
All analyses were conducted using R (R Core Team, 2023). Looking times (accumulated looks on screen in seconds per trial) were analyzed with linear mixed-effect models (using the lme4 package v1.1–26; Bates et al., Reference Bates, Maechler, Bolker and Walker2015). Normality tests indicated that the raw looking times were more normally distributed than the log-transformed looking times. Hence, raw looking times in seconds were used as the dependent measure.
For statistical analysis, we pre-registered a model that included one categorical predictor variable and three continuous predictor variables. The categorical predictor was condition (within-participant, three levels: familiarized disharmonic, familiarized harmonic and novel words). The levels were coded using successive difference contrast coding, an orthogonal contrast (using the MASS package, version 7.3–54, Venables & Ripley, Reference Venables and Ripley2002), implemented as follows: familiarized harmonic (−2/3, −1/3), familiarized disharmonic (1/3–1/3) and novel (1/3, 2/3). As such, the output would present the estimated difference between familiarized disharmonic versus harmonic (1/3, −2/3) and novel words versus familiarized disharmonic (2/3, −1/3). The continuous predictors were ATR-harmony exposure (i.e., amount of exposure to all ATR harmony languages in a % of exposure to all languages, centered), age in days (centered) and trial number (1–12, within-participant factor, centered; a more fine-grained alternative to the predictor block to absorb more variance). The random structure included by-participants and by-target word random intercepts and random slopes by both participant and target word for condition, trial number and their interaction. The R formula for the pre-registered model was as follows:
-
(1) Looking time ~ (condition) × (exposure + age + trial number) + (1 + condition * trial number|participant) + (1 + condition × trial number|target word)
It was pre-registered that we would obtain two ATR-harmony exposure variables: one from the CIME and one from the logbook. However, this would not be possible if they were correlated. An assessment of both measures showed that they were correlated (r(38) = .43, p = .005, see Figure 1).

Figure 1. Correlation of ATR harmony exposure measures obtained from logbook versus the CIME.
To reduce collinearity, we performed a Principal Component Analysis (PCA) on the two measures (as in Omane et al., Reference Omane, Benders and Boll-Avetisyan2024) and used the first Principal Component (PC) as the single predictor exposure in the statistical analyses. Subsequently, we worked with the pre-registered model above.
The pre-registered model did not converge because of its complexity; hence, the random-effects structure was reduced. Correlations between random effects were either zero or approximately 1; hence, they were removed. The following random slopes had variances (close to) zero and were thus removed from the model: the by-participant random slopes for condition, trial and their interaction, as well as the by-target word random slopes for condition. Furthermore, the fixed-effect component of the model was simplified by removing the condition × age and condition × trial number interactions, as including them in the model resulted in a singularity warning. Moreover, these interactions were not directly related to our hypotheses but were included to absorb variance in the data. After this, the model converged with the formula presented in (2):
-
(2) Looking time ~ condition × exposure + age + trial number + (1 | participant) + (1+ trial number |||| target word)
As pre-registered, we continued with model comparisons using the Likelihood Ratio Test to see if the model could be simplified further. This revealed that the model fit was not significantly improved by including either a trial number slope by target word or a random intercept by target word (all p’s = 0.4; see the output of model comparisons in Supplementary Materials S3a and S3b, respectively). The fixed effect trial number improved the model fit (p < .001; see the output of the model comparison in Supplementary Material S3c) and was, hence, retained. We always retained the fixed effects of condition and exposure, as they were essential for hypothesis testing. The final formula was as follows:
-
(3) Looking Time ~ condition × exposure + age + trial number + (1 | participant)
6. Results
For the descriptive analysis, infants’ exposure (in %) to all ATR harmony languages ranged from 6.9% to 93.3% as estimated by the CIME (mean = 57.3%, SD = 20.1%), and from 3.6% to 95.2% as estimated by the logbook (mean = 46.1%, SD = 23.0%). Exposure (in %) to only Akan also ranged from 6.9% to 93.3% as estimated by the CIME (mean = 53.78%, SD = 20.7), and from 3.6% to 95.2% as estimated by the logbook (mean = 40.8, SD = 21.9). Mean looking times to the familiarized target words from the harmonic and disharmonic contexts and the novel target words were computed for each infant. On average, infants’ looking time was highest for familiarized words from the disharmonic context (mean = 8.83 s, SD = 2.38), followed by those that appeared in the harmonic context (mean = 7.76 s, SD = 2.30) and then novel words (mean = 7.49 s, SD = 1.99) (see Figure 2). Twenty-four (60%) infants had the longest average looking times to familiarized words from a disharmonic context, ten (25%) to familiarized words from a harmonic context and six (15%) to novel words.

Figure 2. Averages of individual infants’ mean looking times (s) to familiarized words from harmonic and disharmonic context passages and novel words. Dots indicate the mean for each condition.
The linear mixed-effect model (see Table 4 for results on all fixed-effects parameters) revealed a marginally significant effect of the contrast between familiarized words from disharmonic and harmonic contexts (coefficient: disharmonic versus harmonic, β = 0.71, p = .05, 95% CI [−0.009, 1.433]), with infants looking longer to target words from the disharmonic than the harmonic context. Moreover, the contrast between novel words and familiarized words from the disharmonic context (coefficient: novel versus disharmonic) was significant (β = −0.74, p < .05, 95% CI [−1.472, −0.006]), with infants looking longer to words from the disharmonic context than novel words. There was also an effect of trial number (β = −0.35, p < .001, 95% CI [−0.435, −0.261]), indicating that infants’ looking time to trials decreased over the course of the experiment. The other main effects (i.e., exposure and age) were not significant and no significant interaction of condition and exposure was observed (all p’s > .05).
Table 4. Final model output showing parameters of the linear mixed-effect model

In order to assess the difference in looking time to familiarized words from the harmonic context and novel words, we ran an exploratory model (Table 5) with the same specifications as the planned and fitted model (see model formula 3), except for the following re-coding of levels in the condition contrast: familiarized harmonic versus novel (1/3, −2/3) and familiarized disharmonic versus familiarized harmonic (2/3, −1/3). The model showed no significant effect of the contrast between familiarized words from the harmonic context and novel words (coefficient: harmonic versus novel, β = 0.03, p = .94, 95% CI [−0.70, 0.75]).
Table 5. Exploratory model output showing parameters of the linear mixed-effect model

7. Discussion
The goal of the present study was to investigate whether multilingual infants learning (an) ATR harmony language(s) alongside non-VH languages segment bisyllables from continuous speech using ATR harmony cues. Nine- to 11-month-old infants learning Akan (an ATR harmony language) in addition to at least one non-VH language were tested in Ghana. Using the central fixation procedure (Cooper & Aslin, Reference Cooper and Aslin1990), infants were first familiarized with text passages that included trisyllabic sequences, combinations of a bisyllabic target word plus a monosyllabic modifier, which were either harmonic, providing no segmentation cue, or disharmonic, providing a regressive ATR harmony segmentation cue between the target word and the modifier. In a subsequent test phase, infants’ looking time to novel words, as well as familiarized bisyllabic words that had occurred in either harmonic or disharmonic context, was recorded. An interview assessment and a diary logbook were used to assess infants’ relative exposure to (an) ATR harmony and non-VH language(s). As hypothesized, our findings show that multilingual infants better segment bisyllables from speech when word boundaries are marked by a change in ATR (i.e., disharmony) than when the trisyllabic sequences form a harmonic unit (e.g., tupike). The current study thus provides the first evidence for segmentation with ATR harmony cues and in multilingual infants learning (an) ATR harmony language(s) alongside languages without VH. However, we did not detect an effect of ATR harmony language exposure on infants’ segmentation with ATR harmony cues. We discuss these findings, study limitations and future research considerations below.
Previous artificial language studies found adult speakers of VH languages, but not speakers of non-VH languages, to use VH cues for segmentation (Kabak et al., Reference Kabak, Maniwa and Kazanina2010; Suomi et al., Reference Suomi, McQueen and Cutler1997; Vroomen et al., Reference Vroomen, Tuomainen and De Gelder1998), suggesting that the use of the cue is language-specific in adulthood. Moreover, segmentation with VH cues has previously been reported for infants learning Turkish, a VH language (with backness harmony: van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008), and infants learning English, a language without VH (Mintz et al., Reference Mintz, Walker, Welday and Kidd2018, Experiment 2c).
Our findings extend evidence of infants’ speech segmentation with VH cues to ATR harmony and to a genetically unrelated language (Akan). By this, the present study adds to the existing literature on the cues infants cross-linguistically use for segmentation (Höhle & Weissenborn, Reference Höhle and Weissenborn2003; Jusczyk et al., Reference Jusczyk, Houston and Newsome1999; Saffran et al., Reference Saffran, Aslin and Newport1996; van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). While evidence from three studies now converges on the conclusion that VH is a robust segmentation cue for infants, the question about its language-specificity or potentially universal nature remains open. If this study had shown that exposure to ATR harmony relates to infants’ use of VH, this would have suggested language-specificity or learning (we provide further discussion on the lack of an effect of VH language exposure later in this section). In the absence of evidence for (or against) this effect, we can speculate that infants are universally sensitive to VH patterns, supporting the results from Mintz et al. (Reference Mintz, Walker, Welday and Kidd2018) with an artificial language, as well as a recent study finding auditory preferences for VH in 4-month-old infants learning non-VH languages (Solá-Llonch & Sundara, Reference Solá-Llonch and Sundara2025). Important to note, however, is that Mintz et al. (Reference Mintz, Walker, Welday and Kidd2018) used an artificial language entirely different from the text passages as those used in the present study and Van Kampen et al. (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). To shed more light on the accessibility of VH for infants’ language processing, a next step will be to investigate whether infants learning non-VH languages can use VH cues to segment words from natural language, with all its naturally occurring complexities.
The present results further demonstrate that infants prioritize the VH cue over stimulus-specific transitional probabilities (TPs) between syllables. In our experiment, infants heard trisyllabic sequences with a TP of 1 between all syllables in both the harmonic (e.g., tupi ke) and the disharmonic condition (e.g., dɔbɛ ke). Our findings, however, suggest that infants segmented at the point of disharmony, before the modifier, thus breaking up units that are statistically coherent in terms of TPs between syllables. This indicates that they weighted the ATR harmony cues as more relevant than domain-general statistical cues to detect word units in their native language. This is consistent with previous studies showing that infants weight language-specific prosodic cues over stimulus-specific statistical cues for word segmentation (e.g., 6-month-old German-learning infants: Marimon et al., Reference Marimon, Langus and Höhle2024; 9-month-old English-learning infants: Thiessen & Saffran, Reference Thiessen and Saffran2003, Experiment 1).
At present, the question remains open of whether infants can or when they will learn to segment words when vowel sequences are harmonic. We found no significant effect of the contrast between novel and familiarized words from the harmonic context, providing no evidence for (or against) infants’ recognition of bisyllables from the harmony context, where the harmony cues, as well as TPs of 1 between syllables, are suggestive of a trisyllabic unit. Similarly, Van Kampen et al. (Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008) found no evidence that 9-month-old Turkish-learning infants infer a word boundary within a trisyllabic sequence if the first syllable (or prefix) harmonizes with the subsequent bisyllabic string. Speculatively (as this is a null result), this may reflect infants parsing the trisyllabic unit as a single chunk. Akan- and Turkish-learning infants potentially treating trisyllabic units as single words in the absence of harmony cues to word boundaries resemble 7.5-month-old English-learning infants not recognizing subparts from syllable strings (e.g., not recognizing ‘king’ after familiarization with ‘kingdom,’ Jusczyk et al., Reference Jusczyk, Houston and Newsome1999), and 11-month-old French-learning infants segmenting frequent sequences even when they span word boundaries (Ngon et al., Reference Ngon, Martin, Dupoux, Cabrol, Dutat and Peperkamp2013). These studies, taken together, thus suggest that infants fail to recognize subparts of words or other frequently occurring units. Future studies may want to explore at what age Akan-learning infants learn to segment words from harmonic contexts, as word boundaries can also occur between adjacent harmonic words in Akan. This research should explore the development of word segmentation in the absence of harmony cues and further explore which other segmentation cues, such as top-down lexical and low-level phonetic cues, may then be used by infants growing up in Ghana.
Regarding the direction of the listening preferences at test, infants demonstrated a longer looking time to familiarized bisyllables from a disharmonic context over those from a harmonic context or (importantly) novel bisyllables. This can be interpreted as a familiarity preference for words that were heard and segmented during the familiarization phase. This aligns with reported familiarity preferences in studies that tested the use of other segmentation cues, including phonotactics (VH: van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008; consonant clusters: Mattys & Jusczyk, Reference Mattys and Jusczyk2001), allophonic cues (Jusczyk et al., Reference Jusczyk, Houston and Newsome1999) and rhythmic units (stress pattern: Jusczyk & Aslin, Reference Jusczyk and Aslin1995; syllable-unit: Bosch et al., Reference Bosch, Figueras, Teixidó and Ramon-Casas2013; Nazzi et al., Reference Nazzi, Iakimova, Bertoncini, Frédonie and Alcantara2006), in addition to meta-analytic evidence for a familiarity preference bias in segmentation tasks (Bergmann & Cristia, Reference Bergmann and Cristia2016).
The present study demonstrates that infants use VH cues to segment bisyllables from speech when the segmentation cue is at the end of the target bisyllable. This complements the previous evidence of infants’ segmentation with VH cues marking the target word onset (van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). That infants can track harmony cues for identifying word boundaries both at word onset and offset is akin to their processing of other phonotactic cues like consonant clusters in English (e.g., Mattys & Jusczyk, Reference Mattys and Jusczyk2001). Adult speech segmentation studies have shown that speakers of VH languages can segment bisyllabic words when the VH cue marks the target word onset (Kabak et al., Reference Kabak, Maniwa and Kazanina2010; Suomi et al., Reference Suomi, McQueen and Cutler1997; Vroomen et al., Reference Vroomen, Tuomainen and De Gelder1998) but one study with Finnish listeners did not find evidence for the use of VH cues when it marks the target word offset, suggesting a potential positional restriction on using VH information as a segmentation cue (Suomi et al., Reference Suomi, McQueen and Cutler1997). In our stimuli, target words were assigned to noun positions, with post-nominal modifiers providing the segmentation cue (at target word offset), as this structure is common in Akan. However, as vowels in prefixes must also harmonize with stem vowels in ATR in Akan, our expectation is that infants should successfully segment target words, too, if target words are preceded by a disharmonizing prefix (i.e., target word onset), as was found with Turkish-learning infants (van Kampen et al., Reference van Kampen, Parmaksiz, van de Vijver, Höhle, Gavarró and Freitas2008). Whether the position of harmony cues impacts word segmentation in infants learning Akan and other languages remains to be explored, to broaden our understanding of the language- and age-dependent tracking of directionality in fluent speech.
Our findings do not provide evidence for or against the effect of the quantity of exposure to (an) ATR harmony language(s) on infants’ use of VH cues for segmentation. This prompts speculation that even limited exposure to an ATR harmony language might suffice for multilingual infants to learn how VH disagreements between adjacent syllables align with word boundaries and use this knowledge to segment words from fluent speech. The absence of a detected exposure effect is consistent with a previous study (Omane et al., Reference Omane, Benders and Boll-Avetisyan2024), where multilingual infants exposed to (an) ATR harmony language(s) alongside non-VH languages in Ghana demonstrated a familiarity preference for ATR harmonic syllable sequences in Akan, with no evidence for or against the effect of exposure either. The finding aligns with prior research on other domains of speech perception, showing that, for example, simultaneous German–French bilingual adults’ chunking of continuous speech (Boll-Avetisyan et al., Reference Boll-Avetisyan, Bhatara, Unger, Nazzi and Höhle2020) and German–French bilingual infants’ preference for trochaic patterns was not modulated by language exposure either (Bijeljac-Babic et al., Reference Bijeljac-Babic, Höhle and Nazzi2016). Furthermore, several prior segmentation studies have also not found evidence for (or against) the amount of input in a language impacting bilingual infants’ ability to use native language cues for speech segmentation (Bosch et al., Reference Bosch, Figueras, Teixidó and Ramon-Casas2013; Orena & Polka, Reference Orena and Polka2019; Polka et al., Reference Polka, Orena, Sundara and Worrall2017). Taken together, the absence of a significant exposure effect in the current study leads us to speculate that any input multilingual infants receive in an ATR harmony language allows them to track harmony cues during speech processing and that further input in an ATR harmony language does not further enhance this ability. This suggests the development or maintenance of ATR-based segmentation may require minimal input to an ATR harmony language.
As argued in our previous study (Omane et al., Reference Omane, Benders and Boll-Avetisyan2024), the absence of an exposure effect could also relate to methodological shortcomings. Infants in our study were exposed to diverse languages (between two and five languages per child) with varying numbers of caregivers (between two and six). Consequently, caregivers may not have accurately reported language input for all caregivers, potentially accounting for our inability to find an exposure effect on segmentation. Future research may want to address this issue by comparing data obtained using our CIME and/or logbook with data from daylong recordings (see Orena et al., Reference Orena, Byers-Heinlein and Polka2020, who found correlations between interview assessments and daylong recording data for infants growing up in French–English contexts). Alternatively, more reliable estimates of language exposure in a multilingual context like Ghana could be to survey or interview more than one caregiver per child, although such an approach may be impractical or difficult to realize.
The present study is relevant to multilingual speech segmentation as it presents the first evidence of segmentation abilities in infants learning between two and five languages. It also provides the first evidence of segmentation with phonotactic cues in infants learning more than one language. In the context of previous research on bilingual infants’ speech segmentation, our findings align with studies demonstrating segmentation ability in bilinguals’ native languages (Bosch et al., Reference Bosch, Figueras, Teixidó and Ramon-Casas2013; Mateu & Sundara, Reference Mateu and Sundara2022; Orena & Polka, Reference Orena and Polka2019; Polka & Sundara, Reference Polka and Sundara2003; Polka et al., Reference Polka, Orena, Sundara and Worrall2017; Singh & Foong, Reference Singh and Foong2012; Singh, Reference Singh2018). Unlike prior bilingual studies that tested segmentation in the infant’s two native languages (but see Singh, Reference Singh2018), our study focused on testing infants in only one of their native languages, as Akan, the ATR harmony language that we focused on, was the only language that all infants were exposed to. Our results suggest that multilingual infants can use a phonotactic cue of one of their languages for segmenting speech in that language. It remains to be tested in the future whether infants learning three or more languages have the processing ability to segment words in all their ambient languages, akin to infants with dual language input (e.g., Orena & Polka, Reference Orena and Polka2019; Polka & Sundara, Reference Polka and Sundara2003; Polka et al., Reference Polka, Orena, Sundara and Worrall2017).
In conclusion, this study investigated word segmentation with VH cues in 9–11-month-old multilingual infants learning languages with and without VH in Ghana. Our results demonstrate that infants use ATR harmony cues from one of their languages to segment bisyllabic words in that language, with no evidence for or against the effect of quantity of exposure to an ATR harmony language on infants’ segmentation abilities. These findings prompt further research into multilingual language processing in infants raised in other multilingual contexts. Future studies are encouraged to examine word segmentation in other understudied African languages and multilingual populations to gain deeper insights into their segmentation abilities across all their languages and the cues infants use for such complex tasks.
Supplementary material
The supplementary material for this article can be found at http://doi.org/10.1017/S1366728925100485.
Data availability statement
The data, study materials, CIME, logbook and analysis code will be made available in an online repository: [https://osf.io/h9tvr].
Acknowledgments
The study received approval from the Ethics Committee for Humanities (ECH) at the University of Ghana, Ghana. This research was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) – Project-ID 317633480 – SFB 1287, Project C07 awarded to NB and a CTiMQRES IDEALAB scholarship from Macquarie University (allocation no. 20201820) awarded to PO. We thank Reginald Akuoko Duah for his numerous contributions, including allowing us to use his office as a lab. We thank the University of Ghana Hospital in Legon and an anonymous Hospital for allowing us to recruit and test infants on their premises. We are grateful to the nurses at the children’s ward in both hospitals for their hospitality and for assisting us in the recruitment process. We are thankful to Rebecca Dufie Forson, our research assistant, and Frank Obeng for assisting us in infant recruitment and data collection in Ghana. We also thank Anna Laurinavichyute (SFB 1287, Q-project) for statistical advice. Finally, we thank all parents and infants who participated in the study.
Competing interests
The authors declare none.