Hostname: page-component-77f85d65b8-8wtlm Total loading time: 0 Render date: 2026-04-21T17:42:09.130Z Has data issue: false hasContentIssue false

A learning-based account of local phonological processes

Published online by Cambridge University Press:  22 October 2024

Caleb A. Belth*
Affiliation:
Department of Linguistics, University of Utah, Salt Lake City, UT, USA
Rights & Permissions [Opens in a new window]

Abstract

Phonological processes tend to involve local dependencies, an observation that has been expressed explicitly or implicitly in many phonological theories, such as the use of minimal symbols in SPE and the inclusion of primarily strictly local constraints in Optimality Theory. I propose a learning-based account of local phonological processes, providing an explicit computational model. The model is grounded in experimental results that suggest children are initially insensitive to long-distance dependencies and that as their ability to track non-adjacent dependencies grows, learners still prefer local generalisations to non-local ones. The model encodes these results by constructing phonological processes starting around an alternating segment and expanding outward to incorporate more phonological context only when surface forms cannot be predicted with sufficient accuracy. The model successfully constructs local phonological generalisations and exhibits the same preference for local patterns that humans do, suggesting that locality can emerge as a computational consequence of a simple learning procedure.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Figure 1 (a) The width of PLP’s search expands outward (upward arrows) when and only when an adequate generalisation cannot be formed from a narrower context. (b) and (c) An example of PLP on seven English plural nouns. (b): PLP’s first generalisation (13) is based on only the alternating segment and makes too many wrong predictions; this triggers PLP to expand its attention window. (c) PLP then forms generalisation (15a), which is based on the left-adjacent segment and allows the /z/ $\to $ [s] instances to be isolated.

Figure 1

Figure 2 Plots show model accuracies on test words when trained on each of the artificial languages from Baer-Henney & van de Vijver (2012), as a function of the fraction of the exposure set they are trained on. The $\times $ marks show human generalisation accuracy, at the x-axis point where each model best matches human generalisation behaviour. PLP and MGL’s results are in (a) and (b), while grammars learned by ranking a provided constraint set are in (c)–(e). In (c), the ranked constraints are all phonetically motivated; in (d), all constraints needed for the three languages are included and in (e), each model for each language ranks only the constraints relevant to that language.

Figure 2

Table 1 Model accuracies (with standard deviations) on held-out test data at different training vocabulary sizes. PLP readily learns an accurate generalisation for German syllable-final obstruent devoicing

Figure 3

Table 2 Analysis of the types of errors each of the models that learn an accurate grammar makes in the process. Because it adds generalisations to the grammar only when necessitated by surface alternations, PLP produces no unmotivated errors

Figure 4

Table 3 Model accuracies (with standard deviations) on held-out test data at different training vocabulary sizes. PLP readily learns an accurate grammar for the English processes in (44)

Figure 5

Figure 3 PLP’s accuracy on the plural and past tense nonce words from Berko (1958) as training progressed. The black dashed line denotes plurals that should take [-z] or [-s] and the grey dashed lines those that should take [-әz]. The dotted lines represent the analogues for past tense. The fact that [-z]/[-s] accuracy converges before [-әz] and [-d]/[-t] before [-әd] matches Berko’s finding that children learn [-z]/[-s] and [-d]/[-t] before [-әz] and [-әd].

Figure 6

Table 4 PLP learns precisely the set of processes active in its experience. This provides a straightforward account of how productive phonological processes can be learned even if they operate against apparent phonetic motivation, like devoicing in Tswana following nasals (Coetzee & Pretorius 2010). With PLP, the unmotivated constraint *NC̬ need not be assumed to be universal