Hostname: page-component-89b8bd64d-ksp62 Total loading time: 0 Render date: 2026-05-06T16:13:36.427Z Has data issue: false hasContentIssue false

Is FrameNet usage-based? Predicting frame element coreness with information theory and gradient boosting

Published online by Cambridge University Press:  28 April 2026

Vladimir Buskin*
Affiliation:
English Language and Linguistics, Katholische Universität Eichstätt-Ingolstadt, Germany
*
Rights & Permissions [Opens in a new window]

Abstract

The FrameNet project is a large-scale frame-semantic database with a seemingly usage-based core: It draws on 200,000 annotated sentences from representative corpora and offers the most comprehensive description of semantic valency patterns in English to date. Nevertheless, its empirical validity is weakened by the lack of statistical information on the distribution of lexical units, frames and frame elements. Similarly, the characterisation of frame elements as core, core-unexpressed, peripheral or extra-thematic – intended to indicate their essentiality to a frame – is primarily motivated on theoretical grounds. This raises the question of whether these labels are consistent with actual language use. After exhaustively extracting frequency data from Python’s NLTK FrameNet Corpus for all attested combinations of verbs, frames and frame elements, hierarchical gradient boosting models were trained on information-theoretic measures and word embeddings to predict the coreness of frame elements. The models provide strong usage-based evidence for a general core versus non-core distinction but cast doubt on further subdivisions such as core versus core-unexpressed or peripheral versus extra-thematic. While further validation is necessary, this contribution offers the first statistical perspective on the current state of FrameNet and its compatibility with usage-based approaches.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press
Figure 0

Figure 1. Simplified representation of set relations in the FrameNet database.

Figure 1

Table 1. Frame element distribution of the verb eat in the INGESTION frame with $ N $ = 35

Figure 2

Figure 2. Density plots of observed (log-)frequencies and sample proportions in the FrameNet data. A: Distribution of observed frequencies. B: Distribution of observed frequencies (log-scale). C: Distribution of element probabilities.

Figure 3

Figure 3. Density plot of surprisal in the FrameNet data.

Figure 4

Table 2. Summary of common classification metrics (TP = true positive, TN = true negative, FP = false positive, FN = false negative) based on Chicco and Jurman (2023, pp. 2–3), Alpaydın (2022, pp. 632–636) and James et al. (2021, pp. 148–152)

Figure 5

Table 3. Probability distribution of frame elements conditioned on the verb eat and the frame INGESTION with a Shannon entropy $ H({E}_{eat}^{\mathrm{INGESTION}}) $ of 1.85 bits and a varentropy of 1.38

Figure 6

Figure 4. Surprisal $ {S}_v^f(e) $, Shannon entropy (dashed line) and entropy deviation (solid horizontal lines) of frame element types conditioned on verbs in different frames.

Figure 7

Table 4. Distribution of element types for all verbs in the FrameNet database

Figure 8

Table 5. Summary statistics on the predictor variables ($ n $ = 21,160)

Figure 9

Figure 5. Distribution of information-theoretic measures across frame element types (density plots). A: Distribution of $ {P}_v^f(e) $ by coreness. B: Distribution of $ {S}_v^f(e) $ by coreness. C: Distribution of $ {\Delta}_v^f(e) $ by coreness. D: Distribution of $ [{\Delta}_v^f(e){]}^2 $ by coreness.

Figure 10

Table 6. Dunn test results with Bonferroni-adjusted $ p $-values ($ {p}_{\mathrm{adjusted}} $) for pairwise comparisons between frame element types based on three entropy-derived metrics

Figure 11

Table 7. Feature correlation matrix based on Spearman’s $ \rho $

Figure 12

Table 8. General performance metrics for the full model with fixed and random effects (95% confidence intervals); MCC normalised to [0,1] scale

Figure 13

Table 9. General performance metrics for the random-effects-only model (95% confidence intervals); MCC normalised to [0,1] scale

Figure 14

Figure 6. Summary of general performance metrics (dashed line represents expected values if the model was guessing randomly).

Figure 15

Table 10. Class-specific performance metrics for the Full Model (95% confidence intervals)

Figure 16

Table 11. Class-specific performance metrics for the random-effects-only model (95% confidence intervals)

Figure 17

Figure 7. Summary of class-specific test performance with 95% CIs.

Figure 18

Figure 8. Feature importance scores for the core versus non-core distinction with bootstrapped 95% CIs (‘H_vf’ = $ H\left({E}_v^f\right) $, ‘Se_vf’ = $ {S}_v^f(e) $, ‘Dev_S_H’ = $ {\Delta}_v^f(e) $, ‘Dev_sq_S_H’ = $ {\left[{\Delta}_v^f(e)\right]}^2 $).

Figure 19

Figure 9. Feature importance scores for the extra-thematic versus peripheral distinction with bootstrapped 95% CIs.

Figure 20

Figure 10. Partial dependence of the core versus non-core model (information-theoretic measures).

Figure 21

Figure 11. Partial dependence of the core versus non-core model (word vectors).

Figure 22

Figure 12. Partial dependence of the extra-thematic versus peripheral model.

Figure 23

Figure C1. Unconditional distributions of information-theoretic measures (density plots). A: Distribution of entropy. B: Distribution of entropy deviation. C: Distribution of squared entropy deviation. D: Distribution of varentropy.