Hostname: page-component-89b8bd64d-ksp62 Total loading time: 0 Render date: 2026-05-10T03:00:23.422Z Has data issue: false hasContentIssue false

An experimental study on data augmentation techniques for named entity recognition on low-resource domains

Published online by Cambridge University Press:  24 March 2026

Arthur Elwing Torres*
Affiliation:
Universidade Federal do Amazonas , Manaus, Amazonas, Brazil
Edleno Silva de Moura
Affiliation:
Universidade Federal do Amazonas , Manaus, Amazonas, Brazil Jusbrasil, Salvador, Bahia, Brazil
Altigran Soares da Silva
Affiliation:
Universidade Federal do Amazonas , Manaus, Amazonas, Brazil
Mario A. Nascimento*
Affiliation:
University of Alberta , Edmonton, Alberta, Canada
Filipe Mesquita
Affiliation:
Diffbot Technologies Corp., Menlo Park, California, USA
*
Corresponding authors: Arthur Elwing Torres; Email: arthur.torres@icomp.ufam.edu.br, Mario A. Nascimento; Email: mario.nascimento@ualberta.ca
Corresponding authors: Arthur Elwing Torres; Email: arthur.torres@icomp.ufam.edu.br, Mario A. Nascimento; Email: mario.nascimento@ualberta.ca
Rights & Permissions [Opens in a new window]

Abstract

Named Entity Recognition (NER) is a natural language processing task that traditionally relies on supervised learning and annotated data. Acquiring such data is often a challenge, particularly in specialized fields like medical, legal, and financial sectors. Those are commonly referred to as low-resource domains, which comprise long-tail entities, due to the scarcity of available data. To address this, data augmentation techniques are increasingly being employed to generate additional training instances from the original dataset. In this study, we evaluate the effectiveness of two prominent text augmentation techniques, Mention Replacement and Contextual Word Replacement, on two widely used NER models, Bi-LSTM + CRF and BERT. We conduct experiments on three datasets from low-resource domains, and we explore the impact of various combinations of training subset sizes and the number of augmented examples. We not only confirm that data augmentation is particularly beneficial for smaller datasets, but we also demonstrate that there is no universally optimal number of augmented examples, i.e., NER practitioners must experiment with different quantities in order to fine-tune their projects.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press
Figure 0

Table 1. Text augmentation techniques with examples

Figure 1

Table 2. Datasets used in the experiment

Figure 2

Figure 1. Effects of different amounts of augmented examples (MR) on average F1-score by model and dataset.

Figure 3

Figure 2. Effects of different amounts of augmented examples (CWR) on average F1-score by model and dataset.

Figure 4

Table 3. 0% and best augmentation (MR) average F1-scores and applied paired Student’s t-tests for i2b2 dataset

Figure 5

Table 4. 0% and best augmentation (MR) average F1-scores and applied paired Student’s t-tests for MaSciP dataset

Figure 6

Table 5. 0% and best augmentation (MR) average F1-scores and applied paired Student’s t-tests for BioCreative dataset

Figure 7

Table 6. 0% and best augmentation (CWR) average F1-scores and applied paired Student’s t-tests for i2b2 dataset

Figure 8

Table 7. 0% and best augmentation (CWR) average F1-scores and applied paired Student’s t-tests for MaSciP dataset

Figure 9

Table 8. 0% and best augmentation (CWR) average F1-scores and applied paired Student’s t-tests for BioCreative dataset

Figure 10

Table 9. Summary of models improved by augmentation

Figure 11

Table 10. Summary of models worsened by augmentation (MR)

Figure 12

Table 11. Summary of models worsened by augmentation (CWR)