Hostname: page-component-89b8bd64d-ksp62 Total loading time: 0 Render date: 2026-05-13T23:35:05.224Z Has data issue: false hasContentIssue false

Chinese spelling correction based on Long Short-Term Memory Network-enhanced Transformer and dynamic adaptive weighted multi-task learning

Published online by Cambridge University Press:  07 April 2025

Mingying Xu
Affiliation:
School of Information Science, North China University of Technology, Beijing, China
Jie Liu*
Affiliation:
School of Information Science, North China University of Technology, Beijing, China China Language Intelligence Research Center, Beijing, China
Kui Peng
Affiliation:
School of Information Science, North China University of Technology, Beijing, China
Zhen Li
Affiliation:
Capital Normal University, Beijing, China
*
Corresponding author: Jie Liu; Email: liujie@ncut.edu.cn
Rights & Permissions [Opens in a new window]

Abstract

Chinese spelling correction has achieved significant progress, but critical challenges remain, especially in handling visually and phonetically similar errors within complex syntactic structures. This paper introduces a novel approach combining a Long Short-Term Memory Network (LSTM)-enhanced Transformer for error detection and Bidirectional Encoder Representations from Transformers (BERT)-based correction with a dynamic adaptive weighting scheme. Transformer uses global attention mechanism to capture dependencies between any two positions in the input sequence. By processing each token in the sequence recursively, LSTM is able to more finely capture local context and sequential information within the sequence. Based on adaptive weighting coefficient, weights of multi-task learning are automatically adjusted to help the model better balance the learning process between the detection and correction network, enabling it to converge faster and achieve higher precision. Comprehensive evaluations demonstrate improved performance over existing baselines, particularly in addressing complex error patterns.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - SA
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-ShareAlike licence (https://creativecommons.org/licenses/by-sa/4.0/), which permits re-use, distribution, and reproduction in any medium, provided the same Creative Commons licence is used to distribute the re-used or adapted article and the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Table 1. Examples of Chinese spelling errors

Figure 1

Figure 1. Chinese spelling correction based on LSTM-enhanced Transformer and dynamic adaptive weighted multi-task Learning.

Figure 2

Table 2. Dataset statistics

Figure 3

Table 3. Experimental parameters

Figure 4

Table 4. Experimental parameters

Figure 5

Table 5. The experimental results on SIGHAN14

Figure 6

Table 6. The experimental results on SIGHAN15

Figure 7

Figure 2. Performance of the proposed method under different epochs on SIGHAN13.

Figure 8

Table 7. Ablation experiment results

Figure 9

Figure 3. Performance of the proposed method under different epochs on SIGHAN14.

Figure 10

Figure 4. Performance of the proposed method under different epochs on SIGHAN15.

Figure 11

Table 8. Examples of Chinese spelling correction (CSC) of MDCSpell, DR-CSC, ChatGPT, and our proposed method