Hostname: page-component-77f85d65b8-45ctf Total loading time: 0 Render date: 2026-03-28T11:08:01.580Z Has data issue: false hasContentIssue false

Neural automated writing evaluation for Korean L2 writing

Published online by Cambridge University Press:  07 July 2022

KyungTae Lim
Affiliation:
Hanbat National University, Daejeon 34158, South Korea
Jayoung Song
Affiliation:
Pennsylvania State University, State College, PA 16801, USA
Jungyeul Park*
Affiliation:
The University of British Columbia, Vancouver, BC V6T 1Z4, Canada University of Washington, Seattle, WA 98195, USA
*
*Corresponding author. E-mail: jungyeul@mail.ubc.ca
Rights & Permissions [Opens in a new window]

Abstract

Although Korean language education is experiencing rapid growth in recent years and several studies have investigated automated writing evaluation (AWE) systems, AWE for Korean L2 writing still remains unexplored. Therefore, this study aims to develop and validate a state-of-the-art neural model AWE system which can be widely used for Korean language teaching and learning. Based on a Korean learner corpus, the proposed AWE is developed using natural language processing techniques such as part-of-speech tagging, syntactic parsing, and statistical language modeling to engineer linguistic features and a pre-trained neural language model. This study attempted to determine how neural network models use different linguistic features to improve AWE performance. Experimental results of the proposed AWE tool showed that the neural AWE system achieves high reliability for unseen test data from the corpus, which implies metrics used in the AWE system can help differentiate different proficiency levels and predict holistic scores. Furthermore, the results confirmed that the proposed linguistic features–syntactic complexity, quantitative complexity, and fluency–offer benefits that complement neural automated writing evaluation.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives licence (https://creativecommons.org/licenses/by-nc-nd/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is unaltered and is properly cited. The written permission of Cambridge University Press must be obtained for commercial re-use or in order to create a derivative work.
Copyright
© The Author(s), 2022. Published by Cambridge University Press
Figure 0

Figure 1. Example of the Korean learner corpus: = Level 1, = Chinese, = female, = final examination, = Fall 2013, = my weekend, and = 70. The present example of Korean writing can roughly be translated into I went to the library on Sunday. I went to the library with a friend. There was a book in the library. I studied in the library. I read a book. So it’s fun on Sunday.

Figure 1

Table 1. Examples of most frequent prompts and their number of instances in the learner corpus

Figure 2

Figure 2. Example of Sejong corpus-style POS tagging analysis: MA{J$|$G} are for adverbs, NN{P$|$B$|$G} for nouns, J{KB$|$X$|$KS} for postpositions, E{P$|$F} for verbal endings, VA for adjectives, and SF for punctuations.

Figure 3

Figure 3. Example of phrase-structure analysis.

Figure 4

Figure 4. Example of an M2 file for the Korean learner corpus.

Figure 5

Table 2. Example of features and teir values for te learner’s writing in Figure 1

Figure 6

Figure 5. Distribution of sample features per level.

Figure 7

Figure 6. Distribution of learner’s scores between levels.

Figure 8

Table 3. Statistical AWE system results

Figure 9

Figure 7. System structure of the proposed deep learning model. Three linguistic features are applied: syntactic complexity, fluency, and quantitative complexity, in addition to the sequence of token representations. Each token is transformed into a vector representation based on XLM-RoBERTa.

Figure 10

Table 4. Hyperparameters

Figure 11

Table 5. Experiment results

Figure 12

Figure 8. Visualization of the attention score proposed in Eq (8).

Figure 13

Figure 9. Visualization of the attention scores proposed in (8): [S-COMPLEXITY], [FLUENCY], and [Q-COMPLEXITY] for syntactic complexity, fluency, and qualitative complexity features, respectively.

Figure 14

Figure 10. The usage of verbal endings based on its proficiency levels.

Figure 15

Table 6. Result comparison using different Korean monolingual BERTs

Figure 16

Table 7. Features utilized in previous work