Skip to main content
    • Aa
    • Aa
  • Get access
    Check if you have access via personal or institutional login
  • Cited by 17
  • Cited by
    This article has been cited by the following publications. This list is generated based on data provided by CrossRef.

    Chan, Alice Yin Wa 2016. How much do Cantonese ESL learners know about the English article system?. System, Vol. 56, p. 66.

    Crosthwaite, Peter 2016. L2 English article use by L1 speakers of article-less languages. International Journal of Learner Corpus Research, Vol. 2, Issue. 1, p. 68.

    Liao, Hui-Chuan 2016. Enhancing the grammatical accuracy of EFL writing by using an AWE-assisted process approach. System,

    Oussalah, M. Escallier, B. and Daher, D. 2016. An automated system for grammatical analysis of Twitter messages. A learning task application. Knowledge-Based Systems, Vol. 101, p. 31.

    Oyama, Hiromi Komachi, Mamoru and Matsumoto, Yuji 2016. Hierarchical Annotation and Automatic Error-Type Classification of Japanese Language Learners’ Writing. Journal of Natural Language Processing, Vol. 23, Issue. 2, p. 195.

    Ranalli, Jim Link, Stephanie and Chukharev-Hudilainen, Evgeny 2016. Automated writing evaluation for formative assessment of second language writing: investigating the accuracy and usefulness of feedback as part of argument-based validation. Educational Psychology, p. 1.

    Gamallo Otero, Pablo Garcia, Marcos del Río, Iria and González López, Isaac 2015. Learner Corpora in Language Testing and Assessment.

    Kung, Fan-Wei 2015. Reexamining the NS and NNS Dichotomy in Taiwanese Higher EFL Education. The Asia-Pacific Education Researcher, Vol. 24, Issue. 1, p. 27.

    Lee, John Yeung, Chak Yan Zeldes, Amir Reznicek, Marc Lüdeling, Anke and Webster, Jonathan 2015. CityU corpus of essay drafts of English language learners: a corpus of textual revision in second language writing. Language Resources and Evaluation, Vol. 49, Issue. 3, p. 659.

    Ferrero, Carmen López Renau, Irene Nazar, Rogelio and Torner, Sergi 2014. Computer-assisted Revision in Spanish Academic Texts: Peer-assessment. Procedia - Social and Behavioral Sciences, Vol. 141, p. 470.

    Leacock, Claudia Chodorow, Martin Gamon, Michael and Tetreault, Joel 2014. Automated Grammatical Error Detection for Language Learners, Second Edition. Synthesis Lectures on Human Language Technologies, Vol. 7, Issue. 1, p. 1.

    Tetreault, Joel Chodorow, Martin and Madnani, Nitin 2014. Bucking the trend: improved evaluation and annotation practices for ESL error detection systems. Language Resources and Evaluation, Vol. 48, Issue. 1, p. 5.

    Yang, Ping 2014. Back to basics: Cracking a nut in using English indefinite articles. English Today, Vol. 30, Issue. 04, p. 28.

    Burstein, Jill 2012. The Encyclopedia of Applied Linguistics.

    NAGATA, Ryo and KAWAI, Atsuo 2012. A Method for Detecting Determiner Errors Designed for the Writing of Non-native Speakers of English. IEICE Transactions on Information and Systems, Vol. E95-D, Issue. 1, p. 230.

    Chang, Ru-Yng Wu, Chung-Hsien and Prasetyo, Philips Kokoh 2010. 2010 7th International Symposium on Chinese Spoken Language Processing. p. 401.

    Lee, Yong-Won Gentile, Claudia and Kantor, Robert 2008. ANALYTIC SCORING OF TOEFL®CBT ESSAYS: SCORES FROM HUMANS ANDE-RATER®. ETS Research Report Series, Vol. 2008, Issue. 1, p. i.


Detecting errors in English article usage by non-native speakers

  • DOI:
  • Published online: 22 May 2006

One of the most difficult challenges faced by non-native speakers of English is mastering the system of English articles. We trained a maximum entropy classifier to select among a/an, the, or zero article for noun phrases (NPs), based on a set of features extracted from the local context of each. When the classifier was trained on 6 million NPs, its performance on published text was about 83% correct. We then used the classifier to detect article errors in the TOEFL essays of native speakers of Chinese, Japanese, and Russian. These writers made such errors in about one out of every eight NPs, or almost once in every three sentences. The classifier's agreement with human annotators was 85% (kappa = 0.48) when it selected among a/an, the, or zero article. Agreement was 89% (kappa = 0.56) when it made a binary (yes/no) decision about whether the NP should have an article. Even with these levels of overall agreement, precision and recall in error detection were only 0.52 and 0.80, respectively. However, when the classifier was allowed to skip cases where its confidence was low, precision rose to 0.90, with 0.40 recall. Additional improvements in performance may require features that reflect general knowledge to handle phenomena such as indirect prior reference. In August 2005, the classifier was deployed as a component of Educational Testing Service's Criterion$^{SM}$ Online Writing Evaluation Service.

Recommend this journal

Email your librarian or administrator to recommend adding this journal to your organisation's collection.

Natural Language Engineering
  • ISSN: 1351-3249
  • EISSN: 1469-8110
  • URL: /core/journals/natural-language-engineering
Please enter your name
Please enter a valid email address
Who would you like to send this to? *