Hostname: page-component-77f85d65b8-g98kq Total loading time: 0 Render date: 2026-04-21T22:43:43.399Z Has data issue: false hasContentIssue false

Using the crowd for readability prediction

Published online by Cambridge University Press:  14 December 2012

ORPHÉE DE CLERCQ
Affiliation:
Faculty of Applied Language Studies, University College Ghent, Ghent, Belgium e-mail: orphee.declercq@hogent.be, veronique.hoste@hogent.be, bart.desmet@hogent.be, philip.vanoosten@hogent.be, lieve.macken@hogent.be Department of Applied Mathematics and Computer Science, Ghent University, Ghent, Belgium e-mail: martine.decock@ugent.be
VÉRONIQUE HOSTE
Affiliation:
Faculty of Applied Language Studies, University College Ghent, Ghent, Belgium e-mail: orphee.declercq@hogent.be, veronique.hoste@hogent.be, bart.desmet@hogent.be, philip.vanoosten@hogent.be, lieve.macken@hogent.be Department of Linguistics, Ghent University, Ghent, Belgium
BART DESMET
Affiliation:
Faculty of Applied Language Studies, University College Ghent, Ghent, Belgium e-mail: orphee.declercq@hogent.be, veronique.hoste@hogent.be, bart.desmet@hogent.be, philip.vanoosten@hogent.be, lieve.macken@hogent.be Department of Applied Mathematics and Computer Science, Ghent University, Ghent, Belgium e-mail: martine.decock@ugent.be
PHILIP VAN OOSTEN
Affiliation:
Faculty of Applied Language Studies, University College Ghent, Ghent, Belgium e-mail: orphee.declercq@hogent.be, veronique.hoste@hogent.be, bart.desmet@hogent.be, philip.vanoosten@hogent.be, lieve.macken@hogent.be Department of Applied Mathematics and Computer Science, Ghent University, Ghent, Belgium e-mail: martine.decock@ugent.be
MARTINE DE COCK
Affiliation:
Department of Applied Mathematics and Computer Science, Ghent University, Ghent, Belgium e-mail: martine.decock@ugent.be
LIEVE MACKEN
Affiliation:
Faculty of Applied Language Studies, University College Ghent, Ghent, Belgium e-mail: orphee.declercq@hogent.be, veronique.hoste@hogent.be, bart.desmet@hogent.be, philip.vanoosten@hogent.be, lieve.macken@hogent.be Department of Applied Mathematics and Computer Science, Ghent University, Ghent, Belgium e-mail: martine.decock@ugent.be

Abstract

While human annotation is crucial for many natural language processing tasks, it is often very expensive and time-consuming. Inspired by previous work on crowdsourcing, we investigate the viability of using non-expert labels instead of gold standard annotations from experts for a machine learning approach to automatic readability prediction. In order to do so, we evaluate two different methodologies to assess the readability of a wide variety of text material: A more traditional setup in which expert readers make readability judgments and a crowdsourcing setup for users who are not necessarily experts. To this purpose two assessment tools were implemented: a tool where expert readers can rank a batch of texts based on readability, and a lightweight crowdsourcing tool, which invites users to provide pairwise comparisons. To validate this approach, readability assessments for a corpus of written Dutch generic texts were gathered. By collecting multiple assessments per text, we explicitly wanted to level out readers' background knowledge and attitude. Our findings show that the assessments collected through both methodologies are highly consistent and that crowdsourcing is a viable alternative to expert labeling. This is a good news as crowdsourcing is more lightweight to use and can have access to a much wider audience of potential annotators. By performing a set of basic machine learning experiments using a feature set that mainly encodes basic lexical and morpho-syntactic information, we further illustrate how the collected data can be used to perform text comparisons or to assign an absolute readability score to an individual text. We do not focus on optimising the algorithms to achieve the best possible results for the learning tasks, but carry them out to illustrate the various possibilities of our data sets. The results on different data sets, however, show that our system outperforms the readability formulas and a baseline language modelling approach. We conclude that readability assessment by comparing texts is a polyvalent methodology, which can be adapted to specific domains and target audiences if required.

Information

Type
Articles
Copyright
Copyright © Cambridge University Press 2012 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

Article purchase

Temporarily unavailable