Existing approaches to measuring writing performance are insufficient in terms of both technical adequacy as well as feasibility for use as a screening measure. This study examined the validity and diagnostic accuracy of several approaches to automated text evaluation as well as written expression curriculum-based measurement (WE-CBM) to determine whether an automated approach improves technical adequacy. A sample of 140 fourth grade students generated writing samples that were then scored using traditional and automated approaches and examined in relation to the statewide measure of writing performance. Results indicated that the validity and diagnostic accuracy for the best performing WE-CBM metric, correct minus incorrect word sequences, and the automated approaches to scoring were comparable, with automated approaches offering potentially improved feasibility for use in screening. Averaging scores across three time points was necessary, however, in order to achieve improved validity and adequate levels of diagnostic accuracy across the scoring approaches. Limitations, implications, and directions for future research regarding the use of automated scoring approaches for screening are discussed.

Keller-Margulis, M., Mercer, S., Matta, M. (2021). Validity of automated text evaluation tools for written-expression curriculum-based measurement: a comparison study. READING & WRITING, 34(10), 2461-2480 [10.1007/s11145-021-10153-6].

Validity of automated text evaluation tools for written-expression curriculum-based measurement: a comparison study

Matta M.
2021

Abstract

Existing approaches to measuring writing performance are insufficient in terms of both technical adequacy as well as feasibility for use as a screening measure. This study examined the validity and diagnostic accuracy of several approaches to automated text evaluation as well as written expression curriculum-based measurement (WE-CBM) to determine whether an automated approach improves technical adequacy. A sample of 140 fourth grade students generated writing samples that were then scored using traditional and automated approaches and examined in relation to the statewide measure of writing performance. Results indicated that the validity and diagnostic accuracy for the best performing WE-CBM metric, correct minus incorrect word sequences, and the automated approaches to scoring were comparable, with automated approaches offering potentially improved feasibility for use in screening. Averaging scores across three time points was necessary, however, in order to achieve improved validity and adequate levels of diagnostic accuracy across the scoring approaches. Limitations, implications, and directions for future research regarding the use of automated scoring approaches for screening are discussed.
Articolo in rivista - Articolo scientifico
Automated text evaluation; Curriculum-based measurement; Validity; Written expression;
English
2021
34
10
2461
2480
reserved
Keller-Margulis, M., Mercer, S., Matta, M. (2021). Validity of automated text evaluation tools for written-expression curriculum-based measurement: a comparison study. READING & WRITING, 34(10), 2461-2480 [10.1007/s11145-021-10153-6].
File in questo prodotto:
File Dimensione Formato  
Keller-Margulis-2021-Reading and Writing-VoR.pdf

Solo gestori archivio

Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Tutti i diritti riservati
Dimensione 674.04 kB
Formato Adobe PDF
674.04 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/505640
Citazioni
  • Scopus 8
  • ???jsp.display-item.citation.isi??? 8
Social impact