Paper and Digital Reading Assessments. Exploring aspects of validity using PIRLS and ePIRLS

Abstract

The shift from paper-based to digital assessment formats in education raises fundamental questions about the comparability of reading scores across modes and over time. International large-scale assessments increasingly rely on digital delivery while continuing to report trends on a common scale, making it essential to establish whether paper-based and digital reading assessments support equivalent score interpretations and valid trend comparisons across cycles. This dissertation examines the comparability and validity of paper-based and digital reading assessments using data from PIRLS 2016 and its digital extension, ePIRLS. The thesis consists of an integrative essay and three empirical studies, each addressing a distinct aspect of comparability and validity: score comparability across modes, subgroup differences, and the predictive value of reading scores for later educational outcomes. Results show that paper-based and digital reading assessments share a strong common core of reading comprehension, demonstrating substantial construct overlap across modes. At the same time, systematic mode-related variation is observed, indicating that the two assessment formats are not fully equivalent despite their strong association. Longitudinal analyses further demonstrate that both paper-based and digital reading scores meaningfully predict later academic outcomes, supporting their predictive validity, while small but systematic differences in predictive strength suggest that digital reading captures partly distinct aspects of educationally relevant skills. Taken together, the findings show that paper-based and digital reading assessment scores are similar but not interchangeable. Validity in digital reading assessments should therefore be understood as conditional on assessment mode, student population, and context, rather than assumed a priori. The dissertation concludes that ongoing empirical validation is essential to ensure the quality and interpretability of trend comparisons in international reading assessments as formats and literacy practices continue to evolve.

Description

Keywords

validity/ validation/ assessment/ paper-based reading/ digital reading/ reading comprehension/ PIRLS/ ePIRLS

Citation

ISBN

978-91-7963-255-7 (printed)
978-91-7963-256-4 (pdf)

Articles

1. Baghaei, P., Grammatikopoulou, E., Johansson, S., & Strietholt, R. (2025). Comparing linear and nonlinear reading assessments in PIRLS 2016. Studies in Educational Evaluation, 87, 101513. https://doi.org/10.1016/j.stueduc.2025.101513

2. Grammatikopoulou, E., Johansson, S., & Rosén, M. (2025). Paper-based and digital reading in 14 countries: Exploring cross-country variation in mode effects. Educational Review, 1–19. https://doi.org/10.1080/00131911.2025.2452236

3. Grammatikopoulou, E. (2026). The predictive validity of PIRLS and ePIRLS on later academic achievement: Insights from Sweden,. Studies in Educational Evaluation, 88, 101554. https://doi.org/10.1016/j.stueduc.2025.101554

Department

Department of Education and Special Education ; Institutionen för pedagogik och specialpedagogik

Defence location

Onsdagen den 25 februari 2026, kl. 13:00, Sal AK2 136 K-G Stukát, Hus A, Pedagogen, Västra hamngatan 25

Endorsement

Review

Supplemented By

Referenced By