Abstract
Abstract: The use of digital media in education can bring great benefits and its use in schooling is steadily increasing. Administrating paper- versus computer-based as well as fixed-item versus adaptive tests could create differences in test experience, which can threaten the comparability of test results. This study investigated how the pen-and-paper, computer-based, and computer-adaptive test formats of a standardized reading comprehension test affect test anxiety and motivation among German fourth-grade students. A within-class randomized field trial with 387 fourth graders (aged: 9–10 years; 46.3% female) was conducted. Repeated-measures analysis of covariance (ANCOVA) revealed no differences in state test anxiety between the test formats when controlling for trait test anxiety and pre-test state anxiety, but state reading motivation was initially higher when reading on a screen, controlling for trait reading motivation. However, this difference diminishes over the course of the test. Implications for using digital media in elementary school test situations are discussed.
References
2001). Test motivation in the assessment of student skills: The effects of incentives on motivation and performance. European Journal of Psychology of Education, 16(3), 441–462. https://doi.org/10.1007/BF03173192
(2014). Test anxiety, self-control, and knowledge retrieval in secondary school students. Zeitschrift Für Entwicklungspsychologie Und Pädagogische Psychologie, 46(4), 165–170. https://doi.org/10.1026/0049-8637/a000111
(2023). Supplement to “Effects of test mode and medium on elementary school students’ test experience” https://osf.io/76hc2/?view_only=baf5985b8cb94fe4b78cbce4261aee7a. https://doi.org/10.17605/OSF.IO/76HC2
(2012). Effects of computer-based testing on test performance and testing motivation. Computers in Human Behavior, 28(5), 1580–1586. https://doi.org/10.1016/j.chb.2012.03.020
(2013). Test anxiety, computer-adaptive testing, and the common core. Journal of Education and Training Studies, 1(2), 50–60. https://doi.org/10.11114/jets.vli2.101
(2018). Computer anxiety and interaction: A systematic review. In Proceedings of the 15th International Web for All Conference, (pp. 1–10). https://doi.org/10.1145/3192714.3192825
(2011). Entwicklung der Fünf-Item Kurzskala STAI-SKD zur Messung von Zustandsangst
([Development of the five-item short scale STAI-SKD for the assessment of state anxiety] . Zeitschrift für Gesundheitspsychologie, 19, 173–180. https://doi.org/10.1026/0943-8149/a0000492020). Preparing for life in a digital world: IEA international computer and information literacy study 2018 international report. Springer Nature. https://doi.org/10.1007/978-3-030-38781-5
(2009). Effekte des adaptiven Testens auf die Motivation zur Testbearbeitung am Beispiel des Frankfurter Adaptiven Konzentrationsleistungs-Tests
([Effects of adaptive testing on test-taking motivation with the example of the Frankfurt Adaptive Concentration Test] . Diagnostica, 55(1), 20–28. https://doi.org/10.1026/0012-1924.55.1.202018). Psychological determinants of test motivation in low-stakes test situations: A longitudinal study of singletrait–multistate models in accounting. Empirical Research in Vocational Education and Training, 10(1), 1–34. https://doi.org/10.1186/s40461-018-0074-7
(2017). IGLU 2016. Lesekompetenzen von Grundschulkindern in Deutschland im internationalen Vergleich
([International comparison of German elementary school students’ reading competencies] . Waxmann.2002). PISA 2000: Dokumentation der Erhebungsinstrumente (Materialien aus der Bildungsforschung Nr. 72)
([PISA 2000: Documentation of the survey instruments (Educational research materials No. 72)] . Max-Planck-Institut für Bildungsforschung.2021). Gender differences in text‐based interest: Text characteristics as underlying variables. Reading Research Quarterly, 57(2), 537–554. https://doi.org/10.1002/rrq.420
(2017). Is a computerized adaptive test more motivating than a fixed-item test? Applied Psychological Measurement, 41(7), 495–511. https://doi.org/10.1177/0146621617707556
(2021). Adaptives Testen von Textverständnis in der Grundschule
([Adaptive Testing of Text Comprehension in Primary School. Development of a CAT-Optimized Item Pool] . Diagnostica, 68(1), 39–50.2012). Random generation of response patterns under computerized adaptive testing with the R package catR. Journal of Statistical Software, 48(8), 1–31. https://doi.org/10.18637/jss.v048.i08
(2018). Computer-adaptive testing: Implications for students’ achievement, motivation, engagement, and subjective test experience. Journal of Educational Psychology, 110(1), 27–45. https://doi.org/10.1037/edu0000205
(2011). Consequences of test anxiety on adaptive versus fixed item testing. European Journal of Psychological Assessment, 27(3), 157–163. https://doi.org/10.1027/1015-5759/a000062
(2014). I will probably fail. European Journal of Psychological Assessment, 30(1), 48–56. https://doi.org/10.1027/1015-5759/a000168
(1984). Test anxiety and ineffective test taking: Different names, same construct? Journal of Educational Psychology, 76(2), 279–288. https://doi.org/10.1037/0022-0663.76.2.279
(2014). The impact of ebooks on the reading motivation and reading skills of children and young people: A rapid literature review. National Literacy Trust. Retrieved October 13, 2022 from https://eric.ed.gov/?id=ed560635
(2022). R: A language and environment for statistical computing. R Foundation for Statistical Computing.
. (2021). The effects of modes of test administration on test anxiety and test scores: A study in an Indonesian school. Issues in Educational Research, 31(3), 952–971. https://www.iier.org.au/iier31/sahlan.pdf
(2015). The school age gender gap in reading achievement: Examining the influences of item format and intrinsic reading motivation. Reading Research Quarterly, 50(2), 219–232. https://doi.org/10.1002/rrq.92
(1998). Effects of computer-based test administrations on test anxiety and performance. Computers in Human Behavior, 14(1), 111–123. https://doi.org/10.1016/S0747-5632(97)00035-6
(2019). Beyond novelty effect: A mixed-methods exploration into the motivation for long-term activity tracker use. JAMIA Open, 2(1), 62–72. https://doi.org/10.1093/jamiaopen/ooy048
(1977). Anxiety, learning, and instruction. Routledge. https://doi.org/10.4324/9780203056684
(2002). Computer-adaptive testing: The impact of test characteristics on perceived performance and test takers’ reactions. Journal of Applied Psychology, 87(2), 320–332. https://doi.org/10.1037/0021-9010.87.2.320
(1995). Trait and state motivation and the acquisition of Hebrew vocabulary. Canadian Journal of Behavioural Science/Revue Canadienne des Sciences du Comportement, 27(3), 356–370. https://doi.org/10.1037/0008-400X.27.3.356
(2018). Test anxiety effects, predictors, and correlates: A 30-year meta-analytic review. Journal of Affective Disorders, 227, 483–493. https://doi.org/10.1016/j.jad.2017.11.048
(2008). Eine Kurzform des Prüfungsängstlichkeitsinventars TAI-G
([A short version of the Test Anxiety Inventory TAI-G] . Zeitschrift für Pädagogische Psychologie, 22(1), 73–81. https://doi.org/10.1024/1010-0652.22.1.731989). Weighted likelihood estimation of ability in item response theory. Psychometrika, 54(3), 427–450. https://doi.org/10.1007/bf02294627
(1973). Ability measurement: Conventional or adaptive? (Research Report 73-1). Department of Psychology, University of Minnesota.
(2006). An application of item response time: The effort-moderated IRT model. Journal of Educational Measurement, 43(1), 19–38. https://doi.org/10.1111/j.1745-3984.2006.00002.x
(2019). Introduction of multistage adaptive testing design in PISA 2018 (OECD Education Working Papers No. 209). OECD Publishing. https://doi.org/10.1787/b9435d4b-en
(1998). An additive model of test anxiety: Role of exam-specific expectations. Journal of Educational Psychology, 90(2), 330–340. https://doi.org/10.1037/0022-0663.90.2.330
(