737
Views
0
CrossRef citations to date
0
Altmetric
Editorial

Insight from research across the world: examining student responses, assessment practices, and feedback engagement

ORCID Icon

In this first regular issue of 2024, the original articles span from international large-scale studies to students in the classroom.

The first article, by Steinmann et al. (Citation2024), investigates patterns in student questionnaire responses in the TIMSS 2019 study. More specifically, the researchers examined which students are more likely to respond inconsistently to mixed-worded questionnaire scales, and which country samples have larger shares of inconsistent respondents. At the student level, Steinmann et al. (Citation2024) investigated four predictor variables of inconsistent responding, separately and in a joint model: 1) mathematics achievement, 2) student age, 3) language at home, and 4) gender. Across the different models, countries, and grade levels, mathematics achievement stood out as the strongest and most consistent predictor of inconsistent responding. Interestingly, it was also found that girls in many countries responded more consistently than boys and that students speaking the test language at home are more likely to respond consistently. The authors discuss interpretations of the data noting that the design of the study does not allow for directly testing their interpretation that lack-of-skills seems to be the explanation and not students being careless when responding.

Such interpretations were also offered in similar studies investigating the validity of the PISA questionnaire data, where it was demonstrated that low-performing students on PISA science achievement did not report consistently their learning strategies measured in the PISA student questionnaire (Hopfenbeck & Maul, Citation2011). Further interviews confirmed some of the students did not even understand the items the way the test-developer intended. The study by Steinmann et al. (Citation2024) raises the need for further studies investigating International Large Scale Assessment studies and inconsistency in student response patterns, with both quantitative and mixed methods, particularly considering the large-scale impact these studies and results have globally, both on assessment policy and practice.

In the second article, Chambers et al. (Citation2024) outline how comparative judgement can be used as part of the moderation process when schools have non-exam assessments. Although most of qualifications completed as part of secondary school education (age 16–18) in England are assessed via exams marked by an awarding body, a few qualifications comprise non-exam assessments, internal assessments, such as portfolios, practical demonstrations of skills and performances marked by teachers. Teacher assessments are known for having challenges around reliabilities and standards globally. In England, a moderation process with expert judgers checking sample of teachers’ assessment is taking place to secure standards across schools (Gill, Citation2015). The research team examined whether Comparative Judgement (CJ) could be used as part of this moderation process, by running simulations on datasets. More specifically, they investigated the effects of CJ design parameters on the CJ estimates of script quality and how to assign moderator marks after the CJ procedure. The findings showed that certain design parameters had substantial effects on reliability and suggested minimum values for CJ protocols.

They conclude that the use of CJ for moderation has the potential through technological advances to shape a more efficient moderation process, although they acknowledge their current study did not include any humans and thus effects might differ in a real-life empirical study.

In the third article, Gamlem et al. (Citation2024) investigates the perceptions and argumentations for grading among Norwegian teachers when they assess English as a Foreign Language. The research team examined possible gender differences with respect to how teachers graded their students and how teachers argued and reasoned about their decisions for the grade. Teachers from three schools graded a total of 182 assessments, more specifically essays in English as a Foreign Language, using a 26-point Grading Scale (26-PGS). It was found that female students generally received higher grades compared to male students. These results align with previous research indicating that girls tend to receive higher grades than boys when their performance is assessed by classroom teachers, as teachers often employ subjective criteria in grading, resulting in variability of grading practices both within individual schools and across different schools.

The final paper is authored by Nash and Thomas (Citation2024) on the topic of feedback. The importance of feedback to students for their future learning is well known and has been extensively researched (Hattie & Timperley, Citation2007; Panadero, Citation2023; Winstone & Nash, Citation2023). Most studies have taken a quantitative approach looking at effect sizes or surveyed teachers’ and students’ perception and use of feedback. The current study by Nash and Thomas is the first to our knowledge applying linguistic text analysis to examine how students reflect on their feedback experiences. This novel approach and the study itself offer insight into how students act and engage with the written feedback they are provided by their teachers and are therefore an important contribution to our field. The researchers analysed records from 760 undergraduate single- or joint-honours psychology students at Aston University in the US, who commenced their first year of study between 2017 and 2019. As part of a compulsory introductory social psychology module, students were asked to complete coursework assignments on the topic of how to engage with feedback. A total of three reflective writing samples for each student were analysed by the researchers. More specifically, the research team examined whether certain linguistic features of students’ written reflections in the University’s Blackboard LMS, which students wrote over the course of 1 year, on their experience of receiving and using feedback, would be associated with their track record of accessing vs. ignoring their instructors’ feedback. Nash and Thomas (Citation2024) did not find meaningful associations between feedback-accessing and students’ language use but instead demonstrated in their analysis that students with lower grades have a greater tendency to ignore feedback. Similarly, the analysis of the language revealed that students with lower grades tended to focus relatively more on the past or present in their written reflections than on the future. The main goal of the study was to investigate whether students’ written reflections on feedback experiences could reveal any linguistic cues which would be associated with those students’ tendencies to access their teachers’ feedback in the electronic LMS system. Although no evidence of such association was found in this study, the idea of using language analysis to further enhance our understanding of feedback literacy, use and application, should be encouraged. Moreover, as the authors rightly state:

[…]in order to establish valid and robust theoretical accounts of feedback literacy, it is important that nonsignificant empirical findings feature transparently in the literature alongside significant findings (see e.g. Patall, Citation2021).

With the ongoing replication crisis in psychology (Open Science Collaboration, Citation2015), we urge the assessment community to conduct more replication studies of past studies both in new contexts and with new measures and to report also non-significant findings. The latter is particularly important in an era where researchers and educators are pressured to focus upon ‘what works’ and trust in research is under constant threat by some political leaders globally. Only through high-quality research with openness and integrity can we continue to serve our global education community.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Notes on contributors

Therese Hopfenbeck

Dr. Therese Hopfenbeck is Professor of Educational Assessment, the Director of the Assessment and Evaluation Research Centre at the Faculty of Education, University of Melbourne, Visiting Fellow at Kellogg College, University of Oxford and Adjunct Professor of The Norwegian University of Science and Technology.

References

  • Chambers, L., Vitello, S., & Vidal Rodeiro, C. (2024). Moderation of non-exam assessments: A novel approach using comparative judgement. Assessment in Education Principles, Policy & Practice, 1–24. https://doi.org/10.1080/0969594X.2024.2313237
  • Gamlem, S. M., Segaran, M., & Moltudal, S. (2024). Lower secondary school teachers’ arguments on the use of a 26-point grading scale and gender differences in use and perceptions. Assessment in Education Principles, Policy & Practice, 1–19. https://doi.org/10.1080/0969594X.2024.2325365
  • Gill, T. (2015). The moderation of coursework and controlled assessment: A summary. Research Matters: A Cambridge Assessment Publication, 19, 26–31. https://doi.org/10.17863/CAM.100323
  • Hattie, J., & Timperley, H. (2007). The power of feedback. American Educational Research Association, 77(1), 81–112. https://doi.org/10.3102/003465430298487
  • Hopfenbeck, T. N., & Maul, A. (2011). Examining evidence for the validity of PISA learning strategy scales based on student response processes. International Journal of Testing, 11(2), 95–121. https://doi.org/10.1080/15305058.2010.529977
  • Nash, R., & Thomas, J. M. (2024). Searching students’ reflective writing for linguistic correlates of their tendency to ignore instructors’ feedback. Assessment in Education Principles, Policy & Practice, 1–16. https://doi.org/10.1080/0969594X.2024.2332267
  • Open Science Collaboration. (2015). Estimating the reproducibility of psychological science. Science, 349(6251). https://doi.org/10.1126/science.aac4716
  • Panadero, E. (2023). Towards a paradigm shift in feedback research: Five further steps influenced by self-regulated learning theory. Educational Psychologist, 58(3), 193–204. https://doi.org/10.1080/00461520.2023.2223642
  • Patall, E. A. (2021). Implications of the open science era for educational psychology research syntheses. Educational Psychologist, 56(2), 142–160. https://doi.org/10.1080/00461520.2021.1897009
  • Steinmann, I., Chen, J., & Braeken, J. (2024). Who responds inconsistently to mixed-worded scales? Differences by achievement, age group, and gender, assessment in education: Principles. Assessment in Education Principles, Policy & Practice, 1–28. https://doi.org/10.1080/0969594X.2024.2318554
  • Winstone, N. E., & Nash, R. A. (2023). Towards a cohesive psychological science of effective feedback. Educational Psychologist, 58(3), 111–129. https://doi.org/10.1080/00461520.2023.2224444

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.