1,806
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Are we answering the question that has been set? Exploring the gap between research and practice around examinations in higher education

ORCID Icon
Received 27 Jan 2023, Accepted 10 Nov 2023, Published online: 27 Nov 2023

ABSTRACT

Despite a large amount of critical research literature, traditional examinations continue to be widely used in higher education. This article reviews recent literature in order to assess the role played by the approaches adopted by researchers in the gap between research on exams, and the way exams are used. Viviane Robinson’s ‘problem-based methodology’ focuses on the need for researchers to engage with the challenges and priorities of practitioners. Drawing on Robinson’s approach, the article investigates how the strengths and weaknesses of exams and their alternatives are framed in the literature published between 2016 and 2021. The article concludes that there is an absence of evidence about how and why practitioners make decisions about assessment. This hinders the ability of assessment researchers to appropriately connect their work with the assessment challenges practitioners face. To make a difference, assessment research needs to live in the real world; a world which, at least as far as practitioners’ assessment decisions are concerned, we do not yet sufficiently understand.

1. Introduction

Assessment is an aspect of higher education where the gap between research and practice is often believed to be particularly wide. There is a sense that while the historical shift in focus from what teachers do to what students do (Barr and Tagg Citation1995) has had considerable impact on pedagogy – whether that’s problem-based learning (Savin-Baden Citation2000) or flipped classrooms (O'Flaherty and Phillips Citation2015) – the hold of traditional transmission-focused approaches has lingered longer when it comes to assessment and feedback (Winstone and Carless Citation2020). It is common to see researchers lament the lack of innovation in assessment, and the use of outmoded and unhelpful ways of assessing students and providing feedback. In 1995, David Boud said ‘[t]here is probably more bad practice and ignorance of significant issues in the area of assessment than in any other aspect of higher education’ (Boud Citation1995, 35). Similar things are still said today. For example, relatively low scores on student survey items on assessment are cited – perhaps questionably (Buckley Citation2021a) – as evidence that assessment faces something of a crisis: ‘In contrast to the effective feedback possibilities expounded in the literature, students generally report in surveys that feedback is done poorly in higher education, compared with other aspects of their studies’ (Dawson et al. Citation2019, 25).

Formal examinations probably represent the epitome of this perceived disconnect between assessment research and the assessment methods that are actually used within universities. Chong and McArthur (Citation2023) say that ‘traditional examinations remain commonplace in Western higher education, despite the scholarship on assessment for learning and the work of educational developers in this field’ (5). For anyone in doubt about the status of exams as the traditional bogeyman for those working in the field of assessment, the concerns stretch back a very long way indeed. In 1967, R. Cox (Citation1967) could say ‘[i]t is now thirty years since serious doubts were raised about examinations, yet despite the fact that there has been no serious shortage of critics since then, very little has changed’ (352, quoted in Elton Citation2004, 43). If we want to find out why assessment research generally fails to have the impact on practice that we would like, then examinations seem like a good place to look. That isn’t necessarily to agree with the body of literature criticising the use of exams; the value of exams is a complex topic outside the scope of this paper. This paper is focused on a different issue, namely why there is this gap between the broadly critical literature on examinations and the ongoing and widespread use of examinations in higher education. The status of traditional exams as a standard assessment tool was of course undermined dramatically by the COVID-19 pandemic, which required universities around the world to shift at short notice to alternative forms of assessment that could be undertaken at a distance (Henderson et al. Citation2023). The longer-term implications of that emergency shift for the ongoing use of exams are still unclear; time will tell whether the departure from exams will be a temporary measure or the start of a broader shift. However, while the practices and influences may have changed as a result of COVID-19, there are still lessons to learn about how to close the distance between research and practice.

There has been some limited discussion of the research-practice gap in higher education. In 1985, George Keller infamously attacked HE research, stating that ‘hardly anyone in higher education pays attention to the research and scholarship about higher education’ (7). Colbeck (Citation2000) describes the perverse situation whereby ‘university administrators implement major teaching and curriculum reform initiatives without so much as brief phone consultations with experts on their own campuses’ (35). More recently, Evans et al. (Citation2021) lament the way in which higher education research can end up producing ‘work that is little value on the ground in informing teaching and learning’ (526).

In contrast to the muted debate about the research-practice gap in higher education, there has been vigorous discussion over many decades at school level, going back at least as far as John Dewey (Citation1929). A wide range of explanations have been suggested for the lack of impact of research on learning, teaching and assessment in schools, some of which are unlikely to translate to any great extent into higher education. For example, given the dual research-teaching role of most academics, there is probably not a great divide between a ‘craft culture’ and ‘research culture’ as has been observed in school contexts (Hirschkorn and Geelan Citation2008). The relative lack of status afforded by HE research (Cotton, Miller, and Kneale Citation2018) probably means that there are fewer incentives than in school education for researchers to focus on ‘pure’ rather than practice-focused research (Hemsley-Brown and Sharp Citation2003). On the other hand, some explanations are likely to be relevant to higher education, such as a perception that the findings from educational research lack validity (Vanderline and van Braak Citation2010) and the fact that practitioners lack the time to engage with research (Hirschkorn and Geelan Citation2008).

The issue discussed in this article is whether researchers are using the right kinds of approaches in their work. That is not because the other factors contributing to the gap between research and practice are not important or serious, but because they are in a sense less fundamental. Even if HE researchers solely used words of one syllable, even if HE practitioners had all the time in the world, even if HE research was seen as highly rigorous; even with all those benefits, if the research itself does not engage with the concrete and practical challenges of assessment in the right way, it will not influence practice in any way. Practitioners simply won’t be able to make use of the research as part of their assessment practice.

In order to explore researchers’ approaches as a factor in the research-practice gap in assessment in higher education, the article will describe a theory of impactful educational research, and discuss its application to a particular aspect of assessment. The theory is Viviane Robinson’s ‘problem-based methodology’, and the aspect of assessment is the traditional examination – meaning a high-stakes, summative, in-person, invigilated, individual, written, closed-book assessment.

Over a number of years, Viviane Robinson has developed an influential (Haig Citation2014; McNaughton Citation2014) critique of research methodologies employed by school-focused educational researchers (Robinson, Citation1992, Citation1993, Citation1998, Citation2014; Robinson and Lai Citation1999). In her view, insufficient engagement by researchers with the challenges faced by practitioners is a key factor in the lack of uptake of educational research:

I am making the … radical claim that research may be ignored, regardless of how skilfully it is communicated, because it bypasses the problem-solving processes that sustain the practices that researchers seek to alter. (Robinson Citation1998, 17)

Robinson’s model is worth exploring because its underlying idea – that researchers don’t place enough emphasis on engaging with the problems practitioners take themselves to face – is widely felt among those commenting on higher education research. For example, Evans et al. (Citation2021) caution against research that is ‘ignorant of the situational requirements’, and say that ‘researchers need to have a holistic understanding of issues impacting the field and an in-depth knowledge of the immediate context they are reporting on’ (526). Kezar (Citation2000) is concerned that ‘practitioners’ assumptions and habits of mind are generally not taken into account’ (17) in higher education research. Cross (Citation2000) reports the sense that researchers ‘are not interested in addressing the problems that practitioners must solve’ (65).

Beyond that, there are strong parallels between Robinson’s model and some recent work advocating for a ‘practice perspective’ on assessment (Boud et al. Citation2018). This work shares Robinson’s diagnosis that the gap between research and practice is partly due to researchers’ approaches, or ‘limitations in the ways in which research on assessment is conceived and framed’ (Boud et al. Citation2018, 1108). The proposed remedy is, as for Robinson, to focus on the reality of assessment practice: ‘an unrelentingly grounded picture of the everyday world in which assessment sits’ (Boud et al. Citation2018, 1116). The practice perspective also, like Robinson, has a focus on the decisions that practitioners make about assessment. For example, it has resulted in the development of the ‘Assessment Design Decisions Framework’, which is intended to help practitioners consider a range of different factors when making their decisions; factors such as the purpose and context of the assessment (Bearman et al. Citation2016). In aiming to help practitioners to ‘achieve and expand their aspirations’ (Bearman et al. Citation2017, 51) the approach maps nicely onto Robinson’s idea – discussed below – of addressing both the effectiveness with which assessment approaches satisfy practitioners’ existing constraints, and the coherence of those approaches with wider constraints of relevance to practitioners.

Nevertheless, there are differences between the ‘practice perspective’ and Robinson’s ‘problem-based methodology’. Most fundamentally, there is a disagreement about the value of framing decisions in terms of rationality. While Robinson’s approach ‘requires … a willingness on the part of researchers to recognize the rationality of those practices they seek to alter’ (Robinson Citation1998, 25), Bearman et al. (Citation2017) reject the idea that assessment decisions are ‘based on a rational analysis of competing demands’ (50), and therefore focus on influences that go beyond those that practitioners would consciously consider when weighing up an assessment format – to include, for example, professional identity. So while there are crucial differences between these approaches, there are also key similarities, and the emergence of the ‘practice perspective’ provides additional encouragement that Robinson’s problem-based methodology would find a receptive audience in researchers exploring assessment in higher education.

To say that Robinson’s ‘problem-based methodology’ has value – and more broadly that it is worthwhile exploring how researchers engage with practitioners’ challenges – is not the same as saying that all research is aimed at changing practice; or that all research aimed at changing practice attempts to do so via persuading practitioners that there are better solutions to their challenges. There is much research that is not aimed at changing practice, and there is much research that aims to change practice through changing systems and processes rather than practitioners’ decisions. In those cases, the intended audience of the research will not be practitioners. Indeed, many assessment decisions are out of practitioners’ hands; where accrediting professional bodies have particular expectations, for example, or where institutions have certain kinds of assessment policies. And in understanding and influencing assessment practices, a whole range of factors beyond individual practitioners require exploration: structural, political, financial, etc. However, in many cases practitioners do have a significant element of choice about assessment methods. If those practitioners are going to be persuaded to make different choices (what Robinson calls ‘non-coercive change’) then the research ‘must square with what they already believe, or with what, as a result of dialogue with the researcher, they come to believe’ (Robinson Citation1992, 25). For that kind of research, the intended audience (directly or indirectly) will be practitioners themselves. As evidenced above, there is a concern from higher education researchers that there is insufficient engagement with the challenges practitioners take themselves to face; a concern which Robinson’s approach addresses head-on.

The article will explore how well the recent research literature on exams satisfies the demands of Robinson’s model, in order to develop some suggestions for how exam research – and assessment research in general – could have more impact on assessment practices. The issue of exams is a promising case study for the gap between research and practice in assessment in higher education, simply because the gap appears so large. The widespread and ongoing use of traditional exams seems to stubbornly ignore the volume of negative literature. By focusing on exams, this article attempts to shed some wider light on how the research-practice gap in assessment in higher education can be made at least a little narrower.

2. Closing the research-practice gap with a problem-based methodology

Drawing on the idea of ‘theories of practice’ developed by Chris Argyris and Donald Schon in the 1970s (Argyris and Schon Citation1974), Robinson proposes a ‘problem-based methodology’ for educational research, starting from the idea that educational practices that researchers may wish to challenge and change are not best seen as irrational spasms or unthinking traditions, but as rational attempts to solve particular problems. As a consequence, Robinson believes that researchers need to offer superior solutions to those very same problems if they want to contribute to the improvement of those practices. Just as students can perform poorly in exams because they answer the question they wish they’d been asked rather than the one they are set on the paper, the work of researchers can lack impact because they offer solutions to the problems they see, rather than the problems practitioners take themselves to face: ‘Researchers have little to offer by way of alternative solutions, when the problems they have been studying are not those of the practitioner’ (Robinson Citation1993, 4).

Central to Robinson’s approach is the idea of practitioners’ ‘constraints’, ‘the things that need to be taken into account in coming up with an adequate solution’ (Robinson Citation2014, 343). Constraints can include a very wide range of things: particular values and beliefs, time pressures, available resources, other existing practices, etc. Constraints need not necessarily be transparent to the practitioners themselves. Robinson uses the established idea of ‘espoused theories’ and ‘theories-in-use’ (Argyris and Schon Citation1974) to distinguish between those constraints that practitioners themselves would cite when defining the problem, and those constraints that guide practitioners’ activities without them being aware. The ‘constraint set’, consisting of both kinds of constraints, defines the problem to be solved.

Educational practices are conceptualised as attempts by practitioners to satisfy a relevant constraint set (Robinson Citation1993). Therefore, any research that aims to directly change practice needs to offer a solution that satisfies the constraint set in a superior way: researchers need to offer a better solution to the problem that practitioners are attempting to solve. Robinson offers three criteria for judging the value of such a solution:

  • Accuracy. The factual claims that are part of the solution need to be accurate.

  • Effectiveness. An effective solution needs to satisfy practitioners’ constraint set. This is not as simple as solving the problem as defined by practitioners, due to the inclusion in the constraint set of ‘constraints-in-use’, of which practitioners may not be overtly aware. This criterion captures how well a solution achieves the outcomes anticipated or desired by practitioners.

  • Coherence. Not only does a solution need to satisfy practitioners’ constraint set, it needs to do so in a way that does not make it harder to solve other problems that the practitioners care about. Researchers can make the case that additional considerations are relevant to the problem. This criterion accommodates the need for researchers to ‘get outside’ the constraint set and challenge how practitioners frame the problem: ‘the coherence criterion … goes beyond what problem solvers themselves have currently taken as important and asks whether there are other values which should influence the choice of solution for the problem’ (Robinson Citation2014, 349)

Robinson links a number of familiar concerns about educational research to these three criteria. For example, researchers who privilege one or two constraints rather than addressing the set as a whole – who oversimplify the problem – are failing on the effectiveness criterion. Researchers who offer solutions without considering important knock-on impacts, or who emphasise constraints not included in practitioners’ own constraint set without explaining their relevance to the wider concerns of practitioners’ themselves, are failing on the coherence criterion.

Given that impactful research needs to engage with practitioners’ own ‘problem analysis’, what do we know about the factors that academics are constrained by when they design assessments? On this issue, there is a widely-recognised lack of evidence. Numerous researchers have commented on the absence of attention paid to academics’ decision-making processes when it comes to assessment (Bearman et al. Citation2017; Joughin, Dawson, and Boud Citation2017; Samuelowicz and Bain Citation2002). For example, in his review of the literature comparing exams with coursework, Richardson (Citation2015) ‘found no research on the attitudes of academic staff to different forms of assessment’ (445). Some researchers have contrasted this lack of evidence about how academics develop assessments, with the very extensive research on students’ experiences of those assessments (Adachi, Tai, and Dawson Citation2018; Brady, Devitt, and Kiersey Citation2019). Of the research that does discuss assessment decisions, some focuses on non-rational factors such as the kinds of biases and traditions that can lead academics to ‘resist change’ (Joughin, Dawson, and Boud Citation2017; see also Chong and McArthur Citation2023; Harland and Wald Citation2021). And some of the research is limited in that it explores the experiences of academics who are unrepresentative in crucial ways, such as being particularly passionate about self and peer assessment (Adachi, Tai, and Dawson Citation2018) or because they were identified as using particularly innovative assessment methods (McLean Citation2018). Bearman et al. (Citation2017) report the results of interviews with academic staff and identified a number of constraints (or ‘influences’), such as the need to reduce the time taken up by assessment implementation and marking. Citing the poor understanding of assessment decision-making, Dawson et al. (Citation2013) ‘urge scholars of learning and teaching to move beyond the study of abstracted assessment practices to understand how assessment principles can translate to improved outcomes in local contexts’ (110). It is important to note that this lack of research into practitioners’ assessment decisions doesn’t imply that there is a lack of research into practitioners’ conceptions of assessment. There is a large and longstanding literature on issues such as practitioners’ assessment literacy (e.g. Chan and Luk Citation2022) and the gap between student and staff conceptions of assessment (e.g. MacLellan Citation2001). How practitioners conceptualise the assessment process is related to, but distinct from, the question of how they go about making assessment decisions and the factors they consider when they do so. It is the lack of research on the latter that has been highlighted by researchers.

While there is limited research exploring the constraints that frame the assessment decisions of academic staff, there appears to be effectively none that specifically looks at the constraints that are involved in the decisions around the use of exams. We are limited to glimpses such as an ‘illustrative story’ that Samuelowicz and Bain (Citation2002) use to bring their findings to life, and an interesting but isolated quotation from Bearman et al. (Citation2016):

Well I suppose, yeah, I suppose … Well, I don’t know actually. It depends whether it’s important to test whether they have any basic knowledge, I don’t know. I feel uncomfortable not having exams. (Science lecturer)

I don’t think that probably crossed our mind, not having an exam. (Science lecturer, 550)

Given the widespread concern about the persistence of exams in the face of critical assessment research, it is surprising that research has not explored the decision processes that lead academics to choose examinations as an assessment method. It is surprising in itself, and in the context of employing an appropriate methodology for practice-oriented research, it creates a significant obstacle to influencing practice.

3. Method

Literature on examinations in higher education published in the five-year period between 2016 and 2021 was reviewed to explore the extent of its alignment with Robinson’s problem-based methodology. (Some of the papers reviewed were originally available as advance online publications, and have been published since 2021 in print). In line with the procedures described by Richardson (Citation2015) – who undertook a review of the literature comparing exams and coursework – a number of steps were taken to overcome the issue that the term ‘examination’ can refer to both an assessment format (‘open-book examination’) and as the process of investigation (‘a critical examination of … ’). An initial simple search was undertaken using Google Scholar, with backwards and forwards searching of cited sources. To provide a more systematic literature search this was supplemented by searches within individual journals. In his review of higher education journals, Tight (Citation2018) provides a list of 28 ‘key journals’ based on age, size and ranking, and this list was used as a focus for the search. Within these 28 journals, searches were undertaken using Scopus to find papers published between 2016 and 2021 featuring the terms (and their plurals) ‘exam’ and ‘examination’ in the title, with phrases such as ‘an examination’ and ‘critical examination’ excluded. This yielded 142 papers. This was reduced to 31 papers following the elimination of papers not focused on the investigation of examinations as a way of assessing taught degree programmes in higher education (e.g. focused on school-level or university entry exams, examination of PhDs, physical medical examinations, and where ‘examination’ was used in the sense of an investigation). The number was reduced further to 17 when papers that discussed exams as data sources rather than exams as an assessment format were eliminated. In total (combining the initial and systematic searches) 43 relevant papers were found.

These 43 papers were first reviewed in order to ascertain which assessment constraints they focused on. The most common themes have been used to structure the findings: learning, incorporating both the activity of learning (studying) and the achievement of learning (student performance) (Buckley Citation2021b); student wellbeing (exam anxiety); assessment security and academic integrity; authenticity; inclusivity. The papers were then reviewed to determine how they framed those constraints with regard to practitioners’ concerns. Each paper was categorised as addressing the accuracy of practitioners’ beliefs (accuracy), how well practitioners’ existing constraints are satisfied (effectiveness), or the relevance of additional considerations to practitioners’ existing constraints (coherence). It was also noted where a paper did not provide any framing of practitioners’ concerns, i.e. where there was no mention at all of how the consideration under discussion features in practitioners’ assessment decisions.

In asserting that a particular paper frames a constraint in a certain way, it does not follow that the paper is necessarily focused substantially on addressing practitioners’ concerns. It indicates instead that, where the topic of the paper is discussed (particularly in the introduction and conclusion of papers), some claim is made about why the topic is of relevance to practitioners: because it corrects erroneous practitioner beliefs, because it helps practitioners achieve their existing goals, or because it helps them adjust their priorities to accommodate additional important considerations. In particular, where a concern is simply acknowledged as a consideration for practitioners, with no further discussion, that has been coded as addressing the effectiveness with which the assessment approach/es under discussion satisfy existing constraints.

4. Results

provides a summary of the results of this review. Some papers feature more than once in the table, because they address more than one of the constraints, or because they provide more than one framing.

Table 1. Overview of reviewed papers.

4.1. Learning

The most common constraint considered in recent research on exams is – perhaps predictably – that of ‘learning’, with 17 of the 43 papers reviewed focusing on how well assessments support learning. These are split roughly equally between those that focus on student performance, and those that focus on the impact of assessment on studying. Those who discuss student performance as a constraint on assessment method tend to focus on the difference in student performance on traditional exams as compared to alternatives they advocate, such as ‘bring-your-own-device’ computer-based exams (Nardi and Ranieri Citation2019), two-stage (collaborative) exams (Efu Citation2019) and take-home exams (Spiegel and Nivette Citation2023). Empirical studies use different measures to evaluate student performance, such as average grade (Malone et al. Citation2021; Nardi and Ranieri Citation2019), distribution of marks (Linden and Gonzalez Citation2021), and self-reported grades and performance on a follow-up knowledge test (Spiegel and Nivette Citation2023).

Those who discuss the relationship between exams and studying, rather than performance, address a venerable topic. For many years, exams have been criticised as unevenly distributing student effort, leading to last-minute cramming before the exam period (e.g. Richardson Citation2015); as demotivating students who are interested in more than just passing the exam (e.g. Havnes Citation2004); and as encouraging surface learning and discouraging deep learning (e.g. Scouller Citation1998). In the papers reviewed for this study, there are still concerns about the impact of exams on the depth of student learning (Ardolino et al. Citation2016; Malone et al. Citation2021; Myyry and Joutsenvirta Citation2016; Spiegel and Nivette Citation2023; Villarroel et al. Citation2020), and students’ motivation and engagement more broadly (Guo and Shi Citation2016; Harland and Wald Citation2021; Lindsley et al. Citation2016; Tam Citation2022).

Of the 17 papers that focused on the idea that assessment should support student learning (whether in the activity or achievement sense) nine papers discuss – to different levels of detail – the relationship between learning and practitioners’ constraints. Six of those nine papers present the improvement of learning simply as an existing priority for practitioners. This corresponds to the effectiveness criterion. This can be very brief; for example Levant, Zückert, and Paolo (Citation2018) open their paper with the sentence ‘Question response feedback is widely used in formative and summative testing to facilitate learning, especially in computer-based systems’ (996). This is a very brief indication that they believe that the impact of assessment on learning is a consideration for practitioners. Similarly, in their opening sentence Guo and Shi (Citation2016) state that ‘[t]he quality of student learning and development has been widely accepted as the core of educational quality in the last three decades’ (642).

Three papers frame the consideration in terms of weaknesses in practitioners’ understanding of effective assessment practices – the accuracy criterion. In encouraging practitioners to move away from exams and towards assessments more aligned with the ‘Assessment for Learning’ movement, Chong and McArthur (Citation2023) do partly frame their advice in terms of supporting practitioners to achieve their existing goals: practitioners in both western higher education and Confucian-influenced cultures value assessment for learning (though perhaps not highly enough, in their view). However, they also suggest that practitioners have faulty beliefs about assessment, by ‘not having sufficient knowledge of the varied forms of assessment’ (11). Similarly, Harland and Wald (Citation2021) acknowledge (perhaps sarcastically) that ‘assur[ing] the desired patterns of student behaviour … seems to be a requirement for lecturers’ (114), but are concerned that practitioners possess flawed beliefs about assessment: ‘The new order of graded internal assessment (which is hardly “new” anymore) seems to have been uncritically accepted by most academics’ (115).

A narrow minority of the papers discussing the impact of exams on learning (eight papers) include no discussion of the relevance of student learning to practitioners’ own concerns. For example, Ardolino et al. (Citation2016) explore the impact on student performance of ‘integrated’ exams, covering multiple modules. No discussion is included of how student performance does, or should, feature in practitioners’ decisions about assessment methods. Tam (Citation2022) studied students’ perceptions of the impact of take-home examinations to their study practices, but does not discuss the role of study practices in practitioners’ assessment decisions.

It is perhaps not surprising that some authors omit any discussion of the importance of learning to practitioners’ assessment decisions. It is possible that they simply take it as read that whether or not a change in assessment format would impact on students’ performance, or study practices, is a factor constraining practitioners. In which case, the absence of discussion of the relevance of the constraint to practitioners represents an implicit appeal to the effectiveness constraint: learning is so obviously an existing constraint of practitioners that discussion would be superfluous.

In line with the brief or absent explicit engagement with practitioners’ concerns, few papers offer any evidence about how practitioners perceive the importance of learning as a constraint on assessment methods. Of the 17 papers addressing the relationship between examinations and learning, only four purport to provide references to existing literature on how practitioners make assessment decisions. Chong and McArthur (Citation2023) reference empirical studies about practitioners’ assessment literacy and use of formative assessment. Bay and Pacharn (Citation2017) reference literature about the impact of professional bodies on academics’ assessment choices. In other cases, however, the appearance of evidence is sometimes not matched by reality. The reference that Panadero et al. (Citation2019) provide to support the claim that ‘[t]here is a large literature on teachers’ conceptions about assessments and how they shape what they do in classrooms’ (381) relates to school teachers rather than university teachers. Similarly, while Nsor-Ambala (Citation2020) claims that ‘the literature has been fairly consistent about the views of faculty about exam types’ (33), no reference is provided.

4.2. Stress, anxiety and wellbeing

Another prominent theme in the recent literature on examinations is the impact of assessment on student wellbeing, particularly in terms of the traditional topic of exam anxiety. Nine of the 43 papers reviewed included a focus on this topic. For example, Bulgan et al. (Citation2017) explore levels of anxiety among students at a Turkish university during a set of final examinations, while Spiegel and Nivette (Citation2023) compare the impact on student wellbeing of in-class closed-book exams and take-home open-book exams. As with the impact of exams on learning, a substantial portion of the papers (five of nine) do not include any explicit discussion of how wellbeing features in practitioner assessment choices. For example, in their investigation of students’ emotions in computer-based examination, Harley et al. (Citation2021) do explicitly discuss wellbeing as a focus of assessment researchers (‘test anxiety is one of the most widely investigated areas of academic achievement emotion research’ (956)) but fail to discuss its importance to practitioners. In the case of Monrad et al. (Citation2021) the situation is mixed. The first sentence of the abstract begins: ‘As educators seek to improve medical student well-being … ’ (872). However, that explicit claim about the assessment choices of practitioners is not repeated in the main body of the text (which refers only obliquely to practitioner priorities, by claiming that there is an ‘increasing imperative’ (872) to address negative impacts on student wellbeing).

Of the four papers that do discuss practitioner concerns, three seem to suggest that the importance of wellbeing in the context of exams is important because of its impact on a further consideration, student performance. For example, Lynam and Cachia (Citation2018) state that, because the emotions of assessment have an impact on learning, ‘it is important that academics are aware of the elements of an assessment that are associated with negative or positive emotions and aim to adjust these accordingly’ (232). In framing wellbeing as an important consideration because of its impact on an existing priority, these papers seem to appeal to the coherence of exams as a solution to practitioners’ constraints. One final paper (Nsor-Ambala Citation2020) frames wellbeing as an existing practitioner constraint: ‘this detrimental effect of pre-exam anxiety on student scores is of concern for educators’ (39). This is an appeal to effectiveness.

Only one of the nine papers discussing wellbeing, Nsor-Ambala (Citation2020), purports to provide evidence for its role in practitioner assessment choices. The paper includes a reference for the claim that ‘[t]here is growing interest among educators to choose assessment methods that enhance “active learning” as active learning … reduces pre-exam anxiety’ (35) (though a reference from 1994 perhaps provides limited evidence for a ‘growing interest’).

4.3. Assessment security and academic integrity

Eight of the papers reviewed discussed some aspect of academic misconduct in relation to assessment. For example, Forkuor et al. (Citation2019) explore students’ perceptions of acceptable behaviour in exams, while van Halem, van Klaveren, and Cornelisz (Citation2021) discuss barriers faced by students in the use of virtual proctoring of online exams. Of these eight papers, all but one comment on the relevance of the topic to practitioners. Five of the papers frame student academic misconduct in assessments as an issue of effectiveness, i.e. as an existing concern of practitioners. For example, Manoharan (Citation2019) states that ‘almost all academic institutions [have] in place compulsory tuition on academic honesty’ (140), and Denny et al. (Citation2019) justify focusing on how exams can be made less vulnerable to misconduct on the grounds that ‘[c]heating on tests and exams is viewed, by faculty and students alike, as a particularly serious act’ (463). Only one of those five papers offers evidence that misconduct in assessments is an existing practitioner concern; Denny et al. (Citation2019) cite a study that found that practitioners viewed ‘cheating in tests and exams’ (463) as a particularly severe act of academic dishonesty.

Three of the papers seem to frame misconduct in assessments as something that practitioners should care about because of its impact on other considerations that they do prioritise, i.e. in terms of the coherence of assessments. All three (Baijnath and Singh Citation2019; Forkuor et al. Citation2019; Kapardis and Spanoudis Citation2022) cite academic misconduct in assessments as a risk for institutional reputation. Forkuor et al. (Citation2019) also highlight the socio-economic consequences (in terms of the quality of graduates and national economic productivity), and Baijnath and Singh (Citation2019) similarly flag the potential risks to the ‘moral fabric of the entire society’ (5). Kapardis and Spanoudis (Citation2022) reference a study about practitioners’ perceptions of computer-based versus paper-based exams, but none of these three papers provide evidence that the wider impacts of cheating (on institutional reputation, for example) are existing concerns of practitioners.

4.4. Authenticity

A traditional criticism of exams is the lack of resemblance between a time-constrained, closed-book, handwritten, individual examination, and the kinds of tasks undertaken in contemporary professional workplaces. Going back several decades, Powell and Butterworth (Citation1971) disparaged the belief that ‘[e]mployers are entitled to know if a man [sic] is going to crack up under pressure’ (5) and that ‘[e]xaminations are thus a psychological and moral test as well as an intellectual one’ (10). Of the literature reviewed between 2016 and 2021, five papers addressed the authenticity of examinations. Linden and Gonzalez (Citation2021) and Malone et al. (Citation2021) observe in passing that alternatives (open-book exams and open-book open-web exams, respectively) provide tasks that more closely resemble the world of work. Neither of those papers comment on how that resemblance features in practitioners’ considerations. Myyry and Joutsenvirta (Citation2016), Surry, Torre, and Durning (Citation2017) and Villarroel et al. (Citation2020) discuss the issue in more depth. In the case of Myyry and Joutsenvirta (Citation2016) and Villarroel et al. (Citation2020), authenticity is framed in terms of its positive impact on an additional practitioner constraint (coherence), which in both cases is the benefits of assessment to learning: ‘[Authenticity] has an impact on the quality and depth of learning achieved by the student and the development of higher-order cognitive skills’ (Villarroel et al. Citation2020, 39). For Surry et al. (Citation2017), the issue is one of the effectiveness of exams (specifically clinical-vignette multiple choice question exams) as a solution to the existing need to assess ‘real-world’ clinical reasoning. Only one of the five papers cites any literature that provides evidence about practitioner constraints: Linden and Gonzalez (Citation2021) reference papers exploring staff perceptions of online examinations.

4.5. Inclusivity

Finally, five of the papers reviewed explore the inclusivity of exams and their alternatives. For example, Faisal, Shinwari, and Hussain (Citation2017) analyse the differences in exam performance between male and female students at a medical college in Pakistan. Bygren (Citation2020) studied the impact of student gender and the ‘foreignness’ of student names on non-blind examination marking. All five of the papers comment on the relevance of inclusivity to practitioner constraints. Three of the papers frame inclusivity as an existing concern (effectiveness). Bygren (Citation2020) talks about ‘impartiality and tolerance’ (302) being particular priorities in a university setting, while Faisal, Shinwari, and Hussain (Citation2017) say that ‘[g]ender issues are currently a matter of high concern worldwide, especially among academics and policy formulators’ (204). Both Faisal, Shinwari, and Hussain (Citation2017) and Ellison and Jones (Citation2019) provide some evidence for the importance of inclusivity as a constraint on assessment choices, with the latter citing the policies of governments and the English sector regulator.

Two of the papers frame the concern instead in terms of the coherence of exams: they argue that the importance of diverse assessments lies not in the diversity itself, but in the impact of that diversity on widening participation as a further goal. Both Stegers-Jager, Brommet, and Themmen (Citation2016) and O’Neill and Padden (Citation2022) are concerned about the impact of assessments on the retention of non-traditional students. The latter paper also seems to question the accuracy of practitioner beliefs, by claiming there is a lack of awareness that ‘equitable assessment practices’ (2) are important for supporting diverse student populations. Neither paper provides evidence that practitioners prioritise the widening participation agenda, or – for O’Neill and Padden (Citation2022) – that practitioners have erroneous beliefs about the role of assessment in that agenda.

5. Discussion

Using Robinson’s ‘problem-based methodology’ as a guide to the kind of research that engages with the assessment challenges of practitioners, what lessons can we draw from the research on exams in higher education published between 2016 and 2021? The previous section featured research that matches all three elements of Robinson’s model. There is research that seems aimed at improving the accuracy of practitioner’s factual knowledge, such as Chong and McArthur (Citation2023) who are concerned about the lack of understanding of the varied forms that assessment can take. There is research that attempts to help practitioners improve the effectiveness with which assessments satisfy their existing priorities, such as Manoharan (Citation2019) who offers advice on how practitioners can adapt multiple-choice examinations to better meet their existing desire to reduce student academic misconduct. And there is research that seems aimed at supporting the coherence of practitioners’ assessment choices, by encouraging them to take into account wider considerations; an example of this is Villarroel et al. (Citation2020) who argue for the value of authentic assessment because of its’ positive impact on student learning. While all three approaches are represented in the literature reviewed, they are not represented equally. The dominant strategy is to address the effectiveness of assessment solutions. 17 of the 43 papers reviewed address how well traditional exams, or their alternatives, satisfy the existing constraints of practitioners.

An equal number of papers (17), featured no explicit discussion of how the research relates to practitioners’ assessment decisions. This was common for papers addressing all the themes, except for assessment security, where seven of the eight papers included some information about its relevance to practitioners. In the papers lacking that discussion, a plausible interpretation is that researchers assume that the relevance to practitioners of the strengths and weaknesses they discuss is simply obvious. Of the 43 papers reviewed, six explored the impact of assessment on students’ approaches to their studies, yet failed to explain its relevance to practitioners. It may seem an obvious consideration to some (particularly those professionally involved in the study of assessment) but given the available evidence (e.g. MacLellan Citation2001) the importance of ‘assessment for learning’ to practitioners can’t simply be assumed.

A further issue is that even where researchers do explicitly discuss the relationship between their work and practitioners’ concerns it is rare for relevant evidence about practitioners’ concerns to then be cited; only nine of the 43 papers reviewed do so. That’s not to say that researchers don’t provide evidence for the importance of the issues they discuss: they do, as part of providing a rationale for the research. However what is usually missing is evidence that the issue is of importance to practitioners, rather than other researchers. This is perhaps not surprising, given the limited amount of research about how practitioners make assessment decisions in general; and the almost complete lack of research about how practitioners make decisions about using exams in particular. If we don’t know why practitioners choose to use exams in particular situations, it is hard for those conducting research on exams to provide evidence about how their research engages with those choices. Robinson herself is clear that the kind of practice-focused model she advocates depends on robust information about how practitioners frame their problems and make their choices: ‘the focus of data collection and analysis within problem based methodology is the discovery of the reasoning that sustains the practices under investigation’ (Robinson and Lai Citation1999, 199). More research about the decision-making processes that lead practitioners to use traditional exams, or to switch away to other kinds of assessments, is a prerequisite clearer links between research on exams and the practical concerns of front-line academic staff.

Another likely explanation for researchers failing to provide evidence for the relevance of their considerations to those of practitioners is the dual role of the researchers themselves. The vast majority of those undertaking research on exams will also be practitioners, in the sense that they make decisions about assessment methods in the course of their own teaching. They are likely to be academics in other subjects, educational developers, or other professionals with a teaching load. For assessment researchers rooted in the day-to-day reality of assessment choices, it may seem superfluous to justify how their research relates to that reality. While understandable, this is nevertheless problematic.

6. Conclusion

This article has discussed the research-practice gap in assessment and feedback in higher education. It has focused on the literature about examinations published between 2016 and 2021, and explored how the approaches adopted by researchers may influence the impact of that research. Using Viviane Robinson’s ‘problem-based methodology’ as a guide, it has sought to extract some lessons about how exam research could engage better with the challenges of those practitioners tasked with making decisions about assessments.

An important limitation of this review is that a reasonable proportion of research is not aimed at directly influencing practitioners’ decision-making at all, for which Robinson’s model is not applicable. However, a significant amount of research does have that goal, and as described earlier in the paper there is a clear desire from within HE research to promote that goal further. Another limitation of the review is the narrowness of the literature search. The systematic element of the search was limited to 28 key journals. In addition, the workload challenges created by the dual-meaning of the term ‘examination’ meant that the search for that term was restricted to the papers’ titles, rather than also encompassing the abstracts and keywords as would normally be the case.

Using Robinson’s problem-based methodology as a guide, this article has highlighted some important features of the research on exams. Many of the papers reviewed do not explicitly address how the issue under consideration relates to practitioners’ priorities, and those that do rarely provide any supporting evidence about those priorities. The general sense is that the priorities of the practitioners who ultimately make the key assessment choices are simply not an important part of the picture. There are of course exceptions, and this review has found numerous examples of papers that both explicitly engage with practitioners’ concerns, and provide evidence about those concerns. But it has also found that a substantial minority does not do the former, and a large majority does not do the latter.

This study connects importantly with contemporary ways of thinking about assessment. The approach taken in this paper – applying Robinson’s problem-based methodology to research on assessment in higher education – focuses on how practitioners balance constraints to arrive at assessment decisions. This approach intentionally centres practitioners’ rational decision-making processes. However, this does not imply that only rational processes are involved in those decisions, or that the only factors influencing assessment are practitioners’ decisions. What it does assume is that practitioners’ assessment decisions have an important role in determining assessment practices, and that it is important to engage directly with those decisions. This is common ground with the ‘practice perspective’ of Boud et al. (Citation2018). The focus on practitioners’ decisions is not only compatible with, but complementary to contemporary approaches to assessment research that adopt broader perspectives that may be, for example, socio-material (Tai et al. Citation2023), relational (Gravett, Taylor, and Fairchild Citation2021) or socio-political (Nieminen Citation2022). If important factors relating to social structures, materiality, relationships etc. are revealed, there is value in encouraging practitioners to bring them into consideration in their assessment decisions. Robinson’s model offers a recipe for doing that: such research could indicate that practitioners need to review their assessment decisions because they have erroneous beliefs, because their assessment choices don’t achieve what they want them to, or because they are neglecting additional factors that bear on the goals they’re aiming for.

The findings reported in this paper indicate that, in recent years and in relation to examinations, assessment research has engaged with practitioners’ decisions only intermittently and superficially. For any single piece of research, that is not a flaw: there are many other aims to research than directly exploring or influencing practitioners’ decisions. However, as a judgement about the wider body of literature it is more concerning. And in the context of widespread concerns about a disconnect between the insights of those researching assessment and the assessment practices to be found in universities, it is a fundamental problem. Highlighting how researchers might better engage with practitioners’ decisions is not to lay the blame at researchers’ door for any negative effects of the ongoing use of exams. It is, however, to take seriously the concerns – expressed by researchers – about the gap between research and practice; and to explore ways in which we can improve the situation by bringing research a little closer to the reality of assessment practices.

A clear contributing factor is that more needs to be known about how practitioners see their assessment challenges. As other researchers have highlighted, there is little evidence about how assessment decisions are made; and there is virtually none on the decision-making processes that lead to the use of exams. If research is going to engage with how practitioners attempt to balance constraints, then we need to know what those constraints are and how they are weighted by practitioners. That is not to say that research shouldn’t aim to critique and challenge the way that practitioners frame the problem of assessment design – after all, shifting that framing may be a big part of how assessment can be improved – but there are benefits in motivating that challenge in terms of practitioners’ own wider concerns.

This article has focused on research on exams, but what goes for exams goes for other issues in assessment and feedback. If we want research to influence the decision-making processes of practitioners (and acknowledging that other approaches are legitimate) lamenting the lack of innovation and improvement in assessment while failing to engage with the thought-processes that lead to decisions about assessment is unlikely to succeed. Offering better solutions to the problems that practitioners face is a more promising path to the change we would all like to see.

Disclosure statement

No potential conflict of interest was reported by the author(s).

References

  • Adachi, C., J. Tai, and Phillip Dawson. 2018. “Academics’ Perceptions of the Benefits and Challenges of Self and Peer Assessment in Higher Education.” Assessment & Evaluation in Higher Education 43 (2): 294–306. https://doi.org/10.1080/02602938.2017.1339775.
  • Al-Hakeem, M., and M. Abdulrahman. 2017. “Developing a New e-Exam Platform to Enhance the University Academic Examinations. The Case of Lebanese French University.” International Journal or Modern Education and Computer Science 5 (5): 9–16. https://doi.org/10.5815/ijmecs.2017.05.02.
  • Ardolino, P., S. Noventa, M. Formicuzzi, S. Cubico, and G. Favretto. 2016. “Multiple Integrated Examinations: An Observational Study of Different Academic Curricula Based on a Business Administration Assessment.” Higher Education 71 (1): 59–79. https://doi.org/10.1007/s10734-015-9888-4.
  • Argyris, C., and D. A. Schon. 1974. Theory in Practice: Increasing Professional Effectiveness. San Francisco: Jossey-Bass.
  • Baijnath, N., and D. Singh. 2019. “Examination Cheating: Risks to the Quality and Integrity of Higher Education.” South African Journal of Sciences 115 (11/12): 1–6.
  • Barr, R., and J. Tagg. 1995. “From Teaching to Learning: A New Paradigm for Undergraduate Education.” Change 27 (6): 13–26. https://doi.org/10.1080/00091383.1995.10544672.
  • Bay, D., and P. Pacharn. 2017. “Impact of Group Exams in a Graduate Intermediate Accounting Class.” Accounting Education 26 (4): 316–34. https://doi.org/10.1080/09639284.2017.1292465.
  • Bearman, M., P. Dawson, S. Bennett, M. Hall, E. Molloy, D. Boud, and G. Joughin. 2017. “How University Teachers Design Assessments: A Cross-Disciplinary Study.” Higher Education 74 (1): 49–64. https://doi.org/10.1007/s10734-016-0027-7.
  • Bearman, M., P. Dawson, D. Boud, S. Bennett, M. Hall, and E. Molloy. 2016. “Support for Assessment Practice: Developing the Assessment Design Decisions Framework.” Teaching in Higher Education 21 (5): 545–56. https://doi.org/10.1080/13562517.2016.1160217.
  • Boud, D. 1995. “Assessment and Learning: Contradictory or Complementary.” In Assessment for Learning in Higher Education, edited by P. Knight, 35–48. London: Kogan Page.
  • Boud, D., P. Dawson, M. Bearman, S. Bennett, G. Joughin, and E. Molloy. 2018. “Reframing Assessment Research: Through a Practice Perspective.” Studies in Higher Education 43 (7): 1107–18. https://doi.org/10.1080/03075079.2016.1202913.
  • Brady, M., A. Devitt, and R. Kiersey. 2019. “Academic Staff Perspectives on Technology for Assessment (TfA) in Higher Education: A Systematic Literature Review.” British Journal of Educational Technology 50 (6): 3080–98. https://doi.org/10.1111/bjet.12742.
  • Buckley, A. 2021a. “Crisis? What Crisis? Interpreting Student Feedback on Assessment.” Assessment & Evaluation in Higher Education 46 (7): 1008–19. https://doi.org/10.1080/02602938.2020.1846015.
  • Buckley, A. 2021b. “The Tyranny of ‘Teaching and Learning’.” Studies in Philosophy and Education 40 (4): 415–27. https://doi.org/10.1007/s11217-021-09769-6.
  • Bulgan, C., H. Odabas, S. Dal, E. Mecu, and M. Aydin. 2017. “Assessment of State and Trait Anxiety Level of Physical Education and Sports’ Students Before the Term Final Examinations: Example of Halic University.” Universal Journal of Educational Research 5 (6): 932–8. https://doi.org/10.13189/ujer.2017.050604.
  • Bygren, M. 2020. “Biased Grades? Changes in Grading After a Blinding of Examinations Reform.” Assessment & Evaluation in Higher Education 45 (2): 292–303. https://doi.org/10.1080/02602938.2019.1638885.
  • Chan, C., and L. Luk. 2022. “A Four-Dimensional Framework for Teacher Assessment Literacy in Holistic Competencies.” Assessment & Evaluation in Higher Education 47 (5): 755–69. https://doi.org/10.1080/02602938.2021.1962806.
  • Chirumamilla, A., G. Sindre, and A. Nguyen-Duc. 2020. “Cheating in e-Exams and Paper Exams: The Perceptions of Engineering Students and Teachers in Norway.” Assessment & Evaluation in Higher Education 45 (7): 940–57. https://doi.org/10.1080/02602938.2020.1719975.
  • Chong, D. Y. K., and J. McArthur. 2023. “Assessment for learning in a Confucian-influenced culture: Beyond the Summative/Formative Binary.” Teaching in Higher Education 28 (6): 1395–1411. https://doi.org/10.1080/13562517.2021.1892057.
  • Colbeck, C. 2000. “Reshaping the Forces That Perpetuate the Research-Practice Gap: Focus on New Faculty.” New Directions for Higher Education 2000 (110): 35–47. https://doi.org/10.1002/he.11003.
  • Cotton, D. R. E., W. Miller, and P. Kneale. 2018. “The Cinderella of Academia: Is Higher Education Pedagogic Research Undervalued in UK Research Assessment?” Studies in Higher Education 43 (9): 1625–36. https://doi.org/10.1080/03075079.2016.1276549.
  • Cox, R. 1967. “Resistance to Change in Examining.” Universities Quarterly 21:352–8.
  • Cross, K. P. 2000. “The Educational Role of Researchers.” New Directions for Higher Education 2000 (110): 63–73. https://doi.org/10.1002/he.11005.
  • Dawson, P., M. Bearman, D. Boud, M. Hall, E. Molloy, S. Bennett, and G. Joughin. 2013. “Assessment Might Dictate the Curriculum, but What Dictates Assessment?” Teaching & Learning Inquiry The ISSOTL Journal 1 (1): 107–11. https://doi.org/10.20343/teachlearninqu.1.1.107.
  • Dawson, P., M. Henderson, P. Mahoney, M. Phillips, T. Ryan, D. Boud, and E. Molloy. 2019. “What Makes for Effective Feedback: Staff and Student Perspectives.” Assessment & Evaluation in Higher Education 44 (1): 25–36. https://doi.org/10.1080/02602938.2018.1467877.
  • Denny, P., S. Manoharan, U. Speidel, G. Russello, and A. Chang. 2019. “On the Fairness of Multiple-Variant Multiple-Choice Examinations.” In Proceedings of the 50th ACM Technical Symposium on Computer Science Education, 462–468. https://doi.org/10.1145/3287324.3287357.
  • Dewey, J. 1929. The Sources of a Science of Education. New York: Liveright.
  • Efu, S. 2019. “Exams as Learning Tools: A Comparison of Traditional and Collaborative Assessment in Higher Education.” College Teaching 67 (1): 73–83. https://doi.org/10.1080/87567555.2018.1531282.
  • Ellison, L., and D. Jones. 2019. “First Year Law Students: The Impact of Assessment Type on Attainment.” Assessment & Evaluation in Higher Education 44 (2): 283–93. https://doi.org/10.1080/02602938.2018.1496398.
  • Elton, L. 2004. “A Challenge to Established Assessment Practice.” Higher Education Quarterly 58 (1): 43–62. https://doi.org/10.1111/j.1468-2273.2004.00259.x.
  • Evans, C., C. Kandiko Howson, A. Forsythe, and C. Edwards. 2021. “What Constitutes High Quality Higher Education Pedagogical Research?” Assessment & Evaluation in Higher Education 46 (4): 525–46. https://doi.org/10.1080/02602938.2020.1790500.
  • Faisal, R., L. Shinwari, and S. S. Hussain. 2017. “Academic Performance of Male in Comparison with Female Undergraduate Medical Students in Pharmacology Examinations.” Journal of Pakistan Medical Association 67 (2): 204–8.
  • Forkuor, J. B., J. Amartei, D. O. Attoh, and M. A. Buari. 2019. “Students’ Perception of Cheating and the Best Time to Cheat During Examinations.” The Urban Review 51 (3): 424–43. https://doi.org/10.1007/s11256-018-0491-8.
  • Gravett, K., C. Taylor, and N. Fairchild. 2021. “Pedagogies of Mattering: Re-conceptualising Relational Pedagogies in Higher Education.” Teaching in Higher Education. Advance online publication.
  • Guo, F., and J. Shi. 2016. “The Relationship Between Classroom Assessment and Undergraduates’ Learning Within Chinese Higher Education System.” Studies in Higher Education 41 (4): 642–63. https://doi.org/10.1080/03075079.2014.942274.
  • Haig, B. 2014. “Reflections on Problem-Based Methodology.” In A Companion to Research in Education, edited by A. D. Reid, 353–6. London: Springer.
  • Harland, T., and N. Wald. 2021. “The Assessment Arms Race and the Evolution of a University’s Assessment Practices.” Assessment & Evaluation in Higher Education 46 (1): 105–17. https://doi.org/10.1080/02602938.2020.1745753.
  • Harley, J. M., N. M. Lou, Y. Liu, M. Cutumisu, L. M. Daniels, J. P. Leighton, and L. Nadon. 2021. “University Students’ Negative Emotions in a Computer-Based Examination: The Roles of Trait Test-Emotion, Prior Test-Taking Methods and Gender.” Assessment & Evaluation in Higher Education 46 (6): 956–72. https://doi.org/10.1080/02602938.2020.1836123.
  • Havnes, A. 2004. “Examination and Learning: An Activity-Theoretical Analysis of the Relationship Between Assessment and Educational Practice.” Assessment & Evaluation in Higher Education 29 (2): 159–76. https://doi.org/10.1080/0260293042000188456.
  • Hemsley-Brown, J., and C. Sharp. 2003. “The Use of Research to Improve Professional Practice: A Systematic Review of the Literature.” Oxford Review of Education 29 (4): 449–471. https://doi.org/10.1080/0305498032000153025.
  • Henderson, M., J. Chung, R. Awdry, M. Mundy, M. Bryant, C. Ashford, and K. Ryan. 2023. “Factors Associated with Online Examination Cheating.” Assessment & Evaluation in Higher Education 48 (7): 980–994. https://doi.org/10.1080/02602938.2022.2144802.
  • Hirschkorn, M., and D. Geelan. 2008. “Bridging the Research-Practice Gap: Research Translation and/or Research Transformation.” The Alberta Journal of Educational Research 54 (1): 1–13.
  • Joughin, G., P. Dawson, and D. Boud. 2017. “Improving Assessment Tasks Through Addressing Our Unconscious Limits to Change.” Assessment & Evaluation in Higher Education 42 (8): 1221–32. https://doi.org/10.1080/02602938.2016.1257689.
  • Kapardis, M., and G. Spanoudis. 2022. “Lessons Learned During Covid-19 Concerning Cheating in e-Examinations by University Students.” Journal of Financial Crime 29 (2): 506–518. https://doi.org/10.1108/JFC-05-2021-0105.
  • Kezar, A. 2000. “Understanding the Research-to-Practice gap: A National Study of Researchers’ and Practitioners’ Perspectives.” New Directions for Higher Education 2000 (110): 9–19. https://doi.org/10.1002/he.11001.
  • Kilgour, J. M., and S. Tayyaba. 2016. “An Investigation Into the Optimal Number of Distractors in Single-Best Answer Exams.” Advances in Health Sciences Education 21 (3): 571–85. https://doi.org/10.1007/s10459-015-9652-7.
  • Levant, B., W. Zückert, and A. Paolo. 2018. “Post-exam Feedback with Question Rationales Improves Re-test Performance of Medical Students on a Multiple-Choice Exam.” Advances in Health Sciences Education 23 (5): 995–1003. https://doi.org/10.1007/s10459-018-9844-z.
  • Levy, D., T. Svoronos, and M. Klinger. 2023. “Two-stage Examinations: Can Examinations be More Formative Experiences?” Active Learning in Higher Education 24 (2): 79–94. https://doi.org/10.1177/1469787418801668.
  • Linden, K., and P. Gonzalez. 2021. “Zoom Invigilated Exams: A Protocol for Rapid Adoption to Remote Examinations.” British Journal of Educational Technology 52 (4): 1323–37. https://doi.org/10.1111/bjet.13109.
  • Lindsley, J. E., D. A. Morton, K. Pippitt, S. Lamb, and J. M. Colbert-Getz. 2016. “The Two-Stage Examination: A Method to Assess Individual Competence and Collaborative Problem Solving in Medical Students.” Academic Medicine 91 (10): 1384–7. https://doi.org/10.1097/ACM.0000000000001185.
  • Lynam, S., and M. Cachia. 2018. “Students’ Perceptions of the Role of Assessments at Higher Education.” Assessment & Evaluation in Higher Education 43 (2): 223–34. https://doi.org/10.1080/02602938.2017.1329928.
  • MacLellan, E. 2001. “Assessment for Learning: The Differing Perceptions of Tutors and Students.” Assessment & Evaluation in Higher Education 26 (4): 307–18. https://doi.org/10.1080/02602930120063466.
  • Malone, D., S. Chuang, E. Yuriev, and J. Short. 2021. “Effect of Changing from Closed-Book to Formulary-Allowed Examinations.” American Journal of Pharmaceutical Education 85 (1): 34–42. https://doi.org/10.5688/ajpe7990.
  • Manoharan, S. 2019. “Cheat-resistant Multiple-Choice Examinations Using Personalization.” Computers and Education 130:139–51. https://doi.org/10.1016/j.compedu.2018.11.007.
  • McLean, H. 2018. “This is the way to Teach: Insights from Academics and Students About Assessment That Supports Learning.” Assessment & Evaluation in Higher Education 43 (8): 1228–40. https://doi.org/10.1080/02602938.2018.1446508.
  • McNaughton, S. 2014. “The Research-Practice Gap and How to Fill it.” In A Companion to Research in Education, edited by A. D. Reid, 357–60. London: Springer.
  • Monrad, S. U., M. Wolff, J. Kurtz, N. M. Deiorio, R. Sabo, J. K. Stringer, and S. A. Santen. 2021. “What is the Association Between Student Well-Being and High-Stakes Examination Scores?” Medical Education 55 (7): 872–7. https://doi.org/10.1111/medu.14460.
  • Myyry, L., and T. Joutsenvirta. 2016. “Open-book, Open-web Online Examinations: Developing Examination Practices to Support University Students’ Learning and Self-Efficacy.” Active Learning in Higher Education 16 (2): 119–32. https://doi.org/10.1177/1469787415574053.
  • Nardi, A., and M. Ranieri. 2019. “Comparing Paper-Based and Electronic Multiple-Choice Examinations with Personal Devices: Impact on Students’ Performance, Self-Efficacy and Satisfaction.” British Journal of Educational Technology 50 (3): 1495–506. https://doi.org/10.1111/bjet.12644.
  • Nieminen, J. H. 2022. “Assessment for Inclusion: Rethinking Inclusive Assessment in Higher Education.” Teaching in Higher Education. Advance online publication.
  • Nsor-Ambala, R. 2020. “Impact of Exam Type on Exam Scores, Anxiety, and Knowledge Retention in a Cost and Management Accounting Course.” Accounting Education 29 (1): 32–56. https://doi.org/10.1080/09639284.2019.1683871.
  • O'Flaherty, J., and C. Phillips. 2015. “The use of Flipped Classrooms in Higher Education: A Scoping Review.” The Internet and Higher Education 25: 85–95. https://doi.org/10.1016/j.iheduc.2015.02.002.
  • O’Neill, G., and L. Padden. 2022. “Diversifying Assessment Methods: Barriers, Benefits and Enablers.” Innovations in Education and Teaching International 59 (4): 398–409. https://doi.org/10.1080/14703297.2021.1880462.
  • Panadero, E., J. Fraile, J. Fernández Ruiz, D. Castilla-Estévez, and M. A. Ruiz. 2019. “Spanish University Assessment Practices: Examination Tradition with Diversity by Faculty.” Assessment and Evaluation in Higher Education 44 (3): 379–97. https://doi.org/10.1080/02602938.2018.1512553.
  • Powell, A., and B. Butterworth. 1971. Marked for Life: A Criticism of Assessment at Universities. London: William Cross.
  • Richardson, J. T. E. 2015. “Coursework Versus Examinations in End-of-Module Assessment: A Literature Review.” Assessment & Evaluation in Higher Education 40 (3): 439–55. https://doi.org/10.1080/02602938.2014.919628.
  • Robinson, V. M. J. 1992. “Why Doesn't Educational Research Solve Educational Problems?” Educational Philosophy and Theory 24 (2): 8–28. https://doi.org/10.1111/j.1469-5812.1992.tb00223.x.
  • Robinson, V. M. J. 1993. Problem-Based Methodology: Research for the Improvement of Practice. Oxford: Pergamon Press.
  • Robinson, V. M. J. 1998. “Methodology and the Research-Practice Gap.” Educational Researcher 27 (1): 17–26. https://doi.org/10.2307/1176923.
  • Robinson, V. M. J. 2014. “Reducing the Research-Practice Gap Through Problem-Based Methodology.” In A Companion to Research in Education, edited by A. D. Reid, 341–52. London: Springer.
  • Robinson, V. M. J., and M. K. Lai. 1999. “The Explanation of Practice: Why Chinese Students Copy Assignments.” International Journal of Qualitative Studies in Education 12 (2): 193–210. https://doi.org/10.1080/095183999236259.
  • Samuelowicz, K., and J. Bain. 2002. “Identifying Academics’ Orientations to Assessment Practice.” Higher Education 43 (2): 173–201. https://doi.org/10.1023/A:1013796916022.
  • Savin-Baden, M. 2000. Problem-Based Learning in Higher Education: Untold Stories. London: SRHE & Open University Press.
  • Scouller, K. 1998. “The Influence of Assessment Method on Students’ Learning Approaches: Multiple Choice Question Examination Versus Assignment Essay.” Higher Education 35 (4): 453–72. https://doi.org/10.1023/A:1003196224280.
  • Spiegel, T., and A. Nivette. 2023. “The Relative Impact of in-Class Closed-Book Versus Take-Home Open-Book Examination Type on Academic Performance, Student Knowledge Retention and Wellbeing.” Assessment & Evaluation in Higher Education 48 (1): 27–40. https://doi.org/10.1080/02602938.2021.201660.
  • Stegers-Jager, K. M., F. N. Brommet, and A. P. N. Themmen. 2016. “Ethnic and Social Disparities in Different Types of Examinations in Undergraduate Pre-clinical Training.” Advances in Health Sciences Education 21 (5): 1023–46. https://doi.org/10.1007/s10459-016-9676-7.
  • Surry, L. T., D. Torre, and S. J. Durning. 2017. “Exploring Examinee Behaviours as Validity Evidence for Multiple-Choice Question Examinations.” Medical Education 51 (10): 1075–85. https://doi.org/10.1111/medu.13367.
  • Tai, J., M. Bearman, G. Gravett, and E. Molloy. 2023. “Exploring the Notion of Teacher Feedback Literacies Through the Theory of Practice Architectures.” Assessment & Evaluation in Higher Education 48 (2): 201–13. https://doi.org/10.1080/02602938.2021.1948967.
  • Tam, A. C. F. 2022. “Students’ Perceptions of and Learning Practices in Online Timed Take-Home Examinations During Covid-19.” Assessment & Evaluation in Higher Education 47 (3): 477–492. https://doi.org/10.1080/02602938.2021.1928599.
  • Tight, M. 2018. “Higher Education Journals: Their Characteristics and Contribution.” Higher Education Research & Development 37 (3): 607–19. https://doi.org/10.1080/07294360.2017.1389858.
  • Turner, J., and G. Briggs. 2018. “To see or not to see? Comparing the Effectiveness of Examinations and End of Module Assessments in Online Distance Learning.” Assessment & Evaluation in Higher Education 43 (7): 1048–60. https://doi.org/10.1080/02602938.2018.1428730.
  • Vanderline, R., and J. van Braak. 2010. “The gap Between Educational Research and Practice: Views of Teachers, School Leaders, Intermediaries and Researchers.” British Educational Research Journal 36 (2): 299–316. https://doi.org/10.1080/01411920902919257.
  • van Halem, N., C. van Klaveren, and I. Cornelisz. 2021. “The Effects of Implementation Barriers in Virtually Proctored Examination: A Randomised Field Experiment in Dutch Higher Education.” Higher Education Quarterly 75 (2): 333–47. https://doi.org/10.1111/hequ.12275.
  • Villarroel, V., D. Boud, S. Bloxham, D. Bruna, and Carola Bruna. 2020. “Using Principles of Authentic Assessment to Redesign Written Examinations and Tests.” Innovations in Education and Teaching International 57 (1): 38–49.
  • Winstone, N., and D. Carless. 2020. Designing Effective Feedback Processes in Higher Education. Abingdon: Routledge.
  • Zhan, Y. 2020. “Motivated or Informed? Chinese Undergraduates’ Beliefs About the Functions of Continuous Assessment in Their College English Course.” Higher Education Research & Development 39 (5): 1055–69. https://doi.org/10.1080/07294360.2019.1699029.