1,482
Views
3
CrossRef citations to date
0
Altmetric
Research Articles

Re-examining single-moment-in-time high-stakes examinations in specialist training: A critical narrative review

ORCID Icon & ORCID Icon

Abstract

In this critical narrative review, we challenge the belief that single-moment-in-time high-stakes examinations (SMITHSEx) are an essential component of contemporary specialist training. We explore the arguments both for and against SMITHSEx, examine potential alternatives, and discuss the barriers to change.

SMITHSEx are viewed as the “gold standard” assessment of competence but focus excessively on knowledge assessment rather than capturing essential competencies required for safe and competent workplace performance. Contrary to popular belief, regulatory bodies do not mandate SMITHSEx in specialist training. Though acting as significant drivers of learning and professional identity formation, these attributes are not exclusive to SMITHSEx.

Skills such as crisis management, procedural skills, professionalism, communication, collaboration, lifelong learning, reflection on practice, and judgement are often overlooked by SMITHSEx. Their inherent design raises questions about the validity and objectivity of SMITHSEx as a measure of workplace competence. They have a detrimental impact on trainee well-being, contributing to burnout and differential attainment.

Alternatives to SMITHSEx include continuous low-stakes assessments throughout training, ongoing evaluation of competence in the workplace, and competency-based medical education (CBME) concepts. These aim to provide a more comprehensive and context-specific assessment of trainees’ competence while also improving trainee welfare.

Specialist training colleges should evolve from exam providers to holistic education sources. Assessments should emphasise essential practical knowledge over trivia, align with clinical practice, aid learning, and be part of a diverse toolkit. Eliminating SMITHSEx from specialist training will foster a competency-based approach, benefiting future medical professionals’ well-being and success.

Introduction

Single-moment-in-time high-stakes examinations (SMITHSEx) are often considered the gold standard for assessing competence in specialist training. “Single-moment-in-time” refers to how these are conducted on a specific date or within a short period, often at the end of a defined phase of training, for assessing competence up to that point. The high-stakes nature describes how assessment outcomes carry substantial consequences to the individuals involved, with limited retake options in the proximal period. Written and oral formats are primarily used to measure knowledge while objective structured clinical examinations (OSCEs) have been used to assess behavioural competencies such as professionalism and communication (Jefferies et al. Citation2007; Dwyer et al. Citation2014), in addition to knowledge and skills. Where specialist training is administered by colleges, their raison d’être is to make a categorical decision on progression from one stage of training to the next. Although not the sole determining factor, SMITHSEx are often seen as the most critical element. Progression decisions are unavoidable and should be based on reliable and defensible information (Cizek Citation2005). Despite the incorporation of workplace-based assessments (WBAs), minimum case volumes, time-based clinical experience, and other mandatory activities, it is SMITHSEx that ultimately governs progression in training. In the USA, specialty boards employ SMITHSEx to enable specialist certification. Institutions favour their use as they promise to uphold a universally-accepted standard, offer efficiency in decision-making, are highly visible, and provide economies of scale (Heubert and Hauser Citation1999). However, from the learner’s perspective, SMITHSEx bear little relevance to clinical practice (Craig et al. Citation2010; Weller et al. Citation2014; Kahn et al. Citation2020) and are characterised by inflexibility (Craig et al. Citation2010; Kahn et al. Citation2020). A recent editorial argues that there is no evidence that SMITHSEx improve patient safety in specialist training but instead cause unnecessary stress, directing valuable resources towards the goal of passing an exam rather than preparing for independent practice (Thoma et al. Citation2022).

Practice points

  • Single-moment-in-time high-stakes examinations (SMITHSEx) are often considered the gold standard for assessing competence in specialist training, but they are criticised for being inflexible and irrelevant to clinical practice.

  • The perceived objectivity of SMITHSEx is not apparent when closely scrutinised. At best, it can claim a form of negotiated or shared subjectivity.

  • SMITHSEx focus on knowledge assessment rather than fostering safe and competent clinical practice, driving test preparation rather than genuine learning.

  • Alternatives to SMITHSEx, such as programmatic assessment implemented as part of a competency-based medical education (CBME) program, can provide a more comprehensive and contextual measure of competence, and improve trainee welfare.

  • Barriers to change include institutional inertia, a perceived threat to professional identity, and financial interests.

We conducted a critical narrative review to re-examine the role of SMITHSEx as an essential component of contemporary specialist training. Specifically, we discuss the commonly cited points both for and against SMITHSEx, explore potential alternatives, and examine the barriers to change. Our intention is not to advocate for the total elimination of written, oral, or clinical assessments from specialist training, but rather to highlight the flaws and unnecessary nature of their application as a single high-stakes event in the present day.

Methods

In keeping with the critical narrative review format (Grant and Booth Citation2009), the non-comprehensive search sought to identify the most significant items for our topic. PubMed was searched using free-text ([“high-stakes assessment” OR “high-stakes examinations”] AND [“specialist training” OR “postgraduate medical education”]). A similar search was performed in Google Scholar, followed by backward referencing where applicable. Searches were limited to English-language articles published since 2000. Articles focusing on the topics of high-stakes examinations in specialist training and maintenance of competence were included for full-text review. Articles related to the medical school setting or health professions other than medicine were excluded. The goal was to capture discussion points on the use of SMITHSEx in specialist training, including barriers to change and possible alternatives.

We list the arguments for and against the use of SMITHSEx in specialist training in , acknowledging that this list is by no means exhaustive.

Table 1. Arguments for and against the use of single-moment-in-time high-stakes examinations in specialist training.

Retaining the status quo

Accountability to regulators and public

It is often argued that the primary purpose of SMITHSEx is to provide assurance to the public that a high standard of practice exists among specialist doctors (Maudsley Citation1990). Regulatory bodies responsible for licencing individual specialists and accrediting training programs require evidence of competence but do not dictate the specific means of assessment. Both the General Medical Council (GMC) in the United Kingdom (UK) and the Australian Medical Council (AMC) emphasise multiple forms of assessment within a training program (Specialist Education Accreditation Committee Citation2015; General Medical Council Citation2017). While they acknowledge the requirement for summative assessments to demonstrate the achievement of learning outcomes (General Medical Council Citation2017) and make progression decisions (Specialist Education Accreditation Committee Citation2015), neither explicitly mandate a single high-stakes event nor dismiss the suitability of multiple low-stakes summative assessments. More recently, the AMC has expressed concern at the reliance on “high-stakes barrier assessments” in specialist training programs, particularly the extent of judgements that are made about trainee competence in single-moment-in-time assessments (Langham Citation2022).

Some training jurisdictions omit summative assessments altogether. In the United States (US), completing an accredited residency program is the only requirement for practising as a specialist. SMITHSEx at the completion of training are voluntary, with most specialists choosing to become “board-certified” for credentialing and insurance purposes (Fisher and Schloss Citation2016; Sklar Citation2016). Danish anaesthesia training does not necessitate SMITHSEx (Ringsted et al. Citation2003; Yamamoto et al. Citation2017). Family medicine training in Denmark and the Netherlands does not include SMITHSEx but relies on a range of assessment tools throughout training (Flum et al. Citation2015). When intensive care medicine training was initially introduced in the UK, SMITHSEx were only a voluntary requirement (Robson Citation2022).

Ensuring high standards of practice is a reasonable expectation, achievable through a structured training program with robust forms of assessment that measure learning in various domains. While SMITHSEx has traditionally been used to inform progression decisions, alternatives exist within a program of assessment that do not compromise accountability.

Objective measure of competence

SMITHSEx are often regarded as the ultimate objective measure of competence. However, this perceived objectivity diminishes upon closer examination. In classical test theory, a candidate’s score in any assessment comprises their true score and a measurement error component (De Champlain Citation2010). If an assessment was perfect, the true score would equal the derived score but this rarely happens in practice. Measures of reliability are used to estimate how well derived scores reflect true scores but these measures are sample-dependent, varying based on the candidates being assessed. Previous reviews have shown little published evidence supporting the validity of SMITHSEx for specialist training (Hutchinson et al. Citation2002).

Even when assessments are guided by marking rubrics, judgments of competence in oral and clinical assessments remain subjective. Inter-rater reliability, which assesses the consistency of measurements across examiners, is also sample-dependent. Multiple-choice questions (MCQs) are often considered highly reliable due to their lack of assessor subjectivity, but they are still subjective as they reflect the value judgements of question creators (ten Cate and Regehr Citation2019). Standard-setting conversations among examiners involve negotiations on various aspects, further highlighting the subjective nature of assessment. Some have argued that this consensus among a majority of examiners should be considered a form of negotiated or shared subjectivity rather than labelling it as objective (ten Cate and Regehr Citation2019).

The attempt to standardise outcomes to confer accountability oversimplifies the complex nature of competence (Chin and Lagasse Citation2017), as competence is context-specific (Bates et al. Citation2019). A learner’s performance on one assessment or situation is weakly predictive of their performance on a different assessment or situation (Eva Citation2003). Observed another way, two trainees with the same SMITHSEx score could emerge as entirely different clinicians because they are shaped by their training experience. Truly standardised outcomes may not be achievable (Bates et al. Citation2019). Some suggest embracing the subjective nature of assessment through the use of entrustment concepts and multiple assessment data points while maintaining fairness as an important goal (Hodges Citation2013; ten Cate and Regehr Citation2019).

Assessment drives learning

Despite claims that SMITHSEx enhance learning, they often focus on test preparation rather than fostering competent and safe clinical practice (Bricker Citation2017). The emphasis on passing SMITHSEx has led to the development of an industry geared towards test preparation, with formal teaching often focusing on content required to pass SMITHSEx rather than content required for competent practice.

The detailed knowledge of scientific minutiae may not be necessary for safe and effective clinical practice. Immediate recollection of detailed knowledge is less important in the digital age when information can be readily accessed. Learning should focus on content that facilitates competent and safe clinical practice, and the ability to seek out specific details when needed. An exception to this may be crisis management, a skill best learnt with regular simulation practice and facilitated by task checklists, though not amenable to being assessed by SMITHSEx.

Studies investigating the association between SMITHSEx achievement and clinical or professional outcomes have shown mixed results. A review of national medical licencing examinations showed no evidence that doctors improved as a consequence of their introduction, with claims of improved outcomes linked to correlation rather than causation (Archer et al. Citation2016). Trainees who did not pass the Intercollegiate Membership of the Royal College of Surgeons (MRCS) Part B (clinical) SMITHSEx at their first attempt were more likely to receive unsatisfactory competence progression outcomes during training but no association was found with the Part A (written) SMITHSEx and competence progression (Scrimgeour et al. Citation2018). Similarly, clinical performance scores of final-year residents in one US centre correlated with American Board of Anaesthesiology (ABA) oral and written SMITHSEx scores (Baker et al. Citation2016). If assessment in the workplace is as good a predictor of competence as SMITHSEx, there is an argument that the latter may be redundant. The written MRCS SMITHSEx showed only a weak positive correlation with performance on the national selection interview for entry into specialist training, though this was stronger for the clinical component of the MRCS SMITHSEx (Scrimgeour et al. Citation2017). Another study showed that performance in the MRCS SMITHSEx correlated with success in the Fellowship of the Royal College of Surgeons (FRCS) SMITHSEx (Scrimgeour et al. Citation2019), an indication that people who do well in one exam are likely to do well in another. There is no correlation between MRCS SMITHSEx performance and GMC fitness-to-practice investigations (Ellis et al. Citation2022b). Conversely, specialists who did not complete maintenance of certification (MOC) SMITHSEx requirements were more likely to be disciplined by US state medical boards compared to counterparts who passed (McDonald et al. Citation2018; Zhou et al. Citation2018). This may be interpreted as clinicians who were more likely to obey rules, understood the importance of, and participated in continuing medical education opportunities, were less likely to commit infractions that resulted in disciplinary action. MOC completion among US surgeons, of which SMITHSEx was a component, showed no association with complication rates, with some perceptively suggesting that alternative assessment methods such as video evaluation of technical skills, retraining on state-of-the-art care, or peer review may be more suitable for MOC (Xu et al. Citation2019).

Establishing a professional identity

Success in SMITHSEx is often perceived as a significant achievement, contributing to professional identity formation. This is not surprising, given the hard work and individual sacrifices involved. In the UK and Canada (but not New Zealand and Australia), specialist college fellowships are generally conferred after SMITHSEx success rather than at the completion of training, implicitly valuing the examination over the whole training program. A shared professional identity can be a positive force and is heavily linked to SMITHSEx success, as people are inclined to identify with the shared investment in training time and cognitive ability (Molleman and Rink Citation2015). Some have commented that SMITHSEx are intentionally set at a very high standard by specialties to safeguard their standing (Bricker Citation2017), more so when at risk of role substitution by non-medical health professionals (Fisher and Schloss Citation2016).

The case for change

One of the primary criticisms of SMITHSEx is that they focus excessively on assessing knowledge. While they may evaluate theoretical understanding and applied knowledge in simulated clinical settings, they often fail to capture essential competencies required for competent workplace performance. For example, in procedural specialties like anaesthesia or orthopaedic surgery, core skills such as tracheal intubation or fracture fixation are better assessed informally in the workplace over an extended period. Additionally, factors like crisis management, procedural skills, professionalism, communication, collaboration, lifelong learning, reflection on practice, and judgement are vital aspects of clinical competence that SMITHSEx often overlook. As a result, SMITHSEx performance alone cannot be considered a reliable indicator of workplace competence; at best, it provides a non-contextual measure of knowledge.

Interestingly, the medical community itself sometimes disregards the importance of SMITHSEx. In subspecialty practice, clinicians such as paediatric anaesthesiologists or orthopaedic spine surgeons are often deemed competent without the need for additional SMITHSEx. The Royal Australasian College of Physicians (RACP) SMITHSEx are a prerequisite for entry to advanced or subspecialty training, with no further SMITHSEx required in areas as varied as cardiology, nephrology, neurology, rheumatology, or gastroenterology (Royal Australasian College of Physicians Citation2022). In the UK, the Royal Colleges of Physicians introduced SMITHSEx for advanced or subspecialty training only in 2007 (Federation of the Royal Colleges of Physicians Citation2022). This raises questions about the inconsistency in evaluating competence between different areas of practice. If subspecialists practicing in such diverse areas can be considered competent without SMITHSEx, why do we rely on these examinations for specialist training?

The recent CoVid-19 pandemic further highlighted the incongruence of SMITHSEx, with a number of specialist training colleges altering or suspending certain components of SMITHSEx (ANZCA [@ANZCA], Citation2020; Langham Citation2022). This led to a situation where trainees were not formally assessed on certain clinical skills and allowed to progress through training, yet still considered as competent practitioners. This discrepancy underscores the understanding that competence in clinical practice is derived from real-world performance, not solely from SMITHSEx performance. This prompts us to question which competencies are best observed in the workplace rather than through SMITHSEx.

While most countries assume maintenance of competence through continuing professional development (CPD) activities after training, the US implemented mandatory SMITHSEx for recertification every 10 years (American Board of Medical Specialties Citation2014; Hoyt Citation2017; Nelson and Butterworth Citation2018). However, this approach has faced dissatisfaction among specialists (Cook et al. Citation2016; Hoyt Citation2017), with some even resorting to legal action (Fisher and Schloss Citation2016; Hoyt Citation2017). In response, several specialty boards have introduced alternative assessment formats to replace SMITHSEx. These formats include online open-book assessments (American Board of Internal Medicine Citation2022; American Board of Surgery Citation2022), continuous low-stakes assessment throughout the year (American Board of Anesthesiology Citation2022), and a longitudinal iterative learning assessment (Colson et al. Citation2019). There is cognitive dissonance in specialist doctors perceiving SMITHSEx as irrelevant to their practice (Cook et al. Citation2016), a burden (Cook et al. Citation2016; Howard-McNatt et al. Citation2019), contributing to burnout (Howard-McNatt et al. Citation2019), and an avenue for institutions to generate income (Cook et al. Citation2016), only when it comes to recertification but failing to apply the same logic to initial certification. Good enough for trainees but not for specialists.

The apprenticeship model of medical training allows for ongoing evaluation of competence without relying solely on SMITHSEx. Trainees are constantly assessed based on their performance in the workplace, considering factors such as previous experience, direct observation, and level of training. Supervisors make informed judgments about trainees’ competence, which are not solely dependent on a single examination. This raises doubts about the primacy placed on SMITHSEx and questions the validity of a one-time assessment over the collective judgement of experienced clinicians who have observed a trainee’s clinical practice over an extended period. As specialist training colleges embrace competency-based medical education (CBME) concepts, they cannot simultaneously claim a competency-based system while saying to trainees that “all prior competence assessment does not count if you cannot pass this exam.”

The impact of high-stakes assessments on trainee well-being cannot be ignored. SMITHSEx are often reported as the most stressful aspect of training (Castanelli et al. Citation2017; Looseley et al. Citation2019), leading to feelings of hopelessness, anxiety, and shame (Kahn et al. Citation2020). They adversely affect trainees’ social lives, relationships, mental health, and well-being (Weller et al. Citation2014). The AMC has expressed concern on the detrimental impact of SMITHSEx on trainees’ lives (Langham Citation2022). While some specialist training colleges place limits on attempts or restrict access to trainees, other institutions offer SMITHSEx to any registered doctor prior to entry into a training program. In the period from 2002 to 2010, one candidate was allowed 26 attempts in the Membership of the Royal Colleges of Physicians of the United Kingdom (MRCPUK) Part 1 SMITHSEx (McManus and Ludka Citation2012). Recognising the detrimental effects, the GMC have imposed a limit of six attempts for each part of any national SMITHSEx, effective from January 2022 (General Medical Council Citation2022). However, the question remains whether these assessments truly reflect an individual’s competence and whether they are worth the toll they take on trainees’ well-being.

Another critical concern associated with SMITHSEx is differential attainment, which refers to variations in performance between different groups of doctors. Differential attainment exists in specialist training SMITHSEx (Esmail and Roberts Citation2013; Tiffin and Paton Citation2021; Akinola and Khan Citation2022; Ellis et al. Citation2022; Citation2022a; Luton et al. Citation2023), and is indicative of broader systemic issues within our healthcare and training systems. Machine-marked MCQs eliminate some forms of bias and while some studies show evidence of examiner bias in clinical SMITHSEx (Esmail and Roberts Citation2013; Tiffin and Paton Citation2021), at least one randomised double-blind trial did not (Yeates et al. Citation2017). There is no evidence that differential attainment is due to learner deficits among ethnic minorities as differences persist after controlling for pre-university achievement, socio-economic status, study habits, living arrangements, and personality (Woolf Citation2020). It is likely that differential attainment stems from discrimination or implicit bias in the clinical learning environment, primarily affecting ethnic minorities and women. Merely omitting SMITHSEx from training will not solve this problem but we need to acknowledge the significant punitive role SMITHSEx play because of underlying systemic issues.

What are the alternatives?

We have demonstrated that SMITHSEx are not a prerequisite for showcasing accountability to regulating bodies, nor are they a truly objective measure of competence. While they are strongly tied to the formation of professional identity, there is no compelling reason to believe that completing a structured training program without SMITHSEx would diminish the development of a shared professional identity. Any alternatives to SMITHSEx would need to act as an appropriate driver for learning and improve trainee welfare. In our view, programmatic assessment implemented as part of a competency-based medical education (CBME) program, without SMITHSEx, can fulfil these requirements.

Programmatic assessment refers to the systematic and longitudinal acquisition of performance data to inform progression decisions. It involves the accumulation of numerous low-stakes assessments that contribute data points, which are then used to make a high-stakes decision regarding progression (Wilkinson and Tweed Citation2018). The decision-making process is explicit and transparent, clearly outlining the specific data points used, how they are interpreted, who interprets them, and for what purpose. By shifting from a decision-oriented (pass/fail) SMITHSEx approach to a feedback-orientated model with multiple low-stakes assessments, we move the focus towards providing feedback aimed at improving the quality of care for patients – an aspect that is not currently emphasised in SMITHSEx (Dave et al. Citation2020). Even though SMITHSEx are unnecessary in a programmatic assessment approach aligned with best-practice guidelines, they often continue to be retained, indicating the challenges faced in implementing change. An early version of programmatic assessment was introduced in Danish anaesthesia specialist training in 2000, which involved direct observation of procedural skills, practice volumes, patient feedback, multi-source feedback from supervising specialists, written assignments, and self-directed learning, all without SMITHSEx (Ringsted et al. Citation2003). One Canadian emergency medicine residency program has described its recent implementation (Rich et al. Citation2020). Studies analysing WBAs have shown that a suite of tools can be reliably used to make high-stakes assessment decisions (Hatala et al. Citation2006; Baker Citation2011; Moonen-van Loon et al. Citation2013; Wanderer et al. Citation2018; Castanelli et al. Citation2019; Toale et al. Citation2023). However, it is crucial to strike a balance, as excessive WBA requirements can be a source of stress for learners if poorly implemented. Assessments that are intended to be low-stakes may not be perceived as such by learners unless they are given sufficient agency – that is, their perceived ability to act, control, and make choices in their learning and assessment (Schut et al. Citation2018).

Eliminating SMITHSEx does not imply the complete removal of written assessments, whether administered on paper or online. Formative written assessments can be easily incorporated into programmatic assessment. Low-stakes formative written examinations throughout training have been shown to predict future performance in SMITHSEx (Kearney et al. Citation2000; McClintock and Gravlee Citation2010; Kim et al. Citation2012; Lee et al. Citation2012; Jones et al. Citation2014), suggesting that the latter may be redundant. A systematic review showed the predictive value of low-stakes formative examinations to be stronger for passing SMITHSEx than for failing (McCrary et al. Citation2021). Another method of conducting written assessments is through progress testing, which involves administering the same test to all learners in a program at or around the same time. These longitudinal low-stakes assessments, typically conducted two to three times a year, cover the entire syllabus. They are designed to discourage learners from solely preparing for a specific test and then “discarding” that knowledge, while also demonstrating knowledge growth throughout the program (Chen et al. Citation2015; Albanese and Case Citation2016). At the start of the program, learners are expected to get lower scores compared to their more advanced colleagues, and each test serves a formative purpose. “Failing” a single test does not act as a barrier to progression. One medical school that transitioned from a traditional SMITHSEx model to progress testing found reduced levels of stress among its medical students (Chen et al. Citation2015). Another showed that growth in learning using progress tests was correlated with performance in a national licencing SMITHSEx (Karay and Schauber Citation2018). Although more commonly used in undergraduate medical education, progress testing is well-established in specialist training (Dijksterhuis et al. Citation2013; Noel et al. Citation2016; Pugh et al. Citation2016; Rutgers et al. Citation2018; Halman et al. Citation2020). While multiple-choice or true/false formats are typically used for ease of marking, other formats such as short-answer questions (Rademakers et al. Citation2005), OSCEs (Noel et al. Citation2016; Pugh et al. Citation2016), and simulation-based assessments (Hall et al. Citation2020) have been described. For those who value the oral assessment format, formalised case-based discussions can serve as an excellent alternative.

Formative in-training assessments (ITEs) are widely used in US specialty training and share similarities with progress tests. In one example, the ABA administers the same formative ITE annually to all residents, providing performance feedback in the form of a percentile ranking for their training level (American Society of Anesthesiologists Citation2022). Although not required for progression, residency supervisors may use ITEs to gauge trainees’ progress and preparedness for the ABA SMITHSEx (American Society of Anesthesiologists Citation2022). When the radiology training program in the Netherlands transitioned from a formative ITE to formal progress testing, they found no difference in the trainees’ performance between the two formats (Rutgers et al. Citation2020). Similarly, the transition from SMITHSEx to a low-stakes assessment spread out over time within the ABA’s MOC requirements did not result in any notable changes in outcomes in terms of the likelihood of disciplinary action by a state medical board (Zhou et al. Citation2018, Citation2019).

Possible barriers preventing change

Institutional inertia stands as a prominent barrier to change. The mindset of adhering to established practices with the notion of “if it ain’t broke, don’t fix it” can impede progress. However, our training programs and assessment methods have continuously evolved based on emerging evidence. We have already transitioned away from long-form essays and unstructured oral assessments, as well as favouring the more effective single-best-answer format over true/false multiple-choice questions. The timing of assessments relative to trainee progression is also subject to adjustment. Embracing change and making bold decisions to alter existing rules and structures are essential for the evolution of assessment and training.

Another significant barrier is the perceived threat to professional identity, often intertwined with survivorship bias. It is difficult for individuals to consider the removal of a process that has become intricately tied to their specialist status. Such a move would require tacit acknowledgement that the requirement may have been unnecessary from the start. As “survivors” of this process, decision-makers possess an inherent bias to retain the status quo. Eliminating SMITHSEx could also be seen as a threat to the professional identity of examiners, who often hold esteemed positions in the specialist training college hierarchy. Notably, conflicts of interest may arise when crucial decisions regarding the role of SMITHSEx are made by current or former examiners.

Financial considerations also pose a barrier to change. Some specialist training colleges limit SMITHSEx to trainees, while others extend its availability to registered doctors more broadly. Certain UK specialist training colleges administer SMITHSEx overseas, predominantly in Asia or Africa. Debates have arisen over the merits of this practice (Devakumar and Mandeville Citation2011; Muir and Thacker Citation2011). Critics argue that it undermines local specialist training, provides content that may not be locally relevant, and contributes to the brain drain phenomenon (Devakumar and Mandeville Citation2011). Conversely, proponents contend that these SMITHSEx fulfil local demands as they are often requested by local partners (Muir and Thacker Citation2011), particularly for international medical graduates seeking GMC registration and work opportunities in the UK. The scale of this practice is evident; in 2013, the number of overseas candidates sitting the Part 1 MRCP SMITHSEx surpassed those taking the exam in UK for the first time (McAlpine et al. Citation2014). The revenue generated solely from international candidates undertaking the MRCPUK SMITHSEx could reach a staggering £11.7 million annually (based on current international candidate fees and pre-pandemic candidate numbers; see ) (Royal Colleges of Physicians of the United Kingdom Citation2019b, p. 1, 2019c, p. 2, 2019a). In 2011, the UK Royal College of Psychiatrists made £600,000 profit from its membership SMITHSEx fees, prompting many trainees to request a refund (Jaques Citation2013). These financial drivers are not confined to the UK. In a rather brazen act, the American Board of Internal Medicine (ABIM) surreptitiously formed a separate non-profit foundation, transferring US$60 million in assets and purchasing a US$2.3 million luxury condominium with a chauffeur-driven Mercedes S-class car, without informing its physician diplomates (Fisher and Schloss Citation2016). Meanwhile, ABIM’s own subsequent balance sheet showed it owing $US48 million, balanced by a promise of US$94 million in deferred income from SMITHSEx fees (Fisher and Schloss Citation2016). American specialty boards have been accused of using employees and subcontractors to publish articles supporting their maintenance of competence programs, inadequately disclosing inherent conflicts of interest (Kempen Citation2014). Financial incentives make it challenging to envision impartial deliberation on the topic of SMITHSEx within certain specialist training colleges, specialty boards, or among individuals seeking to improve their prospects through migration to higher-income environments.

Table 2. Potential revenue from international candidates sitting the MRCPUK SMITHSEx.

Conclusion

In conclusion, there is a compelling case for re-examining the role of SMITHSEx in assessing competence in specialist training. SMITHSEx are not an essential requirement for demonstrating accountability to regulating bodies nor are they a truly objective measure of competence. Completing a structured training program without SMITHSEx does not necessarily hinder the formation of a shared professional identity. The limitations of SMITHSEx in measuring workplace competence, the inconsistent reliance on these examinations in different areas of practice, the impact on trainee well-being, and the issues of differential attainment all highlight the need for change. While barriers such as institutional inertia, perceived threats to professional identity, and financial considerations may hinder change, they can be surmounted through a collective commitment to evidence-based evolution, unbiased decision-making processes, and a focus on maintaining the integrity of local training programs. By acknowledging and addressing these obstacles, we can pave the way for meaningful advancements in assessment and training practices.

Specialist training colleges need to move past being “exam providers” to “education providers”. Written, oral, and clinical assessments can still exist, but they don’t have to be a single high-stakes event. All assessment should focus on knowledge essential for competent and safe practice, and less on pure factual content or irrelevant minutiae. They should be clinically relevant, reflect practice in the workplace, be used as assessment for learning to identify areas for improvement, and be part of a suite of tools in programmatic assessment. Eliminating SMITHSEx from specialist training will allow evolution from a knowledge-based system to one that is truly competency-based. By doing so, we can better ensure that assessments align with the realities of clinical practice and promote the well-being and success of future medical professionals.

Ethical approval

Not required.

Author contributions

NS and SF individually conceived the ideas contained in this manuscript. NS drafted the initial manuscript, and SF contributed to its critical revision. NS and SF approved the final manuscript for submission and have agreed to be accountable for all aspects of the work.

Disclaimers

None.

Previous presentations

None.

Acknowledgements

None.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

No direct funding was provided.

Notes on contributors

Navdeep S. Sidhu

Navdeep S. Sidhu, MBChB, PGCertHealSc(Resus), FANZCA, MClinEd, FAcadMEd, is an anaesthesiologist at North Shore Hospital, Auckland and Director of Medical Admissions in the School of Medicine, University of Auckland, New Zealand. He is Chair of the Educators Subcommittee, Australian and New Zealand College of Anaesthetists. He has a master’s degree in clinical education and was awarded the 2019 Distinguished Clinical Teacher Award by the University of Auckland. Twitter/X: @DrNavSidhu

Simon Fleming

Simon Fleming, MBBS, FRCS (Tr&Orth), MSc, PhD, MAcadMEd, MFSTEd, MASE, AFHEA, FRSA, is an orthopaedic surgeon, past president of the British Orthopaedic Trainees” Association, past vice-chair of the Academy of Medical Royal Colleges Trainee Doctors Group, and is a co-chair for the International Conference in Residency Education. He has a PhD in medical education and was recognised as the International Leader of the Year by the Royal College of Physicians and Surgeons of Canada. Twitter/X: @OrthopodReg

References

  • Akinola L, Khan L. 2022. RCOG Differential Attainment Report 2021. https://www.rcog.org.uk/media/ctalnj3f/tef-2021-report-differential-attainment.pdf.
  • Albanese M, Case SM. 2016. Progress testing: critical analysis and suggested practices. Adv Health Sci Educ Theory Pract. 21(1):221–234. doi: 10.1007/s10459-015-9587-z.
  • American Board of Anesthesiology. 2022. Policy Book. Raleigh (NC): American Board of Anesthesiology.
  • American Board of Internal Medicine. 2022. ABIM MOC Requirements. Philadelphia (PA): NEJM. https://knowledgeplus.nejm.org/board-review/abim-moc-requirements/.
  • American Board of Medical Specialties. 2014. Standards for the ABMS program for maintenance of Certification (MOC). Chicago (IL): American Board of Medical Specialties.
  • American Board of Surgery. 2022. Continuous Certification. https://www.absurgery.org/default.jsp?exam-moc.
  • American Society of Anesthesiologists. 2022. Anesthesiology ITE. American Society of Anesthesiologists. https://www.asahq.org/education-and-career/asa-resident-component/study-resources-and-education/anesthesiology-ite.
  • ANZCA [@ANZCA]. 2020. March 6). UPDATE TO MEDICAL VIVAS: Due to concerns over #COVID19 @ANZCA has cancelled the Medical Vivas due to be held in Australian and NZ hospitals on Saturday 14 March. All candidates have been notified. More info at: http://bit.ly/2wtitip. [Tweet]. Twitter. https://twitter.com/ANZCA/status/1235722850259972098.
  • Archer J, Lynn N, Coombes L, Roberts M, Gale T, Price T, Regan de Bere S. 2016. The impact of large scale licensing examinations in highly developed countries: a systematic review. BMC Med Edu. 16(1):. 212. 10.1186/s12909-016-0729-7.
  • Baker K. 2011. Determining resident clinical performance. Getting beyond the noise. Anesthesiology. 115(4):862–878. doi: 10.1097/ALN.0b013e318229a27d.
  • Baker K, Sun H, Harman A, Poon KT, Rathmell JP. 2016. Clinical performance scores are independently associated with the American Board of Anesthesiology certification examination scores. Anesth Analg. 122(6):1992–1999. doi: 10.1213/ANE.0000000000001288.
  • Bates J, Schrewe B, Ellaway RH, Teunissen PW, Watling C. 2019. Embracing standardisation and contextualisation in medical education. Med Educ. 53(1):15–24. doi: 10.1111/medu.13740.
  • Bricker S. 2017. The Anaesthesia Science Viva Book. 3rd ed. Cambridge: Cambridge University Press.
  • Castanelli DJ, Moonen-van Loon JMW, Jolly B, Weller JM. 2019. The reliability of a portfolio of workplace-based assessments in anesthesia training. Can J Anaesth. 66(2):193–200. doi: 10.1007/s12630-018-1251-7.
  • Castanelli DJ, Wickramaarachchi SA, Wallis S. 2017. Burnout and the learning environment of anaesthetic trainees. Anaesth Intensive Care. 45(6):744–751. doi: 10.1177/0310057X1704500615.
  • Chen Y, Henning M, Yielde J, Jones R, Wearn A, Weller J. 2015. Progress testing in the medical curriculum: students” approaches to learning and perceived stress. BMC Med Educ. 15:147. doi: 10.1186/s12909-015-0426-y.
  • Chin M, Lagasse RS. 2017. Assessment of competence: developing trends and ethical considerations. Curr Opin Anaesthesiol. 30(2):236–241. doi: 10.1097/ACO.0000000000000431.
  • Cizek GJ. 2001. More unintended consequences of high-stakes testing. EdM: IPractice. 20(4):19–27. doi: 10.1111/j.1745-3992.2001.tb00072.x.
  • Colson YL, Putnam JB, Yang SC, Fann JI, Vaporciyan AA, Dearani JA, Jones DR, Allen MS, Meyers BF, Wright CD, et al. 2019. American board of thoracic surgery 10-year maintenance of certification exam improves and validates knowledge acquisition. Ann Thorac Surg. 108(6):1895–1900. doi: 10.1016/j.athoracsur.2019.05.074.
  • Cook DA, Blachman MJ, West CP, Wittich CM. 2016. Physician attitudes about maintenance of certification: a cross-specialty national survey. Mayo Clin Proc. 91(10):1336–1345. doi: 10.1016/j.mayocp.2016.07.004.
  • Craig S, Braitberg G, Nicolas C, White G, Egerton-Warburton D. 2010. Assessment and feedback in emergency medicine training: views of Australasian emergency trainees: assessment and feedback in emergency medicine. Emerg Med Australas. 22(6):537–547. doi: 10.1111/j.1742-6723.2010.01353.x.
  • Dave S, Chakravorty I, Menon G, Sidhu K, Bamrah J, Mehta R. 2020. Differential attainment in summative assessments within postgraduate medical education & training: 2020 thematic series on tackling differential attainment in healthcare professions. Sushruta J Health Policy Opinion. 13(3):1–7 doi: 10.38192/13.3.15.
  • De Champlain AF. 2010. A primer on classical test theory and item response theory for assessments in medical education: classical test theory and item response theory. Med Educ. 44(1):109–117. doi: 10.1111/j.1365-2923.2009.03425.x.
  • Devakumar D, Mandeville K. 2011. Should UK membership exams be held overseas? No. BMJ. 343:d4560. doi: 10.1136/bmj.d4560.
  • Dijksterhuis MGK, Schuwirth LWT, Braat DDM, Scheele F. 2013. An exploratory study into the impact and acceptability of formatively used progress testing in postgraduate obstetrics and gynaecology. Perspect Med Educ. 2(3):126–141. doi: 10.1007/s40037-013-0063-2.
  • Dwyer T, Takahashi SG, Hynes MK, Herold J, Wasserstein D, Nousiainen M, Ferguson P, Wadey V, Murnaghan ML, Leroux T, et al. 2014. How to assess communication, professionalism, collaboration and the other intrinsic CanMEDS roles in orthopedic residents: use of an objective structured clinical examination (OSCE). Can J Surg. 57(4):230–236. doi: 10.1503/cjs.018813.
  • Ellis R, Brennan PA, Lee AJ, Scrimgeour DSG, Cleland J. 2022. Differential attainment at MRCS according to gender, ethnicity, age and socioeconomic factors: a retrospective cohort study. J R Soc Med. 115(early online access). (7):257–272. doi: 10.1177/01410768221079018.
  • Ellis R, Cleland J, Scrimgeour D, Lee A, Brennan P. 2022a. The impact of disability on performance in a high-stakes postgraduate surgical examination: a retrospective cohort study. J R Soc Med. 115(2):58–68. doi: 10.1177/01410768211032573.
  • Ellis R, Cleland J, Scrimgeour DSG, Lee AJ, Brennan PA. 2022b. A cross-sectional study examining the association between MRCS performance and surgeons receiving sanctions against their medical registration. Surgeon. 20(4):211–215. doi: 10.1016/j.surge.2021.04.003.
  • Esmail A, Roberts C. 2013. Academic performance of ethnic minority candidates and discrimination in the MRCGP examinations between 2010 and 2012: analysis of data. BMJ. 347(2):f5662. doi: 10.1136/bmj.f5662.
  • Eva K. 2003. On the generality of specificity. Med Educ. 37(7):587–588. doi: 10.1046/j.1365-2923.2003.01563.x.
  • Federation of the Royal Colleges of Physicians. 2022. Specialty Certificate Examinations. Federation of the Royal Colleges of Physicians. https://www.mrcpuk.org/mrcpuk-examinations/specialty-certificate-examinations.
  • Fisher WG, Schloss EJ. 2016. Medical specialty certification in the United States—A false idol? J Interv Card Electrophysiol. 47(1):37–43. doi: 10.1007/s10840-016-0119-4.
  • Flum E, Maagaard R, Godycki-Cwirko M, Scarborough N, Scherpbier N, Ledig T, Roos M, Steinhäuser J. 2015. Assessing family medicine trainees – what can we learn from the European neighbours? GMS Z Med Ausbild. 32(2):Doc21.
  • General Medical Council. 2017. Excellence by design: standards for postgraduate curricula. London: General Medical Council.
  • General Medical Council. 2022. National professional examinations: the expectations. https://www.gmc-uk.org/education/standards-guidance-and-curricula/guidance/national-professional-examinations/the-expectations.
  • Grant MJ, Booth A. 2009. A typology of reviews: an analysis of 14 review types and associated methodologies. Health Info Libr J. 26(2):91–108. doi: 10.1111/j.1471-1842.2009.00848.x.
  • Hall AK, Chaplin T, McColl T, Petrosoniak A, Caners K, Rocca N, Gardner C, Bhanji F, Woods R. 2020. Harnessing the power of simulation for assessment: consensus recommendations for the use of simulation-based assessment in emergency medicine. CJEM. 22(2):194–203. doi: 10.1017/cem.2019.488.
  • Halman S, Fu AYN, Pugh D. 2020. Entrustment within an objective structured clinical examination (OSCE) progress test: bridging the gap towards competency-based medical education. Med Teach. 42(11):1283–1288. doi: 10.1080/0142159X.2020.1803251.
  • Hatala R, Ainslie M, Kassen BO, Mackie I, Roberts JM. 2006. Assessing the mini-clinical evaluation exercise in comparison to a national specialty examination. Med Educ. 40(10):950–956. doi: 10.1111/j.1365-2929.2006.02566.x.
  • Heubert JP, Hauser RM. 1999. High stakes: testing for tracking, promotion, and graduation. Washington (DC): National Academy Press.
  • Hodges B. 2013. Assessment in the post-psychometric era: learning to love the subjective and collective. Med Teach. 35(7):564–568. doi: 10.3109/0142159X.2013.789134.
  • Howard-McNatt M, Sabel M, Agnese D, Amersi F, Ammori J, Arciero C, Berman R, Dominici L, Fleming RY, Han D, et al. 2019. Maintenance of certification and continuing medical education: are they still required? Ann Surg Oncol. 26(12):3820–3823. doi: 10.1245/s10434-019-07702-3.
  • Hoyt DB. 2017. Looking forward. BullAm Coll Surg. 102(8):10–11.
  • Hutchinson L, Aitken P, Hayes T. 2002. Are medical postgraduate certification processes valid? A systematic review of the published evidence: are medical postgraduate certification processes valid? Med Educ. 36(1):73–91. doi: 10.1046/j.1365-2923.2002.01120.x.
  • Jaques H. 2013. Royal College of Psychiatrists offers free CPD in response to fee surplus. BMJ. 346:f1161. doi: 10.1136/bmj.f1161.
  • Jefferies A, Simmons B, Tabak D, Mcilroy JH, Lee K-S, Roukema H, Skidmore M. 2007. Using an objective structured clinical examination (OSCE) to assess multiple physician competencies in postgraduate training. Med Teach. 29(2-3):183–191. doi: 10.1080/01421590701302290.
  • Jones AT, Biester TW, Buyske J, Lewis FR, Malangoni MA. 2014. Using the American board of surgery in-training examination to predict board certification: a cautionary study. J Surg Educ. 71(6):e144–e148. doi: 10.1016/j.jsurg.2014.04.004.
  • Kahn J, Goodman CR, Albert A, Agarwal A, Jeans E, Tye K, Campbell SR, Marcrom S, Colbert LE. 2020. Top concerns of radiation oncology trainees in 2019: job market, board examinations, and residency expansion. Int J Radiat Oncol Biol Phys. 106(1):19–25. doi: 10.1016/j.ijrobp.2019.07.006.
  • Karay Y, Schauber SK. 2018. A validity argument for progress testing: examining the relation between growth trajectories obtained by progress tests and national licensing examinations using a latent growth curve approach. Med Teach. 40(11):1123–1129. doi: 10.1080/0142159X.2018.1472370.
  • Kearney R, Sullivan P, Skakun E. 2000. Performance on ABA-ASA in-training examination predicts success for RCPSC certification. American board of anesthesiology-American society of anesthesiologists. Royal College of physicians and surgeons of Canada. Can J Anaesth. 47(9):914–918. doi: 10.1007/BF03019676.
  • Kempen PM. 2014. Maintenance of certification: ethics and conflicts of interest. J Am Phys Surg. 19(3):85–89
  • Kim PY, Wallace DA, Allbritton DW, Altose MD. 2012. Predictors of success on the written anesthesiology board certification examination. Int J Med Educ. 3:225–235 doi: 10.5116/ijme.5096.5d66.
  • Langham R. 2022. Australian Medical Council statement on the consideration of the Royal Australian and New Zealand College of Psychiatrists Alternate Assessment Pathway. Australian Medical Council. https://www.amc.org.au/wp-content/uploads/2022/03/2022-03-02-AMC-statement-consideration-of-RANZCP-AAP.pdf.
  • Lee AG, Oetting TA, Blomquist PH, Bradford G, Culican SM, Kloek C, Krishnan C, Lauer AK, Levi L, Naseri A, et al. 2012. A Multicenter analysis of the ophthalmic knowledge assessment program and American board of ophthalmology written qualifying examination Performance. Ophthalmology. 119(10):1949–1953. doi: 10.1016/j.ophtha.2012.06.010.
  • Looseley A, Wainwright E, Cook TM, Bell V, Hoskins S, O”Connor M, Taylor G, Mouton R. 2019. Stress, burnout, depression and work satisfaction among UK anaesthetic trainees; a quantitative analysis of the satisfaction and wellbeing in anaesthetic Training study. Anaesthesia. 74(10):1231–1239., & doi: 10.1111/anae.14681.
  • Luton OW, Mellor K, Robinson DBT, Barber Z, James OP, Powell AGMT, Hemington-Gorse S, Walsh L, Egan RJ, Lewis WG. 2023. Differential attainment in higher surgical training: scoping pan-specialty spectra. Postgrad Med J. 99(1174):849–854. postgradmedj-2022-141638. doi: 10.1136/postgradmedj-2022-141638.
  • Maudsley RF. 1990. Medical licensure: let’s not lose sight of the objective. Can Med Assoc J. 143(2):98–100.
  • McAlpine L, Selamaj E, Shannon C, Chis L, Dacre J, Elder A. 2014. UK postgraduate medicine examinations: opportunities for international candidates. Clin Med. 14(5):500–505. doi: 10.7861/clinmedicine.14-5-500.
  • McClintock JC, Gravlee GP. 2010. Predicting success on the certification examinations of the American Board of anesthesiology. Anesthesiology. 112(1):212–219. doi: 10.1097/ALN.0b013e3181c62e2f.
  • McCrary HC, Colbert-Getz JM, Poss WB, Smith BK. 2021. A Systematic review of the relationship between in-training examination scores and specialty board examination scores. j Grad Med Educ. 13(1):43–57. doi: 10.4300/JGME-D-20-00111.1.
  • McDonald FS, Duhigg LM, Arnold GK, Hafer RM, Lipner RS. 2018. The American Board of Internal Medicine Maintenance of Certification Examination and state medical board disciplinary actions: a population cohort study. J Gen Intern Med. 33(8):1292–1298. doi: 10.1007/s11606-018-4376-z.
  • McManus I, Ludka K. 2012. Resitting a high-stakes postgraduate medical examination on multiple occasions: nonlinear multilevel modelling of performance in the MRCP (UK) examinations. BMC Med. 10(1):60. doi: 10.1186/1741-7015-10-60.
  • Molleman E, Rink F. 2015. The antecedents and consequences of a strong professional identity among medical specialists. Soc Theory Health. 13(1):46–61. doi: 10.1057/sth.2014.16.
  • Moonen-van Loon JMW, Overeem K, Donkers HHLM, Van Der Vleuten CPM, Driessen EW. 2013. Composite reliability of a workplace-based assessment toolbox for postgraduate medical education. Adv Health Sci Educ Theory Pract. 18(5):1087–1102. doi: 10.1007/s10459-013-9450-z.
  • Muir G, Thacker J. 2011. Should UK membership exams be held overseas? Yes. BMJ. 343(jul28 2):d4566–d4566. doi: 10.1136/bmj.d4566.
  • Nelson M, Butterworth JF. IV 2018. Maintenance of certification. Has MOC gone amok? Anesthesiology. 129(4):631–633. doi: 10.1097/ALN.0000000000002387.
  • Noel K, Archibald D, Brailovsky C, Mautbur A. 2016. Progress testing in family medicine—A novel use for simulated office oral exams. Med Teach. 38(4):364–368. doi: 10.3109/0142159X.2015.1031733.
  • Pugh D, Bhanji F, Cole G, Dupre J, Hatala R, Humphrey-Murto S, Touchie C, Wood TJ. 2016. Do OSCE progress test scores predict performance in a national high-stakes examination? Med Educ. 50(3):351–358. doi: 10.1111/medu.12942.
  • Rademakers J, ten Cate O, Bär PR. 2005. Progress testing with short answer questions. Med Teach. 27(7):578–582. doi: 10.1080/01421590500062749.
  • Rich JV, Fostaty Young S, Donnelly C, Hall AK, Dagnone JD, Weersink K, Caudle J, Van Melle E, Klinger DA. 2020. Competency‐based education calls for programmatic assessment: but what does this look like in practice? J Eval Clin Pract. 26(4):1087–1095. doi: 10.1111/jep.13328.
  • Ringsted C, Østergaard D, Scherpbier A. 2003. Embracing the new paradigm of assessment in residency training: an assessment programme for first-year residency training in anaesthesiology. Med Teach. 25(1):54–62. doi: 10.1080/0142159021000061431.
  • Ringsted C, Østergaard D, van der Vleuten CPM. 2003. Implementation of a formal in-training assessment programme in anaesthesiology and preliminary results of acceptability. Acta Anaesthesiol Scand. 47(10):1196–1203. doi: 10.1046/j.1399-6576.2003.00255.x.
  • Robson V. 2022. FFICM Chair of Examiners Report – October 2021 FFICM. The Faculty of Intensive Care Medicine. https://www.ficm.ac.uk/sites/ficm/files/documents/2022-01/FFICM%20Chair%27s%20Report%20Oct%202021.pdf.
  • Royal Australasian College of Physicians. 2022. Examinations. Royal Australasian College of Physicians. https://www.racp.edu.au/trainees/examinations.
  • Royal Colleges of Physicians of the United Kingdom. 2019a. PACES 2019 annual performance report. https://www.mrcpuk.org/document/paces-2019-annual-performance-report.
  • Royal Colleges of Physicians of the United Kingdom. 2019b. Part 1 2019 annual performance report. https://www.mrcpuk.org/document/part-1-2019-annual-performance-report.
  • Royal Colleges of Physicians of the United Kingdom. 2019c. Part 2 2019 annual performance report. https://www.mrcpuk.org/document/part-2-2019-annual-performance-report.
  • Rutgers DR, van Raamt F, van Lankeren W, Ravesloot CJ, van der Gijp A, ten Cate TJ, van Schaik JPJ. 2018. Fourteen years of progress testing in radiology residency training: experiences from The Netherlands. Eur Radiol. 28(5):2208–2215. doi: 10.1007/s00330-017-5138-8.
  • Rutgers DR, van Schaik JPJ, Kruitwagen CLJJ, Haaring C, van Lankeren W, van Raamt AF, ten Cate O. 2020. Introducing summative progress testing in radiology residency: little change in residents” test results after transitioning from formative progress testing. Med Sci Educ. 30(2):943–953. doi: 10.1007/s40670-020-00977-2.
  • Schut S, Driessen E, Van Tartwijk J, Van Der Vleuten C, Heeneman S. 2018. Stakes in the eye of the beholder: an international study of learners” perceptions within programmatic assessment. Med Educ. 52(6):654–663. doi: 10.1111/medu.13532.
  • Scrimgeour DSG, Brennan PA, Griffiths G, Lee AJ, Smith FCT, Cleland J. 2018. Does the intercollegiate membership of the Royal College of Surgeons (MRCS) examination predict “on-the-job” performance during UK higher specialty surgical training? Ann R Coll Surg Engl. 100(8):1–7. 65. doi: 10.1308/rcsann.2018.0153.
  • Scrimgeour DSG, Cleland J, Lee AJ, Brennan PA. 2019. Prediction of success at UK specialty board examinations using the mandatory postgraduate UK surgical examination. BJS Open. 3(6):865–871. doi: 10.1002/bjs5.50212.
  • Scrimgeour DSG, Cleland J, Lee AJ, Griffiths G, McKinley AJ, Marx C, Brennan PA. 2017. Impact of performance in a mandatory postgraduate surgical examination on selection into specialty training: performance in a postgraduate surgical examination and selection into specialty training. BJS Open. 1(3):67–74. doi: 10.1002/bjs5.7.
  • Sklar DP. 2016. What happens after medical school? Current controversies about licensure, maintenance of ertification, and continuing professional development. Acad Med. 91(1):1–3. doi: 10.1097/ACM.0000000000001011.
  • Specialist Education Accreditation Committee. 2015. Standards for Assessment and Accreditation of Specialist Medical Education Programs and Professional Development Programs by the Australian Medical Council. Kingston (Australia): Australian Medical Council Limited.
  • ten Cate O, Regehr G. 2019. The power of subjectivity in the assessment of medical trainees. Acad Med. 94(3):333–337. doi: 10.1097/ACM.0000000000002495.
  • Thoma B, Monteiro S, Pardhan A, Waters H, Chan T. 2022. Replacing high-stakes summative examinations with graduated medical licensure in Canada. CMAJ. 194(5):E168–E170. doi: 10.1503/cmaj.211816.
  • Tiffin PA, Paton LW. 2021. Differential attainment in the MRCPsych according to ethnicity and place of qualification between 2013 and 2018: a UK cohort study. Postgrad Med J. 97(1154):764–776. doi: 10.1136/postgradmedj-2020-137913.
  • Toale C, Morris M, O”Keeffe D, Boland F, Ryan DM, Nally DM, Kavanagh DO. 2023. Assessing operative competence in core surgical training: a reliability analysis. Am J Surg. doi: 10.1016/j.amjsurg.2023.06.020.
  • Wanderer JP, de Oliveira Filho GR, Rothman BS, Sandberg WS, McEvoy MD. 2018. Implementation and evaluation of the Z-score system for normalizing residency evaluations. Anesthesiology. 128(1):144–158. doi: 10.1097/ALN.0000000000001919.
  • Weller JM, Henning M, Butler R, Thompson A. 2014. The Impact of the Australian and New Zealand College of Anaesthetists” specialist examinations on trainee learning and wellbeing: A qualitative study. Anaesth Intensive Care. 42(6):736–744. doi: 10.1177/0310057X1404200609.
  • Wilkinson TJ, Tweed MJ. 2018. Deconstructing programmatic assessment. Adv Med Educ Pract. 9:191–197. doi: 10.2147/AMEP.S144449.
  • Woolf K. 2020. Differential attainment in medical education and training. BMJ. 368:m339. doi: 10.1136/bmj.m339.
  • Xu T, Mehta A, Park A, Makary MA, Price DW. 2019. Association between board certification, maintenance of certification, and surgical complications in the United States. Am J Med Qual. 34(6):545–552. doi: 10.1177/1062860618822752.
  • Yamamoto S, Tanaka P, Madsen MV, Macario A. 2017. Comparing anesthesiology residency training structure and requirements in seven different countries on three continents. Cureus. 9(2):e1060. doi: 10.7759/cureus.1060.
  • Yeates P, Woolf K, Benbow E, Davies B, Boohan M, Eva K. 2017. A randomised trial of the influence of racial stereotype bias on examiners” scores, feedback and recollections in undergraduate clinical exams. BMC Med. 15(1):179. doi: 10.1186/s12916-017-0943-0.
  • Zhou Y, Sun H, Macario A, Keegan MT, Patterson AJ, Minhaj MM, Wang T, Harman AE, Warner DO. 2018. Association between performance in a maintenance of certification program and disciplinary actions against the medical licenses of anesthesiologists. Anesthesiology. 129(4):812–820. doi: 10.1097/ALN.0000000000002326.
  • Zhou Y, Sun H, Macario A, Keegan MT, Patterson AJ, Minhaj MM, Wang T, Harman AE, Warner DO. 2019. Association between participation and performance in MOCA minute and actions against the medical licenses of anesthesiologists. Anesth Analg. 129(5):1401–1407. doi: 10.1213/ANE.0000000000004268.