47,591
Views
201
CrossRef citations to date
0
Altmetric
Web Paper AMEE Guide

The Objective Structured Clinical Examination (OSCE): AMEE Guide No. 81. Part I: An historical and theoretical perspective

, , &
Pages e1437-e1446 | Published online: 22 Aug 2013

Abstract

The Objective Structured Clinical Examination (OSCE) was first described by Harden in 1975 as an alternative to the existing methods of assessing clinical performance (Harden et al. 1975). The OSCE was designed to improve the validity and reliability of assessment of performance, which was previously assessed using the long case and short case examinations. Since then the use of the OSCE has become widespread within both undergraduate and postgraduate clinical education. We recognise that the introduction of the OSCE into an existing assessment programme is a challenging process requiring a considerable amount of theoretical and practical knowledge. The two parts of this Guide are designed to assist all those who intend implementing the OSCE into their assessment systems. Part I addresses the theoretical aspects of the OSCE, exploring its historical development, its place within the range of assessment tools and its core applications. Part II offers more practical information on the process of implementing an OSCE, including guidance on developing OSCE stations, choosing scoring rubrics, training examiners and standardised patients and managing quality assurance processes. Together we hope these two parts will act as a useful resource both for those choosing to implement the OSCE for the first time and also those wishing to quality assure their existing OSCE programme.

Introduction

Conducting an Objective Structured Clinical Examination (OSCE) for the first time is a complex and time-consuming task. It requires considerable understanding of the underlying educational principles of an OSCE and the development of academic and administrative structures to support and implement the examination. These supporting structures or entities could be in the form of teams and committees as described in detail in the second part of this Guide. In the institutions where the OSCE is already being used as an assessment tool, quality assurance and continuous improvement are important in order to maintain standards and psychometric rigour.

The current literature appears to be lacking a detailed and comprehensive manual to help institutions with the practicalities of implementing the OSCE for the first time, though a lot of studies covering various aspects of the OSCE have been published in peer reviewed journals. This Guide will present an evidence-based perspective on setting up an OSCE for those new to the approach, and will also provide some guidance and thought to those who would like to revisit their programmes for quality assurance purposes.

The Guide consists of two parts; Part I focuses on the historical background and educational principles of the OSCE. Knowledge and understanding of these principles is essential before moving any further in designing and administering an OSCE. We hope that the contents of Part I will act as a suitable and informative introduction for the readers, enabling them eventually to understand and implement the ideas and practical advice given in Part II, which will describe the organisation and administration of the OSCE.

Assessment of performance and the development of the OSCE

Assessment of performance in health care education is dependent upon the choice of appropriate tools to measure the outcomes in question. According to Stokes (Citation1979) such tools are designed for determining whether one can ‘do things’ rather than simply remember, talk and write about them. Any assessment tool used alone is not capable of assessing the combination of knowledge, skills and behaviours, which influence performance within the various contexts of health care. Epstein (Citation2007) argues for the use of a combination of assessment tools to measure a range of learning domains; such a methodology can be called a ‘test battery’ approach (Hamdy et al. Citation2010). The OSCE plays an important part in this test battery in the assessment of performance in simulated environments (PMETB Citation2007; Khan & Ramachandran Citation2012).

Assessment in health care education has been constantly evolving; from the assessment of knowledge in the period up to the 1960s, to the assessment of performance from the 1970s onwards (Epstein Citation2007; Hays Citation2008). A range of assessment tools have subsequently been developed for the assessment of performance, including the OSCE with standardised patients, workplace based assessments and assessments using computer enhanced simulations (Norcini & McKinley Citation2007).

Prior to the development of the OSCE in the 1970s by Harden (Citation1975), combinations of ‘short case’ and ‘long case’ examinations were used in an attempt to assess performance. Frequently, these methods also included an oral or viva voce examination (Sood Citation2001). These assessment tools are still being used widely in different parts of the world and warrant some consideration before we move on to discuss the OSCE.

Short case examinations

In short case examinations candidates are asked to perform a brief and focused clinical examination of around five or six patients with specific clinical findings. Fifteen to twenty minutes per patient are usually allocated. Following this, the findings and diagnosis are discussed with a pair of examiners. The examiners award marks independently on the candidates’ performance; however, this scoring is often unstructured (Ponnamperuma et al. Citation2009), with the same examiners examining the candidate throughout the entire time.

Although the patients have real signs and symptoms, these patients often change between candidates, affecting the standardisation of the examination (Norcini Citation2002; Wass & van der Vleuten Citation2004). This format of examination has the advantage of having multiple, brief and observed real patient encounters similar to a clinicians’ outpatient experience and are therefore considered as a very good assessment tool for clinical examination skills (Ponnamperuma et al. Citation2009). However, the reproducibility and validity of these examinations is affected by unstructured questioning by the examiners, a lack of standardisation of patients between candidates, and a lack of ability to assess history taking (Walsh Citation2006).

Viva voce examinations

In a typical Viva voce assessment, the candidates are provided with clinical material such as a case transcript, pathological specimen or investigation results. After a given time period to assimilate the information, two examiners question the candidates for 10–15 min. These examinations allow the assessment of recall, depth of knowledge, hypothetical problem solving, judgement, clinical reasoning and analytical skills (Jayawickramarajah Citation1985). Although typically unstructured questioning leads to poor inter-rater reliability (Harden et al. Citation1975); however, it is possible to standardise these items in order to minimise variance (Ponnamperuma et al. Citation2009). As there is no direct observation of the candidates’ clinical performance the Viva voce can only assess this by proxy.

Long case examinations

During long case examinations, candidates are asked to take a history and perform a complete physical examination on a real patient, frequently chosen from the current in-patient or out-patient cohort. Candidates at different examination sittings may be allocated different patients with varying conditions and clinical signs. Candidates are asked to take a history and perform a complete physical examination in the first 30–45 min, often unobserved by the examiners (Ponnamperuma et al. Citation2009). Typically, unstructured questioning of the candidates follows this, which is usually focused on their clinical findings, diagnosis and management plan of the patients’ problems. The candidates’ interaction with the patient, including history taking, general communication and clinical skills is not always observed. Most frequently the discussion is based on the theoretical aspects of the case in question, exploring the depth of candidates’ understanding i.e. their knowledge and the management plan for the patient (Sood Citation2001; Wass et al. Citation2001a; Wass & van der Vleuten Citation2004).

A number of drawbacks with this form of assessment have been identified; for example, the marks awarded for long cases are based upon unstructured questioning rather than checklists or standardised scoring sheets (Norman Citation2002). There is an inherent inter-case variability at separate sittings of the same examination (Norcini Citation2002; Wass & van der Vleuten Citation2004). Further, a single case does not allow the assessment of the candidates’ performance on a range of areas, restricting the sampling of the ‘test universe’. The content specificity of long cases makes the generalisation of the skills assessed using this technique nearly impossible (Wass et al. Citation2001b; Norcini Citation2002).

Despite the disadvantages described above, long case examinations have a high ‘face validity’, as the encounter between the candidate and the patient is realistic and tests the integrated interaction which is very close to actual daily practice (Wass & van der Vleuten Citation2004). lists the advantages and disadvantages of long cases ().

Table 1  Advantages and disadvantages of long cases

To overcome the above mentioned drawbacks, many modifications of the long case examinations have been developed. The major focus of these modifications is on the observation of the candidates’ performance during the long case. For instance, the Observed Long case (Newble Citation1991), the Leicester Assessment Package (Fraser et al. Citation1994), the Objective Structured Long Case Record (OSLER) (Gleeson Citation1997), the Standardised Structured Long Case Examination of Clinical Competence (Troncon et al. Citation2000) and the Direct Observation Clinical Encounter Examination (DOCEE) (Hamdy et al. Citation2003) are described in the literature.

Amongst all of the above, the OSLER has gained a degree of popularity. The reliability and construct validity of the OSLER is better than the long case; this has been achieved by the introduction of different domains to assess the performance of the candidates in each case (Gleeson Citation1997) ().

Table 2  Assessment domains in the OSLER (Gleeson Citation1997)

In order to achieve the high levels of reliability demonstrated in the literature (Cronbach's alpha of 0.8), observed long case examinations require long testing times of up to three and a half hours which may be impractical at some institutions (Wass et al. Citation2001a). The concepts of reliability and validity are described in some detail later.

In short, the main criticisms of the traditional methods of assessing clinical performance are a lack of structure and standardisation. The development of the OSCE by Harden (Citation1975) aimed to address the above issues and improve the quality of the assessment of performance. At the time of its inception, it was described as a tool for the assessment of competence but more recently competence is seen as a point on the spectrum of performance. By this inference the OSCE is a tool used for the assessment of performance in simulated environments (Khan & Ramachandran Citation2012); this concept is explored in some detail later in this Guide.

The original OSCE

Prior to the development of the OSCE, the candidate's assessment could be affected by the patient's performance, examiner bias, a non-standardised marking scheme and the candidate's actual performance. The OSCE was first designed to introduce standardisation and reduce the number of variables that could impact the assessment of performance. Hence, in a well-designed OSCE, the grades of the candidates should predominantly be affected by the performance of the candidates alone, with minimal effect from other sources of variance. In addition, the OSCE was designed as a novel evaluation tool, allowing the assessment of candidates’ clinical skills, attitudes, problem-solving abilities, and their application of knowledge in one examination (Harden et al. Citation1975).

The first OSCE was conducted by Harden in 1972 in Dundee, and described in the literature in 1975 (Harden et al. Citation1975). Since then a healthy body of published literature has developed to support and substantiate an evidence-based use of OSCE techniques, underpinned by sound educational principles (Carraccio & Englander Citation2000; Wass et al. Citation2001b; Hodges et al. Citation2002; Turner & Dankoski Citation2008; Boursicot Citation2010; Gupta et al. Citation2010; Zayyan Citation2011).

In the original version described by Harden, the students moved around 18 testing and 2 rest stations in a hospital ward. Each station was 4.5 min long with a 30 s break between the stations. The total examination time was 100 min. Each station tested a single competency; for instance performance of a procedure, history taking or clinical examination of a patient. The original 18 stations and the sequence as described by Harden (Citation1979) are shown (). The standardised marking scheme used at station 3 in this examination is also reproduced in the subsequent figure ().

Figure 1 Eighteen station OSCE as originally described by Harden (Citation1979). Shaded stations had examiners present.

Figure 1 Eighteen station OSCE as originally described by Harden (Citation1979). Shaded stations had examiners present.

Figure 2 Scoring rubric used at station 3 as described by Harden (Citation1979).

Figure 2 Scoring rubric used at station 3 as described by Harden (Citation1979).

Defining the OSCE

Since the original OSCE, many definitions of this assessment technique have been proposed; Harden (Citation1988) defined it as; “An approach to the assessment of clinical competence in which the components are assessed in a planned or structured way with attention being paid to the objectivity of the examination”.

According to Newble (Citation2004) “The OSCE is not a test method in the same way as an essay or multiple-choice question. It is basically an organization framework consisting of multiple stations around which students rotate and at which students perform and are assessed on specific tasks”. Hodder (Citation1989) and van der Vleuten and Swanson (Citation1990) support this view that many different types of test methods can be incorporated within the OSCE format.

Based on various descriptions of OSCE in the literature we propose a consolidated definition of the OSCE; “An assessment tool based on the principles of objectivity and standardisation, in which the candidates move through a series of time-limited stations in a circuit for the purposes of assessment of professional performance in a simulated environment. At each station candidates are assessed and marked against standardised scoring rubrics by trained assessors”.

Some variants of the OSCE described in use the original format of moving around assessment stations to assess a range of different outcomes ().

Table 3  Variations of the OSCE

Table 4  Affective, cognitive and psychomotor domains of learning from basic to advanced levels

What Does the OSCE Assess?

It is important to understand the relationship of competence and performance before exploring the details of what an OSCE can be used to assess. The OSCE has previously been described in the literature as a tool used for the assessment of competence. This is slightly misleading as according to the Dreyfus & Dreyfus (Citation1980) competence is a point on the spectrum of performance. Any tool used for the assessment of performance should be able to grade the candidates as novice, advanced beginner, competent, proficient or experts in the tasks allocated to them. A key difference between the OSCE and Work Place Based Assessment (WPBA) is the setting in which performance is being assessed, not that the former assesses competence and the latter performance, as is commonly perceived (Boursicot et al. Citation2011). Nevertheless, difference is very significant since the performance of an individual on identical tasks can vary considerably depending on the context of the assessment (ten Cate et al. Citation2010). Therefore, for all practical purposes the OSCE should be seen as a tool for the assessment of performance within simulated environments. Competency is the composite of cognitive, psychomotor and affective skills as appropriate, while competence is an attribute of a person. Assessment of performance on competencies, for example; the ability to communicate or examine the respiratory system allows the assessor to determine if the performer is competent, proficient or expert, etc. A detailed discussion on this topic is beyond the scope of this Guide and we have addressed it in our paper on competency, competence and performance (Khan & Ramachandran Citation2012).

The OSCE and its variants described earlier are only a few from a number of tools available to assess performance in health care. The diagram () describes the use of Workplace Based Assessments and Incognito Patients (Rethans et al. Citation2007) as additional methods for the competency based assessment of performance (CBAP), in different settings.

Figure 3 Model for the competency based assessment of performance and examples of available assessment tools. Reproduced from Khan & Ramachandran (Citation2012).

Figure 3 Model for the competency based assessment of performance and examples of available assessment tools. Reproduced from Khan & Ramachandran (Citation2012).

highlights that assessment at an OSCE station gives a ‘snapshot’ of candidates’ demonstrated performance in a particular area in a simulated environment. With reference to Miller's pyramid of assessment (Miller Citation1990), the OSCE assesses that a candidate is able to ‘show how’ one would perform in a simulated environment (). This contrasts to the other methods, such as ‘incognito patients’, which allows assessment of the candidates’ actual performance in their workplace, at the ‘does’ level of the pyramid.

Figure 4 The OSCE in relation to the Miller's pyramid.

Figure 4 The OSCE in relation to the Miller's pyramid.

The simulated environment itself can influence the performance of candidates; therefore, the performance of the candidates in the OSCE might not be the same as their performance in the workplace on identical tasks (Khan & Ramachandran Citation2012), hence it is important to bear this in mind when interpreting the results. A number of other factors also influence performance in the health care setting as outlined in .

Figure 5 Factors influencing performance. Modified from Khan & Ramachandran (Citation2012).

Figure 5 Factors influencing performance. Modified from Khan & Ramachandran (Citation2012).

In real life, non-clinical skills such as team working, resource management, situational awareness and leadership, etc. also play a very important part in determining the overall performance. It is not very easy to assess these using the OSCE format. Assessment using the OSCE mainly focuses on cognitive, psychomotor and affective skills, described as learning domains. Krathwohl, Bloom and Jewett have described taxonomies of learning in affective, cognitive and psychomotor domains, respectively; each with a range of levels of difficulty. These levels are shown for each of the three learning domains in (Krathwohl Citation1956; Jewett et al. Citation1971; Bloom Citation1974).

It is possible for the OSCE to assess skills within all three learning domains; however, it should be borne in mind that there might be more appropriate methods of assessing certain skills. For instance, the theoretical application of knowledge is better tested by using multiple choice questions (MCQs), and evaluation of knowledge can be assessed better by critical appraisal of a scholarly article. As previously described, an ideal summative assessment programme needs to be based on a combination of assessment tools in a ‘test battery approach’, thus allowing the assessment of cognitive, affective and psychomotor skills, mapped to a range of learning outcomes. It is also important to consider that it is not possible to test all levels of difficulty within each domain with the use of an OSCE, for instance it would be very difficult to plan a station testing the ability of the candidates to ‘evaluate knowledge’ in the cognitive domain.

We believe that, the OSCE should be designed to assess certain competencies or skills which are not possible to be assessed using pen and paper or computer based testing methods. Typical examples of such skills could be the ability of a candidate to take a history or perform a procedure. However, there have been criticisms that the OSCE risks the compartmentalisation of these skills (performance of skills in isolation) for the purposes of assessment (Nestel et al. Citation2011). This results in a loss of ability to assess candidates’ performance holistically. For this reason it is important that performance related skills are not tested in isolation, rather these are blended with skills such as the application of knowledge or formulating a management plan so that performance is assessed more holistically.

Consider stations designed to assess candidates’ ability to interpret an X-ray or to recognise a raised urea and list possible causes; an OSCE station is not required for such assessments. In fact, paper-based methods could assess these cognitive skills reliably, and prove to be less resource intensive. Also consider a station designed to assess candidates’ ability to perform the psychomotor component of a ‘pap smear test’ on a model; such a station risks compartmentalisation as in reality candidates would need to communicate with the patient and apply their knowledge on this procedure whilst they perform it. Poorly structured OSCE stations may lead to candidates learning skills to pass examinations rather than to improve their actual clinical performance. At the same time it is also important to prevent assessment overload by trying to assess too many skill subsets of performance at a single station. The example of such a scenario could be asking the candidates to take a history and examine a patient complaining of haemoptysis, interpret their X-ray findings, and explain the likely diagnosis to the patient. Although this station may reflect ‘real life’ by placing skills within context it is unlikely that assessing all of these skills adequately within the one station is achievable within time constraints.

Educational principles of the OSCE

The two major underlying principles of the OSCE are ‘objectivity’ and ‘structure’. Objectivity predominantly depends on standardised scoring rubrics and the same, trained, examiner asking the same questions to every candidate. A well-structured OSCE station on the other hand has a standardised station design assessing a specific clinical task which is blueprinted against the curriculum. A well-designed OSCE has a high level of validity (Downing Citation2003), which in simple terms means that the OSCE assesses what it is designed to assess. At the same a time well-designed OSCE also has been shown to demonstrate a high degree of reliability (Boursicot Citation2010), i.e. the examination results are reproducible with very little error. The concepts of validity, reliability, feasibility, and educational impact are explored in more detail below.

Validity

Previously different types of validity such as face validity, content validity, and construct validity, etc. were seen as separate entities. More contemporaneously, construct validity is seen as an overarching term covering all types of validity (Downing Citation2010). By this new definition, reliability itself is also seen as an element of construct validity, which was previously seen as a separate entity. It is beyond the scope of this Guide to discuss this in further detail but generally construct validity depends on five sources of evidence (American Educational Research Association Citation1999):

  1. The test content represents what the curriculum needs to assess, the tasks are realistic and the right domains are being assessed.

  2. The responses to the test item are accurately recorded, handled, stored, and analysed.

  3. The test has high reliability.

  4. The test results correlate with other test results, assessing similar domains and show poor correlation with those assessing different domains (convergence and divergence).

  5. The consequences of assessment are sound, i.e. effect on the learning is positive, the penalties for failure are justifiable and the consequences of passing are socially and professionally acceptable.

The above principles are applicable to the OSCE and these are discussed in the second part of this Guide.

Reliability

Many factors influence the reliability of any assessment method and an understanding of these is essential when developing a new assessment programme. The main influences on the reliability of an OSCE are outlined briefly below; each of these areas is discussed in considerable detail in the relevant sections in the second part of this Guide.

  1. The number of stations

Testing candidates across a large sample of clinical cases maximises reliability (Roberts et al. Citation2006) and an appropriate test length has been shown to have the greatest influence in ensuring that candidates’ overall performance is reliably assessed (Swanson et al. Citation1995).

  1. Standardised scoring rubrics

These ensure that examiners are marking candidates against the same criteria thereby improving consistency of scoring between candidates and examiners (Smee Citation2003).

  1. Using trained examiners

There is some evidence suggesting that examiner training reduces examiner variation in scoring (Newble et al. Citation1980; van der Vleuten et al. Citation1989) and improves consistency of examiners’ behaviour. Further using different examiners for different stations can reduce individual assessor bias (Gormley Citation2011).

  1. Standardised patient performance

Poorly standardised patients who vary their performance between candidates can reduce the reliability of the examination (Smee Citation2003).

Feasibility

Feasibility can be defined as the degree of practicality of the assessment intervention. Compared to other assessment methods, the OSCE is seen as more resource intensive and time-consuming to set-up and run. For this reason it is important to ensure that the best use is made of the OSCE by only developing stations that require an OSCE format, i.e. assessment of performance rather than knowledge.

Educational impact

The OSCE can have a positive educational impact as it can drive learning (Boursicot Citation2010). This positive educational impact is dependent on realistic recreation of assessment scenarios at the OSCE stations. If candidates find it difficult to differentiate between the assessment tasks and real life practice then the OSCEs can drive lifelong learning. If the tasks given at OSCE stations are compartmentalised and driven by checklist scoring then the candidates learn to pass exams, decreasing the educational impact of the OSCE (Miller Citation1990; Shumway & Harden Citation2003).

Common uses of the OSCE

In the last three decades the OSCE has seen a steady exponential growth and usage in both undergraduate and postgraduate examinations around the globe. The OSCE is also used for licensure examinations and as a feedback tool in formative settings. Common uses of the OSCE are listed below.

  1. As a performance based assessment tool for testing the minimum accepted standards of students or trainees as barrier (exit) examinations during the undergraduate years in most of the medical schools in the United States, United Kingdom and Canada.

  2. As a postgraduate high stakes assessment tool in Royal College examinations, e.g. the Royal College of Physicians (United Kingdom) uses the Practical Assessment of Clinical Examination Skills (PACES) examination as a component of its membership examinations (Wallace Citation2002).

  3. As a formative assessment tool in undergraduate medical education (Townsend et al. Citation2001).

  4. As a tool for the assessment of graduates seeking high-stakes licensure and certification to practise medicine (Hodges Citation2003). Professional Linguistics and Assessment Board Part II [PLAB] in United Kingdom (Tombleson et al. Citation2000), Medical Council of Canada Qualifying Examination II (Reznick et al. Citation1992), and the Clinical Skills Assessment part of the United States Medical Licensure Examination are based on an OSCE model (Whelan Citation1999).

  5. As an educational tool to provide immediate feedback (Hodder et al. Citation1989; Brazeau et al. Citation2002).

Conclusions

Part I of this Guide has introduced the concept of the OSCE as a tool for performance-based assessment in simulated environments. This part has also described why the OSCE was developed and how it can meet the need for more structure and objectivity within the assessment of performance. We have stressed throughout this Guide that the OSCE should be used only when other methods of assessment cannot assess the competency in question or in conjunction with other assessments of performance. It neither replaces the tools used for assessment of knowledge nor is it a panacea for all the assessment needs in an educational programme. The challenges associated with the academic and practical aspects of implementing an OSCE, particularly when this examination format is novel to the institution are addressed in the Part II of this Guide. Readers equipped with the information in this part of the Guide will find it easy to understand the complex processes needed for the setup of OSCE programmes, described in the second part.

Declaration of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the article.

References

  • American Educational Research Association. Standards for educational and psychological testing. American Educational Research Association, Washington, DC 1999
  • Bloom BS, 1974. Taxonomy of educational objectives: The classification of educational goals handbook 1: Cognitive domain, [S.l.], New York: Longman
  • Bodle JF, Kaufmann SJ, Bisson D, Nathanson B, Binney DM. Value and face validity of objective structured assessment of technical skills (OSATS) for work based assessment of surgical skills in obstetrics and gynaecology. Med Teach 2008; 30: 212–216
  • Boursicot K. Structured assessments of clinical competence. Br J Hosp Med 2010; 71: 342–344
  • Boursicot K, Etheridge L, Setna Z, Sturrock A, Ker J, Smee S, Sambandam E. Performance in assessment: Consensus statement and recommendations from the Ottawa conference. Med Teach 2011; 33: 370–383
  • Brazeau C, Boyd L, Crosson J. Changing an Existing OSCE to a Teaching Tool: The Making of a Teaching OSCE. Acad Med 2002; 77: 932
  • Carraccio C, Englander R. The Objective Structured Clinical Examination. A step in the direction of competency-based evaluation. Arch Pediat Adol Med 2000; 154: 736–741
  • Downing MS. Validity: On the meaningful interpretation of assessment data. Med Educ 2003; 37: 830–837
  • Downing SM, 2010. Validity: Establishing meaning for assessment data through scientific evidence. St George's Advanced Assessment Course, London
  • Dreyfus S, Dreyfus H. A Five stage model of the mental activities involved in directed skill acquisition. Research Paper. Department of Philosophy, University of CaliforniaBerkeley, 1980
  • Epstein RM. Assessment in Medical Education. N Eng J Med 2007; 356: 387–396
  • Fraser RC, Mckinley RK, Mulholland H. Consultation competence in general practice: Establishing the face validity of pritoritised ciriteria in the Leicester assessment package. Br J Gen Prac 1994; 44: 109–113
  • Gleeson F. Assessment of Clinical Competence using the Objective Structured Long Examination Record (OSLER), AMEE Medical Education Guide No. 9. Med Teach 1997; 19: 7–14
  • Gormley G. Summative OSCEs in undergraduate medical education. Ulster Med J 2011; 80: 127–132
  • Gupta P, Dewan P, Singh T. Objective Structured Clinical Examination (OSCE) Revisited. Indian Pediat 2010; 47: 911–920
  • Hamdy H, Prasad K, Williams R, Salih FA. Reliability and validity of the direct observation clinical encounter examination (DOCEE). Med Educ 2003; 37: 205–212
  • Hamdy H, Telmesani AW, Wardy NA, Abdel-Khalek N, Carruthers G, Hassan F, Kassab S, ABU-Hijleh M, AL-Roomi K, O'malley K, et al. Undergraduate medical education in the Gulf Cooperation Council: A multi-countries study (Part 2). Med Teach 2010; 32: 290–295
  • Harden RM. How to Assess Clinical Competence – An overview. Med Teach 1979; 1: 289–296
  • Harden RM. What is an OSCE?. Med Teach 1988; 10: 19–23
  • Harden RM, Cairncross RG. Assessment Of Practical Skills: The Objective Structured Practical Examination (OSPE). Studies High Educ 1980; 5: 187–196
  • Harden RM, Stevenson M, Downie WW, Wilson GM. Assessment of Clinical Competence using Objective Structured Examination. BMJ 1975; 1: 447–451
  • Hays R. Assessment in medical education: Roles for clinical teachers. Clin Teach 2008; 5: 23–27
  • Hodder RV, Rivington RN, Calcutt LE, Hart IR. The effectiveness of immediate feedback during the Objective Structured Clinical Examination. Med Educ 1989; 23: 184–188
  • Hodges B. OSCE! Variations on a theme by Harden. Med Educ 2003; 37: 1134–1140
  • Hodges B, Hanson M, Mchaughton N, Regehr G. Creating, Monitoring, and Improving a Psychiatry OSCE. Acad Psychiat 2002; 26(3)134–161
  • Humphris GM, Kaney S. The Objective Structured Video Exam for assessment of communication skills. Med Educ 2000; 34: 939–945
  • Jayawickramarajah PT. Oral examinations in medical education. Med Educ 1985; 19: 290–293
  • Jewett AE, Jones SL, Luneke SM, Robinson SM. Educational change through a taxonomy for writing physical education objectives. Quest 1971; 15: 32–38
  • Khan K, Ramachandran S. Conceptual Framework for Performance Assessment: Competency, Competence and Performance in the Context of Assessments in Healthcare – Deciphering the Terminology. Med Teach 2012; 34: 920–928
  • Krathwohl DR, 1956. Taxonomy of educational objectives: The classfication of educational goals; handbook 2: Affective domain, [S.l.], New York: Longmans
  • Mcmanus IC, Thompson M, Mollon J. Assessment of examiner leniency and stringency (‘hawk-dove effect') in the MRCP(UK) clinical examination (PACES) using multi-facet Rasch modelling. BMC Medical Education 2006; 6: 42
  • Miller GE. The assessment of clinical skills/competence/performance. Acad Med 1990; 65: S63–S67
  • Nestel D, Kneebone R, Nolan C, Akhtar K, Darzi A. Formative Assessment of Procedural Skills: Students' Responses to the Objective Structured Clinical Examination and the Integrated Performance Procedural Instrument. Assess Eval Higher Educ 2011; 36: 171–183
  • Newble D. Techniques for measuring clinical competence: Objective structured clinical examinations. Med Educ 2004; 38: 199–203
  • Newble DI. The observed long-case in clinical assessment. Med Educ 1991; 25: 369–373
  • Newble DI, Hoare J, Sheldrake PF. The selection and training of examiners for clinical examinations. Med Educ 1980; 14: 345–349
  • Norcini JJ. Death of a long case. BMJ 2002; 324: 408–409
  • Norcini JJ, Mckinley DW. Assessment methods in medical education. Teaching Teach Educ 2007; 23: 239–250
  • Norman G. The long case versus objective structured clinical examinations. BMJ 2002; 324: 748–749
  • PMETB. 2007. Developing and maintaining an assessment system - a PMETB guide to good practice [Online]. London: PMETB. [Accessed 10 June 2012] Available from http://www.gmc-uk.org/assessment_good_practice_v0207.pdf_31385949.pdf
  • Ponnamperuma GG, Karunathilake IM, Mcaleer S, Davis MH. The long case and its modifications: A literature review. Med Educ 2009; 43: 936–941
  • Rethans J-J, Gorter S, Bokken L, Morrison L. Unannounced standardised patients in real practice: A systematic literature review. Med Educ 2007; 41: 537–549
  • Reznick R, Smee S, Rothman A, Chalmers A, Swanson D, Dufresne L, Lacombe G, Baumber J, Poldre P, Levasseur L. An objective structured clinical examination for the licentiate: Report of the pilot project of the Medical Council of Canada. Acad Med 1992; 67: 487–494
  • Roberts C, Newble D, Jolly B, Reed M, Hampton K. Assuring the quality of high-stakes undergraduate assessments of clinical competence. Med Teach 2006; 28: 535–543
  • Shumway JM, Harden RM. AMEE Guide No. 25: The assessment of learning outcomes for the competent and reflective physician. Med Teach 2003; 25: 569–584
  • Singleton A, Smith F, Harris T, Ross-Harper R, Hilton S. An evaluation of the Team Objective Structured Clinical Examination (TOSCE). Med Educ 1999; 33: 34–41
  • Smee S. ABC of learning and teaching in medicine ‘Skill based assessment’. BMJ 2003; 326: 703–706
  • Smith LJ, Price DA, Houston IB. Objective structured clinical examination compared with other forms of student assessment. Arch Dis Child 1984; 59: 1173–1176
  • Sood R. Long Case Examination - Can it be Improved. Indian Acad Clin Med 2001; 2: 251–255
  • Stokes JF. Take a Clinical Examination. BMJ 1979; 1: 98–99
  • Swanson DB, Norman GR, Linn RL. Performance-Based Assessment: Lessons From the Health Professions. Educ Res 1995; 24: 5–11
  • Ten Cate TJO, Snell L, Carraccio C. Medical competence: The interplay between individual ability and the health care environment. Med Teach 2010; 32: 669–675
  • Tombleson P, Fox RA, Dacre JA. Defining the content for the OSCE component of the PLABoard examination development of a blueprint. Med Educ 2000; 34: 566–572
  • Townsend AH, Mcllvenny S, Miller CJ, Dunn EV. The use of an objective structured clinical examination (OSCE) for formative and summative assessment in a general practice clinical attachment and its relationship to final medical school examination performance. Med Educ 2001; 35: 841–846
  • Troncon LA, Dantas RO, Figueiredo JFC, Ferriolli E, Moriguti JLC, Martinelli ALC, Voltarelli JLC. A standardized structured long case examination of clinical competence of senior medical students. Med Teach 2000; 22: 380–385
  • Turner JL, Dankoski ME. Objective Structured Clinical Exams: A Critical Review. Fam Med 2008; 40: 574–578
  • Van Der Vleuten CPM, Swanson DB. Assessment of clinical skills with standardized patients: State of the art. Teach Learn Med 1990; 2: 58–76
  • Van Der Vleuten CPM, Van Luyk SJ, Van Ballegooijen AMJ, Swansons DB. Training and experience of examiners. Med Educ 1989; 23: 290–296
  • Van Hove PD, Tuijthof GJM, Verdaasdonk EGG, Stassen LPS, Dankelman J. Objective assessment of technical surgical skills. Br J Surg 2010; 97: 972–987
  • Vlantis AC, Lee WC, Van Hasselt CA. The objective structured video examination of medical students. Med Educ 2004; 38: 1181–1182
  • Wallace J. Simulated patients and objective structured clinical examinations: Review of their use in medical education. Adv Psychiat Treat 2002; 8: 342–348
  • Walsh K, 2006. OSCE. BMJ Learning. Available from group.bmj.com/products/learning/medical-presentations-2/OSCEs.ppt [Accessed 13 August 2013]
  • Wass V, Jones R, Van Der Vleuten C. Standardized or real patients to test clinical competence? The long case revisited. Med Educ 2001a; 35: 321–325
  • Wass V, Van Der Vleuten C. The long case. Med Educ 2004; 38: 1176–1180
  • Wass V, Van Der Vleuten C, Shatzer J, Jones R. Assessment of clinical competence. Lancet 2001b; 357: 945–949
  • Whelan GP. Educational commission for Foreign Medical Graduates - clinical skills assessment prototype. Med Teach 1999; 21: 156–160
  • Wilson GM. Examination of clinical examiners. Lancet 1969; 4: 37–40
  • Zayyan M. Objective Structured Clinical Examination: The Assessment of Choice. Oman Med J 2011; 26(4)219–222

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.