529
Views
4
CrossRef citations to date
0
Altmetric
ORIGINAL ARTICLE

Regulatory end-point assessment of the consultation competence of family practice trainees in Kuwait

, , &
Pages 100-107 | Received 18 Aug 2004, Published online: 20 Mar 2010

Abstract

Background: No single approach to the regulatory assessment of global consultation competence has been shown to possess the required levels of validity, reliability and feasibility. Objective: To evaluate the approach adopted in Kuwait to the regulatory end-point assessment of the global consultation competence of family practice trainees with particular reference to validity, reliability and feasibility. Methods: Family practice trainees in Kuwait were individually and directly observed for 3 hours in consultation with a minimum of 10 patients by a pair of examiners. Performance was judged against the explicit criteria of consultation competence as contained in the Leicester Assessment Package (LAP). Results: The marks independently allocated by the pairs of examiners to 126 trainees between 1994 and 2001 were within five percentage points on 91% of occasions. A reliability coefficient of 0.82 was achieved when two examiners independently marked candidates consulting with 10 real patients; this rose to 0.95 at the critical 50% pass–fail margin. The main sources of variance contributing to the reliability of marks allocated were candidate performance (42%) and the interaction of candidate performance across cases, i.e., case specificity (30%). The clinical challenges presented by the patients were judged by both examiners to be sufficient to enable performance to be assessed across the seven LAP consultation categories as follows: behaviour and relationship with patients (100% of consultations), interviewing/history taking (100%), record keeping (99%), patient management (99%), problem solving (98%), physical examination (95%), and anticipatory care (86%). Each assessment involved a pair of examiners and lasted approximately 3.5 hours.

Conclusion: The Kuwait clinical examination achieves high content validity and authenticity as it uses direct observation of performance, validated and explicit criteria against which performance is judged, and real patient challenges. It can discriminate between different levels of consultation performance and satisfies the recognized reliability threshold for regulatory examinations (0.82 vs 0.80). Accordingly, we recommend the use of such an approach in the regulatory end-point assessment of the global consultation competence of trainees in family practice. Such an approach is more valid, and is likely to be more feasible, than simulated surgeries or the short-case OSCE format.

Introduction

Since the core activity in clinical practice is the consultation between doctor and patient Citation[1], it follows that the focal point of the assessment of clinical competence must be the systematic observation and analysis of the performance of a clinician in the consultation Citation[2]. To date, no single procedure for the regulatory assessment of global consultation performance has been shown to possess the required levels of validity, reliability and feasibility.

We now report a study whose aims were to evaluate the regulatory end-point assessment of the global consultation competence of trainees in the Kuwait Family Practice Training Programme with particular reference to content validity and reliability (contributions to variance and generalizability).

Background

The Kuwait Family Practice Training Programme Citation[3], which began in 1984, is a collaborative venture between the Royal College of General Practitioners (RCGP) and the Kuwait Ministry of Health. Following a selection process, all trainees undergo a full-time speciality training programme lasting 3 years.

Trainees initially spend 3 months in a family practice training centre, followed by 24 months in a rotation of hospital posts, during which they attend a weekly day-release programme supervised by family practice trainers. The final 9 months are spent in a family practice training centre.

Before being eligible to undertake independent clinical practice, trainees have to pass an end-point regulatory assessment which consists of three written papers (45% of available marks) and clinical (30%) and oral (25%) examinations. The overall pass mark is 60%, but candidates must also achieve a minimum mark of 50% in each of the three components. Internal and external examiners (the latter nominated by the RCGP and approved by the Kuwait Examination Board) share examining duties. Successful candidates are awarded the Diploma in Family Practice (RCGP/Kuwait).

The assessment tool used was the Leicester Assessment Package (LAP) Citation[4], Citation[5], which was developed to facilitate educational and regulatory assessment and enhancement of the global consultation competence of a doctor in the context of general practice. The LAP is an integrated assessment tool which contains seven prioritized categories of consultation competence and 39 component competences (see Box 1). The LAP criteria, against which consultation performance is judged, have been demonstrated to be valid for general practice in the United Kingdom Citation[6] and Hong Kong Citation[7]. The LAP also contains descriptive criteria to assist assessors in the allocation of marks (see Box 2). Use of the LAP requires consulting doctors at various stages in the consultation to answer certain questions to enable a teacher or examiner to become more aware of the reasoning which underpins the doctor's actions (see Methods).

Box 1. Diploma in Family Practice (RCGP/Kuwait): consultation competences assessed in the clinical examination.

Box 2. Criteria for the allocation of marks.

The LAP can be used with “real” or simulated patients and with “live” or videotaped consultations. It can also facilitate the achievement of reliable assessments of doctors consulting with simulated patients in an experimental situation Citation[8] and in identifying medical students with unsatisfactory levels of consultation performance Citation[9].

Methods

The procedure for the Kuwait clinical examinations is laid down in an Examiners’ Code of Practice by the Kuwait Family Practice Examination Board, and all examiners are required to follow it.

The clinical component consists of 3 hours of direct and independent observation by two examiners of candidates in consultation with a minimum of 10 consenting but unselected patients in a Family Practice Training Centre. There is one internal and one external examiner. The criteria against which clinical performance is assessed are specified in Box 1, and marks are allocated according to the criteria specified in Box 2. Whenever possible—and with the additional permission of the patient—examiners directly observe candidates conducting physical examinations and/or using instruments, and personally verify positive physical findings.

In each consultation, the examiner is required to seek clarification from the candidate by asking (in English) the following questions:

  1. At the completion of (initial) history taking (the candidate to inform the examiner when this has occurred):

    • What are your diagnostic hypotheses at this stage?

    • Why have you erected these hypotheses?

    • What physical examination do you intend to carry out, and why?

  2. After any physical examination:

    • What did you find on examination of the patient?

    • How have these findings affected your thoughts?

  3. After the patient has left:

    • Why did you choose your management plan?No comments are allowed, and examiners must not enter into discussion with candidates.

Since the consultations are conducted mainly in Arabic, candidates are required to articulate a brief summary in English of the salient points of the history prior to the asking of the first LAP questions. Furthermore, an external examiner has the discretion to consult his internal examiner colleague to seek clarification on matters of fact relating to candidate/patient interactions. In order to maintain the independence of judgement of external examiners, internal examiners are not permitted to provide opinions.

According to the procedures described above, the examiners independently allocated marks to reflect the candidates’ performance in every LAP consultation category challenged in individual consultations. The sum of the consultation category marks represented the global performance in individual consultations. If a consultation category was not challenged in a particular consultation, the denominator was altered appropriately. For example, if anticipatory care had not been challenged, the denominator would be 90 instead of 100 in that consultation. All marks for individual consultations had to be allocated before the next consultation began.

Once the required number of consultations had been observed, the allocated marks were transferred to a master mark sheet. Examiners then awarded final marks to reflect overall performance in all seven categories of competence in turn. Since examiners were required to take account of the nature and difficulty of the clinical challenges presented, the overall mark did not automatically represent the average of marks awarded for individual consultations. The final mark was the sum of the marks allocated by each examiner divided by two.

To estimate reliability, a generalizability analysis Citation[10], Citation[11] was carried out with candidates (P), consultations (C) and examiners (A) as factors and sources of variance. Since real (i.e., unstandardized) patients were used, different candidates were assessed dealing with differing sorts of patient challenges. Thus, case variance was nested within candidate variance (C:P). Furthermore, the same examiners were used across all cases for an individual candidate, but different examiners were allocated to different candidates (A:P). Variance components were estimated using this design and subsequently generalizability coefficients were computed for a number of different samples of assessors and cases. (The complexity of clinical assessment procedures creates problems in both testing and determining their reliability. Although further explanation is provided in the Discussion, readers interested in more details of the educational and statistical principles underpinning the complex methodology involved can consult Fraser et al. Citation[8].)

Content validity was investigated by determining the extent to which the LAP categories of consultation competence were sufficiently challenged by the real patients encountered to enable the assessors to arrive at judgments of actual clinical performance of candidates.

The local examiners were all senior trainers and experienced clinicians who had received formal training in the use of the LAP for regulatory purposes, occupying a minimum of 7 days. The external examiners were all experienced examiners and GPs in the UK. All had received a minimum of 2 days’ specific training, and all had undergone a programme to familiarize them with the organization and delivery of training and clinical care within Kuwait and with local culture and tradition. Several of the examiners had made multiple visits to Kuwait. All examiners also attended a formal briefing session immediately prior to every examination diet.

Results

Between 1994 and 2001, 126 clinical assessments were conducted, involving a total of 1328 patient consultations. There was a range of 6–21 candidates per diet. The assessments were performed by 11 internal and nine RCGP external examiners who examined a range of 1–28 candidates and 3–29 candidates, respectively. The mean global consultation score was 57% (range 45–72%, SD 6.8). The standard deviations for the scores allocated by internal and external examiners were 6.4 and 7.1, respectively. Nineteen candidates (15%) were allocated scores below 50%, and three candidates (2.4%) achieved scores above 70%. The scores independently and individually allocated by paired examiners were identical in five assessments (4%), within two percentage points of each other in 57% of assessments and within five percentage points in 91%. The maximum difference in scores allocated to an individual candidate was 9.4 percentage points.

sets out the contribution of each source of variance to the reliability of the marks allocated. The variance calculations are based on the case-specific data which were collected from 1997 onwards involving 73 candidates. More than two-fifths (42%) of all varience were attributed to the variance between candidates, while less than one-third was associated with the cases (30%). Only 10% were the confounded effect of the case and the examiners.

Table I.  Contribution of each source of variance to the reliability of the marks allocated.

shows reliability (generalizability) coefficients as a function of the number of consultations and examiners. The actual numbers used were 10 consultations with two examiners, leading to a reliability of 0.82. The same reliability can be achieved by reducing the number of consultations to six if the number of examiners is increased to three. The reliability coefficient increased to 0.95 at the critical pass–fail level of 50%.

Table II.  Generalizability coefficients as a function of the number of consultations (cases) and examiners, using the same examiners rating across all consultations for a single candidate.

The clinical challenges presented by the patients were judged by both assessors to be sufficient to enable performance to be assessed across the seven consultation categories as follows: behaviour and relationship with patients (100% of consultations), interviewing/history taking (100%), record keeping (99.2%), patient management (99.2%), problem solving (97.8%), physical examination (94.8%), and anticipatory care (85.9%).

Each candidate consumed approximately 7 hours of assessor time: 3 hours per examiner of direct observation plus up to half-an-hour for collating component marks and arriving at final mark allocations. To this has to be added travelling time to the candidate's health centre and examiner training and briefing (see Methods).

Discussion

“In high stakes (regulatory) assessments, credibility of the method is of major importance” Citation[12]. Primarily, this means that the assessment must be both valid and reliable Citation[13], although compromises usually have to be made in the “real world” on the grounds of feasibility Citation[13]. Nevertheless, “validity is the sine qua non of assessment” Citation[14], and “all assessments in medical education require evidence of validity” Citation[14].

For an assessment to be valid, it must measure what it is supposed to measure. In the context of the Kuwait assessments, this means that the criteria against which consultation competence is judged should be professionally important and relevant, and the nature of the clinical challenges encountered should be suitable, reflecting day-to-day practice as closely as possible Citation[13]. Furthermore, the assessment process must allow any examiner to directly observe the consultation performance of the candidate at all times Citation[15]. The Kuwait examination satisfies all these validity criteria.

Although the criteria of consultation competence against which performance in the Kuwait clinical examinations was judged have been formally validated in the UK Citation[6] and Hong Kong Citation[7], this has not been specifically done in Kuwait. Nevertheless, despite the differences between the healthcare systems, “the close similarity in the acceptance of the LAP criteria in Hong Kong and the UK support the conceptualisation of the LAP as a generic assessment tool that can be applied to consultations in widely differing settings” Citation[7]. Furthermore, the LAP has proved both acceptable and useful since the early days of the Kuwait Family Practice Training Programme for both formative and regulatory purposes Citation[3]. Trainees, trainers and local examiners have made repeated use of the LAP and are very familiar with, and have made no changes to, the 39 LAP criteria of consultation competence. Furthermore, because the LAP criteria are generic, i.e., not case specific, this makes it very unlikely that any important aspect of the consultation would be overlooked. Consequently, it is reasonable to claim that the criteria of consultation competence used in the Kuwait clinical examinations are also valid in that context, since they command such consensual support from local experts. After all, “validity is judgemental … to hunt for validity in statistical procedures is to pursue a will o’ the wisp” Citation[16].

Direct observation of actual performance in daily practice with real patients is the most authentic and valid approach to the assessment of consultation competence as it most closely represents real clinical practice Citation[12], Citation[17], Citation[18]. Indeed, a systematic review Citation[15] of published articles (1966–2001) on the validity and reliability of measures of clinical competence of physicians, medical students and residents identified the direct observation of trainees “in real-life situations” as the most valid form of assessment. Nevertheless, unlike Kuwait, “few assessments observe trainees in real-life situations” Citation[15].

Although using real patients is highly authentic, it has the potential to compromise the content validity of the clinical challenge Citation[19]. Perhaps the optimum method for overcoming this potential problem is the use of a blueprint with specific criteria for selecting a representative sample of patient challenges Citation[12]. Our approach was to record the proportion of consultations in which each of the seven LAP categories of consultation competence was deemed by the assessors to have been sufficiently challenged for them to make a judgement of the consulting doctor's performance. This condition was satisfied in over 97.8% of consultations for five of the seven consultation categories. In respect of the other two categories, it is well recognized and accepted that a physical examination does not need to be conducted in every consultation, and that anticipatory care opportunities do not exist in all consultations. Accordingly, it is reasonable to conclude that 10 patients can provide a valid set of assessable challenges, although a minimum of eight patients produces reliable results. Nevertheless, it cannot be guaranteed that every candidate in the Kuwait examination encountered a totally representative sample of clinical challenges or identical degrees of difficulty. However, examiners were required to take account of the nature and difficulty of the clinical challenges in making their judgements (see Methods). Although the extent to which this occurred was not quantified, the difficulty of the cases has been factored into the generalizability analysis (see below).

“Reliability is defined as the extent to which a result reflects all possible measurements of the same construct” Citation[10]. In the Kuwait examination, the “construct of interest” Citation[10] is the candidate's consultation competence. The analysis of variance components facilitate measurement of all the possible sources of error in the assessment process (see ). With a reliable assessment instrument, differences in scores should reflect true differences between candidates (P). The other sources of variance in any assessment of consultation performance are the influence of the examiners (A) and the nature and difficulty of the clinical challenges (C), i.e., case specificity. Combinations of any of these sources of variance can cause distortion of true scores.

Determining true variations in candidate performance is the principal function of a regulatory assessment in order to accurately identify candidates who deserve to pass or fail. In Kuwait, over two-fifths (42%) of all the variance in the reliability of marks allocated can be attributed to the differences in performance between candidates. Accordingly, the Kuwait examination succeeded satisfactorily in discriminating between candidate performance. The influence of the nature and difficulty of the patient challenge (C) and the effects of the interaction between candidates across cases (C:P) was less than one-third (30%). Although both effects cannot be disentangled because of the nesting of cases within candidates, the reported influence of case specificity on performance is frequently much larger Citation[13]. Indeed, it is accepted wisdom that “professional behaviour is highly dependent upon the nature and details of the problem being faced” Citation[10]. On the other hand, the examiner contribution to variance is minimal, both nested within candidates (A:P = 10%) and in the confounded influence of the case by examiners within candidates (plus all other non-explained residual variance) (CA:P = 18%). Overall, the relative contribution of these variance components indicates that the Kuwait examination is able to discriminate very well in the assessment of consultation competence of candidates. Unfortunately, comparing variances across studies with different designs is extremely difficult, if not impossible. Nevertheless, the small differences in marks independently allocated to individual candidates by the examiners were virtually identical to those achieved in assessments of clinical competence using the Objective Structured Long Examination Record (OSLER) Citation[20], Citation[21] and when using the LAP in a regulatory assessment in Hong Kong Citation[22].

Although “the relative size of the separate sources of variation (variance) provide rich information in their own right, they can also be combined using equations … to express the extent to which the result reflects all possible measurements of the construct of interest. The result is a fraction between zero and one called the generalisability coefficient (G). It integrates the discriminating ability of the test and the reproducibility of the result. Essentially, it provides a measure of how confident you can be that any differences detected between assesses are real differences … simply because it takes account of all possible sources of error at the same time” Citation[10]. By mathematically modelling G in different hypothetical scenarios, it is possible to estimate G with different numbers of observers and cases (analogous to the power calculation in an intervention trial) Citation[10].

By convention, a regulatory assessment should achieve a reliability coefficient (G) of 0.8. The Kuwait examination is able to achieve this with a minimum of eight consultations and two examiners. It is likely that the following factors contributed to the achievement of this level of reliability:

  • As a consequence of their systematic programme of preparation, all examiners were familiar with the content and application of the assessment package and the local context.

  • The LAP contains explicit descriptors of performance linked to mark allocations to aid calibration in the allocation of marks by examiners (see Box 2).

These two measures have been demonstrated to reduce subjectivity and the potential for bias in examiners Citation[23].
  • The responses to the LAP questions provide examiners with major insights into the thinking processes underpinning the actions of the consulting doctors. This advantage is unique to the LAP.

  • The inclusion of 10 different patient challenges increases both the reliability and validity of the assessments. Increasing the number of cases makes it more likely that examiners will be exposed to the true performance of the consulting doctor, and minimizes the impact of the well-recognized variations in physician performance which occur from case to case Citation[24].

  • The LAP contains seven categories of consultation competence and 39 component competences (see Box 1); the inclusion of so many explicit and different aspects of consultation competence also adds to increased validity and reliability Citation[25].

Concerns about the reliability of assessments involving unstandardized “long cases” have focussed on three main influencing factors: “case specificity, examiner stringency and the aspects of a competence evaluated” Citation[26]. Although these concerns have principally been addressed to long cases in a hospital setting, they equally apply in the setting of family practice. If all these areas are addressed, as they have been in the Kuwait clinical examinations, problems of reliability can be successfully overcome without resorting to less valid methods of assessing consultation competence, such as simulated surgeries Citation[27], Citation[28] and objective structured clinical examinations (OSCEs) Citation[29]. Our study shows that reliable inferences can be achieved in assessments involving real (i.e., unstandardized) patients, providing adequate sampling is undertaken. This confirms previous research in a hospital setting Citation[30] and provides further evidence that concerns over the reliability of long-case assessments can be overcome. Indeed, in terms of reliability, a recent review of the relevant evidence has concluded that “… for equal testing time, the long case turns out a bit better than the OSCE” Citation[31].

There are also some well-documented concerns about the use of simulated patients and OSCE stations in assessments which render them less valid than direct observation of the global consultation competence of physicians in daily practice with real patients Citation[15], Citation[32]. Although both techniques can be reliable Citation[15], OSCE stations usually test fragmented components of consultation competence Citation[29], and short OSCE stations, in particular, risk “trivialisation of isolated elements of what should be a coherent whole” Citation[15]. It has also been demonstrated that “simulated patients are limited in the number and complexity of the medical problems they can portray” Citation[18]. Furthermore, “standardised patient assessments … under-emphasise important domains of professional competence”, such as “content of care” and “patient–physician relationships” Citation[15]. There are also potential problems of feasibility in the use of simulated patients. For example, “[a]lthough few cases are needed to assess straightforward skills, up to 27 cases may be necessary to assess inter-personal skills reliably in high stakes examinations” Citation[15]. This can be achieved with as few as eight patient encounters using the approach adopted in Kuwait. Accordingly, we would argue in favour of using the most authentic form of assessment, i.e., real-life consultations being directly observed and holistically judged against a sufficient number of relevant criteria.

The approach to consultation assessment in Kuwait is also feasible to conduct as each assessment requires approximately 3.5 hours×2 of examiner time per candidate, excluding travelling time. As the assessments are based on largely unselected real patients, time required for their development and maintenance is minimal compared with OSCE stations or simulated surgeries. Examiner training and briefing would be required for all three approaches.

Conclusion

The direct observation and assessment of a trainee family practitioner in consultation with 10 real patients by two independent examiners using the LAP has content validity, produces reliable results and is feasible to conduct. Accordingly, we would recommend the use of such an approach in the regulatory end-point assessment of the global consultation competence of trainees in family practice. This approach is more valid, and is likely to be more feasible, than simulated surgeries or short-case OSCE stations, although all three approaches can achieve satisfactory levels of reliability.

We gratefully thank all external and internal examiners for their cooperation in collecting the additional data required for this study, and Ron Hoogenboom for his contribution to the data processing.

References

  • Spence J. The need for understanding the individual as part of the training and function of doctors and nurses. National Association for Mental Health. The purpose and practice of medicine. Oxford University Press, Oxford 1960; 271–80
  • Hager P, Gonczi A, Athanasou J. General issues about assessment of competence. Assessment & Evaluation in Higher Education 1994; 19: 3–15
  • Fraser RC. Developing family practice in Kuwait. Br J Gen Pract 1995; 45: 102–6
  • Fraser RC, McKinley RK, Mulholland H. Assessment of consultation competence in general practice: the Leicester Assessment Package. Approaches to the assessment of clinical competence. Part 1, RM Harden, IR Hart, H Mulholland. Centre for Medical Education, Dundee 1992; 192–8
  • Fraser RC. Assessment of consultation competence. The Leicester Assessment Package2nd ed. Glaxo Medical Fellowship, Macclesfield 1994
  • Fraser RC, McKinley RK, Mulholland H. Consultation competence in general practice: establishing the face validity of prioritized criteria in the Leicester assessment package. Br J Gen Pract 1994; 44: 109–13
  • Lau J, Fraser RC, Lam CLK. Establishing the content validity in Hong Kong of the prioritised criteria of consultation competence in the Leicester Assessment Package. The Hong Kong Practitioner 2003; 25: 596–602
  • Fraser RC, McKinley RK, Mulholland H. Consultation competence in general practice: testing the reliability of the Leicester assessment package. Br J Gen Pract 1994; 44: 293–6
  • McKinley RK, Fraser RC, Van der Vleuten C, Hastings AM. Formative assessment of the consultation performance of medical students in the setting of general practice using a modified version of the Leicester Assessment Package. Med Educ 2000; 34: 573–9
  • Crossley J, Davies H, Humphris G, Jolly B. Generalisability: a key to unlock professional assessment. Med Educ 2002; 36: 972–8
  • Colliver JA, Verhulst SJ, Williams RG, Norcini JJ. Reliability of performance testing on standardised patients: a comparison of consistency measures based on generalisability theory. Teach Learn Med 1983; 1: 31–7
  • Ram P, Grol R, Rethans JJ, Schouten B, VD Vleuten C, Kester A. Assessment of general practitioners by video observation of communicative and medical performance in daily practice: issues of validity, reliability and feasibility. Med Educ 1999; 33: 447–54
  • Van der Vleuten CPM. The assessment of professional competence: developments, research and practical implications. Adv Health Sci Educ Theory Pract 1996; 1: 41–67
  • Downing SM. Validity: on the meaningful interpretation of assessment data. Med Educ 2003; 37: 830–7
  • Epstein RM, Hundert EM. Defining and assessing professional competence. JAMA 2002; 287: 226–35
  • Stones E. Assessment of a complex skill: improving teacher education. Assessment in Education 1994; 1: 235–51
  • Campbell LM, Howie JGR, Murray TS. Use of videotaped consultations in summative assessment of trainees in general practice. Br J Gen Pract 1995; 45: 137–41
  • Norcini J. The death of the long case?. Br Med J 2002; 324: 408–9
  • Caulford PG, Lamb SB, Kaigas TB, Hanna E, Norman GR, Davis DA. Physician incompetence: specific problems and predictors. Acad Med 1994; 69(Suppl 10)16–8
  • Gleeson F. AMEE medical education guide No. 9. Assessment of clinical competence using the Objective Structured Long Examination Record (OSLER). Med Teach 1997; 19: 7–14
  • Gleeson F. Pers. comm, 1994.
  • Fraser RC, Lee RSY, Yiu YK, Lam CLK, McKinley RK, van der Vleuten CPM. Regulatory assessment of the consultation competence of family physicians in Hong Kong. The Hong Kong Practitioner 2004; 26: 5–15
  • Kane MT. The assessment of professional competence. Eval Health Prof 1992; 15: 163–82
  • Elstein AS, Shulman LS, Sprafka SA. Medical problem solving. An analysis of clinical reasoning. Harvard University Press, Cambridge, MA 1978
  • Ramsay PG, Wenrich Carline JD, Inui TS, Larson EB, LoGerfo JP. Use of peer ratings to evaluate physician performance. JAMA 1993; 269: 1655–60
  • Norcini J. The validity of long cases. Med Educ 2001; 35: 720–1
  • Allen J, Evans A, Foulkes J, French A. Simulated surgery in the summative assessment of general practice training: results of a trial in the Trent and Yorkshire regions. Br J Gen Pract 1998; 48: 1219–23
  • Burrows PJ, Bingham L. The simulated surgery – an alternative to video-taped submission for the consulting skills component of the MRCGP examination: the first year's experience. Br J Gen Pract 1999; 49: 269–72
  • Harden RM. What is an OSCE?. Med Teach 1988; 10: 19–22
  • Wass V, Jones R, Van der Vleuten C. Standardized or real patients to test clinical competence? The long case revisited. Med Educ 2001; 35: 321–5
  • Norman G. The long case versus objective structured clinical examination. Br Med J 2002; 324: 748–9
  • Ram P. Comprehensive assessment of general practitioners: a study on validity, reliability and feasibility. PhD Thesis, University of Maastricht, 1998.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.