Connect
MJA
MJA

The mini clinical evaluation exercise (mini-CEX) for assessing clinical performance of international medical graduates

Balakrishnan R Nair, Heather G Alexander, Barry P McGrath, Mulavana S Parvathy, Eve C Kilsby, Johannes Wenzel, Ian B Frank, George S Pachev and Gordon G Page
Med J Aust 2008; 189 (3): 159-161. || doi: 10.5694/j.1326-5377.2008.tb01951.x
Published online: 4 August 2008

Assessing the performance of junior doctors in the workplace is important but challenging. The optimum assessment is by direct observation of doctors’ interactions with patients and comprises multiple assessments by multiple examiners on a variety of patient problems. Clinical supervisors are best suited to observe and certify trainees, but often do not observe them directly.1 Performance assessment is not done well in most instances, as it requires multiple sampling over time.2 In-training assessments done at the end of a term introduce a “halo effect”.3

Most of these problems can be overcome by the mini clinical evaluation exercise (mini-CEX), developed by the American Board of Internal Medicine.4 The mini-CEX involves direct observation of a trainee in a focused clinical encounter, followed by immediate feedback. The assessment is recorded on a rating form that has been shown to have high internal consistency and reliability among internal medicine trainees, giving scores comparable with a high-stake clinical examination.5,6 The mini-CEX has higher fidelity than other formats.7

International medical graduates (IMGs) comprise about 25% of the medical workforce in developed countries.8 Their certification for registration is a major task of the medical boards and registration authorities in Australia and other countries.9 The Australian Medical Council (AMC) has conducted clinical examinations to assess IMGs since 1978.10 Successful candidates undertake 12 months of supervised practice before obtaining full registration. Despite having passed the current AMC clinical examination, IMGs’ competence and performance in the workplace have been criticised.11

We evaluated the feasibility, reliability and acceptability of the mini-CEX as a performance assessment tool for IMGs in the Australian workplace.

Methods

The study was conducted in three large metropolitan teaching hospitals in Australia, one each in New South Wales, Queensland and Victoria, as part of a larger international collaborative study with the Medical Council of Canada. The ethics committee in each centre approved the study.

Statistical analysis

Reliability was assessed using generalisability theory analysis. Generalisability analyses (G studies) allow estimation of the variance components associated with the different examination conditions (eg, types of tasks, number of tasks, and number of markers). The ratio of the variance component for the object of measurement (in this case, differences between IMGs) to the total error variance yields an estimate of reliability: the generalisability (G) coefficient, with values ranging from 0 to 1. The effect of changes in the examination conditions (eg, increase in the number of tasks to be performed by the IMG, or change in the number of markers for each task) can be modelled to inform decisions on optimising the measurement (decision [D] studies).

We used a G study followed by a D study to evaluate the reliability of the measurement, and sought to determine the number of observed clinical encounters needed to attain a G coefficient of 0.90. This value was set by taking into account the high stakes of the assessment for the IMGs. The average rating for the different assessments within each encounter (eg, history taking, examination) was the outcome measure for these analyses. As each IMG interacted with different patients and a different number of patients, the “patient-encounters” factor was treated as nested within IMGs in the G-study design.

We used urGENOVA statistical software (Robert Brennan, Center for Advanced Studies in Measurement and Assessment, University of Iowa, Iowa, USA) for the analyses.

Results

All 28 IMGs who had passed the AMC examinations within the previous 12 months and 35 examiners volunteered to participate in the study. Twenty-two examiners were trained in assessing the mini-CEX; the remaining 13 examiners participated without training. The examiners included specialists and specialist trainees in internal medicine, surgery and emergency medicine.

Discussion

Under the conditions and settings used, the mini-CEX reliably assessed the clinical performance of IMGs with eight to 10 encounters. This is consistent with the results of other studies.7 As the mini-CEX is conducted within the workplace with real patients, it has high fidelity and it is acceptable to both IMGs and examiners. A fail rate of 9% (19/209 encounters) across 12 IMGs is concerning, given these IMGs had passed the AMC clinical examination.

As we have demonstrated, the mini-CEX appears to be a reliable and acceptable assessment tool of clinicians in the workplace, and is a valuable method of identifying which candidates may have problems in a clinical situation. Most IMGs were satisfied with their feedback. Examiners reported that this was the most important part of the mini-CEX. However, they would have preferred more training in this task. Overall, examiners felt this tool assesses clinical performance better than the conventional methods.

There are limitations to this study. Only patients from medical, surgery and emergency departments were included, and the results cannot automatically be extended to other specialties. The IMGs were a self-selected group, introducing a possible bias, and only a small group of experienced clinicians took part, limiting the generalisability of the study.

Further analyses will need to consider the effect of examiner training, and the relationship between performance on the mini-CEX with that in other assessments.

The mini-CEX is a feasible, reliable and high-fidelity instrument for workplace-based assessment of IMGs with the strong advantage of providing ongoing observation and feedback. It has the potential to be used for summative assessment of IMGs and other medical trainees. The AMC, in collaboration with some licensing bodies, has already introduced the mini-CEX as a workplace assessment tool for some IMGs.

  • Balakrishnan R Nair1,2
  • Heather G Alexander3
  • Barry P McGrath4
  • Mulavana S Parvathy2
  • Eve C Kilsby3
  • Johannes Wenzel5
  • Ian B Frank6
  • George S Pachev7
  • Gordon G Page7

  • 1 University of Newcastle, Newcastle, NSW.
  • 2 John Hunter Hospital, Newcastle, NSW.
  • 3 Griffith Institute for Higher Education, Griffith University, Brisbane, QLD.
  • 4 Monash University, Melbourne, VIC.
  • 5 Southern Health, Melbourne, VIC.
  • 6 Australian Medical Council, Canberra, ACT.
  • 7 University of British Columbia, Vancouver, British Columbia, Canada.


Correspondence: kichu.nair@newcastle.edu.au

Acknowledgements: 

The study was supported by a grant from the AMC. We thank all the supervisors, international medical graduates and patients.

Competing interests:

None identified.

  • 1. Holmboe ES. Faculty and the observation of trainees’ clinical skills: problems and opportunities. Acad Med 2004; 79: 16-22.
  • 2. Williams RG, Dunnington GL, Klamen DL. Forecasting residents’ performance — partly cloudy. Acad Med 2005; 80: 415-422.
  • 3. Wilkinson TJ, Wade WB. Problems with using a supervisor’s report as a form of summative assessment. Postgrad Med J 2007; 83: 504-506.
  • 4. Norcini JJ, Blank LL, Arnold GK, Kimball HR. The mini-CEX (clinical evaluation exercise): a preliminary investigation. Ann Intern Med 1995; 123: 795-799.
  • 5. Durning SJ, Cation LJ, Markert RJ, Pahgaro LN. Assessing the reliability and validity of the mini-clinical evaluation exercise for internal medicine residency training. Acad Med 2002; 77: 900-904.
  • 6. Hatala R, Ainslie M, Kassen BO, et al. Assessing the mini-Clinical Evaluation Exercise in comparison to a national specialty examination. Med Educ 2006; 40: 950-956.
  • 7. Norcini JJ, Blank LL, Duffy FD, Fortna GS. The mini-CEX: a method for assessing clinical skills. Ann Intern Med 2003; 138: 476-481.
  • 8. Mullan F. The metrics of the physician brain drain. N Engl J Med 2005; 353: 1810-1818.
  • 9. McGrath BP. Integration of overseas medical graduates into the Australian medical workforce. Med J Aust 2004; 181: 640-642. <MJA full text>
  • 10. Breen K, Frank I, Walters T. Australian Medical Council: the view from inside. Intern Med J 2001; 31: 243-248.
  • 11. Van Der Weyden MB, Chew M. Arriving in Australia: overseas trained doctors [editorial]. Med J Aust 2004; 181: 633-634. <MJA full text>

Author

remove_circle_outline Delete Author
add_circle_outline Add Author

Comment
Do you have any competing interests to declare? *

I/we agree to assign copyright to the Medical Journal of Australia and agree to the Conditions of publication *
I/we agree to the Terms of use of the Medical Journal of Australia *
Email me when people comment on this article

Online responses are no longer available. Please refer to our instructions for authors page for more information.