Medical education has finally got political, a fact that informs Brown & Doshi's analysis but is never named (Reference Brown and DoshiBrown & Doshi, 2006, this issue). What they omit to mention is that doctors and the medical Royal Colleges are positioning themselves to deal with some of the most radical changes since the National Health Service (NHS) began, involving the restructuring of what we expect of doctors, the redefining of boundaries with other healthcare staff and some redistribution of power from the Royal Colleges to the newly created Postgraduate Medical Education and Training Board (PMETB). Changing the medical profession has been on the political agenda at least since the Thatcher era, but it took a series of high-profile cases of medical bad behaviour (Bristol, Alder Hey, Shipman, etc.) to create the head of steam necessary to bring it about. The most significant change is the long overdue shift to defining doctors in terms of their competencies or skills. These will now become the outcome measure of training, supported by relevant curricula and assessment programmes. Incidentally, in this we seem a little behind Canada and the USA (Reference Scheiber, Kramer and AdamowskiScheiber et al, 2003), but ahead of Europe and Australia. Hence, the actual performance of doctors now lies at the heart of the new training arrangements (Postgraduate Medical Education and Training Board, 2004) – something that many of us would say has always been the case, but sometimes in spite of, rather than because of, the curricula and assessment systems to date.
A special specialty
Despite being an enthusiast for many of the changes, I have reservations that I will push further than Brown & Doshi were inclined to. To start with, how appropriate is it to assume that we can adopt American assessment tools designed for non-psychiatric medical situations that have never been evaluated in UK psychiatry? Arguably, for example, specialist psychiatric assessment is less dominated by the differential diagnostic and treatment choices that characterise medical assessment, and is more focused on the fine-grained interpretation of complex biopsychosocial data, weaving this into case formulations that inform treatment and care management decisions. This process of critical analysis and judgement can't be readily assessed using the mini-CEX (which focuses on what the fly on the wall sees of the consultation) or the case-based discussion (which allows the trainee to cram up on the case notes that the assessor selects). Common sense tells us that case formulation competency in response to newly presenting clinical situations is one of the most fundamental of our skills. Yet existing workplace assessments seem to miss this out, thus undermining the validity of the process when applied to psychiatry. A redesigned and properly evaluated psychiatric mini-CEX might get round this problem, with carefully thought through operational criteria on the skills being evaluated; however, time is running out, with ‘run through’ specialist training due to go live in August 2007.
Observed and observer
Much has been made of workplace assessment measuring performance, as though this alleged gold standard were actually possible. But as we know from second-order cybernetics (Reference Von FoersterVon Foerster, 1981), the very presence of the observer changes what is observed. Assessments in the workplace may give the illusion of testing what goes on all the time, but they are open to much the same cramming and ‘special performance‘ factors as existing OSCE or individual case assessment exams. True, the clinical context is more everyday, but to attribute a gold standard authenticity to workplace assessment seems in part at least to be an act of political spin.
Brown & Doshi stress the benefits of using multiple assessments and assessment methods in order, among other things, to diminish sampling and interrater reliability problems. They seem to accept the given that even if individual assessments have problems of reliability, taken as a whole the assessment programme will make them go away. But an assessment programme can only be as good as its parts, and in respect to the current versions of the mini-CEXand case-based discussion used in the foundation years, these have weaknesses. Examples include the absence of written text required from the assessor to justify rating scores, the vagueness of the scoring criteria, the absence of a second assessor or observer, and the fact that assessments are done mainly by people well known to the trainee. This degree of nepotism may be fine for ongoing formative assessments used as a vehicle for learning and development, but lacks sufficient objectivity to carry much weight in any publicly accountable summative assessment. The lesson here seems to be, either tighten up the reliability of workplace assessments, for example by adding an independent assessor, or complement them with centrally organised tests of clinical skill so that there are sufficiently robust checks and balances in place to reassure everyone.
What role for national exams?
Foundation training began in earnest in August 2005, and to some extent is a rehearsal for what specialist training could look like 2 years down the line. This is particularly so in relation to workplace assessments, which PMETB sees as a core part of the assessment process throughout medical training. However, it is important to reflect on the differences between foundation and specialist training. Foundation training is generic in nature and intent, and replaces a stage of a doctor's career that has never been subject to any final assessment. This has changed, in as much as foundation training is now topped off by a summative assessment, which comprises the successful completion of the 2-year programme of formative workplace assessments. At no stage is there any attempt to make an independent assessment free from potential bias.
By contrast, independent central assessment in the form of the Royal College of Psychiatrists’ membership examination (the MRCPsych) has always played a key role in the first half of specialist psychiatric training, with local ongoing assessment by trainers being a rather poorly regulated and less influential affair. But as Brown & Doshi say, the way forward is undoubtedly to embrace workplace assessments (PMETB gives us no choice), although in so doing we create a dilemma that needs to be addressed: where do these changes leave national examinations? What should their purpose now be, and what can they do that workplace assessments cannot? Clearly national examinations will need to continue to play an important role in testing knowledge, knowledge application and critical thinking. But given the reliability problems associated with workplace assessment, any national exam is going to need to retain a hefty portion of clinical skills testing. Just how great a portion is a matter for interesting debate; but certainly enough to reassure a demanding public, a predatory press and a College highly motivated to maintain high standards in a changing world where clinical governance, accountability and legal defensibility matter.
Has anyone asked the patients?
One final area not mentioned by Brown & Doshi is the patient and carer perspective (Department of Health, 1997, 1999; British Medical Association, 2000). What aspects of the new ways of assessing clinical competency would most reassure service users? What would most concern them? What might they have to say about the balance between workplace and national assessment? Would they think that workplace assessment was independent enough to protect their interests, rather than the interests of the doctors concerned? And of course, where would they see the patient's voice in giving feedback to the trainee on their performance? An odd omission in the new user-focused NHS.
Declaration of interest
None.
eLetters
No eLetters have been published for this article.