As medical director of Procedure Service at one of the nation’s largest academic medical centers, I was keenly interested to read the medical education article by Hughes et al
1 in the June 2016 issue of
The Journal of the American Osteopathic Association. The authors offered 10 tips to educators for designing an effective emergency medicine procedure laboratory, and they did a very good job in delineating some of the underpinnings of a successful simulation-based educational procedure experience. The program at the University of Miami-Jackson Memorial Hospital, previously described in this journal,
2 subscribes to many of their tips.
With respect to employing instructors with formal training in adult learning concepts (tip 1), the University of Miami-Jackson Memorial Hospital uses a multifaceted approach that progresses gradually, from a show-and-tell demonstration of our program to faculty participation as a learner to mentoring (ie, faculty observe a teaching session led by a senior instructor) and proctoring (ie, faculty teach while being observed). Then they are prepared to directly supervise our team of residents as they perform clinically warranted procedures on patients. Even at this last stage, a senior instructor accompanies the team (ie, residents and a new instructor) into the clinical realm and is available as back-up.
Although I agree with most of the tips recommended by Hughes et al,
1 I have some comments on others. In developing learning objectives that are well defined, specific, and measurable (tip 2), I would add the importance of identifying the role of the exercise. For example, in our program, we recognized the lack of uniformity in previous apprenticeship-based foundational training. By laying the foundation from the outset, students, house staff, and faculty alike benefit from a standardized training method. I disagree that providing learners with training materials before they arrive at procedure laboratories (tip 3) evens the playing field. In doing so, we would not be able to adequately obtain preintervention baseline data, which are necessary to demonstrate the effectiveness of the intervention.
Hughes et al
1 recommend using mastery learning strategies (tip 6). Based on my experience, I disagree. In our paradigm, skills are perfected under direct attending supervision with pre-, peri-, and postprocedural feedback. This interaction occurs during learners’ performance on a dedicated procedure service rotation, the experiential piece that translates skill in the simulation laboratory to the clinical bedside. Because learners progress at different rates, some will never achieve technical proficiency. True competency includes, but is not defined by, skills performed on a mannequin. Rather, competency is determined by patient-based outcomes, such as procedural completion, complications, and “forward needle passes.”
3
I applaud Hughes et al
1 for their work and believe that their tips will serve the medical community well.