Randomised controlled trial of clinical decision support tools to improve learning of evidence based medicine in medical studentsBMJ 2003; 327 doi: https://doi.org/10.1136/bmj.327.7423.1090 (Published 06 November 2003) Cite this as: BMJ 2003;327:1090
- Gabriel M Leung, clinical assistant professor1,
- Janice M Johnston, assistant professor ()1,
- Keith Y K Tin, research assistant1,
- Irene O L Wong, senior research assistant1,
- Lai-Ming Ho, computer officer1,
- Wendy W T Lam, lecturer1,
- Tai-Hing Lam, chair professor1
- 1Department of Community Medicine, University of Hong Kong, Faculty of Medicine Building, 21 Sassoon Road, Pokfulam, Hong Kong, China
- Correspondence to: J M Johnston
Objective To assess the educational effectiveness on learning evidence based medicine of a handheld computer clinical decision support tool compared with a pocket card containing guidelines and a control.
Design Randomised controlled trial.
Setting University of Hong Kong, 2001.
Participants 169 fourth year medical students.
Main outcome measures Factor and individual item scores from a validated questionnaire on five key self reported measures: personal application and current use of evidence based medicine; future use of evidence based medicine; use of evidence during and after clerking patients; frequency of discussing the role of evidence during teaching rounds; and self perceived confidence in clinical decision making.
Results The handheld computer improved participants' educational experience with evidence based medicine the most, with significant improvements in all outcome scores. More modest improvements were found with the pocket card, whereas the control group showed no appreciable changes in any of the key outcomes. No significant deterioration was observed in the improvements even after withdrawal of the handheld computer during an eight week washout period, suggesting at least short term sustainability of effects.
Conclusions Rapid and convenient access to valid and relevant evidence on a portable computing device can improve learning in evidence based medicine, increase current and future use of evidence, and boost students' confidence in clinical decision making.
Despite the widespread dissemination of evidence based medicine into curricula, much of what is known about the outcomes of teaching methods relies on observational data, mostly before and after intervention. Although evaluation of the quality of research evidence is a core competency of evidence based medicine, the quantity and quality of such evidence for teaching evidence based medicine are poor.1
Information available to clinicians at the point of care in an “evidence cart” increased the extent to which evidence was sought and incorporated into decisions on patient care.2 Such a cart would only be useful during team ward rounds, however, and could not be easily transplanted to ambulatory or other settings where most doctor-patient interactions occur. Handheld computers (personal digital assistants), containing concise summaries of evidence and decision making tools, can perform more functions than an evidence cart and hold promise in further promoting the practice and learning of evidence based medicine anywhere.3
We conducted a randomised controlled trial to test whether providing medical students with a handheld computer clinical decision support tool coupled with a brief teaching intervention could improve learning in evidence based medicine.
Participants and methods
All 169 fourth year undergraduates attending the University of Hong Kong consented to participate in the study during their senior clerkship in 2001. It was emphasised that participation was voluntary and would not affect academic records, and they were assured of anonymity and confidentiality.
The senior clerkship is organised into three teaching blocks: block A (internal medicine), block B (surgery), and block C (multidisciplinary, comprising microbiology, obstetrics and gynaecology, pathology, paediatrics, psychiatry, public health, and radiology). Students were assigned to start in one block and rotated through all three blocks for eight weeks each.
The sample size required under the assumptions of a medium effect size of 0.25, α and β levels of 0.05 and 0.2, respectively, and using one way analysis of variance to test for differences between three independent groups yielded a sample size of 159, divided equally into three intervention arms.4
The students were randomised in two stages. Firstly, they were divided into three groups of about equal size using random numbers generated by computer at the faculty's Medical Education Unit, independently from the research team. Secondly, the groups were randomly allocated to start the clerkship in one of the three teaching blocks (groups A to C). Assignment and randomisation were concealed from the students and investigators.
We used a crossover design with three intervention arms. InfoRetriever software (InfoPOEMs; Charlottesville, VA) loaded onto a personal digital assistant, a pocket card containing guidelines on clinical decisions, and a control arm. InfoRetriever is designed for rapid access to relevant, current, best medical evidence at the point of care. It contains seven evidence databases (including abstracts from the Cochrane Systematic Reviews Database and patient oriented evidence that matters), clinical decision rules and practice guidelines, risk calculators, and basic information on drugs.5 This software as well as a digital version of the pocket card were loaded onto a personal digital assistant (iPAQ 3630, 2001 release; Compaq, Hong Kong).
The pocket card was similar in format to a pharmaceutical pocket card and contained information such as the evidence based decision making cycle, levels and sources of evidence, and abbreviated guidelines on appraising the relevance and validity of articles about diagnostic tests, prognosis, treatment, and practice guidelines.6 7 The card was designed to remind and prompt students to apply evidence based medicine techniques in their clinical learning.
Each active intervention (InfoRetriever and pocket card) was accompanied by two interactive sessions lasting two hours and conducted by the same clinician in groups of about 20 students. During the first session, the principles and practice of evidence based medicine were revisited and reinforced (the students had already received 20 hours teaching for the first three years of the curriculum). The students were given specific guidance on how evidence based medicine techniques could and should be used in the clinical setting. The application of the pocket card (in conjunction with traditional library facilities such as Medline searches) or InfoRetriever were shown through a simulated patient case. For the next session, the students were asked to apply the same techniques of identifying an answerable clinical question, searching for and appraising the evidence, and applying the newly acquired information to making clinical decisions to a particular patient they had recently clerked. They shared their findings and experience with the rest of the group, where interactive discussion, led by the clinician, was encouraged.
For the group that was assigned the InfoRetriever intervention for rotations 1 and 2, this active intervention was withdrawn during rotation 3, the washout period.
Measurement of outcomes
We defined a priori five key outcomes most relevant to the learning needs and objectives of the students, as assessed by self report through a locally validated, standardised questionnaire.8 The items gauged respondents' current and future behavioural outcomes such as personal application and current use of evidence based medicine, future use of evidence based medicine, self reported actual use of evidence in clinical learning, and self perceived confidence in clinical decision making. These outcomes were measured by two summary factor scores and three individual item scores derived from the questionnaire (see table A on bmj.com). We excluded knowledge or attitudinal items because these have been often criticised as unreliable: knowledge items should be tested in an objective examination format, and attitudinal items rarely translate into actual behaviour.1
The students completed the questionnaire at baseline and after rotation through each of the teaching blocks. Response rates were 100% throughout. All responses were anonymous and linked to individual respondents through a unique identifier, where only independent research assistants were able to match identifiers to students. Investigators had access only to aggregate results and were blinded to data at the individual level.
We tested baseline equivalence of the three study arms using analysis of variance for continuous variables and χ2 tests for categorical variables. Effects of the educational interventions in each arm were assessed on an intention to treat basis by changes in mean scores for the five outcomes using analysis of covariance.9 10 We also used paired t tests to look for changes in mean scores within groups as the same groups of students progressed through different interventions longitudinally. All analyses were performed with SPSS version 10.
The figure shows the progress of students through the trial. One student initially randomised to block A withdrew from medical school and therefore dropped out of the study during the first rotation.
The characteristics of the students at baseline were similar between intervention arms for age, sex, educational background, and examination scores by year 3 (table 1). Imbalances were, however, evident in the three item scores, which we accounted for by using analysis of covariance.9 4
Table 2 shows changes in mean scores from baseline to the end of the first rotation. Both the InfoRetriever and pocket card group showed improvements in scores for personal application and current use of evidence based medicine. They also reported more frequent use of evidence during clinical teaching rounds, when we observed a significant dose-response gradient in relation to the intensity of the intervention (P for linear trend 0.001). The InfoRetriever group only showed significant gains in self reported use of evidence based medicine and self perceived confidence in clinical decision making. The control group showed no statistically significant changes in any of the outcome measures.
During the second rotation, the InfoRetriever and pocket card groups continued with the interventions whereas the control group received the pocket card. We observed similar changes in mean scores from baseline to second assessment (table 3). The control group showed improvements for the two factor scores representing both current and future use of evidence based medicine. The InfoRetriever group showed slightly higher scores in all five outcomes, whereas those for the pocket card group were not appreciably different from the previous rotation.
To exclude confounding by teaching block (medicine, surgery, and multidisciplinary), we performed a sensitivity analysis by comparing changes in mean scores for the five outcomes between the control group (second assessment minus baseline assessment scores, where the first rotation was control) and the InfoRetriever group (first assessment minus baseline assessment scores). The superior effects of InfoRetriever were evident for four of the five outcomes, and the lack of benefit in future use of evidence based medicine is also consistent with data (see tables 2 and 3 and table B on bmj.com).
To better study the effect of progressing from control to pocket card, we pooled the results of the control group (first to second assessment) and the pocket card group (baseline to first assessment; table 4). The two factor scores showed significant improvements whereas the mean score for frequency that evidence was considered during bedside rounds achieved only borderline significance. The remaining two outcome measures did not change significantly. Both groups from second to third assessments were similarly combined to reflect the effect of progressing from pocket card to InfoRetriever. The two factor scores showed additional increases, and the three individual item scores all showed significant improvements.
To exclude confounding by period effects–that is, increase in experience rather than the educational intervention was responsible for progression–we carried out a sensitivity analysis to compare changes in mean outcome scores between the InfoRetriever group (first assessment minus baseline assessment) and the pocket card group (third assessment minus baseline assessment) where both groups received the InfoRetriever intervention for the duration of one block during the multidisciplinary teaching block in the first rotation for the InfoRetriver group and in the third rotation for the pocket card group. Period effects did not appreciably or significantly influence four out of the five outcomes–that is, the effect of InfoRetriever did not change with time as students matured and progressed through the year(see table C on bmj.com). One possible remaining confounder was when the pocket card group received the pocket card intervention for two blocks (plus one additional tutorial briefing consisting of two sessions for two hours each), which could have accounted for the inconsistencies of the higher factor scores.
The InfoRetriever showed improvements in all five outcomes. Such gains continued on an upward trend with time in three outcomes (table 5). During the washout period there was no significant deterioration in any outcome measure, suggesting a sustained effect, at least in the short term, that persisted after the withdrawal of the personal digital assistant.
Providing students with a decision support tool on a personal digital assistant has the potential to improve their educational experience with evidence based medicine. Furthermore, this improvement was in addition to that achieved by the pocket card, suggesting that neither the pocket card itself nor the two accompanying teaching sessions could wholly account for the observed gains. Although the results preclude the assessment of the long term effects of InfoRetriever, there did not seem to be any significant deterioration in the improvements attained after withdrawal of the personal digital assistant during an eight week washout period, suggesting short term sustainability of effects. We caution, however, that the potential for confounding by teaching block, intervention timing, and progression, as well as the effect of student experience and maturity, should be considered in interpreting the findings. Sensitivity analyses did, however, support our hypothesis of the educational effectiveness of InfoRetriever (see tables B and C on bmj.com).
To our knowledge this study is the first to evaluate, through a randomised trial design, the educational effectiveness of a handheld clinical decision support tool as measured by five predetermined self reported outcomes. It is also one of only a few trials on evidence based medicine learning in the undergraduate setting, the traditional focus being on postgraduate residency programmes.10 11
Improving access to the medical literature through a personal digital assistant at the point of care yielded considerable positive changes in students' self reported actual use of evidence while clerking patients and seemed to boost their confidence in clinical decision making generally. These findings were consistent with a study using an evidence cart during ward rounds.2 Evidence from other studies suggests that improved patient outcomes may result from this evidence based approach.12–14
In addition, ready access to current best evidence prompted the students with personal digital assistants in our study to raise the role of evidence during ward rounds or bedside teaching more often, as ascertained through self reports. Such activity promotes active learning, which is crucial to achieve deep learning and retained knowledge and is a cornerstone of adult learning theory.15 These attributes are hallmarks of a modern medical curriculum that have been incorporated into many reformed curricula internationally.
Those students who were assigned the InfoRetriever intervention reported that they were more likely to be willing to adopt evidence based medicine techniques in their future clinical practice. This is important given the emphasis on lifelong learning and continuous professional development recognised and espoused by most medical organisations worldwide.
The prerequisites for a valid evaluation of educational interventions have been previously outlined.16 They include a rigorous study design with control groups, standardised and validated instruments for measuring outcomes, and high participation and response rates.16 Our study satisfies all three core criteria, thereby ensuring the internal validity of our findings and conclusions.
Limitations of study
Several caveats bear mention. Firstly, we did not have expressly objective or evaluative type outcome measures. All five outcomes were self reported behaviour and future intentions. Also, the mean score differences between groups did not give a clear indication of the absolute magnitude of benefit, although in the context of educating undergraduates in evidence based medicine, standard measures of outcome are limited. Patient related outcomes are not directly applicable to junior clinical clerks whereas commonly adopted tests for abilities in evidence based medicine such as critical appraisal skills may not be better proxies of the desired educational benefits.8 In the absence of better alternatives, nevertheless, we measured changes as a result of the educational interventions under study using a locally validated, standardised questionnaire to ensure consistency and applicability. In the context of routine practice more generally, the impact of educational initiatives on evidence based medicine can be measured through standardised clinical audits using a before and after intervention or a crossover randomised design similar to our present strategy to estimate the proportion of interventions that are grounded in current, best evidence.17 Secondly, our study design was complex and the duration of intervention was relatively brief in the overall context of the medical curriculum. These pitfalls are often encountered in medical education research, when schools are often reluctant to allow students to be selectively exposed to “unproven” new interventions. Therefore we adopted a complicated crossover design for our study. In addition, medical students in their clinical years rotate through a series of teaching blocks rapidly, and it is often difficult to arrange the necessary logistics that would allow the undertaking of a trial. The window period for intervention and measurement is therefore often shorter than optimal. Thirdly, cross contamination of interventions between study arms is a potential limitation in educational trials of this kind.15 16 Such a phenomenon would, however, only dilute the observed effects, and our findings might therefore have underestimated the true magnitude of benefit. Fourthly, part of the observations could have been influenced by differences between the core content and timing of the three teaching blocks, whereby a particular block might have presented relatively more or fewer opportunities for evidence based medicine learning. Some of these effects would likely have been buffered by the crossover design. Future research could further examine this issue through multiple trials of an intervention at different study sites or by replication. Finally, it is difficult to predict the generalisability of our findings; whether other schools with different institutional arrangements and student bodies would achieve similar improvements with InfoRetriever is unknown.16 18 Therefore, we repeat the calls for collaborative, multicentred efforts in such kinds of educational research in evidence based medicine, which will broaden the representativeness of study results and enhance the statistical power of trials to detect subtle changes in relevant outcomes.8
What is already known on this topic
Information in an evidence cart at the point of care increased the extent to which evidence was sought and incorporated into decisions about patient care
The potential benefits of portable clinical decision tools to support students learning evidence based medicine are unknown
What this study adds
Rapid access to evidence on a portable computing device can improve learning of evidence based medicine in medical students
Other benefits are increased current and future use of evidence and more confidence in clinical decision making
Additional tables appear on bmj.com
We thank Compaq Computers (Hong Kong) and InfoPOEMs for technical support; Dorothy Lam, Jacqueline Wong, and Matthew Tse for informatics support and assistance with data collection; Marie Chi for secretarial assistance in the preparation of the manuscript; the Undergraduate Education Committee of the Faculty of Medicine; and the medical students who participated in the trial.
Contributors GML designed the study in consultation with THL and obtained funding for the project. GML, JMJ, KYKT, and WWTL were responsible for implementing the protocol. KYTK and WWTL collected and collated all data items. JJ, GML, IOLW, KYTK, and LMH analysed the data. JJ wrote the initial draft of the manuscript and all authors contributed to the revision of subsequent drafts and approved the final version. GML will act as guarantor for the paper.
Funding The University of Hong Kong provided a teaching development grant; Compaq Computers (Hong Kong) supplied 140 iPAQ pocket PC computers; and InfoPOEMs provided a training site license for the InfoRetriever software at a discounted rate. The guarantor accepts full responsibility for the conduct of the study, had access to the data, and controlled the decision to publish.
Competing interests None declared.
Ethical approval Ethical approval for the study was granted from the Faculty of Medicine Research Ethics Committee, University of Hong Kong.