- Study protocol
- Open Access
- Open Peer Review
Clinical evidence continuous medical education: a randomised educational trial of an open access e-learning program for transferring evidence-based information – ICEKUBE (Italian Clinical Evidence Knowledge Utilization Behaviour Evaluation) – study protocol
© Moja et al; licensee BioMed Central Ltd. 2008
- Received: 15 December 2007
- Accepted: 17 July 2008
- Published: 17 July 2008
In an effort to ensure that all physicians have access to valid and reliable evidence on drug effectiveness, the Italian Drug Agency sponsored a free-access e-learning system, based on Clinical Evidence, called ECCE. Doctors have access to an electronic version and related clinical vignettes. Correct answers to the interactive vignettes provide Continuing Medical Education credits. The aims of this trial are to establish whether the e-learning program (ECCE) increases physicians' basic knowledge about common clinical scenarios, and whether ECCE is superior to the passive diffusion of information through the printed version of Clinical Evidence.
All Italian doctors naïve to ECCE will be randomised to three groups. Group one will have access to ECCE for Clinical Evidence chapters and vignettes lot A and will provide control data for Clinical Evidence chapters and vignettes lot B; group two vice versa; group three will receive the concise printed version of Clinical Evidence. There are in fact two designs: a before and after pragmatic trial utilising a two by two incomplete block design (group one versus group two) and a classical design (group one and two versus group three). The primary outcome will be the retention of Clinical Evidence contents assessed from the scores for clinical vignettes selected from ECCE at least six months after the intervention. To avoid test-retest effects, we will randomly select vignettes out of lot A and lot B, avoiding repetitions. In order to preserve the comparability of lots, we will select vignettes with similar, optimal psychometric characteristics.
- Educational Intervention
- Continue Medical Education
- Clinical Vignette
- Balance Incomplete Block Design
- Continue Medical Education Program
Continuing Medical Education for health professionals
In the last decade many countries have legislated within their health systems the revalidation and recertification of medical practitioners . Two principles underlie Continuing Medical Education (CME): professional development is a process of lifelong learning in practice, and professionals must be able to demonstrate they are clinically competent in certain roles. Many countries are facing the challenge of building a formal, sustainable assessment of skills through education programs, accredited providers, acquisition of credit points, and voluntary or mandatory regulation. Although it has been claimed that CME should also include managerial, psychosocial and communication skills to reflect the multidisciplinary context of patient care, usually professional development is mainly based principally on advanced clinical subjects. Traditional knowledge transfer formats employ live interactive classes supported by paper-based text and graphic materials; online learning is increasing popular, supported by screen-based text and graphics with different levels of interactivity.
CME in Italy
A compulsory system of CME was introduced in Italy for all health professionals in 1998. During the five-year cycle from 2002 to 2006, health professionals have been required to earn a total of 150 credits . A national education committee defines the educational standards, accredits educational events, and sets the number of credits earned. A credit involves one hours' learning, i.e., one hour of education is worth one credit, but there is a weighting towards activities that present evidence-based (EB) information and are free of commercial interests. The traditional form of acquiring CME credits is to attend lectures and conferences; it is more rare for doctors to be exposed to small group interactive events. Between 2004 and 2005, the committee approved and evaluated the provision of distant learning programmes: 136,208 health professionals (46,952 doctors, 34%) earned 737,603 credits, a result welcomed as a large success .
Dissemination of independent information in Italy
Interest in EB health care is growing in Italy, although its impact upon clinical practice is still not clear. In 1999, the first pilot free distribution of 50,000 copies of Clinical Evidence was assessed through a survey exploring doctors' judgement of its validity, relevance, and usability. The findings indicated that the compendium had been well-received, and confirmed doctors' preference for problem-driven information and the key role of strong endorsement from health authorities for its implementation . However, another survey funded by the Smith Kline Foundation showed that Italian GPs were satisfied with the information provided by pharmaceutical sales representatives, who they considered complete and sufficiently reliable .
The potential role of e-learning from transferring information to modifying clinical practice
Several conceptual models have been developed for assessing knowledge and competence. A popular one proposed by Miller identifies four stages of development: 'knows, knows how, shows how, and does' that are the essential facets of clinical competence . An individual progresses through cognitive and behavioural steps, from acquiring knowledge to performing a task in practice. Miller's theory assumes that competence predicts performance. Other proposed models consider other influences on a doctor's performance , including system-related factors (government incentives, guidelines) and individual-related ones (patient's expectation, relationship with peers) with increasing levels of complexity.
ECCE is thought to directly affect superficial learning (ability to reproduce facts) and, with decreasing impact, deep learning (ability to apply concepts and skills in the workplace) . ECCE's impact on clinical behaviour and patients' outcomes is less straightforward: we prefer to adopt a conservative framework considering phases that necessarily follow one another . Therefore, in this study we want first and foremost to show an improvement in physicians' knowledge, which may or may not affect competence and, finally, on behaviours and patients' outcomes.
Aim of the study
This trial will test the effectiveness of ECCE e-learning program for transferring EB information to medical doctors after a median of three months of ECCE usage, and retention of the information after a median period of six months. The main hypotheses to be tested are: 1) Does an e-learning CME program based on Clinical Evidence and clinical vignettes (ECCE) increase physicians' basic knowledge about epidemiology, therapy, prognosis, and risk factors in a clinical scenario? 2) Do physicians retain the knowledge from the ECCE for more than six months? 3) Is this educational intervention superior to the passive diffusion of information through the printed version of Clinical Evidence?
The secondary hypotheses are: Does ECCE modify physicians' attitudes to EB behaviours in a simulated realistic clinical scenario?
ECCE: An E-learning CME Program
ECCE is an e-learning CME tool that uses interactive clinical vignettes based on chapters in Clinical Evidence and a predefined sequence of questions. ECCE has four components: 1) the Clinical Evidence chapter (e.g., headache, chronic tension-type); 2) a clinical vignette from the Clinical Evidence chapter that presents a plausible medical scenario (e.g., Margaret says to her family doctor: "This time I didn't come for me, but to talk about Rachel, my 25-year-old daughter....); 3) questions addressing the recall of Clinical Evidence facts or their application to the medical scenario, from which the doctor is to select the correct answer; 4) the potential answers (e.g., a list of potential efficacy descriptors for a therapeutic regimen relevant to the theme); and 5) instructions on what to do (e.g., "more than one answer may be correct").
Vignettes are intended to replicate real-life circumstances as seen by an ordinary GP in everyday practice. Whilst GPs are the primary target of ECCE, many vignettes are also relevant to specialists. Each vignette has a narrative with events and clinical details presented in chronological order: the history evolves with new information from diagnostic tests or additional information reported by the patient. All vignettes use news media techniques, sometimes with fictional or interactive elements (e.g., mystery fiction, the possibility to order tests and obtain results in real time, test appropriateness and cost). Users solve the single steps though a question and answer decision system. They gain credits upon completing all steps where they reach a score of 80% or more of the total. Vignettes provide one or two credits depending on the number of questions. Users are required to finish the vignette started regardless of the score reached. If a user fails, the vignette is then not used again for 24 hours. In 2005, we posted 120 vignettes, and in 2006, 52 new ones.
Both clinical vignettes and related questions will be carefully planned against Clinical Evidence chapters by ECCE authors who are all specialised medical practitioners. Standardised guidelines will be developed by the ECCE editors to help authors. Author guidelines will consider system rules (i.e., each question has five multiple-choice answers), writing style (avoid misleading constructs such as double negatives), medical style (avoid excessive technical jargon, such as rigor nucalis) and provide an example framework. Each vignette will be revised by two editors to ensure it is appropriate for the Clinical Evidence chapter tested. All vignettes will be copy-edited to high editorial standards. Although this would be a highly standardised process, a skilled clinical vignette writer is essential for valid assessment of doctors' abilities and competencies .
ECCE has all the standard advantages of e-learning. Users select what and when they want to learn, and at what pace. The system is easy to use and works with basic computer requirements (e.g., low speed connection). The contents of Clinical Evidence can be read on-screen or printed and interactively managed along the steps of each vignette. The system tracks learning content and the learner's progress.
This is a before and after pragmatic randomised controlled trial utilising a two-by-two incomplete block design. When evaluating educational interventions aimed at improving clinical practice, a number of non-specific effects may influence estimates of the effect of an intervention, grouped together under the term Hawthorne effect [13–17]. These include positive attention effects, caused by participants knowing that they are the subject of a study, but also negative and demotivating effects, caused by being allocated to a control rather than an intervention group. If these non-specific effects are imbalanced across study groups in a quality improvement trial, the estimates may be biased. Randomised controlled trials using balanced incomplete block designs should balance such non-specific effects [13, 15, 17–20].
Designs, hypotheses per trial, and possibility of ensuring equal Hawthorne effect across the arms.
Arm I vs. II Complete intervention on lot A. Using lot B as control.
If the test scores related to lot A vignettes increase in accordance with Clinical Evidence chapters A and there is no change in test scores of B chapters, the intervention has a genuine effect.
Arm II vs. I Complete intervention on lot B. Using lot A as control.
If the test scores related to lot B vignettes increase in accordance with Clinical Evidence chapters B and there is no change in test scores of A chapters, the intervention has a genuine effect.
Arm I vs. III Complete intervention for lot A (ECCE). Minimal intervention on A-B chapters (Clinical Evidence) (control).
If the test scores related to A vignettes in the intervention arm increase in accordance with the Clinical Evidence chapters A and there is no change in the control arm the intervention (ECCE) has a favourable effect, but without controlling for the Hawthorne effect.
Arm II vs. III Complete intervention for lot B (ECCE). Minimal intervention on A-B chapters (Clinical Evidence) (control).
If the test scores related to lot B vignettes in the intervention arm increase in accordance with the Clinical Evidence chapters B and there is no change in the control arm the intervention (ECCE) has a favourable effect, but without controlling for the Hawthorne effect.
Equivalence of clinical vignettes
We will develop two comparable lots of clinical vignettes focusing on different medical topics. In order to preserve comparability we will select vignettes with similar, optimal psychometric characteristics [21–23]. They will be ranked for the following psychometric properties: validity of users' judgement, content validity, internal consistency, and responsiveness.
To avoid contamination, we will make sure that the vignettes in lots A and B are different so that learning about the management of lot A conditions does not influence management of lot B conditions. For example, there may be contamination if lot A focuses on ischemic heart disease (and highlights the need to optimize diabetes control in diabetics) and lot B focuses on diabetes (and highlights the need to optimize control of ischemic heart disease risk factors).
This study will involve Italian doctors naïve to ECCE who voluntarily agree to participate. Doctors can learn about this research program through advertisements in medical journals and websites. New users who enter the ECCE website will be automatically invited to participate. There will be no exclusion criteria.
Allocation to experimental arms
Doctors randomised to ECCE will have access to their clinical vignettes for three months after enrolment, or until they finish all vignettes. All participants will have a logbook to indicate how many times they accessed the platform and how much time they spent solving clinical vignettes. Doctors randomised to receive a printed copy of Clinical Evidence may use it as much as they want over the whole trial period and may have access to the on-line version of Clinical Evidence. The ECCE platform will always be inaccessible to them during the intervention period. After the intervention period, a sample of contamination-controlled clinical vignettes will be accessible to arm 1 and 2 participants until the end of the trial.
Outcomes measures and measurements
The primary outcome will be basic knowledge of the Clinical Evidence contents assessed through the scores from the clinical vignettes. The test will consist of fixed and multiple-choice questions from the selected valid, reliable vignettes and will be administered before (pre-test), immediately after (approximately 16 weeks after enrolment, post-test one), and six months after the intervention (approximately 36 weeks after enrolment, post-test two) (see Figure 6).
To reduce the risk of a test-training effect (practice effects or memory of the first administration will influence post-test performance) at each test section ('pre-test', 'post-test one' and 'post-test two'), we will randomly select two vignettes from lot A and two from lot B to avoid repetition. Therefor, each participant will have only a limited probability (approximately zero) of test/re-testing the same vignette from one assessment to another. The questionnaire will investigate the same Clinical Evidence knowledge for intervention and control chapters. This random sampling of outcomes with high psychometric standards will eliminate any interactions between the causal relationship and the classes of outcomes studied or not studied within the same group of vignettes . To prevent participants from guessing which vignettes will be evaluated, we will add a number of "distracter" vignettes to lots A and B. All groups will also rate their satisfaction with the information source and its perceived value for their medical education and clinical practice.
Study procedures and data collection
The three arms will complete a pre-activity demographic survey, the knowledge tests, and surveys of participation at other educational events. There is growing recognition of EBM in Italy and independent educational initiatives to improve general practice are becoming more common. During the trial, this additional education could be an important threat to internal validity. We will record these concurrent educational experiences in our sample and assess their effects in exploratory post hoc analyses.
Data will be collected from all participants using online instruments. The ECCE platform will manage and track all data collection, following up by e-mail the participants after 16 and 36 weeks. Although researchers are not blinded for the trial group allocation, they will not be able to interfere in collection.
The demographic survey will record sex, age, and years of practice; practice characteristics; preferred CME format; computer skills; ease in using the Internet; familiarity with EB information; frequency of attending residential and distant CME activities.
Sample size and calculation of power
Based on a preliminary examination of test scores, we established that the smallest useful difference for this intervention was a 20% absolute improvement. Based on a preliminary test of 300 doctors, we found that a 28% absolute improvement was due to the intervention. We standardised the absolute improvement across different type of vignettes (five or ten questions). Therefore we calculated our sample size to detect a 0.7 standardized difference in the primary outcome, set the α error rate at 0.05 (two-sided), and the β error at 0.10 (90% power). This yielded a sample size of 45 practitioners per study arm. If the accrual period is six months, and the maximum follow-up period is six months with a loss during follow-up of 20% at the end of the study, the total number of practitioners to be randomised has been adjusted upwards to 162 (54 per intervention group).
Knowledge test data will be analysed using repeated-measure analyses of variance (ANOVA), reporting the partial omega squared (ù2) effect size with corresponding 95% confidence interval. Scores for the knowledge test scale will be subjected totwo-by-three [TO TWO-BY-THREE] repeated-measures ANOVA having: 1) one between-subjects factor (ECCE lot A and ECCE lot B) and one within-subject factor (pre-test, post-test one and post-test two) for the incomplete block design trial; 2) one between-subjects factor (ECCE set and control) and one within-subject factor (pre-test, post-test one and post-test two) for the classical trial. Tests of Sidak-adjusted simple main effects will be used for post-hoc mean comparisons as needed. Orthogonal planned contrasts will be formulated for the knowledge test data to verify knowledge retention. We do not expect the data sets to follow normal distributions and assumptions completely. If this is the case, we will do nonparametric and robust analyses to explore differences in medians or distributional shapes. If nonparametric analysis yields similar interpretations and conclusions, we will focus on the parametric results.
Additional analyses will include examination of the relationship between knowledge test outcomes and potential moderator variables including sex, age, and years of practice, speciality (e.g., general internal medicine versus general surgery); preferred CME format; computer skills; ease using the internet; familiarity with and readiness to use EB information; study drop-out; satisfaction with learning experience, and frequency of attending residential and distant CME activities. For these analyses, we will use parametric methods (regression, repeated-measures ANOVA, analyses of covariance (ANCOVA)) and nonparametric methods (Wilcoxon-Mann-Whitney test, χ2 test of independence and γ measure of association) as needed, based on the data. Holm's modified Bonferroni corrections will be applied to check experimental error (e.g., in exploratory analyses). Analyses will be done based on an intention-to-treat basis, whether or not doctors logged and took credits. We will restrict the explanatory analysis to those doctors who took credits. All the analyses will consider p = 0.05 as significant (two-sided).
SAS version 9.0 statistical software (SAS Institute, Inc, Cary NC) will be used.
The study has been approved by the Research Ethics Board Azienda Sanitaria Locale "Città di Milano", Milano (file number 43-06 SO) and the participating institutions (Italian Drug Agency, the Mario Negri Institute for Pharmacological Research and the Italian Cochrane Centre). This study is funded by a grant from the Italian Drug Agency, a government agency that aims to provide credible, impartial advice, and EB information about the effectiveness of drugs. The Italian Drug Agency approved the design and the methods but had no role in its conduct, analysis, interpretation, or reporting, and will not have access to the data. This trial is completely independent from the BMJ Publishing Group, which publishes the original version of Clinical Evidence.
This protocol of an incomplete block design randomised control trial aims to determine the effect of a large-scale online educational intervention using vignettes based on Clinical Evidence. The educational intervention, ECCE, comprises of:Clinical Evidence, the source of the content; clinical vignettes as the educational model; the e-learning tool as information and test vehicle. Our implementation of these elements may reduce the generalizability of this intervention. In addition, differences in the CME system regulation, the availability CME programmes, and access to EB health care information are likely to increase the differences in many industrialized countries. However, other elements favour the transferability of ECCE: the worldwide diffusion and appreciation of Clinical Evidence, including the standardization of its contents and of the clinical vignettes, its basic computer requirements, and the low cost for each credit provided.
A recent randomised controlled trial tested an internet-based CME program compared with a live course . A cholesterol guideline package was delivered in these two ways and the trial assessed the long-term outcomes. Knowledge immediately before and 12 weeks after the intervention was assessed. The percentage of high-risk patients who had appropriate lipid panel screening and pharmacotherapy according to guidelines was documented with chart audits. Both interventions produced similar and significant 12-week knowledge gains. Chart audits indicated high baseline screening rates in all patient groups (93%) with no significant post-intervention change. However, the internet-based intervention was associated with a significant increase in the percentage of high-risk patients treated with drugs according to cholesterol guidelines. Our trial does not consider behavioural outcomes because this would be rather complicated due to the large number and types of diseases considered in each Clinical Evidence chapter. It is clearly easier to measure an educational intervention designed to change a single behaviour than one targeting general management of conditions or patients. In particular, without evidence of prior efficacy or knowledge, it may be injudicious and precipitous to consider behavioural outcomes. Indeed, our trial's positive results may reflect physicians' competence more than appropriate clinical practice. The teaching properties of case histories are known [27–29], and in a recent study vignette scores appeared to be highly correlated to physicians' practice in outpatient settings and were a valid overall measure of the care provided . Furthermore, one of the greatest barriers to reading Clinical Evidence was boredom. If our intervention is effective, the written case simulation will turn the passive reading into a more interactive experience in which doctors search for the right information for specific situations, increasing their knowledge.
Many thanks to Christine Costantino and Sabrina Bidoli that provided support and constructive criticism during the preparation of this study. Funding/support: Italian Drug Agency (AIFA). Lorenzo Moja is funded by the Monzino Foundation and the Mario Negri Institute. Jeremy Grimshaw holds a Canada Research Chair in Health Knowledge Transfer and Uptake.
- Peck C, McCall M, McLaren B, Rotem T: Continuing medical education and continuing professional development: international comparisons. Bmj. 2000, 320: 432-435. 10.1136/bmj.320.7232.432.View ArticlePubMedPubMed CentralGoogle Scholar
- Epstein RM, Hundert EM: Defining and assessing professional competence. Jama. 2002, 287: 226-235. 10.1001/jama.287.2.226.View ArticlePubMedGoogle Scholar
- Ministero della Salute. Il programma nazionale di Educazione Continua in Medicina. Roma. 2007,(Accessed 5 March 2007)., [http://ecm.sanita.it/presentazione/programma.htm]
- Linetti M: Il programma ECM: risultati raggiunti e risultati attesi. Monitor. 2005, 14-Google Scholar
- Formoso G, Moja L, Nonino F, Dri P, Addis A, Martini N, Liberati A: Clinical Evidence: a useful tool for promoting evidence-based practice?. BMC Health Serv Res. 2003, 3: 24-10.1186/1472-6963-3-24.View ArticlePubMedPubMed CentralGoogle Scholar
- Velo G, Bassi M, Apolone G, Rossi M, Buscone N, Scarpino V, Rizzini P: Informazione scientifica sul farmaco: risultati di un indagine pilota su un campione di medici di medicina generale. Economia politica & del farmaco. 2006, 7: 13-19.Google Scholar
- Moja L, Moschetti I, Liberati A, Manfrini R, Deligant C, Satolli R, Addis A, Martini N, Dri P: Using Clinical Evidence in a national continuing medical education program in Italy. PLoS Med. 2007, 4 (5): e113-10.1371/journal.pmed.0040113.View ArticlePubMedPubMed CentralGoogle Scholar
- Miller GE: The assessment of clinical skills/competence/performance. Acad Med. 1990, 65: S63-67. 10.1097/00001888-199009000-00045.View ArticlePubMedGoogle Scholar
- Rethans JJ, Norcini JJ, Baron-Maldonado M, Blackmore D, Jolly BC, LaDuca T, Lew S, Page GG, Southgate LH: The relationship between competence and performance: implications for assessing practice performance. Med Educ. 2002, 36: 901-909. 10.1046/j.1365-2923.2002.01316.x.View ArticlePubMedGoogle Scholar
- Fritsche L, Greenhalgh T, Falck-Ytter Y, Neumayer HH, Kunz R: Do short courses in evidence based medicine improve knowledge and skills? Validation of Berlin questionnaire and before and after study of courses in evidence based medicine. Bmj. 2002, 325 (7376): 1338-1341. 10.1136/bmj.325.7376.1338.View ArticlePubMedPubMed CentralGoogle Scholar
- Campbell M, Fitzpatrick R, Haines A, Kinmonth AL, Sandercock P, Spiegelhalter D, Tyrer P: Framework for design and evaluation of complex interventions to improve health. Bmj. 2000, 321 (7262): 694-696. 10.1136/bmj.321.7262.694.View ArticlePubMedPubMed CentralGoogle Scholar
- Walsh K: Advice on writing multiple choice questions (MCQs). BMJ Career Focus. 2005, 330: 25-a-27Google Scholar
- Cook TD, Campbell DT: Quasi-experimentation: Design and Analysis Issues for Fields Settings. 1979, Chicago: Rand McNallyGoogle Scholar
- Eccles M, McColl E, Steen N, Rousseau N, Grimshaw J, Parkin D, Purves I: Effect of computerised evidence based guidelines on management of asthma and angina in adults in primary care: cluster randomised controlled trial. Bmj. 2002, 325 (7370): 941-10.1136/bmj.325.7370.941.View ArticlePubMedPubMed CentralGoogle Scholar
- Eccles M, Steen N, Grimshaw J, Thomas L, McNamee P, Soutter J, Wilsdon J, Matowe L, Needham G, Gilbert F, Bond S: Effect of audit and feedback, and reminder messages on primary-care radiology referrals: a randomised trial. Lancet. 2001, 357 (9266): 1406-1409. 10.1016/S0140-6736(00)04564-5.View ArticlePubMedGoogle Scholar
- Shadish WR, Cook TD, Campbell DT: Construct validity and external validity. Experimental and quasi-experimental designs for generalized causal inference. Edited by: Shadish WR, Cook TD, Campbell DT. 2002, Boston New York: Houghton Mifflin Company, 79n-Google Scholar
- Verstappen WH, Weijden van der T, ter Riet G, Grimshaw J, Winkens R, Grol RP: Block design allowed for control of the Hawthorne effect in a randomized controlled trial of test ordering. J Clin Epidemiol. 2004, 57 (11): 1119-1123. 10.1016/j.jclinepi.2004.03.009.View ArticlePubMedGoogle Scholar
- Eccles M, Grimshaw J, Campbell M, Ramsay C: Research designs for studies evaluating the effectiveness of change and improvement strategies. Qual Saf Health Care. 2003, 12 (1): 47-52. 10.1136/qhc.12.1.47.View ArticlePubMedPubMed CentralGoogle Scholar
- Shadish WR, Cook TD, Campbell DT: Statistical conclusion validity and internal validity. Experimental and quasi-experimental designs for generalized causal inference. Edited by: Shadish WR, Cook TD, Campbell DT. 2002, Boston New York: Houghton Mifflin Company, 45-48.Google Scholar
- Verstappen WH, Weijden van der T, Sijbrandij J, Smeele I, Hermsen J, Grimshaw J, Grol RP: Effect of a practice-based strategy on test ordering performance of primary care physicians: a randomized trial. Jama. 2003, 289: 2407-2412. 10.1001/jama.289.18.2407.View ArticlePubMedGoogle Scholar
- Shaneyfelt T, Baum KD, Bell D, Feldstein D, Houston TK, Kaatz S, Whelan C, Green M: Instruments for evaluating education in evidence-based practice: a systematic review. Jama. 2006, 296 (9): 1116-1127. 10.1001/jama.296.9.1116.View ArticlePubMedGoogle Scholar
- Vleuten van der C: Validity of final examinations in undergraduate medical training. Bmj. 2000, 321: 1217-1219. 10.1136/bmj.321.7270.1217.View ArticlePubMedPubMed CentralGoogle Scholar
- Wass V, Vleuten Van der C, Shatzer J, Jones R: Assessment of clinical competence. Lancet. 2001, 357: 945-949. 10.1016/S0140-6736(00)04221-5.View ArticlePubMedGoogle Scholar
- Altman DG, Schulz KF, Moher D, Egger M, Davidoff F, Elbourne D, Gotzsche PC, Lang T: The revised CONSORT statement for reporting randomized trials: explanation and elaboration. Ann Intern Med. 2001, 134: 663-694.View ArticlePubMedGoogle Scholar
- Shadish WR, Cook TD, Campbell DT: Construct validity and external validity. Experimental and quasi-experimental designs for generalized causal inference. Edited by: Shadish WR, Cook TD, Campbell DT. 2002, Boston New York: Houghton Mifflin Company, 64-102.Google Scholar
- Fordis M, King JE, Ballantyne CM, Jones PH, Schneider KH, Spann SJ, Greenberg SB, Greisinger AJ: Comparison of the instructional efficacy of Internet-based CME with live interactive CME workshops: a randomized controlled trial. Jama. 2005, 294 (9): 1043-1051. 10.1001/jama.294.9.1043.View ArticlePubMedGoogle Scholar
- Fox R: Writing a case report: an editor's eye view. Hosp Med. 2000, 61 (12): 863-864.View ArticlePubMedGoogle Scholar
- Naldi L, Manfrini R, Martin L, Deligant C, Dri P: Feasibility of a web-based continuing medical education programme in dermatology. The DermoFAD experience in Italy. Dermatology. 2006, 213: 6-11. 10.1159/000092830.View ArticlePubMedGoogle Scholar
- Vandenbroucke JP: In defense of case reports and case series. Ann Intern Med. 2001, 134: 330-334.View ArticlePubMedGoogle Scholar
- Peabody JW, Luck J, Glassman P, Dresselhaus TR, Lee M: Comparison of vignettes, standardized patients, and chart abstraction: a prospective validation study of 3 methods for measuring quality. Jama. 2000, 283: 1715-1722. 10.1001/jama.283.13.1715.View ArticlePubMedGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.