The United States Medical Licensing Examination (USMLE) Step 1 (“Step 1”) is a crucial milestone for medical students. Passing Step 1 is necessary for progression through most medical schools, and for U.S. medical licensure. Moreover, residency program directors cite Step 1 score as the most commonly used factor when selecting applicants to interview, rating it more important than core clerkship grades, leadership/volunteerism, and the Medical Student Performance Evaluation.1 Thus, while there are convincing arguments for reduced reliance on Step 1 scores in residency candidate selection,2,3 optimizing student performance on this high-stakes exam—even above the passing score threshold—remains critical for students and medical schools alike.
Towards this end, numerous studies have identified predictors of improved Step 1 performance, including prior academic performance (e.g., increased undergraduate grade point average,4,5 Medical College Admission Test [MCAT] scores,6,7 medical school preclinical performance,8–10 USMLE practice exam scores11,12); student demographics (e.g., younger age,13 male gender,14 white or Asian race,7 decreased premedical debt15); and curricular factors (e.g., problem-based learning,16,17 system-based education,18 content integration19,20). While such factors are helpful in predicting students’ expected Step 1 performance and identifying those at risk of exam failure,9,21,22 in our experience, medical students beginning their preparations for Step 1 take little comfort in such nonmodifiable factors. Rather, many students understandably desire advising on modifiable behaviors, such as study strategies or use of specific resources, to improve their Step 1 performance.
There is a comparative paucity of evidence regarding student-modifiable exam preparation approaches: Practice question usage,23–25 spaced repetition activities,26,27 and increased study time23,28 have correlated with improved Step 1 scores, while exam timing,29 use of commercial preparation courses,30 and process-oriented preparation31 have shown no clear association with scores. Unfortunately, many of these studies are compromised by dated samples, low power, or inadequate control of confounding covariates (e.g., student demographics, prior academic performance), further limiting evidence-based advising.
In this student-initiated study, we sought to identify students’ self-selected study tools and behaviors when preparing for the USMLE Step 1 and to examine their association with exam performance. On the basis of limited existing literature, we hypothesized that certain study behaviors and tools—such as earlier initiation of dedicated Step 1 studying, spending more time studying, and increased use of Step 1-focused review books and practice questions—are positively associated with Step 1 scores. Furthermore, we hypothesized that these exam-directed study behaviors predict Step 1 scores even when relevant covariates are statistically controlled.
Local curriculum and USMLE Step 1
At the time of this study (2014, 2015), our medical school, a large public research institution, used a systems-based two-year pass/fail preclinical curriculum that did not overtly emphasize Step 1 preparation. Rather, students were encouraged to focus on their preclinical course work and examinations, which were created in-house by faculty. Students were then provided up to six weeks of protected Step 1 study time (the “study period”) at the conclusion of their second year, and our alumni association purchased First Aid for the USMLE Step 132 (hereafter “review book”) for all students.
All medical students at the study institution completing their first Step 1 attempt within academic years 2014 or 2015 were included in the study, which received exemption from the University of Michigan Health Sciences Institutional Review Board. Among all students in this two-year cohort (n = 332), mean MCAT score was 34.7 (standard deviation [SD] 3.2), mean Step 1 score was 235.6 (SD 17.3), and overall Step 1 passing rate was 98.8%. To encourage response to our voluntary survey, a monetary incentive was provided to a randomly selected subset of respondents.
We created a study-behaviors survey (Supplemental Digital Appendix 1, https://links.lww.com/ACADMED/A479) based on input from diverse stakeholders (faculty, counselors, learning specialists), review of previous studies,25,28,31,33 and a local needs assessment of second-year students (n = 121) who had not yet taken the exam—all of which enhanced content validity. Respondents reported their approach to course work, timing of Step 1 study, use of study tools and resources, and “score goal”—the self-identified minimum score they wished to achieve when they began their protected study period. Practice-question usage was self-reported as a completion fraction of UWorld,34 USMLE-Rx,35 or Kaplan36 question banks or self-assessments; unique questions (first-time completion) were differentiated from repeated questions. To ensure that we captured the scope of Step 1 study behaviors, we piloted the survey on five medical students who had recently completed Step 1, which promoted content and response process validity. Pilot respondents felt that score goal was an important determinant of overall study intensity, which led to its inclusion in this work. We distributed the survey electronically in September 2014 and June 2015 using Qualtrics software (Provo, Utah).
Demographic and academic performance data
We retrieved students’ sex, most recent MCAT score, “preclinical score” (a weighted average of exam scores across preclinical organ-based sequences), study period duration, and first-attempt USMLE Step 1 score from internal institutional databases. Using National Institutes of Health guidelines,37 students underrepresented in medicine (URiM) were identified on the basis of self-reported race and ethnicity. The lead investigator (J.P.) paired this information with survey respondent and nonrespondent cases, and deidentified data for subsequent analysis.
We performed statistical analysis using SPSS version 24.0 (SPSS Inc., Chicago, Illinois). Chi-square tests were used to assess differences in count data. Independent samples t tests were used to compare means. One-way analysis of variance with contrasts was used to test linear trends in Step 1 score across categorical groups. Bivariate correlations used Pearson or point–biserial correlation. For bivariate data tables, predictor groupings were created at natural cut points for illustrative purposes and were not used for subsequent regressions. All P values were two sided, with the alpha level set at 0.05.
We implemented multiple linear regression analyses based on our a priori conceptual model. Control variables included sex, URiM status, MCAT score, preclinical score, and score goal. Variables of interest included “early study” (any dedicated Step 1 study before the protected study period), study period duration and hours per day studied, review book passes (number of times reading cover-to-cover), and unique and repeated questions. Variables with significant zero-order correlations were entered simultaneously in each block. Cases with missing data were excluded listwise. We assessed violations of linearity, normality, noncollinearity, and homoscedasticity.
Of 332 medical students, 274 (82.5%) responded to the study-behaviors survey. There were no significant differences in baseline characteristics between the 2014 and 2015 respondents (Table 1). Compared with respondents’ scores, nonrespondents’ Step 1 scores were 12.2 points lower (P < .001) and nonrespondents’ preclinical scores were 2.2 points lower (P < .001). Nevertheless, 20.1% (n = 55) of respondents’ scores fell within the lowest overall score quartile (≤ 224).
Study timing and intensity
While most respondents (n = 235; 90.4%) reported focusing primarily on their preclinical course work during the academic year, with Step 1 a secondary concern, 77.0% (n = 211) conducted dedicated Step 1 studying prior to their study period (“early study”). Some students (n = 38; 13.9%) began Step 1 studying in their first year, while most (n = 173; 63.1%) began in their second year. During the study period, students on average studied 11.0 hours per day (SD 2.1, range 4.5–17.5) over a period of 35.3 days (SD 6.2, range 11–98).
A small number of students (n = 49; 17.9%) used lecture notes or video recordings from their preclinical courses to study for Step 1, using them lightly and finding them only somewhat useful (Table 2). In contrast, students reported predominant use of third-party Step 1 study resources: The review book and UWorld Step 1 question bank (“UWorld”) were used by over 99% of students. Pathoma,38 a pathology textbook with accompanying online lectures, was used heavily by many. Other common resources included Goljan audio lectures (a 30-hour series of lecture recordings of Dr. Edward Goljan, Professor of Pathology, Oklahoma State University Center for Health Sciences), the USMLE-Rx and Kaplan question banks, and Firecracker39 spaced-repetition review. These latter resources were used less heavily and were identified as less useful to most students compared with the review book, UWorld, and Pathoma. Finally, students reported taking an average of 2.8 self-assessment exams (sponsored by either UWorld40 or the National Board of Medical Examiners41).
Students read the review book on average 2.1 times (SD 0.8; 95% confidence interval [CI]: 2.0–2.2), while completing 3,597 total practice questions (SD 1,611; 95% CI: 3,404–3,790), of which 2,961 completed questions were unique (SD 1,026; 95% CI: 2,838–3,084) and 636 were repeated (SD 871; 95% CI: 531–740).
Bivariate association between behaviors and Step 1 score
In bivariate analyses, the control variables male sex, MCAT score, preclinical score, and score goal were positively associated with Step 1 score (Table 3). URiM status was not significantly associated with Step 1 score. Study behaviors showing positive associations with Step 1 scores included early study, greater review book usage, and increased completion of unique or repeated practice questions. Study period hours per day studying and duration were not associated with Step 1 scores. Several variables associated with Step 1 score were also associated with preclinical score, a hypothesized confounder. Bivariate regression models (Table 4 [top]) identified the same significant associations noted in Table 3.
Multiple linear regression for Step 1 score
Our baseline control model including sex, MCAT score, preclinical score, and score goal accounted for 47.7% of Step 1 performance variation (Table 4 [bottom], P < .001). In this model, sex was no longer significantly associated with Step 1 scores. Adding early study, review book passes, unique question usage, or repeat question usage separately to the control model accounted for an additional 1.9% to 7.1% of variance, with each variable remaining statistically significant (Supplemental Digital Table 1, https://links.lww.com/ACADMED/A479). When added simultaneously to the control model, these variables combined to account for an additional 9.2% of Step 1 performance variation (Table 4 [bottom], master model, P < .001). Early study, review book passes, and unique questions completed remained significant in the master model, while sex and repeat questions were not significant. Control variable beta coefficients were largely unchanged from the baseline control model. When controlling for the other covariates, early study was associated with a 4.2-point Step 1 score increase (95% CI: 0.6–7.9), each additional cover-to-cover review book reading a 2.3-point increase (95% CI: 0.3–4.3), and completion of 286 additional unique questions a 1.0-point increase (95% CI: 0.5–1.5). URiM status, study period hours per day, and study period duration were not included in the models, as their zero-order correlations were not statistically significant; sensitivity analyses including these variables did not substantially alter the model (data not shown), nor did exclusion of self-identified score goal (Supplemental Digital Table 2, https://links.lww.com/ACADMED/A479).
“Early” self-directed Step 1 studying the norm
Prior to the protected Step 1 study period, most students at our institution focused primarily on their preclinical course work, yet also engaged in self-directed Step 1 studying. This “early” study phenomenon—which was associated with improved Step 1 performance in multiple regression models—may be a response to the increasing importance of Step 1 in resident selection, in conjunction with a preclinical curriculum that does not “teach to the test.” The effect of early study on Step 1 outcomes has not been previously described, although similar effects were seen for the Comprehensive Osteopathic Medical Licensing Examination of the United States.33
Minimal use of course work review, ubiquitous use of third-party Step 1 resources
Few students reviewed formal course work (e.g., notes or lectures) to prepare for Step 1. Rather, students preferred third-party resources at additional monetary cost. Survey comments noted that a nearly ubiquitous 700-page review book,32 while imperfect, served as a unifying “content outline” covering “high-yield” concepts across the otherwise overwhelming scope of Step 1. Students also used a practice-question-heavy study strategy, replicating the actual Step 1 exam with case-based vignettes requiring multistep reasoning.
Question banks also provide detailed feedback and facilitate spaced repetition and repeated testing that may enhance concept recall.42 On the basis of this work, prior studies,24,31 and conversations with peer programs, we suspect that intense use of third-party Step 1 preparation resources is common practice among medical students at most U.S. medical schools. Additional work is needed to understand the appeal of these resources and to explore the feasibility and appropriateness of formal incorporation of these or related tools into core medical curricula.
MCAT, preclinical performance, and score goal are important covariates
Similar to previous studies,6–10 we found that MCAT scores and preclinical performance were associated with Step 1 performance. We extended prior work in demonstrating the significance of preclinical performance even when including multiple covariates, supporting our medical school’s emphasis that learning content throughout the preclinical years builds an important scientific foundation that also pays dividends in Step 1 performance.
We also described a novel covariate, students’ self-identified score goal (i.e., the score above which students hoped to score, as conceived before beginning the intensive study period). Students piloting our study emphasized that their score goal shaped their overall study intensity; its persistence as a significant covariate merits additional research and consideration when advising students.
Early study and test preparation resource use are associated with Step 1 performance
We confirmed our hypothesis that early study and use of exam-specific resources were positively associated with Step 1 scores, even when controlling for likely covariates. For example, our model estimates that a student who studied early, completed an additional review book pass, and went through an additional question bank in its entirety would score 13.5 points higher than a student who did not engage in these behaviors. These effect sizes are statistically meaningful, as a score increase greater than 10 points for an individual test taker is significant at 95% confidence.43 Moreover, these effect sizes are meaningful in the context of an increasingly Step 1–reliant residency application process, where small score differences can put an applicant above a screening cutoff. Our effect size for unique question usage—a 1.0-point Step 1 score increase per 286 additional questions—is comparable to prior studies showing a 1-point increase per 200 to 445 questions.23,25,26
Contrary to our hypothesis, amount studied during the dedicated study period was not significantly associated with Step 1 performance, suggesting that identifying approaches for “smarter” study may be more important than increased study amounts. It is possible that study efficiency or the amount of early study before the dedicated study period are associated with Step 1 performance, but we were unable to assess these behaviors accurately in this study.
We distinguished between unique and repeated study questions for the first time in the Step 1 literature. Emerging evidence suggests that time spent testing (e.g., question bank usage) is superior to repeated studying (e.g., reading review books) for long-term content retention.44,45 However, the distinction between testing on unique versus repeated identical questions is not well described. Here, repeated testing on identical questions was inferior to testing on new questions covering similar Step 1 content, suggesting an important area for further learning science research, as this approach is broadly applicable in medical education.
Finally, we showed that rigorously controlling for multiple likely confounders—absent from many prior Step 1 studies—is critical for such cross-sectional studies. Our control model accounted for the majority of explained Step 1 performance variance, and the effect sizes of study behaviors were attenuated when considering control variables.
This was a single-institution study, which may limit generalizability to institutions with differing curricular approaches, although studies suggest that students engage in similar behaviors at peer institutions.23–25 The study was cross-sectional, limiting conclusions on causality and directionality. Behaviors were self-reported retrospectively, raising the possibility of recall bias; however, pilot students reported that their Step 1 study behaviors were intentional and regimented, aiding accurate recall. Sampling bias is possible, as nonrespondents had lower Step 1 scores than respondents, although our high response rate and sampling of many students in the lowest score quartile attenuate the impact of this bias. One hourlong information session for students was conducted between study years detailing findings from the first year of our study, such as the most commonly used resources. We speculate that the effect of this session on behaviors in our second cohort was minimal, as we did not detect substantial heterogeneity across study years. Finally, as our analysis was planned a priori, we did not adjust P values for multiple comparisons; however, type I errors may exist, especially for P values near the significance threshold.
Implications and future work
This student-initiated study describes in detail the nature and impact of student self-directed USMLE Step 1 study behaviors, providing preliminary evidence to guide students and medical schools alike. Many medical schools, including our own, evidence an institutional culture of not “teaching to the test.” Importantly, student self-directed co-curricular Step 1 studying and use of third-party resources suggest a “parallel” Step 1 curriculum, with preclinical course work perceived as insufficient alone for preparation.
While the USMLE Step 1 serves an important purpose in licensure, it correlates poorly with resident clinical skills,2 predominantly assessing the “medical knowledge” competency.46 Thus, the monetary and time investment of Step 1 preparation—in our sample averaging nearly five 80-hour weeks during the study period alone—represent a significant opportunity cost for both students and educators. Like others,3 we advocate reassessing the use of the USMLE Step 1 exam in residency candidate selection, which appears to be driving students to seek scores well above the pass/fail threshold. In the meantime, Step 1 performance remains an important outcome for students, and this work provides insight into self-directed study behaviors that may benefit performance. Aligning core preclinical curricula with students’ parallel Step 1 curriculum may yield benefits to students and educators, but should be balanced against overemphasizing Step 1 performance as a meaningful marker of overall trainee quality.
In the future, we plan to explore the relationship between Step 1 preparation behaviors and performance in core clinical clerkships, as we do not know how Step 1 study approaches interact with longer-term measures of medical knowledge retention. Additionally, we hope to conduct a prospective multi-institutional study incorporating local curricular factors to determine more broadly generalizable value-added study behaviors. Finally, the medical education research community should work toward identifying alternative approaches for screening residency applicants for interview, which might reign in students’ Step 1 performance stress and “parallel” curriculum.
The authors wish to thank Dr. Linda Li and Mr. Joey Linzey for their helpful review of the survey instrument; Dr. Eric Middleton, Ms. Amy Tschirhart, and Ms. Tania Reis for input on academic advising factors; Dr. James Cranford for helpful discussions of the statistical approach; Dr. Paula Ross for critical reading; and the numerous students who participated in the study.
1. National Residency Matching Program. Results of the 2016 NRMP Program Director Survey. http://www.nrmp.org/wp-content/uploads/2016/09/NRMP-2016-Program-Director-Survey.pdf
. Accessed July 19, 2017.
2. McGaghie WC, Cohen ER, Wayne DB. Are United States Medical Licensing Exam Step 1 and 2 scores valid measures for postgraduate medical residency selection decisions? Acad Med. 2011;86:4852.
3. Prober CG, Kolars JC, First LR, Melnick DE. A plea to reassess the role of United States Medical Licensing Examination Step 1 scores in residency selection. Acad Med. 2016;91:1215.
4. Julian ER. Validity of the Medical College Admission Test for predicting medical school performance. Acad Med. 2005;80:910917.
5. Basco WT Jr, Way DP, Gilbert GE, Hudson A. Undergraduate institutional MCAT scores as predictors of USMLE step 1 performance. Acad Med. 2002;77(10 suppl):S13S16.
6. Donnon T, Paolucci EO, Violato C. The predictive validity of the MCAT for medical school performance and medical board licensing examinations: A meta-analysis of the published research. Acad Med. 2007;82:100106.
7. White CB, Dey EL, Fantone JC. Analysis of factors that predict clinical performance in medical school. Adv Health Sci Educ Theory Pract. 2009;14:455464.
8. Haight SJ, Chibnall JT, Schindler DL, Slavin SJ. Associations of medical student personality and health/wellness characteristics with their medical school performance across the curriculum. Acad Med. 2012;87:476485.
9. Coumarbatch J, Robinson L, Thomas R, Bridge PD. Strategies for identifying students at risk for USMLE Step 1 failure. Fam Med. 2010;42:105110.
10. Gohara S, Shapiro JI, Jacob AN, et al. Predictors of success on the United States Medical Licensing Examinations (USMLE). Learn Assist Rev (TLAR). 2011;16(1):1120.
11. Morrison CA, Ross LP, Fogle T, Butler A, Miller J, Dillon GF. Relationship between performance on the NBME Comprehensive Basic Sciences Self-Assessment and USMLE Step 1 for U.S. and Canadian medical school students. Acad Med. 2010;85(10 suppl):S98S101.
12. Sawhill A, Butler A, Ripkey D, et al. Using the NBME self-assessments to project performance on USMLE Step 1 and Step 2: Impact of test administration conditions. Acad Med. 2004;79(10 suppl):S55S57.
13. Kleshinski J, Khuder SA, Shapiro JI, Gold JP. Impact of preadmission variables on USMLE Step 1 And Step 2 performance. Adv Health Sci Educ Theory Pract. 2009;14:6978.
14. Cuddy MM, Swanson DB, Clauser BE. A multilevel analysis of examinee gender and USMLE step 1 performance. Acad Med. 2008;83(10 suppl):S58S62.
15. Andriole DA, Jeffe DB. Prematriculation variables associated with suboptimal outcomes for the 1994–1999 cohort of US medical school matriculants. JAMA. 2010;304:12121219.
16. Hecker K, Violato C. How much do differences in medical schools influence student performance? A longitudinal study employing hierarchical linear modeling. Teach Learn Med. 2008;20:104113.
17. Hoffman K, Hosokawa M, Blake R Jr, Headrick L, Johnson G. Problem-based learning outcomes: Ten years of experience at the University of Missouri–Columbia School of Medicine. Acad Med. 2006;81:617625.
18. Lieberman SA, Ainsworth MA, Asimakis GK, et al. Effects of comprehensive educational reforms on academic success in a diverse student body. Med Educ. 2010;44:12321240.
19. Wilkerson L, Wimmers P, Doyle LH, Uijtdehaage S. Two perspectives on the effects of a curriculum change: Student experience and the United States Medical Licensing Examination, Step 1. Acad Med. 2007;82(10 suppl):S117S120.
20. Yoshida H, Sims KL. Education initiatives for improved United States Medical Licensing Examination performance. Med Sci Educ. 2013;23(4):637647.
21. Holtman MC, Swanson DB, Ripkey DR, Case SM. Using basic science subject tests to identify students at risk for failing step 1. Acad Med. 2001;76(10 suppl):S48S51.
22. Burns ER, Garrett J. Student failures on first-year medical basic science courses and the USMLE Step 1: A retrospective study over a 20-year period. Anat Sci Educ. 2015;8:120125.
23. Kumar AD, Shah MK, Maley JH, Evron J, Gyftopoulos A, Miller C. Preparing to take the USMLE Step 1: A survey on medical students’ self-reported study habits. Postgrad Med J. 2015;91:257261.
24. Bonasso P, Lucke-Wold B, Reed Z, et al. Investigating the impact of preparation strategies on USMLE Step 1 performance. MedEdPublish. 2015;4(1):5.
25. Johnson JH, Jordan EJ, Burton W, Silbiger S. Are questions the answer? The effect of popular study resources on USMLE Step 1 performance. Oral abstract presented at: Association of American Medical Colleges Annual Meeting; November 1–7, 2012; San Francisco, CA. https://members.aamc.org/eweb/upload/Are%20Questions%20the%20Answer%20PPT%2011-6-12%2010PM.pdf
. Accessed July 19, 2017.
26. Deng F, Gluckstein JA, Larsen DP. Student-directed retrieval practice is a predictor of medical licensing examination performance. Perspect Med Educ. 2015;4:308313.
27. Kerfoot BP, DeWolf WC, Masser BA, Church PA, Federman DD. Spaced education improves the retention of clinical knowledge by medical students: A randomised controlled trial. Med Educ. 2007;41:2331.
28. Thadani RA, Swanson DB, Galbraith RM. A preliminary analysis of different approaches to preparing for the USMLE step 1. Acad Med. 2000;75(10 suppl):S40S42.
29. Pohl CA, Robeson MR, Hojat M, Veloski JJ. Sooner or later? USMLE Step 1 performance and test administration date at the end of the second year. Acad Med. 2002;77(10 suppl):S17S19.
30. Werner LS, Bull BS. The effect of three commercial coaching courses on Step One USMLE performance. Med Educ. 2003;37:527531.
31. Strowd RE, Lambros A. Impacting student anxiety for the USMLE Step 1 through process-oriented preparation. Med Educ Online. 2010;15(1). http://www.tandfonline.com/doi/full/10.3402/meo.v15i0.4880
. Published February 24, 2010. Accessed July 19, 2017.
32. Le T, Bhushan V, Sochat M. First Aid for the USMLE Step 1 Exam, 2015 edition. 2015.New York, NY: McGraw-Hill Education.
33. Vora A, Maltezos N, Alfonzo L, Hernandez N, Calix E, Fernandez MI. Predictors of scoring at least 600 on COMLEX-USA Level 1: Successful preparation strategies. J Am Osteopath Assoc. 2013;113:164173.
34. UWorld. USMLE Step 1 Qbank. https://www.uworld.com/step1/step1_qbank.aspx
. Accessed July 19, 2017.
35. USMLE-Rx. Step 1 QMax. https://www.usmle-rx.com/content/step-1-qmax
. Accessed July 19, 2017.
37. National Institutes of Health. Notice of NIH’s interest in diversity. http://grants.nih.gov/grants/guide/notice-files/NOT-OD-15–053.html
. Published January 12, 2015. Accessed July 19, 2017.
38. Sattar HA. Pathoma: Fundamentals of pathology. http://www.pathoma.com/
. Accessed July 19, 2017.
39. Firecracker. http://www.firecracker.me/
. Accessed July 19, 2017.
40. UWorld. USMLE Step 1 self-assessment exam. https://www.uworld.com/step1/step1_exam.aspx
. Accessed July 19, 2017.
41. National Board of Medical Examiners. National Board of Medical Examiners (NBME) self-assessment services. https://nsas.nbme.org/home
. Accessed July 19, 2016.
42. Delaney PF, Verkoeijen PPJL, Spirgel A. Ross B. Spacing and testing effects: A deeply critical, lengthy, and at times discursive review of the literature. In: The Psychology of Learning and Motivation. 2010:Vol 53. Burlington, VT: Academic Press; 63148.
43. United States Medical Licensing Examination. USMLE score interpretation guidelines. http://www.usmle.org/pdfs/transcripts/USMLE_Step_Examination_Score_Interpretation_Guidelines.pdf
. Updated June 1, 2016. Accessed September 1, 2016.
44. Roediger HL 3rd, Butler AC. The critical role of retrieval practice in long-term retention. Trends Cogn Sci. 2011;15:2027.
45. Larsen DP, Butler AC, Roediger HL 3rd. Repeated testing improves long-term retention relative to repeated study: A randomised controlled trial. Med Educ. 2009;43:11741181.
46. United States Medical Licensing Examination. USMLE Step 1 content outline and specifications. http://www.usmle.org/step-1/#content-outlines
. Accessed July 19, 2017.