The Consumer Assessment of Healthcare Providers and Systems (CAHPS) program, funded by the Agency for Healthcare Research and Quality, aims to develop standardized surveys to assess patient experiences with their healthcare providers.1 Over the years, the CAHPS program has produced surveys designed to collect information across a range of health care settings, including both ambulatory and facility-based settings.
The CAHPS Hospital Survey (HCAHPS), developed in partnership with the Centers for Medicare and Medicaid Services (CMS), is a standardized survey of the experiences of adult patients with hospital care and services. HCAHPS was created using the standard CAHPS survey development process, including a public call for measures, an extensive environmental scan, cognitive testing in English and Spanish, patient focus groups, input from stakeholders, and multiple field tests in both English and Spanish. CMS provided opportunities for the public to comment on HCAHPS.2 The National Quality Forum has endorsed the survey and the federal Office of Management and Budget approved national implementation for public reporting purposes. HCAHPS was implemented nationally in the United States by CMS on a voluntary basis, starting in October 2006. CMS includes this survey in its Hospital Quality Reporting program requirements and has been reporting the survey results on a public Web site at http://www.hospitalcompare.hhs.gov.3
HCAHPS includes items that capture communication with nurses and doctors and communication about medicines and discharge information. The HCAHPS Item Set for Addressing Health Literacy was developed as a supplement to the core HCAHPS to provide hospitals with actionable information they can use to “drill down” to identify areas for quality improvement. The development of this item set builds on earlier CAHPS work to develop an item set to address health literacy issues in an ambulatory care setting.4 Stakeholders who participated in these earlier development efforts identified the need for a CAHPS Health Literacy item set specifically for the HCAHPS. At the same time, CMS reported that stakeholders were requesting more detailed questions about discharge coordination that they could use for quality improvement efforts.
This new Health Literacy item set reflects both the ascendency of health literacy as a public health concern, and recognition that addressing health literacy issues has the potential to mitigate the negative sequelae of limited health literacy.5 A recent systematic review of the literature found that poor health literacy among patients is associated with more hospitalizations, greater use of emergency care, lower receipt of preventive care, poorer ability to demonstrate taking medications appropriately, poorer ability to interpret labels and health messages, and—among elderly persons—worse overall health status and higher mortality rates.6 National studies document that the complexity of health information frequently exceeds American’s health literacy skills.7–10
In 2007, the Joint Commission called for health care organizations to make effective communication a priority across the care continuum.11 Transitions from hospital to home represent a particularly vulnerable time. Many patients leave the hospital without understanding their diagnosis or how to take their medicines.12 Surgical patients often do not remember postoperative instructions.13 Patients with low literacy reported hospital staff becoming frustrated or angry when someone could not complete a form or read instructions and recounted serious medication errors resulting from their inability to read labels.14 Reengineering the discharge process so that patient education occurs throughout the hospital stay and health literacy techniques are used have been shown to be able to reduce rehospitalizations by 30%.15 With the recent reduction in reimbursement to hospitals with excessive readmission rates, hospitals are eager to make changes to improve how they communicate with patients about tests, self-care, and medicines.16 The new Health Literacy item set can help them assess their efforts.
Domain and Item Development
The HCAHPS Item Set for Addressing Health Literacy was developed using the survey development approach used to develop other CAHPS surveys.2 First, we identified the domains for the item set through an extensive environmental scan and by interviews with content experts. We held meetings with stakeholders to obtain input into the domains identified by the environmental scan, prioritize candidate domains, review draft survey items, and obtain input on how best to disseminate the new Health Literacy item set and promote its use. Two stakeholder meetings were held in March 2009 that included representatives from various government agencies (including CMS), hospital representatives, clinicians and other health providers, health literacy experts and advocates, and consumers.
Seven health literacy domains were identified through this process: (1) communication with nurses; (2) communication with doctors; (3) communication about tests; (4) communication about caring for yourself at home; (5) communication about medicines; (6) interpreter services; and (7) communication about forms. Although some of these domains were completely new (eg, interpreter services, communication about forms), others represented an expansion of domains already included in the core HCAHPS but, in stakeholders’ opinions, not sufficiently addressed.
We reviewed the survey items that had been collected through the environmental scan and mapped the survey items to the domains. Agency for Healthcare Research and Quality issued a call for measures through the Federal Register, but very few responses were submitted. In developing the draft item set, we modified or adapted items in the public domain to conform to the HCAHPS structure and format. In addition, new survey items were drafted for domains for which the team was unable to identify existing items.
The core HCAHPS was developed and evaluated concurrently in both English and Spanish. Findings from the original HCAHPS pilot tests demonstrate that responses to both language versions have similar patterns with respect to item-scale correlations, factor structure, content validity and associations between the reporting measures and the overall rating of the hospital. Overall, the survey items were generally equivalent across language versions.17
Like the core HCAHPS, we developed and tested the supplemental Health Literacy item set concurrently in both English and Spanish. The supplemental items were translated into Spanish using the CAHPS guidelines for selection of translators and reviewers and for translation.18–19 The translation team aimed to produce a Spanish version that was conceptually equivalent to the English version, that was at the appropriate reading level for the target population, and that could be understood by Spanish speakers throughout the continental United States. Using a translation approach that involves multiple translators and bilingual reviewers from different Spanish-speaking countries (often referred to as the “translation by committee” approach) ensures that the translation is understood by a wide range of Spanish speakers and has been shown to produce more culturally appropriate translations.20–22
Cognitive interviewing23–24 is a key step in the CAHPS survey development process and has traditionally been used to evaluate CAHPS survey items before field-testing.25–27 We conducted 3 rounds of cognitive interviews in the spring and summer of 2009. In total, 36 interviews were conducted, half in English and half in Spanish. All interviews were conducted with adults with a recent overnight hospital stay. Cognitive interview respondents included a mix of men and women in terms of age and race and ethnicity, insurance coverage, and patients who had been admitted to the hospital through the emergency room as well as those who had scheduled surgeries.
To test the Health Literacy item set with respondents with limited health literacy skills, we recruited subjects with less than a high school education. The 2003 National Assessment of Adult Literacy found that adults who had not gone beyond a high school education had lower average health literacy than adults with higher levels of education.28 Although we did not assess respondents’ literacy or health literacy skills, two thirds of the cognitive interviews were conducted with respondents who had no more than a high school education.
We used the cognitive interviews to revise and refine the item set iteratively and to “weed out” survey items that were problematic, did not capture meaningful information, or focused on issues that did not resonate strongly with respondents. Of the 84 items that we initially tested, 22 items were dropped based on findings from the cognitive interviews.
A composite is composed of ≥2 survey items that are closely related conceptually and statistically. Composites summarize a large amount of survey data, thus making it easier to understand.29 For each of the 7 health literacy domains in the item set, we identified the subset of items that could be used to create a composite measure for each domain (excluding screener questions that are used to skip respondents out of questions that do not apply to them). Table 1 lists the hypothesized composites and the subset of items that comprise them.
We evaluated the extent to which items correlated together into multi-item composites and calculated a measure of internal consistency, Cronbach α, to estimate the internal consistency reliability of the composites. We deemed those as acceptable items that were correlated as ≥0.30 with their hypothesized composite (correcting for item overlap with the total score). We also examined the correlations of each item with composites they were not hypothesized to represent to assess whether the composites represent unique aspects of communication to improve health literacy. Through an iterative process, we revised the placement of items into composites, taking into account correlations between the items as well as item content, and reran the correlation analysis to assess the fit of each individual item into the composite it was hypothesized to represent. We also conducted categorical confirmatory factor analysis in Mplus30 to assess the fit of the final composite structure.
The purpose of the field test was to assess the reliability and validity of the supplemental Health Literacy item set and to identify items that could be combined to create composite measures that could be used for both internal quality improvement and public reporting. The field test was conducted between December 2010 and February 2011. The 62 health literacy items were appended to the 27-item core HCAHPS, in accordance with the technical specifications for implementing the HCAHPS provided by CMS,31 creating an 89-item field test survey.
Recruiting field test partners proved challenging. Although several hospitals expressed interest in participating in the field test, in the end only 1 hospital was able to secure all the necessary permissions in time to take part in the field test. This hospital serves a largely suburban population located in a midwestern metropolitan area with a population of approximately 145,000 persons. The majority of patients served by the hospital are non-Hispanic white patients who are English speaking. In 2009, among in-patient discharges, 35% were covered by Medicare, 36% were commercially insured, 23% were covered by Medicaid, 5% were uninsured, and 1% was covered by other types of health insurance.32
The sample frame for the field test included 3772 adult patients drawn by the participating hospital using sample selection specifications adapted from version 4.0 of the HCAHPS Quality Assurance Guidelines (published in February 2009). The participating hospital was instructed to create a sampling frame of adult patients (age 18 y or older) with at least 1 overnight hospital stay in the previous 6 weeks. Excluded from the sample frame were patients who were less than 18 years old at the time of their admission, had a psychiatric diagnosis, were discharged to a hospice facility, died during the hospitalization, patients who requested not to be contacted, court/law enforcement patients (ie, prisoners), patients with a foreign home address, and patients who should be excluded based on state regulations. From the sample frame provided by the hospital, RAND then randomly selected 2000 patients to participate in the field test. Although we wanted to field test the Health Literacy item set in both English and Spanish, the sample provided by our field test partner did not include sufficient Spanish speakers.
The survey was field-tested using a combination of mail and phone survey administration, which has been shown to produce higher response rates than mail surveys alone and is a less expensive option than conducting the survey either in person or by telephone alone.33,34 Conducting phone follow-up of sampled individuals that did not respond to the mailed survey also offers a means to reach individuals whose low literacy presents a barrier to self-administering a survey. The data collection protocol included 4 mailings before phone follow-up [(1) advance notification letter; (2) first survey mailing; (3) reminder letter; (4) second survey mailing]. Up to 10 attempts to complete the survey by phone were made over a 4-week period.
To maximize response rates among those with limited literacy skills, the survey and all survey materials were written using simple, plain language. As described above, the survey was assessed for comprehension through extensive cognitive testing. The letters used as part of the survey were evaluated using the Flesch-Kincaid readability tool available through Microsoft Office Word. The readability ease score was 71.0 (indicating that in theory, the letters could be understood by average 13–15 y olds) and the Flesch-Kincaid grade level score was 6.9 (indicating that in theory, the letters could be understood by an average sixth or seventh grader).35
We examined item distributions (ceiling and floor effects), item missing data, internal consistency reliability of the multi-item scales or composites, correlations of the new health literacy composites and the core HCAHPS composites, and correlations of the health literacy composites with a global rating of the hospital on a scale of 0–10 (where 0=worse possible hospital and 10=best possible hospital).36–38
Field Test Response
A total of 1013 surveys were completed for a response rate of 55%. Overall, more surveys were completed by mail (790 or 78%) than by phone (223 or 22%), with most of the surveys (62%) coming in after the first mailing. All but one of the surveys was completed in English. Table 2 shows the demographic characteristics of survey respondents. Twenty-two percent reported that their health was fair or poor, whereas 30% reported it was good, 43% reported it was very good or excellent. Ten percent of the sample had less than a high school education, whereas 21% had graduated from college or had more than a college degree. Whites were by far the largest racial group and Hispanics constituted only 2% of respondents. Almost all respondents reported that English is the main language they speak at home and a majority was female. Over half the sample was 65 years or older. Demographic characteristics were not available for those who did not respond to the survey; therefore nonresponse bias could not be estimated.
We conducted analyses to evaluate the response distribution of survey items. The percentage of ceiling effects for the items ranged from 11% (q32) to 99.6% (q48) with a median of 73%. Question 48 appears to be an outlier in terms of response distributions; however, we opted to retain it pending further evaluation of the item set. The percentage of floor effects ranged from 0.1% (q36) to 67% (q66) with a median of 2%.
Evaluation of Hypothesized Multi-Item Composites
With a few exceptions, correlations between items and scales revealed that the data were consistent with the hypothesized multi-item composites. A few items correlated just as highly (or in some cases more highly) with composites they were not hypothesized to belong to. On the basis of item content, however, we opted to keep them in the composite that was conceptually the best fit. For example, item 23 correlated more highly with the communication with doctor composite than with the hypothesized communication with nurses composite. However, this item measure experiences with nurses and therefore, in terms of content, does not fit into the communication with doctor composite. In addition, items 54, 56, 58, and 59 in the communication about medicines composite correlated more highly with the original composite on communication about how to care for yourself at home. Given that these items refer to information about medicines that the patient would take at home upon discharge, we combined the 2 composites into 1 larger composite that includes items 46, 47, 49, 54, 56, 58, and 59.
Of the original 7 domains, 2 of the hypothesized composites (communication with doctors and communication with nurses) were similar to existing HCAHPS composites and a decision was made not to publish these until such time that composite labels could be tested to ensure consumers would not be confused. A third composite (interpreter services) was not published because there was insufficient number of responses to these items to conduct psychometric testing. As noted above, 2 of the hypothesized composites (communication about medicines and communication about how to care for yourself at home) were combined, resulting in a total of 3 final composites. The item-scale correlation matrix for the 3 final composites is presented in Table 3: (1) communication about tests (4 items with item-scale correlations ranging from 0.64 to 0.77); (2) communication about how to care for yourself and medicines (7 items with item-scale correlations ranging from 0.33 to 0.54); and (3) communication about forms (4 items with item-scale correlations ranging from 0.34 to 0.57).
A 3-factor categorical confirmatory factor analysis model representing the final item configuration fit the data well (Comparative Fit Index=0.957; Tucker-Lewis Index=0.951; Root Mean Square Error of Approximation=0.057). Factor loadings were all statistically significant and ranged from 0.776 to 0.898 for communication about tests, 0.612 to 0.893 for communication about how to care for yourself and medicines, and 0.514 to 0.835 for communication about forms. The estimated correlations between factors ranged from 0.392 (communication about how to care for yourself and medicines) to 0.780 (communication about forms and communication about tests). Thus, the confirmatory factor analysis provided support for the 3 new CAHPS composites. Table 4 presents scale means, SDs, and internal consistency reliability estimates for the final composites.
Eleven of the 62 health literacy items that were field-tested (items 24, 26, 32, 35 36, 37, 38, 39, 48, 50, and 55) did not fit into their hypothesized composite. Although these items had low item-scale correlations for their hypothesized composite, they still provide useful content and can provide information that survey users can use for quality improvement purposes. For this reason, we opted to keep them as part of the final version of the Health Literacy item set.
The Health Literacy item set that was field-tested included 17 items that collect information on interpreter services. However, given the fact that almost all of the respondents that participated in the field test were English speaking, the vast majority screened out of these questions and thus we do not have sufficient data to evaluate these items. Given the importance of this domain in a hospital setting, however, we have opted to keep some of these items in the final version of the Health Literacy item set. The items were modified based on similar interpreter service items that were developed and tested as part of other CAHPS efforts and have been included in Table 5. The items are available for public use as users may find them useful for quality improvement purposes. However, further research is needed to fully test and evaluate these items.
Associations of New Composites With Global Ratings of the Hospital and the Core HCAHPS Composites
The largest correlations of the new composites with the global rating of the hospital and whether the respondent would recommend the hospital to family or friends were observed for the items on communication with nurses (0.50 and 0.43, respectively). This is consistent with results for the HCAHPS core composites.1 Not surprisingly, the new health literacy items on communication with nurses correlated most highly with the core communication with nurses composite (r=0.62). Similarly, the largest correlation of the new health literacy items on communication with doctor were with the core communication with doctor composite (r=0.69) and the largest correlation of the new communication about discharge and medicines composite was with the core discharge composite (r=0.47).
In contrast, the largest correlation of the new communication about tests composite was with the core communication about medicines composite (r=0.49). The new communication about forms composite, which does not have a core composite counterpart, correlates most highly with the core communication about medicines composite (r=0.42). Although these correlations are sizable, they suggest that the composites provide unique information beyond what would be provided by the core composite alone. A multiple regression of the recommend to family and friends item on the HCAHPS core and new composites had an adjusted R 2 of 52% and significant unique associations with the core nursing composite (β=0.01, t=5.14, P<0.0001), the core pain management composite (β=0.01, t=5.68, P<0.0001), and the new items on communication with nurses (β=0.01, t=2.02, P<0.05). None of the new composites were uniquely associated with the global rating of the hospital item.
After the field test was completed, we made further revisions to the item set based on the results of additional testing conducted as part of other CAHPS initiatives. In addition, CAHPS has recently undertaken an effort to harmonize various CAHPS supplemental item sets, which resulted in further revisions to the item set. The final version of the item set includes 58 items and can be found in Table 5.
Hospitals are increasingly being called to do a better job communicating with patients, as exemplified by the Joint Commission’s new patient-centered communication standard on hospitals communicating effectively with patients when providing care, treatment, and services.39 Hospitals need to be able to monitor how well they are meeting expectations about better communication. For example, the American Hospital Association’s Improving Communication with Patients and Families: A Blueprint for Action 40 includes on its checklist of leadership strategies to enhance communication communicating with patients and families at all stages of their hospital experience. The HCAHPS Item Set for Addressing Health Literacy is 1 tool hospitals can use to make sure they are taking advantage of the patient’s time in the hospital to provide clear, understandable information. This comprehensive item set includes topics that are of interest to a wide range of stakeholders, with 3 multi-item composites that can be reported. Although the item set includes more survey items than any 1 hospital is likely to use, users have the flexibility of picking and choosing individual survey items to “drill down” to get detailed, actionable information, or can use one of the 3 composites to report a composite score on a particular topic. Additional information on how CAHPS composite measures are calculated can be found at http://www.cahps.ahrq.gov/About-CAHPS/FAQs. Although there are no plans for CMS to publicly report the survey results from these items, the new Health Literacy item set is designed to provide information that can be used for quality improvement and for consumers to assess key aspects of hospital quality of care.
The study has several limitations. First, the Health Literacy item set was field-tested with a predominantly white, adult patient sample from 1 hospital in the midwest. Further research is needed to test the Health Literacy item set with a mix of patients in terms of race/ethnicity, in a broader range of hospital settings in various geographic locations. Second, the study was primarily limited to an English-speaking population. Although the Health Literacy item set is available in Spanish (and indeed was extensively cognitively tested in Spanish), the sample for the field test did not include sufficient numbers of Spanish speakers to adequately assess the Spanish version and in particular, to assess the interpreter services items. Further research is needed to fully test the Health Literacy item set with Spanish speakers and other non–English-speaking populations. Despite these limitations, the HCAHPS Item Set for Addressing Health Literacy can serve as a tool to measure, from the patient’s perspective, how well hospital providers’ are meeting their patients’ health literacy needs and to use this information for quality improvement purposes. Additional information on quality improvement strategies using CAHPS data can be found at https://www.cahps.ahrq.gov/Quality-Improvement/Improvement-Guide.aspx.
2. Goldstein E, Farquhar M, Crofton C, et al. Measuring hospital care from the patients’ perspective: an overview of the CAHPS Hospital Survey development process. Health Serv Res. 2005;40(6 pt 2):1977–1995
3. Giordano LA, Elliott MN, Goldstein E, et al. Development, implementation, and public reporting of the HCAHPS survey. Med Care Res Rev. 2010;67:27–37
4. Weidmer B, Brach C, Hays RH. Development and evaluation of CAHPS®
Survey Items assessing how well healthcare providers address health literacy. Med Care. 2012;50(suppl 2):S3–S11
5. Koh HK, Berwick DM, Clancy CM, et al. New federal policy initiatives to boost health literacy can help the nation move beyond the cycle of costly “crisis care.” Health Aff. 2012;31:434–443
6. Berkman ND, Sheridan SL, Donahue KE, et al. Low health literacy and health outcomes: an updated systematic review. Ann Intern Med. 2011;155:97–107
7. Helitzer D, Hollis C, Cotner J, et al. Health literacy demands of written health information materials: an assessment of cervical cancer prevention materials. Cancer Control. 2009;16:70–78
8. Goodfellow GW, Trachimowicz R, Steele G. Patient literacy levels within an inner-city optometry clinic. Optometry. 2008;79:98–103
9. Badarudeen S, Sabharwal S. Readability of patient education materials from the American Academy of Orthopaedic Surgeons and Pediatric Orthopaedic Society of North America web sites. J Bone Joint Surg Am. 2008;90:199–204
10. Hill-Briggs F, Smith AS. Evaluation of diabetes and cardiovascular disease print patient education materials for use with low-health literate populations. Diabetes Care. 2008;31:667–671
11. What Did the Doctor Say?: Improving Health Literacy to Protect Patient Safety. 2007 Oakbrook Terrace, IL The Joint Commission
12. Makaryus AN, Friedman EA. Patients’ understanding of their treatment plans and diagnosis at discharge. Mayo Clin Proc. 2005;80:991–994
13. Atchison KA, Black EE, Leathers R, et al. A qualitative report of patient problems and postoperative instructions. J Oral Maxillofac Surg. 2005;63:449–456
14. Baker DW, Parker RM, Williams MV, et al. The health care experience of patients with low literacy. Arch Fam Med. 1996;5:329–334
15. Jack BW, Chetty VK, Anthony D, et al. A reengineered hospital discharge program to decrease rehospitalization: a randomized trial. Ann Intern Med. 2009;150:178–187
16. Advancing Effective Communication, Cultural Competence, and Patient- and Family-Centered Care: A Roadmap for Hospitals. 2010 Oakbrook Terrace, IL The Joint Commission
17. Hurtado MP, Angeles J, Blahut SA, et al. Assessment of the equivalence of the Spanish and English versions of the CAHPS® hospital survey. Health Serv Res. 2005;40:2140–2161
20. Harkness JA., van de Vijver FJR, Mohler P.Ph, et al. Cross-Cultural Survey Methods. 2003 Hoboken, NJ John Wiley & Sons Inc.
21. . Translating Questionnaires and Other Research Instruments: Problems and Solutions. Sage University Papers Series on Quantitative Applications in the Social Sciences. 2000 Thousand Oaks, CA Sage:07–131
22. US Census Bureau Methodology and Standards Council. The translation of surveys: an overview of methods and practices and the current state of knowledge. Census Bureau Guideline: Language Translation of Data Collection Instruments and Supporting Materials
. Washington, DC
: Census Bureau; 2000. Working Paper
23. Hughes KA. Comparing Pretesting Methods: Cognitive Interviews, Respondent Debriefing, and Behavior Coding. RESEARCH REPORT SERIES (Survey Methodology #2004-02), Statistical Research Division U.S. Bureau of the Census Washington D.C. 20233. Report Issued: September 30, 2004
24. Willis G Cognitive Interviewing: A Tool for Improving Questionnaire Design. 2004 Thousand Oaks, CA Sage Publications
25. Harris-Kojetin LD, Fowler FJ, Brown JA, et al. The use of cognitive testing to develop and evaluate CAHPS 1.0 core survey items. Med Care. 1999;37:MS10–MS21
26. Levine RE, Fowler FJ Jr, Brown JA. Role of cognitive testing in the development of the CAHPS®
Hospital Survey. Health Serv Res. 2005;40(6 pt 2):2037–2056
27. Weech-Maldonado R, Weidmer B, Morales L, et al.Cynamon ML, Kulka RA Cross-cultural adaptation of survey instruments: the CAHPS experience. Seventh Conference on Health Survey Research Methods (DHSS Publications No. [PHS] 01-1013). 2001 Hyattsville, MD
29. McGee J, Kanouse DE, Sofaer S, et al. Making survey results easy to report to consumers: how reporting needs guided survey design in CAHPS. Med Care. 1999;37(suppl 3):MS32–MS40
30. Muthén LK, Muthén BO Mplus User’s Guide. 2009 Los Angeles, CA Muthén & Muthén
33. Dillman DA, Smyth JD, Christian LM Internet, Mail, and Mixed-Mode Surveys: The Tailored Design Method. 2009 Hoboken John Wiley & Sons Inc.
34. Fredrisckshon DD, Jones TL, Molgaard CA, et al. Optimal design features for surveying low income populations. J Health Care Poor Underserved. 2005;16:677–690
35. Kincaid JP, Fishburne RP, Rogers RL, et al. Derivation of New Readability Formulas (Automated Readability Index, Fog Count, and Flesch Reading Ease formula) for Navy Enlisted Personnel. Research Branch Report 8-75. Chief of Naval Technical Training: Naval Air Station Memphis; 1975
36. Hays RD, Shaul JA, Williams VS, et al. Psychometric properties of the CAHPS 1.0 survey measures consumer assessment of health plans study. Med Care. 1999;37(suppl 3):MS22–MS31
37. Hargraves JL, Hays RD, Cleary PD. Psychometric properties of the Consumer Assessment of Health Plans Study (CAHPS) 2.0 adult core survey. Health Serv Res. 2003;38:1509–1527
38. Weech-Maldonado R, Weidmer BO, Morales LS, et al.Cynamon ML, Kulka RA Cross-cultural adaptation of survey instruments: the CAHPS experience. Seventh Conference on Health Survey Research Methods. Conference Proceedings. 2001 Hyattsville, Md US Dept of Health and Human Services:75–82. Publication PHS 01-1013
39. New & Revised Standards & EPs for Patient Centered Communication. Accreditation Program: Hospital. ©The Joint Commission. 2010. Pre-publication version. Available at: http://www.imiaweb.org/uploads/pages/275.pdf
. Accessed May 15, 2012