EVALUATING THE EVIDENCE FOR EFFECTIVENESS, APPROPRIATENESS, MEANINGFULNESS, AND FEASIBILITY
Effectiveness, a term used similarly by both Evans (2003) and the JBI, refers to the reliability with which a given practice or intervention results in the desired outcome. This is the only domain included in the medical grading systems referred to above, and it is certainly a critical component of any grading system. Meta-analyses, systematic reviews, and multicenter trials are considered the strongest evidence in this domain. The RCT is generally considered the next strongest in demonstrating the efficacy of any single treatment. Because of the rigor involved in conducting a well-designed RCT, the outcome achieved can be reasonably attributed to the intervention (e.g., a medication) and not to some other factor that may have been present, such as the patient's physical condition, other medications or treatment the patient may be undergoing, or variation in the way the intervention or medication was delivered. Other research designs used to establish effectiveness are rated lower than the RCT by all three nursing examples included here. Advanced emergency nurses should keep in mind, however, that although the RCT is an excellent design choice to test interventions in controlled settings (i.e., the efficacy* of the intervention), it is not a particularly good design to study effectiveness in clinical settings where multiple patient-level, provider-level, and institution-level variables interact simultaneously to affect outcomes (Horn & Gassaway, 2007).
Horn and Gassaway (2007) propose relying more on practice-based evidence (PBE) when testing an intervention for effectiveness. Studies using PBE draw on the expertise of care providers to identify multiple patient-level variables and processes of care. Detailed data are collected on all possible patient and process variables that could affect the outcome(s) and then analytic tools are used to isolate a given process or intervention from the “noise” of other variables. Using this method, it is possible to evaluate more than one intervention at a time and determine the relative benefits of each for different types of patients, for example, men versus women; patients older than 65 years versus those younger than 65 years; those with higher acuity scores versus those with lower scores. In the traditional hierarchies of evidence, these observational trials are not generally considered as strong as RCTs and have been accorded a grade of low or very low in some grading recommendations (Grade Working Group, 2004; Guyatt et al., 2006). Evans (2003) places well-designed observational studies alongside RCTs as good evidence of efficacy, whereas the JBI and the AACN still rate RCTs more highly in terms of effectiveness. It is important to note that the emphasis the JBI places on the narrowness of confidence intervals in RCTs and the consistency of findings across multiple studies. Both of these criteria add to the user's confidence that the intervention being evaluated is predictably effective in a variety of settings.
Appropriateness, Meaningfulness, and Feasibility
Studies that address the appropriateness of their findings to various patient populations are especially helpful to emergency nurses. Different emergency departments (EDs) see different types of patients and thus evidence from a study in one ED may or may not be applicable to others. Evans (2003) uses appropriateness to address the psychosocial aspects of care, relating to the patients' experiences, their understanding of health and illness, and the outcomes they hope to achieve with their ED visit. The JBI (JBI, 2008) includes many of these considerations in its domain of meaningfulness, along with ethical considerations. Both Evans and the JBI appreciate that these dimensions of care are best addressed with nonexperimental research designs, including both quantitative and qualitative methods. The JBI emphasizes metasyntheses and consistency of findings across studies to add confidence that the results point to a true understanding of the phenomenon under study.
Feasibility refers to the reality that change is difficult and that change in environments such as EDs and hospitals is difficult even in the face of strong evidence of effectiveness and appropriateness. Here, as with effectiveness and appropriateness, systematic reviews and multicenter trials provide excellent sources of evidence, and RCTs, observational studies, and interpretive studies may yield good evidence, especially related to aspects of organizational culture that could affect both the ease of acceptance of the new practice and help determine how best to implement it (Evans, 2003).
RECOMMENDING EVIDENCE FOR PRACTICE
Evaluating evidence and making recommendations for practice changes are two separate, albeit closely related, processes. Evidence that is recommended for practice must not only meet standards for good internal and external validity, that is, should be considered “strong” enough to instill confidence in those ready to apply the practice, but also be appropriate to both the patients and practice setting, as well as feasible to implement. Using a systematic approach and making explicit the criteria for assigning recommendations gives users of that evidence confidence that their recommendations for practice change reflect the current evidence (Grade Working Group, 2004). The premise is that stronger evidence from high-quality sources is more likely to truly represent best practices than that from lesser quality sources (Newhouse, Dearolt, Poe, Pugh, & White, 2007). Some organizations have chosen simple approaches such as labeling their recommendations “high,” “good,” or “low/major flaw” (Newhouse et al., 2007, p. 90), whereas others such as the AHA (2005) and McGinn et al. (2000) have somewhat more complex levels of recommendation, although they are still focused solely on evidence of effectiveness. Among the nursing resources reviewed here, only the JBI makes explicit recommendations for practice based on the results of the evidence review and they do this across the four domains of effectiveness, appropriateness, meaningfulness, and feasibility. Table 2 shows how the JBI applies its grades to the evidence.
Advanced emergency nurses who are considering a systematic practice change should agree on a grading scheme for use with their review of published evidence. It does not have to be a complex scheme but should contain the domains included by Evans (2003) and the JBI (2008). Although evidence of effectiveness is necessary before recommending or adopting a change in practice, it is not sufficient to address the multiple dimensions that must be considered when nurses apply interventions in their practice setting. Advanced emergency nurses must also consider more than just the design of the studies or recommendations they are evaluating. Studies must be both well designed and well conducted for them to add meaningfully to the scientific basis for practice; “studies of poor methodological quality” provide poor evidence of effectiveness, appropriateness, meaningfulness, and feasibility, and it remains up to the advanced emergency nurse to critically appraise studies with this in mind (Evans, 2003, p. 79).
Many other grading schemes have been proposed both in medicine (AHRQ, 2002; Guyatt et al., 2006; Uhlig et al., 2006) and in nursing (Newhouse et al., 2007), as well as for reviewing clinical guidelines or recommendations (The AGREE Collaboration, 2001). Table 3 provides an overview of some of these schemes that advanced practice nurses may find especially useful. In applying an evidence rating system or assigning a graded recommendation for practice, advanced emergency nurses should rely on the principle of transparency, making explicit to their readers the ways in which evidence has been evaluated and grades assigned.
Advanced emergency nurses are expected to be able to review current and emerging evidence for emergency care and recommend whether or not to implement any practice changes on the basis of this review. Whether they participate in actually reviewing the evidence and assigning the recommendations, or confine themselves to evaluating the recommendations of their professional associations and colleagues, they need to understand and appreciate the meaning of the levels of evidence and associated recommendations for adoption.
American Heart Association. (2005). Part 1: Introduction. Circulation, 112
(22, Suppl.), III-1–III-4.
Armola, R. R., Bourgault, A. M., Halm, M. A., Board, R. M., Bucher, L., Harrington, L., ... Medina, J. (2009). Upgrading the American Association of Critical-Care Nurses' evidence-leveling hierarchy. American Journal of Critical Care, 18
Evans, D. (2003). Hierarchy of evidence: A framework for ranking evidence evaluating healthcare interventions. Journal of Clinical Nursing, 12
Evans, K. D., & Boyce, K. E. (2008). Focusing on the issues. Fostering a hierarchy of evidence within the profession. Journal of Diagnostic Medical Sonography, 24
Grade Working Group (2004). Grading quality of evidence and strength of recommendations. BMJ, 328
Guyatt, G., Gutterman, D., Baumann, M. H., Addrizzo-Harris, D., Hylek, E. M., Phillips, B., ... Schünemann, H. (2006). Grading strength of recommendations and quality of evidence in clinical guidelines. Chest, 129
Horn, S. D., & Gassaway, J. (2007). Practice-base evidence study design for comparative effectiveness research. Medical Care, 45
McGinn, T. G., Guyatt, G. H., Wyer, P. C., Naylor, C. D., Stiell, I. G., & Richardson, W. S. (2000). Users' guides to the medical literature XXII: How to use articles about clinical decision rules. JAMA, 284
Newhouse, R. P., Dearolt, S. L., Poe, S. E., Pugh, L. C., & White, K. M. (2007). Johns Hopkins nursing evidence-based practice model and guidelines
. Indianapolis, IN: Sigma Theta Tau.
Rycroft-Malone, J., Seers, K., Titchen, A., Harvey, G., Kitson, A., & McCormack, B. (2004). What counts as evidence in evidence-based practice? Journal of Advanced Nursing, 47
Shapiro, S. E. (2007). Evidence-based practice for advanced practice emergency nurses. Advanced Emergency Nursing Journal, 29
Shapiro, S. E., & Donaldson, N. E. (2008a). Evidence-based practice for advanced practice emergency nurse, Part III: Planning, implementing, and evaluating an evidence-based small test of change. Advanced Emergency Nursing Journal, 30
Shapiro, S. E., & Donaldson, N. E. (2008b). Evidence-based practice for advanced practice emergency nurses, Part II. Critically appraising the literature. Advanced Emergency Nursing Journal, 30
The AGREE Collaboration. (2001). Appraisal of Guidelines for Research & Evaluation (AGREE) instrument
. Retrieved from www.agreecollaboration.org
Uhlig, K., MacLeod, A., Craig, J., Lau, J., Levey, A. S., Levin, A., ... Eknoyan, G. (2006). Grading evidence and recommendations for clinical practice guidelines in nephrology. A position statement from Kidney Disease: Improving Global Outcomes (KDIGO). Kidney International, 70
*Efficacy means that the intervention works in the controlled research environment; effectiveness means that the intervention works in “real-life” clinical environments. Cited Here...
Keywords:© 2010 Lippincott Williams & Wilkins, Inc.
evidence-based practice; grading evidence