You could be reading the full-text of this article now if you...

If you have access to this article through your institution,
you can view this article in

Agreement and Coverage of Indicators of Response to Intervention: A Multimethod Comparison and Simulation

Fletcher, Jack M.; Stuebing, Karla K.; Barth, Amy E.; Miciak, Jeremy; Francis, David J.; Denton, Carolyn A.

Topics in Language Disorders:
doi: 10.1097/TLD.0000000000000004
Original Articles
Abstract

Purpose: Agreement across methods for identifying students as inadequate responders or as learning disabled is often poor. We report (1) an empirical examination of final status (postintervention benchmarks) and dual-discrepancy growth methods based on growth during the intervention and final status for assessing response to intervention and (2) a statistical simulation of psychometric issues that may explain low agreement.

Methods: After a Tier 2 intervention, final status benchmark criteria were used to identify 104 inadequate and 85 adequate responders to intervention, with comparisons of agreement and coverage for these methods and a dual-discrepancy method. Factors affecting agreement were investigated using computer simulation to manipulate reliability, the intercorrelation between measures, cutoff points, normative samples, and sample size.

Results: Identification of inadequate responders based on individual measures showed that single measures tended not to identify many members of the pool of 104 inadequate responders. Poor to fair levels of agreement for identifying inadequate responders were apparent between pairs of measures. In the simulation, comparisons across 2 simulated measures generated indices of agreement (κ) that were generally low because of multiple psychometric issues inherent in any test.

Conclusions: Expecting excellent agreement between 2 correlated tests with even small amounts of unreliability may not be realistic. Assessing outcomes based on multiple measures, such as level of curriculum-based measure performance and short norm-referenced assessments of fluency, may improve the reliability of diagnostic decisions.

Author Information

Texas Institute of Measurement, Evaluation, and Statistics and Department of Psychology, University of Houston, Houston (Drs Fletcher, Stuebing, Barth, Miciak, and Francis); and Children's Learning Institute, University of Texas Health Science Center at Houston (Dr Denton).

Corresponding Author: Jack M. Fletcher, PhD, Department of Psychology, University of Houston, 2251 W. Holcombe Blvd, 222TMC Annex, Houston TX 77204 (Jack.fletcher@times.uh.edu).

This research was supported by grants P50 HD052117, Texas Center for Learning Disabilities, and K08 HD068545-01A1, Language, Cognitive, and Neuropsychological Processes in Reading Comprehension, from the Eunice Kennedy Shriver National Institute of Child Health and Human Development. The content is solely the responsibility of the authors and does not necessarily represent the official views of the Eunice Kennedy Shriver National Institute of Child Health and Human Development or the National Institutes of Health.

The authors have indicated that they have no financial and no nonfinancial relationships to disclose.

© 2014Wolters Kluwer Health | Lippincott Williams & Wilkins