Anesthesia in the United States is provided by anesthesiologists and nurse anesthetists. There are broad differences in how the 2 groups are used, with nurse anesthetists being more likely to practice in rural areas and more likely to provide monitored anesthesia care, as opposed to general anesthesia.1,2 Generally, nurse anesthetists are supervised by an anesthesiologist, but occasionally they are supervised by a proceduralist (e.g., the surgeon performing the case). Concerns over access to anesthesia care, as well as predicted future shortages of anesthesiologists,1 have led policy makers to consider loosening the degree to which nurse anesthetists must be supervised, up to and including independent practice.
The degree to which nurse anesthetists in the United States require supervision is influenced by several entities including hospitals, state laws, and insurer regulations. Until 2001, the United States Centers for Medicare and Medicaid Services (CMS) required nurse anesthetists to be supervised by a physician (either an anesthesiologist or the proceduralist) in order for providers to receive payment from Medicare for the case. That year, in an effort to increase access to anesthesia care, CMS issued a rule allowing states to “opt-out” of this requirement. Specifically, the rule allows health care providers to receive Medicare payments without requiring physician supervision of nurse anesthetists in states where the governor has issued a letter attesting that: (1) s/he has consulted with the state medical and nursing boards about access to and quality of anesthesia services in the state, (2) it is in the interests of the state’s citizens to opt-out of the federal requirement, and (3) opt-out is consistent with state law. By 2013, 17 states had decided to opt-out of the Medicare regulations requiring supervision of nurse anesthetists, which could, in theory, allow for independent nurse anesthetist practice.3
Debate over the merits of opt-out has focused on 2 issues: whether it has affected quality and whether it has affected access to care. Although a larger literature has explored the issue of quality,4–13 we are unaware of studies examining whether opt-out has improved access to care. This omission is surprising since proponents of opt-out argue that it increases access to carea in underserved or rural areas, by expanding the pool of anesthesia providers. Moreover, the CMS rule states that governors must consider the availability of anesthesia care before opt-out, and in choosing to opt-out, governors have typically cited the necessity of improving access to critical services in rural areas. In theory, opt-out could increase access by facilitating nurse anesthetists’ independent practice and thereby directly increasing the pool of available anesthesia providers. Whether it has done so in practice remains an open question.
In this article, we use data from the National Inpatient Sample (NIS), a nationally representative sample of inpatient discharges, to examine whether opt-out was associated with increased access to anesthesia care for urgent procedures by considering 2 measures of access. First, we examine whether opt-out was associated with an increase in the percentage of patients admitted for appendicitis, intestinal obstruction, choledocholithiasis, and hip fracture (conditions often requiring urgent access to anesthesia care) who received a procedure requiring anesthesia. Many factors, such as surgeon and patient preferences, will affect the number and percentage of patients who ultimately receive surgery for these conditions, particularly since there are nonoperative alternatives in some cases (e.g., intestinal obstruction). Nonetheless, the percentage of patients receiving surgery is typically used as a proxy for access to surgery.14,15 Second, we examine whether opt-out was associated with a decrease in the number of acute appendicitis patients presenting with a ruptured appendix, because the incidence of ruptured appendicitis has also been used to measure access to surgery.16
Data and Sample Selection
This study used deidentified data and did not require IRB approval. We used data from the 1998 to 2010 NIS, produced by the Healthcare Cost and Utilization Project. The NIS is a nationally representative sample of inpatient visits drawn from community hospitals in the United States.b Each year, community hospitals from participating states are classified into strata based on rural/urban location, number of beds, region, teaching status, and ownership. A 20% random sample of hospitals is then selected from each stratum, and all admissions to these hospitals in the given year are recorded. The NIS provides detailed patient-level data for each admission, including demographic variables such as age, sex, and insurance status, as well as International Classification of Diseases, Ninth Revision (ICD-9) procedure and diagnosis codes obtained from the discharge abstract. The diagnoses can be the admitting diagnosis, as well as diagnoses present on discharge. In addition, the NIS provides data on the hospital, such as its state and whether it is located in a rural or urban area. The NIS has been extensively used to examine trends in health care utilization,17 as well as the effect of policies at the state level.18,19
We extracted all patients with a diagnosis code (either primary or secondary diagnosis), indicating appendicitis, bowel obstruction, choledocholithiasis, or hip fracture. Our inclusion/exclusion criteria for each condition were adopted from previous studies and are shown in Table 1.15,17,20–23 In all, we identified 802,892 cases of appendicitis, 724,475 cases of bowel obstruction, 293,863 cases of choledocholithiasis, and 808,888 cases of hip fracture. We then excluded patients younger than 18 years, patients with more than one of these diagnoses, and patients with missing data on our covariates of interest (see below). With these exclusions, our final sample consisted of 574,807 cases of appendicitis, 693,257 cases of bowel obstruction, 286,196 cases of choledocholithiasis, and 798,870 cases of hip fracture. Of note, the methods used to identify choledocholithiasis exclude chronic biliary obstruction (a nonurgent scenario), while the methods used to identify hip fracture include low-velocity hip fractures and exclude high-velocity hip fractures, such as those associated with polytrauma.
Our outcomes consisted of 2 measures of access. The first was whether an individual patient underwent a procedure (typically a surgery), which we inferred by the presence of the ICD-9 procedure codes listed in Table 1. Using these codes, we generated a dummy variable, which equaled 1 if a patient underwent a procedure and 0 otherwise. For choledocholithiasis, patients could receive 3 types of procedures (i.e., surgery, endoscopic procedures, and percutaneous procedures), and we considered any of the 3. It is important to note that the NIS does not follow a patient across hospitals, so if a patient was admitted and then transferred to another hospital for the procedure, we would have no record of the subsequent admission. However, we would observe that the patient did not receive a procedure at the initial hospital. Since transfer represents a lack of access, both in terms of patient inconvenience and because it delays the receipt of a potentially urgent procedure, our approach counted these transferred patients as not having received a procedure. Our second outcome, limited to patients with appendicitis, was a dummy variable equaling 1 if the patient presented with a ruptured appendix (ICD-9 540.0 or 540.1) and zero otherwise.
Our analysis included 2 independent variables of interest. The first indicated whether opt-out was in effect. Specifically, a dummy variable equaled 1 in the opt-out states starting the year following the opt-out year (Table 1). For non-opt-out states and for opt-out states in the years before and including the opt-out year, the variable equaled zero. It is important to note that the number of states participating in the NIS has grown over time from 22 in 1998 to 46 in 2010. Therefore, our analysis of the NIS characterized the experience of opt-out states that (a) reported data to the NIS and (b) chose to opt-out at least 1 year after initially providing data to the NIS. For example, Colorado opted-out in 2010 and so was not examined. Table 2 provides a list of opt-out states, the year they chose to opt-out, and the year they began providing data to the NIS.
We also obtained additional control variables including patient age and sex. In addition, we controlled for comorbidities using a version of the Charlson Index that has been modified for use with administrative databases.24 Although the NIS does record race, we did not include controls for race in our analysis because it was inconsistently reported across states in the NIS, and indeed was missing for 23% of our sample.
Our analysis used a difference-in-differences approach to further minimize confounding.25 Even after controlling for the various patient factors listed above (e.g., Charlson Index and age), a simple comparison of outcomes between opt-out and non-opt-out states is likely to face additional confounding due to unobservable differences between patients in these 2 groups of states. With a difference-in-differences approach, state-specific controls are used to adjust for unobservable state-level factors. Therefore, rather than comparing outcomes across states, the difference-in-differences approach identifies the effect of opt-out by comparing outcomes in the pre- and post-opt-out period within the opt-out states. The difference-in-differences approach also controls for year effects to control for secular trends at the national level, such as general changes in surgical practice.
We implemented our difference-in-differences approach by using a multivariable linear regression. Our dependent variable in this regression was an indicator variable indicating whether the patient received a procedure/presented with a ruptured appendix. Our independent variable of interest was a dummy variable indicating whether opt-out is in effect. The coefficient on this variable can be interpreted as the difference in the percentage of patients undergoing a procedure/presenting with appendiceal rupture following opt-out in urban areas. For example, if the coefficient equaled 0.0500, this would imply that opt-out was associated with a 5.00% (absolute) increase in the probability of surgery/appendiceal rupture.
Our baseline analysis combines patients with appendicitis, bowel obstruction, choledocholithiasis, or hip fracture when examining how opt-out affected the probability of receiving a procedure. In addition, we analyzed the effect of opt-out on each diagnosis separately. In the case of appendiceal rupture, our analysis was necessarily limited to patients presenting with appendicitis. In addition to the difference-in-differences approach described above, we incorporated the use of state-specific linear and quadratic year trends to further reduce confounding by controlling for secular time trends at the state level. For example, these trends would control for changes in surgical practice occurring at a state level. Since the NIS is a multistage, stratified sample of admissions, all our analyses incorporate appropriate sampling weights (the TRENDWT variable), and our standard errors are adjusted for clustering by state.26 We performed our analyses with STATA 13.0 (College Station, TX). We adjusted our confidence intervals for each of the 4 individual procedures composing our composite end point using the Bonferroni method. Further details are found in Appendix 1.
The baseline analyses we described above examined the degree to which opt-out affected outcomes at across the entire state. However, these analyses may understate the effect of opt-out if its effects were limited to particular parts of the state. For example, suppose that, as a result of opt-out, more hospitals were opened in rural areas so that, instead of being transferred to an urban hospital, rural patients now received their procedure closer to home. Then at the statewide level, there would be no change in the overall number of procedures (or probability of receiving a procedure). Procedures that had been done in urban areas would now be done in rural areas. However, this finding would underestimate the true effect of opt-out because rural patients are now able to receive their procedure closer to home, and access has increased.
To address this scenario, we tested for the possibility that opt-out had differential effects in urban and rural areas. Our approach was built on the baseline analyses described above by incorporating a second independent variable of interest consisting of an interaction term between our opt-out variable and a dummy variable that equaled 1 if a hospital were located in a rural area and 0 otherwise. Our interaction term was produced by multiplying our opt-out variable by our variable indicating rural status. Thus, it equaled 1 if opt-out were in effect and the hospital were located in a rural area, and 0 otherwise. By adding this interaction term, the interpretation on our original opt-out variable now changes slightly, because it now represents the effect of opt-out in urban areas, while the coefficient on this new interaction term represents the effect of opt-out in rural areas, relative to its effect in urban areas. For example, suppose that the coefficient on our opt-out variable were 0.0500 and the coefficient on our interaction term −0.0200. In tandem, these results would indicate that opt-out was associated with a 5.00% (absolute) increase in procedure rates/appendiceal rupture in urban areas and that the effect of opt-out was 2.00% lower in rural areas, relative to urban areas. This, the effect of opt-out in rural areas overall would be a 3.00% (=5.00% − 2.00%) increase. In short, the coefficient on the interaction term is a test of the hypothesis that opt-out had a different effect in rural areas.
We used the NIS variable HOSP_LOCATION to identify rural hospitals. Before 2004, the NIS defined urban hospitals as those located in a Metropolitan Statistical Area (MSA) as delineated by the United States Office of Management and Budget (OMB). Hospitals not located in an MSA were defined as rural.c In defining an MSA, the OMB first looks for an urban area with a population of at least 50,000 people. The MSA is then defined as the set of counties containing the urban area, as well as the set of counties with close economic and social ties to the urban area, as measured by commuting patterns. In 2004, the definition of “rural” used by the NIS changed slightly primarily because it now defined rural counties based on a newer OMB entity, the metropolitan Core-Based Statistical Area (CBSAs). Functionally, metropolitan CBSAs and MSAs are similarly defined. However, in identifying metropolitan CBSAs, the NIS used data from the 2000 United States census, while it had used data from the 1990 United States census to define MSAs. This change resulted in a slight decrease in the number of hospitals classified as rural, with a corresponding increase the number of hospitals classified as urban.
Table 3 shows summary statistics for our sample. With regard to age, sex, and Charlson Index, there were some differences between the opt-out and non-opt-out states that were small in magnitude (standardized differences scores <0.05) and generally always of no statistical significance. There were no significant differences in the percentage of patients receiving a procedure in the case of hip fracture, choledocholithiasis, and appendicitis. We found that a slightly smaller percentage of patients with bowel obstruction received a procedure in opt-out states (27.1% vs 28.9%, P = 0.029).
Figure 1 plots trends in surgical rates for patients with bowel obstruction, appendicitis, choledocholithiasis, and hip fracture. The figure demonstrates 2 important facts. First, as indicated from previous literature, the percentage of patients undergoing a procedure was smaller in rural areas. Second, while the percentage of patients undergoing a procedure overall declined during this period, the decline qualitatively “looks” sharper among the opt-out states, particularly in rural areas. While this result provides a basis for suggesting that opt-out did not increase access to surgery, the goal of our difference-in-differences approach is to analyze this result more rigorously by imposing additional controls for case mix, patient characteristics, and secular time trends at the national and state level. Similar to Figure 1, Figure 2 plots the percentage of patients with appendicitis who presented with a ruptured appendix. Overall, there appears to have been a downward trend in the incidence of appendiceal rupture during this period. Qualitatively, the trend does not appear to be different in opt-out versus non-opt-out states, again providing some basis to suggest that opt-out did not decrease rates of appendiceal rupture.
Table 4 presents the results of our analysis in 3 columns. The first column presents our baseline results, which show the effect of opt-out at the state level. Across all 4 diagnoses (bowel obstruction, appendicitis, choledocholithiasis, and hip fracture), opt-out was not associated with a significant increase in the percentage of patients who received a procedure in urban hospitals (0.0315 percentage point increase, P = 0.926). From the confidence interval, there could be an (absolute) increase as large as 0.906%. When considered by procedure, opt-out was associated with statistically insignificant changes in the percentage of patients undergoing a procedure for bowel obstruction (0.511 percentage point decrease, P = 0.455), choledocholithiasis (−2.78 percentage point increase, P = 0.144), and hip fracture (0.291 percentage point increase, P = 0.713). In the case of appendicitis, we observed a small but statistically significant increase in the percentage of patients undergoing a procedure (0.876 percentage point increase, P = 0.008). Opt-out was also associated with a statistically insignificant change in the percentage of appendicitis patients who presented with a ruptured appendix (0.914 percentage point decrease, P = 0.119).
In addition to our baseline results, Table 4 presents the results of a subanalysis examining whether opt-out had differential effects in rural and urban areas. Column 2 shows the effect of opt-out in urban areas, while column 3 shows the impact of opt-out in rural areas relative to its effect in urban areas. The values in column 3, therefore, expressed the extent to which opt-out affected rural areas separately from urban areas. Across all our outcomes, the values shown in column 3 are statistically insignificant, suggesting that the effects of opt-out were not different across rural and urban areas.
Although many states have chosen to opt-out with the stated intention of increasing access to anesthesia care, whether the decision to do so has actually increased access is an open question. Using 2 measures of access: the percentage of patients receiving surgery for 4 urgent conditions (appendicitis, small bowel obstruction, hip fracture, and choledocholithiasis) and the percentage of patients with appendicitis who presented with a ruptured appendix, we found that opt-out was not associated with increased access to anesthesia care. Our results also suggest that opt-out did not have heterogeneous effects across urban and rural areas.
These results may not necessarily be surprising. First, as previously discussed, opt-out merely means that the federal government will continue to pay providers when nurse anesthetists practice independently. This does not require states, hospitals, insurers, or surgeons to accept independent nurse anesthetist practice; each of these groups is free to restrict the scope of nurse anesthetist practice. Indeed, 14 of the 17 opt-out states do not allow true independent nurse anesthetist practice, because statutes and regulations at the state level require nurse anesthetists to work in a relationship with a physician, which may involve physician supervision, collaboration, consultation, agreement, accountability, or discretion over nurse anesthetists providing anesthesia services. Second, even if opt-out expands the supply of anesthesia providers, this expansion will not result in an increase in anesthesia services if the supply of proceduralists is insufficient. In this light, our results should not imply that opt-out could never increase the supply of anesthesia services. Rather, our results suggest that opt-out is unlikely to increase the supply of anesthesia services in the absence of policies addressing the issues outlined above. These features are not limitations of our study, however, because our goal was to examine the effects of opt-out as it was actually implemented.
Our study should be viewed in light of its limitations. We cannot exclude the possibility that unobserved factors, such as changes in patient population or surgical practice, could explain our results. However, our difference-in-differences approach does minimize the possibility of confounding. For a factor to confound our results, it would have to be associated with the timing of opt-out among each opt-out state and be independent of yearly effects at a national level, as well as secular time trends at the state level. In addition, as previously noted, our analysis does not cover the entirety of opt-out states, because not all states submitted data to the NIS. In addition, our findings only address the issue of whether opt-out has affected access to care; whether it has affected quality remains an area for further study. Finally, our study is limited to the case of urgent surgery for 4 conditions. As previously noted, nurse anesthetists and anesthesiologists tend to practice anesthesia in different settings. It is possible that opt-out could have increased access to other types of anesthesia care, such as monitored anesthesia care for elective cases. Whether it may have done so, and the social value of these increases, are subjects for further research, using a dataset different from the NIS that we used.
In sum, our results suggest that opt-out has not significantly increased access to anesthesia care for urgent inpatient procedures. Although we did not consider elective and/or outpatient procedures, presumably one of the primary goals of opt-out would have been to increase access to urgent inpatient procedures, both because of their acuity and because, as demonstrated here, rates of surgery are markedly lower in rural areas, with potential implications for patient health. While increasing access to surgery, particularly in underserved areas, is a laudable goal, our results suggest that opt-out alone is unlikely to address this issue.
In this technical appendix, we describe in detail the implementation of our difference-in-differences approach. As discussed in the main body of the paper, a simple comparison of outcomes between opt-out and non-opt-out states may suffer from confounding due to differences in observable and unobservable factors between the 2 groups, such as patient type and surgical practice patterns. Although we did control for many plausible differences—such as Charlson scores, race, and sex—these controls might not fully account for all plausible confounders.
The difference-in-differences approach, which has been extensively used to evaluate the effects of policy,25 provides one way to minimize confounding. At heart, the difference-in-differences approach estimates the effect of opt-out by performing 2 calculations. First, it evaluates the change in outcomes among “opt-out” states following the decision to opt-out rather than simply comparing opt-out states to non-opt-out states. By focusing on the changes in outcomes before and after opt-out within a given state, the difference-in-differences approach eliminates much of the confounding that may occur due to differences in patients, surgeons, and hospitals across states. However, this simple “before-after” comparison can also be confounded by secular trends (such as changes in practice patterns) at the national level. The difference-in-differences approach addresses this issue by including year-specific controls to control for year-to-year variability in surgical rates at the national level.
We can implement a basic difference-in-differences approach using the following linear regression:
where surgery ijt is an indicator variable indicating whether patient i received surgery in state j in year t. surgery ijt was set to equal 1 if the patient received surgery and 0 otherwise. f i is a state effect for state i and d t is a year effect for year t. Γ i is a vector of patient characteristics: age, sex, and Charlson score. The year effects d t, state effects f i, age, and Charlson scores were modeled as categorical variables. In other words, they were composed of vectors of indicator variable equaling 1 if the patient met the relevant criteria and 0 otherwise. For example, the state effects f i consisted of a vector of indicator variables, one for each state in our analysis, equaling 1 if the hospital was located in the given state and 0 otherwise. Similarly, for age, we incorporated a vector of indicator variables equaling 1 if the patient was of the given years of age (e.g., age = 62 years) and 0 otherwise. opt-out ijt is an indicator variable which equals one if opt-out was in effect in state j during year t and 0 otherwise. ε ijt is the error term in the regression. Our coefficient of interest is β, which represents the absolute (percentage point) change in the percentage of patients who received a procedure associated with a state’s decision to “opt-out.”
We expanded the basic regression shown in Equation (1) in 2 ways. First, d t, our year effect, was used to control for secular time trends at the national level. In our final regression model, we incorporated additional controls for secular time trends occurring at the state level, leading to the following regression equation:
Equation (2) is similar to Equation (1), except that we included the terms year × f j, which controls for secular linear trends occurring at the state level, and year 2 × f j, which controls for secular quadratic trends occurring at the state level. Equation (2) formed the basis for our baseline analyses, which examined how opt-out affected our outcomes at the state level. As we noted, however, it is also important to test the potential that opt-out may have affected urban and rural areas differentially. To test this hypothesis, we used the following regression equation:
Equation (3) is similar to Equation (2), except that we included the term opt-out jt × rural it, which was our interaction term. This term was obtained by multiplying our opt-out variable, opt-out jt, by another indicator variable, rural it, equaling one if the patient was seen in a rural hospital and zero otherwise. Thus, our interaction term, opt-out jt × rural it, equaled one only if opt-out was in effect and if the patient was seen in a rural hospital. As can be seen from Equation (3), β 1 represented the effect of opt-out in urban areas, while β 2 represented the effect of opt-out in rural areas relative to its effect in urban areas. Therefore, testing the hypothesis that opt-out had a different effect in rural areas was a test of the hypothesis that β 2 = 0. In addition, note that the total effect of opt-out in rural areas was β 1 + β 2. Although not shown in Equation (3) for the purposes of exposition, in estimating Equation (3), we also interacted rural it with each of the other covariates (e.g., f i, d t, (year × f j), (year 2 × f j), and X i) to allow for the possibility that the effects of these variables could differ in urban and rural areas. Given the large number of terms in our regression, we only reported the results for β 1 and β 2. Results for the other regression coefficients are available upon request.
Because the National Inpatient Sample employs a complex sampling strategy, we incorporated sampling weights (the TRENDWT variable) in our analyses. In addition, we clustered our errors at the state level. Our sample consists of 2,353,130 observations, each with a corresponding sampling weight. A simple ordinary least squares regression would tend to underestimate our standard errors (and therefore overestimate the statistical significance of our regression coefficients) because these observations were not independent, since observations within a given state were likely to be correlated. Calculating clustered standard errors is an appropriate approach to deal with this issue.27 In essence, clustering adjusts the standard errors based on the observed level of correlation within a given unit (cluster) defined by the investigator. Since we were primarily concerned with correlation within a given state, we clustered our standard errors at the state level.28
By using a linear regression, we estimated a linear probability model, in contrast to a probit or logistic regression. We chose a linear probability model for several reasons. First, implementing our difference-in-differences model, along with the controls for linear and quadratic trends, required the use of numerous indicator variables, which can lead to computational difficulties with logistic or probit regression.29 Second, the coefficients from the linear probability model are easier to interpret and directly correspond to the increase in probability that a patient received surgery. By contrast, it is harder to translate the coefficients from a logistic or probit regression into the change in probability for receiving surgery. For example, odds ratios are easily estimated from a logistic regression and are an unbiased estimator for changes in probability for rare events.30,31 However, since around 50% of patients receive a procedure, odds ratios would be a biased measure of probability changes in our case. Finally, since our analysis incorporated an interaction term between 2 dummy variables to estimate how opt-out affects the probability of surgery in rural areas, we chose a linear probability model as the interpretation of this interaction term is more straightforward.32
Name: Eric Sun, MD, PhD.
Contribution: This author helped design the study, conduct the study, analyze the data, and write the manuscript.
Conflicts: Eric Sun reported no conflicts of interest.
Attestation: Eric Sun has seen the original study data, reviewed the analysis of the data, approved the final manuscript, and is the author responsible for archiving the study files.
Name: Franklin Dexter, MD, PhD.
Contribution: This author helped design the study, analyze the data, and write the manuscript.
Conflicts: Franklin Dexter reported no conflicts of interest.
Attestation: Franklin Dexter has reviewed the analysis of the data and approved the final manuscript.
Name: Thomas R. Miller, PhD, MBA.
Contribution: This author helped design the study and write the manuscript.
Conflicts: Thomas R. Miller is employed by the American Society of Anesthesiologists.
Attestation: Thomas R. Miller has seen the original study data, reviewed the analysis of the data, and approved the final manuscript.
Applicable as of the accept date of this article (November 20, 2015), Dr. Dexter is the Statistical Editor for Anesthesia & Analgesia. This manuscript was handled by Dr. Jean-Francois Pittet, Incoming Editor-in-Chief, and Dr. Dexter was not involved in the editorial process or decision.
a See “February Marks 10-Year Anniversary of Landmark Decision Ensuring Nebraskans Access to Safe, Cost-Effective Anesthesia Care,” available at http://www.aana.com/newsandjournal/News/Pages/022412-February-Marks-10-Year-Anniversary-of-Landmark-Decision-Ensuring-Nebraskans.aspx.
b In essence, “community hospitals” exclude Federal hospitals (e.g., hospitals run by the United States Veterans Administration or the military), as well as hospitals that are part of an institution (e.g., a prison hospital).
c See https://www.hcup-us.ahrq.gov/db/vars/hosp_location/nisnote.jsp.
1. Daugherty L, Fonseca R, Kumar KB, Michaud P-C. An Analysis of the Labor Markets for Anesthesiology. 2010Santa Monica, CA: RAND Corporation.
2. Bayman EO, Dexter F, Laur JJ, Wachtel RE. National incidence of use of monitored anesthesia care. Anesth Analg 2011;113:165–9.
3. Miller TR, Abouleish A, Halzack NM. Anesthesiologists are affiliated with many hospitals only reporting anesthesia claims using modifier QZ for Medicare claims in 2013. A A Case Rep2016;6:217–9.
4. Lewis SR, Nicholson A, Smith AF, Alderson P. Physician anaesthetists versus non-physician providers of anaesthesia for surgical patients. Cochrane Database Syst Rev 2014;7:CD010357.
5. Needleman J, Minnick AF. Anesthesia provider model, hospital resources, and maternal outcomes. Health Serv Res 2009;44:464–82.
6. Kane M, Smith AF. An American tale—professional conflicts in anaesthesia in the United States: implications for the United Kingdom. Anaesthesia 2004;59:793–802.
7. Fleisher LA, Anderson GF. Perioperative risk: how can we study the influence of provider characteristics? Anesthesiology 2002;96:1039–41.
8. Abenstein JP, Warner MA. Anesthesia providers, patient outcomes, and costs. Anesth Analg 1996;82:1273–83.
9. Pine M, Holt KD, Lou YB. Surgical mortality and type of anesthesia provider. AANA J 2003;71:109–16.
10. Silber JH, Kennedy SK, Even-Shoshan O, Chen W, Koziol LF, Showan AM, Longnecker DE. Anesthesiologist direction and patient outcomes. Anesthesiology 2000;93:152–63.
11. Simonson DC, Ahern MM, Hendryx MS. Anesthesia staffing and anesthetic complications during cesarean delivery: a retrospective analysis. Nurs Res 2007;56:9–17.
12. Smith AF, Kane M, Milne R. Comparative effectiveness and safety of physician and nurse anaesthetists: a narrative systematic review. Br J Anaesth 2004;93:540–5.
13. Dulisse B, Cromwell J. No harm found when nurse anesthetists work without supervision by physicians. Health Aff (Millwood) 2010;29:1469–75.
14. Francis ML, Scaife SL, Zahnd WE. Rural-urban differences in surgical procedures for Medicare beneficiaries. Arch Surg 2011;146:579–83.
15. Shelton J, Kummerow K, Phillips S, Griffin M, Holzman MD, Nealon W, Pinson CW, Poulose BK. An urban-rural blight? Choledocholithiasis presentation and treatment. J Surg Res 2012;173:193–7.
16. Paquette IM, Zuckerman R, Finlayson SR. Perforated appendicitis among rural and urban patients: implications of access to care. Ann Surg 2011;253:534–8.
17. Mancini GJ, Petroski GF, Lin WC, Sporn E, Miedema BW, Thaler K. Nationwide impact of laparoscopic lysis of adhesions in the management of intestinal obstruction in the US. J Am Coll Surg 2008;207:520–6.
18. Shetty KD, DeLeire T, White C, Bhattacharya J. Changes in U.S. hospitalization and mortality rates following smoking bans. J Policy Anal Manage 2010;30:6–28.
19. Safavi A, Rhee P, Pandit V, Kulvatunyou N, Tang A, Aziz H, Green D, O’Keeffe T, Vercruysse G, Friese RS, Joseph B. Children are safer in states with strict firearm laws: a National Inpatient Sample study. J Trauma Acute Care Surg 2014;76:146–50.
20. Schraufnagel D, Rajaee S, Millham FH. How many sunsets? Timing of surgery in adhesive small bowel obstruction: a study of the Nationwide Inpatient Sample. J Trauma Acute Care Surg 2013;74:181–7.
21. Masoomi H, Mills S, Dolich MO, Ketana N, Carmichael JC, Nguyen NT, Stamos MJ. Comparison of outcomes of laparoscopic versus open appendectomy in children: data from the Nationwide Inpatient Sample (NIS), 2006-2008. World J Surg 2012;36:573–8.
22. Hagendorf BA, Liao JG, Price MR, Burd RS. Evaluation of race and insurance status as predictors of undergoing laparoscopic appendectomy in children. Ann Surg 2007;245:118–25.
23. Patorno E, Neuman MD, Schneeweiss S, Mogun H, Bateman BT. Comparative safety of anesthetic type for hip fracture surgery in adults: retrospective cohort study. BMJ 2014;348:g4022.
24. Quan H, Li B, Couris CM, Fushimi K, Graham P, Hider P, Januel JM, Sundararajan V. Updating and validating the Charlson comorbidity index and score for risk adjustment in hospital discharge abstracts using data from 6 countries. Am J Epidemiol 2011;173:676–82.
25. Buckley J, Shang Y. Estimating policy and program effects with observational data: the “differences-in-differences” estimator. Practical Assessment, Research & Evaluation 2003;8. Available at: http://PAREonline.net/getvn.asp?v=8&n=2
. Accessed January 12, 2016
26. Bertrand M, Duflo E, Mullainathan S. How much should we trust differences-in-differences estimates? Q J Econ 2004;119:249–75.
27. Wooldridge JM. Cluster-Sample Methods in Applied Econometrics. Am Econ Rev 2003;93:133–8.
28. Bertrand M, Duflo E, Mullainathan S. How much should we trust differences-in-differences estimates? Q J Econ 2004;119:249–75.
29. Allison P. Convergence failures in logistic regression. SAS Global Forum, 2008;Paper 360
30. Prasad K, Jaeschke R, Wyer P, Keitz S, Guyatt G; Evidence-Based Medicine Teaching Tips Working Group. Tips for teachers of evidence-based medicine: understanding odds ratios and their relationship to risk ratios. J Gen Intern Med 2008;23:635–40.
31. Prasad V, Jena AB. Prespecified falsification end points: can they validate true observational associations? JAMA 2013;309:241–2.
© 2016 International Anesthesia Research Society
32. Ai CR, Norton EC. Interaction terms in logit and probit models. Econ Lett 2003;80:123–9.