Abstract
This article accounts for multivariate dependence of the variable of policy interest in dynamic panel data models by disentangling the two sources of intertemporal dependence: one from the effect of the policy variable and the other from mean reversion. In a situation where intensity of the policy varies over time, we estimate the unconditional mean in the autoregressive process as a function of the agent’s characteristics and the policy intensity. Comparison of the fitted values of the unconditional mean under different values of the policy intensity enables identification of the policy effect cleared of mean reversion. The approach is relevant for measuring the effect of reforms, which use an intertemporal incentive where intensity of the reform varies over time. The empirical part of the article assesses the effect of hospital financing reform based on incentive contracts, related to the observed quality of services at Medicare hospitals in 2013–2019. We find a direct association between prior quality and quality improvement owing to the reform. Our result reassesses a stylized fact in the literature, which asserts that a pay-for-performance incentive leads to greater improvements at hospitals with lower baseline quality.
1 Introduction
The phenomenon of regression toward the mean (mean reversion) is observed in case of longitudinal observations of a variable, which is susceptible to random variations. In this case, exceptionally low or high values of the variable in initial measurement tend to be closer to the center of the distribution in subsequent measurements [24]. In short, mean reversion is an inherent part of the stationary process and implies the return of the process to its mean value [25,31].[1]
Historically, the appearance of the term “mean reversion” is associated with the seminal works by Galton, who discovered an inverse relationship between the height of parents and children [30] and hence framed the term “regression” as the tendency of the dependent variable to revert to the mean value. Recent examples of the analysis of processes which exhibit mean reversion in various fields of economics include the current account of countries [81] and their productivity [29], profitability of banks [48], housing prices [31], tax avoidance by companies [3], blood pressure and cholesterol level of patients [5], and birthweight of children in successive pregnancies of the same mother [79].
Mean reversion contaminates judgment about the time profile of the dependent variable in case of groupwise estimations. If the value of the dependent variable for a certain observation is lower than average in period
The purpose of this article is to model multivariate dependence of the variable of policy interest by disentangling the two sources of intertermporal dependence: one from the effect of the policy of interest per se and the other from mean reversion. Specifically, we show a way of separating the effect of mean reversion from the policy effect when evaluating the impact of an incentive scheme with intertemporal stimuli and intertemporal variation of the parameter of the reform intensity.
Although mean reversion is inherent to any stationary process, it is most often noted in the analysis of dynamic panels. The dynamic panel data model is a generalization of the panel data fixed effect regression when the dynamic structure of the process needs to be introduced. In our article we use the example of Medicare’s incentive contract applied to the observed quality of services, which has to be described as an autoregressive process. Hence, in evaluating the effect of this incentive scheme on hospital quality, we follow a handful of articles which deal with mean reversion in dynamic panels [25,31,48,81].
We focus on the pay-for-performance mechanism – an innovative method of remuneration, which originally emerged in corporate finance and managerial economics, and has since been much used in the public sector (civil service, education, social work, and healthcare). In order to quantify the unobserved quality of work, the incentive scheme computes the performance level using imprecisely measured proxies for various dimensions of quality. Next, the regulator imposes an incentive contract, which relates remuneration to performance, so that agents with higher performance in the current period receive higher payment for their services in future periods than agents with lower performance. The reform intensity parameter in this context is the share of the agent’s income, which is “at risk” under the incentive contract.
Assuming a direct association between demand for services and quality of work, higher payment to agents with high performance incentivizes agents to improve their level of quality in order to raise demand for their services. In such a setting, if the unobserved quality could be measured precisely, each agent would have sustained their fixed level of performance.
However, performance is in fact a noisy signal. First, there is an imprecision in measuring performance, since it is only a proxy for true quality. Second, in case of healthcare, the unobserved true quality of services is itself subject to a random variation, due, for instance, to patient non-compliance with medical treatment [62]. So it is plausible to assume that performance contains a random error. Hence, performance may unexpectedly be valued as having improved in period
What therefore happens is that performance of the economic agent becomes a process with serial correlation. So the evolution of the variable of policy interest when such incentives are applied can be viewed as an autoregressive process. In a situation where the policy variable changes over time, we estimate the unconditional mean in the autoregressive process as a function of the agent’s characteristics and of policy intensity. Comparison of the fitted values of the unconditional mean under different values of the reform intensity enables us to identify the reform effect cleared of mean reversion. For instance, we contrast the unconditional means estimated under the values of the policy variable in two consecutive time periods. Alternatively, we compare the fitted value of the unconditional mean in period
It should be noted that our identification strategy is close to difference-in-difference analysis in a non-binary treatment: the intensity of reform is the analogue of the treatment variable and the share of Medicare’s patients at the hospital is the analogue of the variable for the treatment/control groups.[2]
We use the example of Medicare’s value-based purchasing, implemented at national level in the US since 2013 on the basis of a reward function that relates the aggregate measure of hospital performance to remuneration. Overall, applications of pay-for-performance are very numerous in healthcare, since healthcare is the classic example of an industry with asymmetric information where sustained quality of service is extremely important. It should be noted that the research in health economics is vulnerable to random shocks in the dependent variable and hence, to the phenomenon of mean reversion. Yet, as regards incentive schemes, to the best of our knowledge, only one article explicitly discusses the impact of random variation of quality [62] and only a few articles point to the need for reassessing the impact of Medicare’s pay-for-performance incentive mechanisms in view of the potential impact of mean reversion [58,63].
Our estimations of the association between the observed level of prior quality and measured quality improvement employ nationwide data for 2,984 acute-care Medicare hospitals which are financed according to the quality-incentive mechanism in 2013–2019. The empirical approach uses annual variation in the size of quality incentives in order to estimate the effect of pay-for-performance cleansed of mean reversion. We control for other potential channels of quality improvement by Medicare hospitals, using data on the Hospital Readmissions Reduction Program (HRRP) and on the meaningful use of Electronic Health Records (EHR).
We find that the higher the quintile of the composite quality measure at Medicare hospitals, the larger the estimated effect of the reform. Our empirical results suggest that the stylized fact of inverse relationship between improvement owing to the incentive scheme and the baseline performance should be revisited. This inverse relationship has been found by most empirical assessments of the impact of incentive contracts on healthcare quality and seems to hold for various designs of pay-for-performance: it is observed for general practitioners in the UK; physician groups in California, Chicago, and Ontario; US hospitals in Michigan, New York, and Wisconsin; and hospitals involved in Medicare’s pilot project for quality improvement [19,26,34,40,42,51,52,63,67,76]. However, we argue that the finding of an inverse relationship may be incorrect when the empirical approach fails to account for the impact of the random shocks on the time profile of quality under the intertemporal incentive scheme.
The remainder of the article is structured as follows. Section 2 reviews the design of Medicare’s quality incentive and sets up the framework for evaluating its outcomes. Section 3 outlines the empirical methodology, and Section 4 describes the data for Medicare hospitals. The results of the empirical analysis are presented in Section 5. Section 6 contains a discussion of our approach in view of conventional methods for policy evaluation, and Section 7 supports the quantitative findings of our analysis by suggesting potential channels used for quality improvement at hospitals.
2 Medicare’s incentive contract
2.1 Policy setting
The mechanism provides an incentive proportional to measured quality and has been applied to discharges in the inpatient prospective payment system at acute-care Medicare hospitals since 2013.[3] The scheme reduced Medicare’s base payment[4] to each hospital by a factor
The accumulated saving from reduction in base payment is redistributed across hospitals according to an adjustment coefficient, which is computed as a linear function of the composite quality measure:
The TPS is a weighted sum of scores for measures in several domains: timely implementation of recommended medical interventions (clinical process of care), quality of healthcare as perceived by patients (patient experience of care), survival rates for AMI, heart failure and pneumonia patients and other proxies for outcome of care, healthcare-associated infections and other measures of safety of care, and spending per beneficiary as a measure of efficiency of care.[5]
A hospital’s intertemporal incentive in Medicare’s scheme is based on the expectation that the quality payments will continue over a long term, so the hospital’s executives and physicians realize that demand is proportionate to quality and that their current policies toward quality of care will influence future reimbursement [46,73].
2.2 Autoregressive process and quality convergence
The evolution of the measured quality constitutes a process with serial correlation. If the process for the measured quality is stationary, then it may be treated as an autoregressive process
Using definitions in [29], we can disentangle a permanent component in
The reason for the phenomenon of mean reversion is the existence of the random error
The autoregressive specification can be taken as equivalent to convergence of the measured quality toward the value
The hospital receives higher profits for improvement of performance under higher values of
2.3 Expected outcomes of the reform and time profiles of the quality measure
2.3.1 Mean effect of the reform
The payment schedule makes the hospital adjustment coefficient a linear function of TPS, so each hospital has an incentive to raise the value of the observed composite quality measure. Hence, the introduction of pay-for-performance is expected to have a positive effect on mean value of the composite quality measure. Indeed, the mean level of hospital performance was improved even in case of a continuous reward function applied to hospitals above the threshold values of quality indicators (Medicare’s pilot program, Phase I) [18,34,37,52,68]. Specifically, the value of the composite performance score in Medicare’s pay-for-performance hospitals was higher than in the control group of hospitals [52,78]. Moreover, sociological evidence points to the fact that hospitals participating in incentive schemes are likely to improve performance as they implement a larger number of quality improving activities that non-incentivized hospitals do not carry out [41].
The higher the value of
Accordingly, the expected mean effects of the reform may be formulated as follows:
Hypothesis
Hypothesis
Hypothesis
2.3.2 Groupwise effects of the reform
We assume that the effect of Medicare’s reform will be larger at hospitals with higher quality, based on findings in the health policy literature that emphasis on quality improvement in incentive schemes is greater at high-quality hospitals or among high-quality physicians in comparison with low-quality hospitals and physicians [21,37,69,77,78].
For instance, [77] conducted structural surveys at hospitals in the top two and bottom two deciles of performance measure in Medicare’s pilot program and discovered stronger involvement in quality improving activities among top performing hospitals. The statistically significant differences between top- and bottom-performing hospitals were observed in case of the numerical values, assigned to the following components of quality improvement: organizational culture, multidisciplinary teams, “adequate human resources for projects to increase adherence to quality indicators” and “new activities or policies related to quality improvement” (Tables 3 and P on pp. 836–837).
Interviews with the leaders of California physician organizations [21] similarly discovered that physicians with high performance placed higher emphasis on the support that “the organization dedicates to addressing quality issues” than medium- and low-performing physicians (Exhibit 3, p. 521).
Moreover, papers that use policy evaluation techniques applied to assessment of the effect of the pilot pay-for-performance program at Medicare hospitals report that hospitals in the top two deciles of quality measures showed the fastest improvement, while hospitals in the lowest deciles raised their quality to a much lesser extent or may even have failed to improve [69,78].
To sum up, the hypothesis on groupwise effects of pay-for-performance is as follows:
Hypothesis
2.3.3 Net total effect over time at groups of hospitals
Consider the multivariate dependence of the variable of interest on two sources of intertemporal dependence: the policy reform and mean reversion. The effect of mean reversion implies a differential time profile of measured quality: measured quality increases at hospitals in low percentiles of the quality distribution and decreases at hospitals in high percentiles. Combined with the positive effect of pay-for-performance on the mean value of measured quality (Hypothesis
Hypothesis
Hypothesis
If
Hypothesis
3 Empirical approach
3.1 Specification
The dependent variable
where
We use two hospital control variables which affect quality improvement and allow us to mitigate potential biases, which might occur if the pay-for-performance effect is identified based only on the variation of
Eq. (1) can be estimated using the generalized method of moments: the [2] and [12] methodology for dynamic panel data. Examples of use of the methodology in health economics include analysis of the quality of care at Medicare’s hospitals in [56], study of the length of stay at Japanese hospitals in [10], investigation of labor supply by Norwegian physicians in [4], and of health status of individuals in the US in [57].
The first set of moment conditions in GMM comes from the approach of [2] and [12]. We take the first difference of the right-hand side and left-hand side of Eq. (1):
Since
where
Another set of moment conditions comes from [12] for the level Eq. (1):
So
More specifically, lagged value of TPS and other hospital control variables in
It should be noted that the use of dynamic panel data methodology requires justification on economic grounds. This is because the approach uses lags and lagged differences as instruments, and there are potential problems with using lags as instruments even though they pass the Arrelano-Bond tests. Specifically, lags may prove to be weak and invalid [7]: the weakness may occur when lags are distant [59], and invalidity happens due to overfitting of the endogenous variable under large
The validity of instruments is assessed through statistics of the Arellano-Bond test. We employ [80] robust standard errors for estimation.[11] But formal tests are insufficient for establishing the causal relationship in models, which use an instrumental variable approach [1,7]. Accordingly, it is necessary to provide an economic justification for the assumption of the exclusion restriction of the instruments, i.e., that the instruments are exogenous and impact the dependent variable through no channels other than the endogenous variable and, possibly, also through exogenous covariates. An example of such justification on theoretical grounds can be found in [6], who uses lags of GDP and lags of the inflation rate as instruments for GDP and inflation. Another way of arguing for the exclusion restriction is given in [38], which estimates per capita output in various countries as a function of social infrastructure. Owing to endogeneity of social infrastructure, variables related to exposure to Western culture are used as instruments, and there is a discussion of the absence of any direct channels through which these variables could impact a country’s per capita output.
We follow the latter approach to provide an economic justification for the validity of instruments in the dynamic panel data model for the composite quality measure at Medicare hospitals. Our arguments below, which advocate the applicability of lagged first differences as instruments for the level Eq. (1) and first lagged levels as instruments for the difference Eq. (2), are based on the plausible assumption of a short adjustment period in the values of the dependent variable. Specifically, we assume that hospital managers take prompt action upon learning the TPS in year
As regards our formal analysis, Eq. (1) has TPS as a dependent variable and its first and second lags as explanatory variables.
Another equation is (2) and it models first differences, i.e., changes in quality. The dependent variable is
Note that [56] used similar arguments in discussing applicability of the dynamic panel data model to analyze in-hospital mortality and the complication rate, which are used as measures of hospital quality in US Medicare hospitals. They write: “We believe our approach is appropriate because (i) changes to in-hospital mortality and complications should be immediately affected by changes in staffing levels, not after a long adjustment period, and (ii) the influence of the past is incorporated through the lagged value of the dependent variable.” (p. 296, Footnote 3).
A related study applying dynamic panel data models to hospital performance indicators deals with average length of stay at Japanese acute-care hospitals that plan to introduce a prospective payment system [10]. The variable is treated in Japan as a proxy for hospital efficiency. It is regularly monitored and analyzed by the regulator and by hospital management, with feedback actions by hospital personnel in response to annual updates on levels of the variable [9,10,43,45,75]. Accordingly, the assumption of a short adjustment period for the length of stay is likely to hold at Japanese hospitals and the use of lagged levels and lagged differences as instruments is justified.
Note that potential violations of the exclusion restriction may occur in instances where the quality measure requires long periods to adjust. In such instances, causal impact of the Medicare reform on the quality of care cannot be established [1,7].
We note other limitations of our approach. First, the analysis deals with the composite quality measure. While quality-related efforts of a hospital and the TPS composite quality measure are multi-dimensional, we do not touch upon multi-tasking in the empirical estimations. Our approach considers a one-dimensional effort, a one-dimensional true quality, and its measurable proxy.[13]
Second, we do not touch on the rules for computing the scores of each dimension of the composite measure or on aggregation of dimension scores. It is important to note that Medicare uses whichever is highest, improvement points or achievement points, as the score for each dimension. The choice between achievement and improvement points stimulates low-performing hospitals, and the uniform formula assumes that all groups of hospitals have equal margin for improvement. A minor exception is protection of hospitals above the benchmark value of the 95th percentile of a corresponding measure score: these hospitals receive 10 points for their achievement on a
Third, weighting of scores across domains is another feature of the design of the incentive mechanism which is not analyzed in our article. So the dichotomous variables for annual periods in the empirical specification capture time effects unrelated to Medicare’s value-based purchasing as well as time effects not associated with the size of incentives but potentially linked to changes in other elements of the reform design (i.e., changes in weights).
Finally, conventional policy evaluation using a control group of hospitals is not possible because quality measures for non-Medicare hospitals are not available.[15] The empirical part of the article therefore focuses solely on pay-for-performance hospitals and identifies the effect of quality incentives based on variation in
3.2 Multivariate dependence of the quality variable
3.2.1 Calculation of the mean in the autoregressive process
We interpret the second-order dynamic panel (1) as a second-order autoregressive process. The coefficients for the first and the second lags of
To test the hypotheses which concern the mean value of the measured quality
For a fixed value of
where
Since
The estimate of
Note that the expression for
3.2.2 Intertemporal dependence due to the policy reform
The policy parameter
Accordingly, we examine the difference between
The null hypothesis is as follows:
and it is tested against the positive alternative.
Equivalently, we compute the difference between
Note that
The null hypothesis is as follows:
and it is tested against the positive alternative.
In conjunction with hypothesis
Now consider hypothesis
([39], Section 2.3). Again, for a fixed value of
Then we replace the expected values by sample means and solve this quadratic equation to obtain the following formula for
where
An alternative approach considers the value of the autocorrelation function (ACF(1)) (the correlation coefficient between
([39], Section 3.4).
Testing
is tested against the positive alternative.
Alternatively, we assess whether
To assess
Testing
3.2.3 Estimation of the multivariate effect due to policy reform and mean reversion
Evaluation of
4 Data
4.1 Data sources and variables
The analysis uses data for Medicare hospitals in 2011–2019 from several sources. We use Hospital Compare data archives (January 2021 update) for quality measures, hospital ownership, and geographic location. The medical school affiliation of a hospital, the number of hospital beds, nurses, and physicians come from Provider of Service files. Other hospital control variables are taken from the Final Rules, which are Medicare’s annual documents on reimbursement rates in the inpatient prospective payment system. Specifically, we use information from the Impact Files, which accompany the Final Rules and estimate the impact of the reimbursement mechanism on hospital characteristics. The variables taken from the Impact Files are the share of Medicare’s discharges, ownership, and urban location.
Patient characteristics are also taken from the Impact Files. The casemix variable reflects the relative weight of each DRG in financial terms and is adjusted for transfers of patients between hospitals.[16] Casemix makes it possible to control for the composition of patient cases taking account of the objective link between severity of illness and hospital resources. The disproportionate-share index accounts for the share of low-income patients and makes it possible to proxy a patient’s income.
To account for other major channels of quality improvement by Medicare hospitals over the observed time period, we use the data for two programs run by the Centers for Medicare and Medicaid Services. One of them is the HRRP, which applies to Medicare hospitals since fiscal year 2013 and penalizes them for excess readmissions. Specifically, the payment reduction which may equal from 0 to 3% is applied to hospital’s Medicare remuneration, higher values of the percentage for the penalty represent more excess readmissions at the hospital. Using the HRRP Supplemental data files, which accompany annual Final Rules on acute inpatient PPS (June 2020 update), we find the HRRP penalty for 2013–2019 and use it as one of the control variables in the empirical analysis.
We also consider the EHR Incentive Program, which was in force since 2011. The program establishes hospital attestation on the use of EHR. The adoption of quality-improving information technology requires substantial fixed cost, so the binary variable for hospital attestation within EHR makes it possible to control for the fixed cost in the empirical analysis. The EHR promotion program consists of three stages (sequentially introduced in 2011, 2014, and 2017). Using data from The Eligible Hospitals Public Use Files on the EHR incentive program (February 2020 update), we set the EHR attestation dummy equal to one if the hospital passed its attestation for the given year at any stage. Owing to non-availability of data on the third stage of the program, we extend the second stage data from year 2016 to years 2017–2019. Use of an attestation dummy lets us control for the fact of incurring the fixed cost of quality-improvement efforts. Owing to the small size of the non-EHR group (only 8–10% of the sample), we do not analyze whether quality goes up faster in the group of the hospitals (for instance, we do not interact the attestation dummy with
4.2 Sample
The non-anonymous character of the data sources allows us to merge them by year and hospital name. Our analysis focuses on acute-care Medicare’s hospitals, as the pay-for-performance incentive contract applies exclusively to this group. We restrict the sample by considering only hospitals with share of Medicare cases greater than 5%.
The specification with second-order lag enables estimation of the fitted values of
Variable | Definition | Obs | Mean | St.Dev | Min | Max |
---|---|---|---|---|---|---|
Hospital performance | ||||||
TPS | Hospital TPS | 18,545 | 37.265 | 11.468 | 2.727 | 98.182 |
Patient characteristics | ||||||
Casemix | Transfer-adjusted casemix index | 18,545 | 1.599 | 0.298 | 0.834 | 3.972 |
Dsh | Disproportionate share index, reflecting the prevalence of low-income patients | 18,545 | 0.307 | 0.165 | 0 | 1.232 |
Hospital characteristics | ||||||
Nurses/beds | Nurse-to-bed ratio | 18,545 | 1.312 | 3.849 | 0 | 170.479 |
Physicians/beds | Physician-to-bed ratio | 18,545 | 0.099 | 0.947 | 0 | 70.992 |
Beds | Number of beds | 18,545 | 272.158 | 241.538 | 3 | 2,891 |
log(beds) | Number of beds (in logs) | 18,545 | 5.283 | 0.819 | 1.099 | 7.969 |
Medicare share | Share of Medicare cases | 18,545 | 0.378 | 0.118 | 0.050 | 0.983 |
HRRP penalty | Percentage reduction of the Medicare payments under HRRP | 18,545 | 0.498 | 0.590 | 0 | 3.000 |
MUEHR | =1 if passed attestation for meaningful usage of EHR | 18,545 | 0.924 | 0.265 | 0 | 1 |
Urban | =1 if an urban hospital | 18,545 | 0.711 | 0.453 | 0 | 1 |
Public | =1 if managed by federal, state or local government, or hospital district or authority | 18,545 | 0.147 | 0.354 | 0 | 1 |
Teaching | =1 if hospital has medical school affiliation | 18,545 | 0.364 | 0.481 | 0 | 1 |
Hospital location | ||||||
New England | =1 if located in Connecticut, Maine, Massachusetts, New Hampshire, Rhode Island, or Vermont | 18,545 | 0.046 | 0.210 | 0 | 1 |
Mid-Atlantic | =1 if located in New Jersey, New York, or Pennsylvania | 18,545 | 0.123 | 0.328 | 0 | 1 |
East North Central | =1 if located in Illinois, Indiana, Michigan, Ohio, or Wisconsin | 18,545 | 0.168 | 0.374 | 0 | 1 |
West North Central | =1 if located in Iowa, Kansas, Minnesota, Missouri, Nebraska, North Dakota, or South Dakota | 18,545 | 0.081 | 0.272 | 0 | 1 |
South Atlantic | =1 if located in Delaware, Florida, Georgia, Maryland, North Carolina, South Carolina, Virginia, District of Columbia, or West Virginia | 18,545 | 0.177 | 0.381 | 0 | 1 |
East South Central | =1 if located in Alabama, Kentucky, Mississippi, or Tennessee | 18,545 | 0.087 | 0.282 | 0 | 1 |
West South Central | =1 if located in Arkansas, Louisiana, Oklahoma, or Texas | 18,545 | 0.129 | 0.335 | 0 | 1 |
Mountain | =1 if located in Arizona, Colorado, Idaho, Montana, Nevada, New Mexico, Utah, or Wyoming | 18,545 | 0.069 | 0.253 | 0 | 1 |
Pacific | =1 if located in California, Oregon, Washington, Alaska, or Hawaii | 18,545 | 0.115 | 0.319 | 0 | 1 |
Note: Section 401 hospitals are treated as rural hospitals.
4.3 Flow of quality and evidence of mean reversion
Descriptive analysis of the values of TPS offers suggestive evidence in support of some of the main hypotheses generated by the model. Specifically, we focus on the flow of hospitals between quintiles of TPS in different years. The Sankey diagrams in Figures 1 and 2 use the width of arrows as the intensity of flow rates and demonstrates how hospitals change their position in quintiles of the composite quality measure after the introduction of pay-for-performance (e.g., from 2012 to 2013).
As can be inferred from Figure 1, there is considerable movement of hospitals between quintiles. For instance, consider hospitals which in 2012 belonged to the fifth quintile of TPS (quintile with the highest performance). Fewer than half of these hospitals remained in the fifth quintile of TPS in 2013, and the rest saw a decline of their position relative to other hospitals by moving to quintiles one through four. Similar tendencies are observed for hospitals in any other given quintile of TPS in 2012: only a small share of hospitals continue to belong to the same quintile in the subsequent year. This can be viewed as graphic support for the phenomenon of mean reversion since hospitals would rarely change their quintile from year to year in the absence of mean reversion.
It is plausible to assume that mean reversion becomes weaker when there is an increase of
5 Empirical results
The first set of our results is reported in Table 2 and concerns the mean effect of pay-for-performance at Medicare hospitals.
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
|
1.00 | 1.25 | 1.50 | 1.75 | 2.00 | 2.00 | 2.00 |
|
|
|
|
|
|
|
|
(0.973) | (0.678) | (0.385) | (0.453) | (0.970) | (0.961) | (0.956) | |
|
|
|
|
|
|
|
|
(0.737) | (1.009) | (1.349) | (1.800) | (2.355) | (2.309) | (2.266) | |
|
|
|
|
|
|
0.277 | 0.157 |
(0.737) | (0.375) | (0.474) | (0.542) | (0.634) | (0.240) | (0.237) |
Notes: Standard errors calculated using the delta-method are in parentheses.
Measured as
Note that the mean value of
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
|
1.00 | 1.25 | 1.50 | 1.75 | 2.00 | 2.00 | 2.00 |
|
|
|
|
|
|
|
|
(0.112) | (0.032) | (0.020) | (0.017) | (0.016) | (0.016) | (0.016) | |
|
|
|
0.076 |
|
|
|
|
(0.151) | (0.069) | (0.055) | (0.048) | (0.044) | (0.044) | (0.045) | |
|
|
|
|
|
|
|
|
(0.151) | (0.082) | (0.016) | (0.009) | (0.007) | (0.005) | (0.005) | |
|
|
|
|
|
|
|
|
(0.018) | (0.015) | (0.013) | (0.013) | (0.015) | (0.014) | (0.014) | |
|
|
|
|
|
|
|
|
(0.019) | (0.024) | (0.029) | (0.034) | (0.039) | (0.038) | (0.038) | |
|
|
|
|
|
|
|
|
(0.019) | (0.005) | (0.006) | (0.006) | (0.006) | (0.003) | (0.003) |
Notes: Standard errors calculated using the delta-method are in parentheses.
The persistence parameter
Since the values of
The heterogeneous changes in hospital quality owing to pay-for-performance are given in Tables 4, 5, 6 where hospitals are divided into quintiles according to the values of their TPS. Note that the change in hospital quality is a function of the regression coefficient and the mean values of covariates. So its standard error consists of two parts: the error of the estimated regression coefficient and the error of the mean values of covariates. Only the second part of this error depends on sample size and should go up approximately
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
Quintile 1 |
|
|
|
|
|
|
|
(0.741) | (1.007) | (1.334) | (1.727) | (2.264) | (2.162) | (2.202) | |
Quintile 2 |
|
|
|
|
|
|
|
(0.745) | (1.013) | (1.358) | (1.803) | (2.375) | (2.293) | (2.250) | |
Quintile 2 minus |
|
|
0.247 | 0.155 |
|
|
|
Quintile 1 | (0.138) | (0.124) | (0.174) | (0.233) | (0.328) | (0.327) | (0.313) |
Quintile 3 |
|
|
|
|
|
|
|
(0.738) | (1.023) | (1.378) | (1.849) | (2.346) | (2.381) | (2.278) | |
Quintile 3 minus | 0.037 |
|
|
|
|
0.539 |
|
Quintile 2 | (0.078) | (0.114) | (0.155) | (0.232) | (0.321) | (0.356) | (0.355) |
Quintile 4 |
|
|
|
|
|
|
|
(0.739) | (1.027) | (1.392) | (1.850) | (2.391) | (2.338) | (2.279) | |
Quintile 4 minus | 0.110 | 0.146 |
|
0.165 | 0.194 |
|
|
Quintile 3 | (0.074) | (0.111) | (0.230) | (0.235) | (0.368) | (0.419) | (0.349) |
Quintile 5 |
|
|
|
|
|
|
|
(0.741) | (1.030) | (1.418) | (1.849) | (2.506) | (2.476) | (2.421) | |
Quintile 5 minus |
|
0.043 | 0.393 |
|
0.094 | 0.022 | 0.414 |
Quintile 4 | (0.101) | (0.251) | (0.503) | (0.433) | (0.568) | (0.503) | (0.497) |
Notes: quintile 1 denotes the lowest quality and quintile 5 – the highest. The table reports the effect at each corresponding quintile and the differences in the effects at consecutive quintiles.
Standard errors (calculated using the delta-method for the difference of the reform effects across the corresponding two categories of each time-invariant hospital characteristic) are in parentheses.
There are two sources of errors in the estimates shown in the table: the error of the regression coefficient and the error of the mean values of covariates. The first part of the error does not vary across all result tables, while the second part of the error depends on the group size and is approximately
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
Quintile 1 |
|
|
|
|
|
0.536 |
|
(0.741) | (0.517) | (0.497) | (0.582) | (0.656) | (0.440) | (0.434) | |
Quintile 2 |
|
0.678 |
|
|
|
|
0.655 |
(0.745) | (0.489) | (0.546) | (0.626) | (0.736) | (0.465) | (0.505) | |
Quintile 2 minus |
|
|
|
|
0.667 |
|
0.869 |
Quintile 1 | (0.138) | (0.522) | (0.501) | (0.545) | (0.585) | (0.629) | (0.676) |
Quintile 3 |
|
|
|
|
|
0.743 | 0.104 |
(0.738) | (0.458) | (0.530) | (0.669) | (0.723) | (0.501) | (0.531) | |
Quintile 3 minus | 0.037 | 0.255 |
|
|
|
0.868 |
|
Quintile 2 | (0.078) | (0.490) | (0.514) | (0.545) | (0.609) | (0.677) | (0.758) |
Quintile 4 |
|
|
|
0.929 |
|
0.505 |
|
(0.739) | (0.464) | (0.626) | (0.702) | (0.789) | (0.552) | (0.510) | |
Quintile 4 minus | 0.110 | 0.722 |
|
|
0.458 |
|
|
Quintile 3 | (0.074) | (0.500) | (0.565) | (0.630) | (0.654) | (0.745) | (0.748) |
Quintile 5 |
|
|
|
|
|
|
0.666 |
(0.741) | (0.620) | (0.864) | (0.875) | (0.892) | (0.583) | (0.567) | |
Quintile 5 minus |
|
|
|
|
|
|
0.711 |
Quintile 4 | (0.101) | (0.599) | (0.829) | (0.876) | (0.780) | (0.803) | (0.762) |
Notes: quintile 1 denotes the lowest quality and quintile 5 – the highest. The table reports the effect at each corresponding quintile and the differences in the effects at consecutive quintiles.
Standard errors calculated using the delta-method are in parentheses.
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
Quintile 1 |
|
|
|
|
|
|
0.226 |
(0.458) | (0.329) | (0.298) | (0.271) | (0.263) | (0.273) | (0.266) | |
Quintile 2 |
|
|
|
|
|
|
|
(0.288) | (0.249) | (0.236) | (0.223) | (0.212) | (0.224) | (0.238) | |
Quintile 2 minus |
|
|
|
|
|
|
|
Quintile 1 | (0.367) | (0.285) | (0.254) | (0.215) | (0.207) | (0.221) | (0.228) |
Quintile 3 |
|
|
|
|
|
|
|
(0.282) | (0.236) | (0.237) | (0.218) | (0.216) | (0.228) | (0.235) | |
Quintile 3 minus |
|
|
|
|
|
|
|
Quintile 2 | (0.276) | (0.243) | (0.212) | (0.201) | (0.193) | (0.214) | (0.231) |
Quintile 4 |
|
|
|
|
|
|
|
(0.321) | (0.273) | (0.280) | (0.250) | (0.256) | (0.272) | (0.270) | |
Quintile 4 minus |
|
|
|
|
|
|
|
Quintile 3 | (0.289) | (0.243) | (0.247) | (0.230) | (0.228) | (0.267) | (0.229) |
Quintile 5 |
|
|
|
|
|
|
|
(0.439) | (0.402) | (0.476) | (0.388) | (0.381) | (0.381) | (0.396) | |
Quintile 5 minus |
|
|
|
|
|
|
|
Quintile 4 | (0.353) | (0.330) | (0.406) | (0.336) | (0.330) | (0.323) | (0.330) |
Notes: quintile 1 denotes the lowest quality and quintile 5 – the highest. The table reports the effect at each corresponding quintile and the differences in the effects at consecutive quintiles.
Standard errors calculated using the delta-method are in parentheses.
The estimates of the effect of pay-for-performance in terms of
Table 6 gives estimates of the net total effect, i.e., the expected change in hospital quality over time, measured as the difference between the predicted TPS and the lagged TPS. The net total effect is the sum of the impact of mean reversion and the effect of pay-for-performance.
Note that the estimation of the fitted value of TPS includes time effects which account both for time trend and for important changes in the incentive mechanism not captured by variation in
The values of net total effect reveal an increase of quality in the groups of low-quality hospitals, while quality deteriorates in high-quality groups. Negative total effect is less prevalent or is smaller in absolute terms at high-quality hospitals in 2016–2017. The result can be attributed to the weakening of mean reversion with increase in
Finally, we focus on the effect of pay-for-performance for groups of Medicare hospitals according to their ownership, teaching status, urban location, and geographic region. The mean effect increases in
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
Public |
|
|
|
|
|
|
|
(0.723) | (0.984) | (1.304) | (1.739) | (2.254) | (2.203) | (2.148) | |
Private |
|
|
|
|
|
|
|
(0.740) | (1.015) | (1.360) | (1.814) | (2.377) | (2.333) | (2.289) | |
Private minus | 0.174 | 0.269 |
|
|
|
|
|
Public | (0.118) | (0.168) | (0.256) | (0.323) | (0.432) | (0.437) | (0.414) |
Urban |
|
|
|
|
|
|
|
(0.739) | (1.003) | (1.314) | (1.721) | (2.203) | (2.146) | (2.104) | |
Rural |
|
|
|
|
|
|
|
(0.741) | (1.070) | (1.494) | (2.054) | (2.802) | (2.706) | (2.578) | |
Rural minus |
|
|
|
0.396 | 1.136 |
|
|
Urban | (0.227) | (0.333) | (0.493) | (0.599) | (0.832) | (0.749) | (0.615) |
Teaching |
|
|
|
|
|
|
|
(0.755) | (1.013) | (1.334) | (1.735) | (2.224) | (2.198) | (2.156) | |
Non-teaching |
|
|
|
|
|
|
|
(0.732) | (1.018) | (1.373) | (1.854) | (2.467) | (2.407) | (2.362) | |
Non-teaching minus |
|
|
0.140 | 0.219 | 0.487 | 0.404 | 0.301 |
Teaching | (0.179) | (0.236) | (0.366) | (0.430) | (0.609) | (0.578) | (0.560) |
Notes: Standard errors (calculated using the delta-method for the difference of the reform effects across the corresponding two categories of each time-invariant hospital characteristic) are in parentheses.
2013 | 2014 | 2015 | 2016 | 2017 | 2018 | 2019 | |
---|---|---|---|---|---|---|---|
New England |
|
2.080 |
|
|
|
|
|
(0.815) | (1.264) | (1.752) | (2.105) | (3.078) | (2.976) | (2.888) | |
Mid-Atlantic |
|
|
|
|
|
|
|
(0.737) | (0.999) | (1.321) | (1.739) | (2.253) | (2.219) | (2.146) | |
Mid-Atlantic minus |
|
0.470 | 0.012 |
|
|
|
|
New England | (0.328) | (0.702) | (0.978) | (0.574) | (1.082) | (0.997) | (0.981) |
East North Central |
|
|
|
|
|
|
|
(0.751) | (1.036) | (1.404) | (1.865) | (2.393) | (2.387) | (2.347) | |
East North Central minus | 0.215 | 1.004 | 1.168 |
|
|
|
|
New England | (0.316) | (0.694) | (0.955) | (0.451) | (0.905) | (0.801) | (0.775) |
West North Central |
|
|
|
|
|
|
|
(0.741) | (1.052) | (1.431) | (2.025) | (2.842) | (2.797) | (2.765) | |
West North Central minus | 0.455 |
|
1.549 | 0.483 |
|
0.069 | 0.238 |
New England | (0.332) | (0.709) | (1.001) | (0.520) | (0.808) | (0.780) | (0.757) |
South Atlantic |
|
|
|
|
|
|
|
(0.761) | (1.040) | (1.422) | (1.887) | (2.481) | (2.407) | (2.374) | |
South Atlantic minus | 0.384 | 1.131 | 1.341 |
|
|
|
|
New England | (0.326) | (0.699) | (0.949) | (0.434) | (0.826) | (0.788) | (0.760) |
East South Central |
|
|
|
|
|
|
|
(0.780) | (1.065) | (1.492) | (2.042) | (2.803) | (2.597) | (2.460) | |
East South Central minus | 0.432 | 1.038 | 1.357 | 0.054 |
|
|
|
New England | (0.356) | (0.720) | (0.989) | (0.517) | (0.719) | (0.705) | (0.726) |
West South Central |
|
|
|
|
|
|
|
(0.735) | (0.990) | (1.325) | (1.755) | (2.241) | (2.227) | (2.175) | |
West South Central minus | 0.413 | 1.146 | 1.588 |
|
|
|
|
New England | (0.347) | (0.723) | (1.022) | (0.583) | (1.061) | (0.969) | (0.933) |
Mountain |
|
|
|
|
|
|
|
(0.647) | (0.863) | (1.119) | (1.415) | (1.809) | (1.819) | (1.861) | |
Mountain minus |
|
0.729 | 0.698 |
|
|
|
|
New England | (0.371) | (0.756) | (1.073) | (0.869) | (1.468) | (1.345) | (1.242) |
Pacific |
|
|
|
|
|
|
|
(0.716) | (0.957) | (1.238) | (1.613) | (2.101) | (2.067) | (2.066) | |
Pacific minus |
|
1.245 | 0.671 |
|
|
|
|
New England | (0.388) | (0.765) | (1.072) | (0.827) | (1.315) | (1.245) | (1.189) |
Notes: Standard errors (calculated using the delta-method for the difference of the reform effects across New England hospitals and hospitals in each corresponding geographic region) are in parentheses.
The effect of pay-for-performance is greater for private hospitals than for public hospitals, which corresponds to findings in [13] and [78]. The result can be explained by a greater emphasis on financial incentives at these healthcare institutions. These profit constraints, combined with the altruistic character of healthcare services, induce more effective quality competition at non-public hospitals [16]. The difference in the effect for private and public hospitals is statistically significant in most years.
As for teaching status, quality improvement owing to the incentive scheme is often higher at non-teaching hospitals, which may be because they can devote all of their labor resources to patient treatment, while teaching hospitals lose some efficiency due to their educational activities [64]. Also, teaching hospitals may be treating more difficult cases. This complexity could not be fully captured by the casemix variable in our analysis and may cause a downward bias of the estimated effect at teaching hospitals, explaining the lower value of the effect at teaching than at non-teaching hospitals. Yet, the difference in the values at teaching and non-teaching hospitals is statistically insignificant in each year.
Statistically significant differences in the effect of pay-for-performance for urban and rural hospitals are observed only in the last 2 years: the effect is larger at rural hospitals.
As regards geographic location, there is practically no variation in the effect across groups of hospitals in the early years of pay-for-performance. The differences are present mainly in the later few years: for instance, the mean effect of pay-for-performance is greater in New England than in Mid-Atlantic in 2016–2019 and than in East North Central and West South Central regions in 2017–2019.
6 Discussion
In this article, we focused on exclusion of mean reversion in evaluating the response of TPS at Medicare hospitals to an incentive contract. Since TPS under this contract becomes an autoregressive process, our analysis deals with dynamic panels.
It should be noted that dynamic panel data models are prevalent in various fields of economics. Examples in macroeconomics include the analysis of a country’s growth [11,50] or its current account [81]. Application in corporate finance deals with the study of such firm-level variables as size [33,61], profit [54], leverage [32,36], and such proxies of firm performance as return on asset and Tobin’s Q [49,65]. In the banking sphere dynamic panels are applied to ROE and profitability [35,48] while in finance they are used for housing prices [31] and fuel prices [71]. Papers in the economics of labor, health, and welfare employ dynamic panel data models to analyze physician labor supply [4], hospital staffing intensity [82], wealth of households and health status of individuals [57], and quality and efficiency of hospitals (e.g., mortality ratio in [56], and average length of stay in [10]).
The approach used in our study estimates the unconditional mean of the dependent variable in the dynamic panel data model and employs it for policy evaluation. Specifically, the comparison of the fitted values of the unconditional mean at different values of policy intensity offers a measure of the effect of reform. The advantages of the approach are twofold. First, it excludes the impact of mean reversion in groupwise estimations (e.g., in lower and in higher quantiles of hospitals according to their TPS). Second, the approach may also be used in the analysis of the mean effect of the reform if we focus on effects in the long run. Indeed, the unconditional mean in dynamic panel data analysis is sometimes called the long-term mean as it reflects the mean value in the long run. It should be noted that an alternative approach that uses the estimated coefficient for the policy variable as a measure of the mean effect of reform does not suffer from the problem of mean reversion. But in dynamic panel data models it evaluates only the short-term impact of policy.
As regards exclusion of mean reversion in dynamic panels, we note a limitation on the character of mean reversion, imposed by the nature of the dynamic panel model where the unconditional mean is the long-term mean. Mean reversion is not instantaneous: if a deviation from the mean is observed in period
It may be noted that our approach is similar to difference-in-difference estimations. The long-run effect of reform under our approach is the difference in the fitted value of the long-term mean under the value
As regards policy evaluation based on panel data fixed effects methodology, our approach of computing the unconditional mean as a function of the policy variable
Finally, we note the prerequisites for identification of the unconditional mean which are similar to the assumptions in difference-in-difference estimations. Two requirements apply both to the static and dynamic panels. First, time variation in the policy variable is required for identification of the coefficient for the policy variable in the unconditional mean function. Second, if there is only time variation in the policy variable
7 Conclusion
Studies of incentive contracts usually focus on the mean tendency and give scant attention to potentially heterogeneous response to the policy of interest by agents at different percentiles of the distribution of the dependent variable. But insufficient analysis of such heterogeneity may lead to speculation on ceiling effects and belief among agents with better values of the variable of interest that there are no ways of making further financial gains by further improvements.
This article highlights the fact that there is a multivariate dependence of the variable of interest in such incentive contracts. Specifically, a part of intertemporal dependence can be attributed to the policy reform and a part to mean reversion. So the article proposes a method to help model such multivariate dependence by excluding the impact of mean reversion. As mean reversion contaminates judgment regarding the time profile of the dependent variable, and this contamination is different for agents in lower and higher percentiles of the variable of interest, clearing out the reform effect of mean reversion makes the method suitable for assessing heterogeneity of incentive schemes.
In an application to the longitudinal data for Medicare’s acute-care hospitals taking part in the nationwide quality incentive mechanism (“value-based purchasing”), we find that the higher the quintile of quality in the prior period, the larger the increase in the composite quality measure owing to the reform. Quality improvement in each quintile increases with the increase in size of the quality incentive.
Our results reveal that increase in the quality measure owing to pay-for-performance is greater at hospitals with higher levels of quality. The finding suggests stronger emphasis on quality activities at high-quality hospitals, and this is indeed discovered in a number of works. For instance, top-performing hospitals in the US pilot program paid more attention to quality enhancement than bottom-performing hospitals [77]. Under the proportional pay-for-performance mechanism in California, high-quality physicians similarly placed more emphasis on an organizational culture of quality and demonstrated stronger dedication to addressing quality issues than low-quality physicians [21]. The desire of high-quality hospitals, which have reached top deciles of hospital performance, to pursue quality improvement by means additional to those proposed by the policy regulator is further evidence in support of our research [37].
Directions for future work in health economics applications may include analysis of heterogeneous hospital response to quality incentives by considering different dimensions of the composite quality measure. A related field of research is the study of potential sacrifice of quality of non-incentivized measures in favor of measures incentivized by pay-for-performance. This has been analyzed at the mean level [27,47] and may be expanded to account for different behavior by high-quality and low-quality hospitals.
Acknowledgements
This article was prepared in the framework of the Basic Research Program at the National Research University Higher School of Economics (HSE). The authors are grateful to the editor and two anonymous referees for helpful comments.
-
Conflict of interest: The authors state no conflict of interest.
Appendix A Estimation with the dynamic panel
TPS | ||
---|---|---|
|
|
(0.045) |
|
|
(0.040) |
|
|
(0.070) |
|
|
(0.065) |
Medicare share |
|
(5.530) |
|
|
(4.904) |
Casemix |
|
(1.522) |
Physicians/beds |
|
(0.187) |
|
|
(0.220) |
Nurses/beds | 0.007 | (0.030) |
|
|
(0.052) |
Dsh | 4.210 | (3.695) |
|
|
(0.653) |
|
|
(0.854) |
HRRP penalty |
|
(0.227) |
MUEHR |
|
(1.061) |
|
|
(0.749) |
|
|
(0.416) |
|
|
(0.182) |
|
|
(0.201) |
|
0.274 | (0.372) |
|
|
(0.379) |
|
|
(0.390) |
Constant |
|
(5.317) |
Observations | 18,545 | |
Hospitals | 2,984 | |
Arellano–Bond test statistic |
|
|
|
0.494 | |
|
0.781 |
Notes: The sum of coefficients for annual dummies is normalized to zero.
Robust standard errors are in parentheses.
The Sargan statistic is not applicable to the specification with robust standard errors.
B Data sources
Total performance scores and other Hospital Compare data were downloaded from https://data.medicare.gov/data/hospital-compare (Table A2).
Variable | Source |
---|---|
TPS | Hospital compare |
Casemix | Impact files |
Dsh | Impact files |
Medicare share | Impact files |
Urban | Impact files |
Public | Hospital compare |
Physicians | Provider of service files |
Nurses | Provider of service files |
Teaching | Provider of service files |
Beds | Provider of service files |
Regional dummies | Hospital compare |
HRRP penalty | HRRP supplemental files to acute inpatient PPS final rules |
MUEHR | EHR incentive program eligible hospitals public use files |
TPSs and other Hospital Compare data were downloaded from https://data.medicare.gov/data/hospital-compare (Table A3).
Provider of Service data come from https://www.cms.gov/Research-Statistics-Data-and-Systems/Downloadable-Public-Use-Files/Provider-of-Services (Table A4).
Impact Files data are taken from https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/AcuteInpatientPPS (Table A5).
The Hospitals Readmissions Reduction Program data were downloaded from https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/AcuteInpatientPPS/Archived-Supplemental-Data-Files (Table A6).
The EHR (Electronic Hospital Records) Incentive Program (renamed to the Promoting Interoperability (PI) Program) data are taken from https://www.cms.gov/Regulations-and-Guidance/Legislation/EHRIncentivePrograms/PUF (Table A7).
References
[1] Angrist, J. D., & Pischke, J.-S. (2009). Mostly harmless econometrics. Princeton: Princeton University Press. 10.1515/9781400829828Search in Google Scholar
[2] Arellano, M., & Bond, S. (1991). Some tests of specification for panel data: Monte Carlo evidence and an application to employment equations. Review of Economic Studies, 58, 277–297. 10.2307/2297968Search in Google Scholar
[3] Armstrong, C. S., Blouin, J. L., Jagolinzer, A. D., & Larcker, D. F. (2015). Corporate governance, incentives, and tax avoidance. Journal of Accounting and Economics, 60(1), 1–17. 10.1016/j.jacceco.2015.02.003Search in Google Scholar
[4] Baltagi, B. H., Bratberg, E., & Holmås, T. H. (2005). A panel data study of physicians’ labor supply: The case of Norway. Health Economics, 14(10), 1035–1045. 10.1002/hec.991Search in Google Scholar PubMed
[5] Barnett, A. G., van der Pols, J. C., & Dobson, A. J. (2004). Regression to the mean: What it is and how to deal with it. International Journal of Epidemiology, 34(1), 215–220. 10.1093/ije/dyh299Search in Google Scholar PubMed
[6] Barro, R. J. (2013). Inflation and economic growth. Annals of Economics and Finance, 14(1), 85–109. 10.3386/w5326Search in Google Scholar
[7] Bazzi, S., & Clemens, M. A. (2013). Blunt instruments: Avoiding common pitfalls in identifying the causes of economic growth. American Economic Journal: Macroeconomics, 5(2), 152–86. 10.1257/mac.5.2.152Search in Google Scholar
[8] Beaulieu, N. D., & Horrigan, D. R. (2005). Putting smart money to work for quality improvement. Health Services Research, 40(5p1), 1318–1334. 10.1111/j.1475-6773.2005.00414.xSearch in Google Scholar PubMed PubMed Central
[9] Besstremyannaya, G. (2011). Managerial performance and cost efficiency of Japanese local public hospitals: A latent class stochastic frontier model. Health Economics, 20(S1), 19–34. 10.1002/hec.1769Search in Google Scholar PubMed
[10] Besstremyannaya, G. (2016). Differential effects of declining rates in a per diem payment system. Health Economics, 25(12), 1599–1618. 10.1002/hec.3128Search in Google Scholar PubMed
[11] Bleaney, M., Gemmell, N., & Kneller, R. (2001). Testing the endogenous growth model: Public expenditure, taxation, and growth over the long run. Canadian Journal of Economics, 34(1), 36–57. 10.1111/0008-4085.00061Search in Google Scholar
[12] Blundell, R., & Bond, S. (1998). Initial conditions and moment restrictions in dynamic panel data models. Journal of Econometrics, 87(1), 115–143. 10.1920/wp.ifs.1995.9517Search in Google Scholar
[13] Borah, B. J., Rock, M. G., Wood, D. L., Roellinger, D. L., Johnson, M. G., & Naessens, J. M. (2012). Association between value-based purchasing score and hospital characteristics. BMC Health Services Research, 12(1), 464. 10.1186/1472-6963-12-464Search in Google Scholar PubMed PubMed Central
[14] Bousquet, F., Bisiaux, R., & Chi, Y.-L. (2014). France: Payment for public health services. In C. Cashin, Y.-L. Chi, P. C. Smith, M. Borowitz, & S. Thomson (Eds.), Paying for performance in health care (pp. 141–156). Berkshire, England: Open University Press. Search in Google Scholar
[15] de Brantes, F. S., & d’Andrea, B. G. (2009). Physicians respond to pay-for-performance incentives: Larger incentives yield greater participation. American Journal of Managed Care, 15(5), 305–10. Search in Google Scholar
[16] Brekke, K. R., Siciliani, L., & Straume, O. R. (2012). Quality competition with profit constraints. Journal of Economic Behavior and Organization, 84(2), 642–659. 10.1016/j.jebo.2012.09.006Search in Google Scholar
[17] Centers for Medicare and Medicaid Services. (2007). Medicare hospital value-based purchasing plan development. 1st public listening session, January 17, 2007. Issues paper. U.S. Department of Health and Human Services.Search in Google Scholar
[18] Christianson, J. B., Leatherman, S., & Sutherland, K. (2008). Lessons from evaluations of purchaser pay-for-performance programs. Medical Care Research and Review, 65(6 Suppl), 5S–35S. 10.1177/1077558708324236Search in Google Scholar PubMed
[19] Coleman, K., Reiter, K. L., & Fulwiler, D. (2007). The impact of pay-for-performance on diabetes care in a large network of community health centers. Journal of Health Care for the Poor and Underserved, 18(5), 966–983. 10.1353/hpu.2007.0090Search in Google Scholar PubMed
[20] Dafny, L., & Dranove, D. (2008). Do report cards tell consumers anything they donat already know? The case of Medicare HMOs. The Rand Journal of Economics, 39(3), 790–821. 10.1111/j.1756-2171.2008.00039.xSearch in Google Scholar PubMed
[21] Damberg, C. L., Raube, K., Teleki, S. S., & delaCruz, E. (2009). Taking stock of pay-for-performance: A candid assessment from the front lines. Health Affairs, 28(2), 517–525. 10.1377/hlthaff.28.2.517Search in Google Scholar PubMed
[22] Damberg, C. L., Raube, K., Williams, T., & Shortell, S. M. (2005). Paying for performance: Implementing a statewide project in California. Quality Management in Healthcare, 14(2), 66–79. 10.1097/00019514-200504000-00002Search in Google Scholar PubMed
[23] Damberg, C. L., Sorbero, M. E., Lovejoy, S. L., Martsolf, G. R., Raaen, L., & Mandel, D. (2014). Measuring success in health care value-based purchasing programs: Findings from an environmental scan, literature review, and expert panel discussions. Research report. Santa Monica, CA: RAND Corporation. Search in Google Scholar
[24] Davis, C. E. (1976). The effect of regression to the mean in epidemiologic and clinical studies. American Journal of Epidemiology, 104(5), 493–498. 10.1093/oxfordjournals.aje.a112321Search in Google Scholar PubMed
[25] Dias, D. A., & Marques, C. R. (2010). Using mean reversion as a measure of persistence. Economic Modelling, 27(1), 262–273. 10.1016/j.econmod.2009.09.006Search in Google Scholar
[26] Doran, T., Fullwood, C., Kontopantelis, E., & Reeves, D. (2008). Effect of financial incentives on inequalities in the delivery of primary clinical care in England: Analysis of clinical activity indicators for the quality and outcomes framework. The Lancet, 372(9640), 728–736. 10.1016/S0140-6736(08)61123-XSearch in Google Scholar
[27] Eggleston, K. (2005). Multitasking and mixed systems for provider payment. Journal of Health Economics, 24(1), 211–223. 10.1016/j.jhealeco.2004.09.001Search in Google Scholar PubMed
[28] Eijkenaar, F., Emmert, M., Scheppach, M., & Schöffski, O. (2013). Effects of pay for performance in health care: A systematic review of systematic reviews. Health Policy, 110(2), 115–130. 10.1016/j.healthpol.2013.01.008Search in Google Scholar PubMed
[29] Friedman, M. (1992). Do old fallacies ever die? Journal of Economic Literature, 30(4), 2129–2132. Search in Google Scholar
[30] Galton, F., & Dickson, H. (1886). Family likeness in stature. Proceedings of the Royal Society of London, 40, 42–73. 10.1098/rspl.1886.0009Search in Google Scholar
[31] Gao, A., Lin, Z., & Na, C. F. (2009). Housing market dynamics: Evidence of mean reversion and downward rigidity. Journal of Housing Economics, 18(3), 256–266. 10.1016/j.jhe.2009.07.003Search in Google Scholar
[32] Gaud, P., Jani, E., Hoesli, M., & Bender, A. (2005). The capital structure of Swiss companies: An empirical analysis using dynamic panel data. European Financial Management, 11(1), 51–69. 10.1111/j.1354-7798.2005.00275.xSearch in Google Scholar
[33] Geroski, P. A., Machin, S. J., & Walters, C. F. (1997). Corporate growth and profitability. The Journal of Industrial Economics, 45(2), 171–189. 10.1111/1467-6451.00042Search in Google Scholar
[34] Glickman, S. W., Ou, F.-S., DeLong, E. R., Roe, M. T., Lytle, B. L., Mulgund, J. …, Peterson, E. D. (2007). Pay for performance, quality of care, and outcomes in acute myocardial infarction. JAMA, 297(21), 2373–2380. 10.1001/jama.297.21.2373Search in Google Scholar PubMed
[35] Goddard, J., Molyneux, P., & Wilson, J. O. (2004). The profitability of European banks: A cross-sectional and dynamic panel analysis. The Manchester School, 72(3), 363–381. 10.1111/j.1467-9957.2004.00397.xSearch in Google Scholar
[36] González, V. M., & González, F. (2012). Firm size and capital structure: Evidence using dynamic panel data. Applied Economics, 44(36), 4745–4754. 10.1080/00036846.2011.595690Search in Google Scholar
[37] Grossbart, S. R. (2006). What’s the return? Assessing the effect of “pay-for-performance” initiatives on the quality of care delivery. Medical Care Research and Review, 63(1 Suppl), 29S–48S. 10.1177/1077558705283643Search in Google Scholar
[38] Hall, R. E., & Jones, C. I. (1999). Why do some countries produce so much more output per worker than others? The Quarterly Journal of Economics, 114(1), 83–116. 10.3386/w6564Search in Google Scholar
[39] Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press. 10.1515/9780691218632Search in Google Scholar
[40] Hannan, E. L., Kumar, D., Racz, M., Siu, A. L., & Chassin, M. R. (1994). New York State’s cardiac surgery reporting system: Four years later. The Annals of Thoracic Surgery, 58(6), 1852–1857. 10.1016/0003-4975(94)91726-4Search in Google Scholar
[41] Hibbard, J. H., Stockard, J., & Tusler, M. (2003). Does publicizing hospital performance stimulate quality improvement efforts? Health Affairs, 22(2), 84–94. 10.1377/hlthaff.22.2.84Search in Google Scholar PubMed
[42] Hibbard, J. H., Stockard, J., & Tusler, M. (2005). Hospital performance reports: Impact on quality, market share, and reputation. Health Affairs, 24(4), 1150–1160. 10.1377/hlthaff.24.4.1150Search in Google Scholar PubMed
[43] Higuchi, S. (2010). Chuusho jichitai byouin-no genjo-to kadai (current situation and tasks for small and medium local public hospitals). Journal of Japan Hospital Association, 5, 95–101. Search in Google Scholar
[44] Hillman, A. L., Pauly, M. V., Kerman, K., & Martinek, C. R. (1991). HMO managers’ views on financial incentives and quality. Health Affairs, 10(4), 207–219. 10.1377/hlthaff.10.4.207Search in Google Scholar PubMed
[45] Hisamichi, S. (2010). Byouin keiei koto hajime. Byouin jigyou kanri-no tachiba kara (Starting hospital management. Point of view of a hospital manager). Journal of Japan Hospital Association, 2, 98–119. Search in Google Scholar
[46] Jones, N. B. (2014). Health care executives participating in value-based purchasing: A qualitative phenomenological study. (Ph.D. thesis). Phoenix: University of Phoenix. Search in Google Scholar
[47] Kaarbøe, O. M., & Siciliani, L. (2011). Multi-tasking, quality and pay for performance. Health Economics, 20(2), 225–238. 10.1002/hec.1582Search in Google Scholar PubMed
[48] Knapp, M., Gart, A., & Chaudhry, M. (2006). The impact of mean reversion of bank profitability on post-merger performance in the banking industry. Journal of Banking & Finance, 30(12), 3503–3517. 10.1016/j.jbankfin.2006.01.005Search in Google Scholar
[49] Kyereboah-Coleman, A. (2008). Corporate governance and firm performance in Africa: A dynamic panel data analysis. Studies in Economics and Econometrics, 32(2), 1–24. 10.1080/10800379.2008.12106447Search in Google Scholar
[50] Laeven, L., Levine, R., & Michalopoulos, S. (2015). Financial innovation and endogenous growth. Journal of Financial Intermediation, 24(1), 1–24. 10.1016/j.jfi.2014.04.001Search in Google Scholar
[51] Li, J., Hurley, J., DeCicca, P., & Buckley, G. (2014). Physician response to pay-for-performance: Evidence from a natural experiment. Health Economics, 23(8), 962–978. 10.3386/w16909Search in Google Scholar
[52] Lindenauer, P. K., Remus, D., Roman, S., Rothberg, M. B., Benjamin, E. M., Ma, A., & Bratzler, D. W. (2007). Public reporting and pay for performance in hospital quality improvement. New England Journal of Medicine, 356(5), 486–496. 10.1056/NEJMsa064964Search in Google Scholar PubMed
[53] Ma, C.-t.A., & Mak, H. Y. (2015). Information disclosure and the equivalence of prospective payment and cost reimbursement. Journal of Economic Behavior and Organization, 117, 439–452. 10.1016/j.jebo.2015.07.002Search in Google Scholar
[54] Machin, S., & Van Reenen, J. (1993). Profit margins and the business cycle: Evidence from UK manufacturing firms. The Journal of Industrial Economics, 41, 29–50. 10.2307/2950616Search in Google Scholar
[55] Manary, M., Staelin, R., Kosel, K., Schulman, K. A., & Glickman, S. W. (2015). Organizational characteristics and patient experiences with hospital care: A survey study of hospital chief patient experience officers. American Journal of Medical Quality, 30(5), 432–440. 10.1177/1062860614539994Search in Google Scholar PubMed
[56] Mark, B. A., Harless, D. W., McCue, M., & Xu, Y. (2004). A longitudinal examination of hospital registered nurse staffing and quality of care. Health Services Research, 39(2), 279–300. 10.1111/j.1475-6773.2004.00228.xSearch in Google Scholar PubMed PubMed Central
[57] Michaud, P.-C., & van Soest, A. (2008). Health and wealth of elderly couples: Causality tests using dynamic panel data models. Journal of Health Economics, 27(5), 1312–1325. 10.1016/j.jhealeco.2008.04.002Search in Google Scholar PubMed PubMed Central
[58] Morton, V., & Torgerson, D. J. (2003). Effect of regression to the mean on decision making in health care. British Medical Journal, 326(7398), 1083–1084. 10.1136/bmj.326.7398.1083Search in Google Scholar PubMed PubMed Central
[59] Murray, M. P. (2006). Avoiding invalid instruments and coping with weak instruments. Journal of Economic Perspectives, 20(4), 111–132. 10.1257/jep.20.4.111Search in Google Scholar
[60] Ogundeji, Y. K., Bland, J. M., & Sheldon, T. A. (2016). The effectiveness of payment for performance in health care: A meta-analysis and exploration of variation in outcomes. Health Policy, 120(10), 1141–1150. 10.1016/j.healthpol.2016.09.002Search in Google Scholar
[61] Oliveira, B., & Fortunato, A. (2006). Firm growth and liquidity constraints: A dynamic analysis. Small Business Economics, 27(2), 139–156. 10.1007/s11187-006-0006-ySearch in Google Scholar
[62] Oxholm, A. S., Kristensen, S. R., & Sutton, M. (2018). Uncertainty about the effort-performance relationship in threshold-based payment schemes. Journal of Health Economics, 62, 69–83. 10.1016/j.jhealeco.2018.09.003Search in Google Scholar
[63] Pai, C.-W., Finnegan, G. K., & Satwicz, M. J. (2002). The combined effect of public profiling and quality improvement efforts on heart failure management. The Joint Commission Journal on Quality Improvement, 28(11), 614–624. 10.1016/S1070-3241(02)28065-7Search in Google Scholar
[64] Pauly, M. (1980). Doctors and their workshops: Economic models of physician behavior. A National Bureau of Economic Research Monograph. Chicago and London: The University of Chicago Press. 10.7208/chicago/9780226650463.001.0001Search in Google Scholar
[65] Pérez-Calero, L., delMarVillegas, M., & Barroso, C. (2016). A framework for board capital. Corporate Governance, 16, 452–475.10.1108/CG-10-2015-0146Search in Google Scholar
[66] Roodman, D. (2009). A note on the theme of too many instruments. Oxford Bulletin of Economics and Statistics, 71(1), 135–158. 10.1111/j.1468-0084.2008.00542.xSearch in Google Scholar
[67] Rosenthal, M. B., Fernandopulle, R., Song, H. R., & Landon, B. (2004). Paying for quality: Providers’ incentives for quality improvement. Health Affairs, 23(2), 127–141. 10.1377/hlthaff.23.2.127Search in Google Scholar PubMed
[68] Ryan, A. M. (2009). Effects of the premier hospital quality incentive demonstration on medicare patient mortality and cost. Health Services Research, 44(3), 821–842. 10.1111/j.1475-6773.2009.00956.xSearch in Google Scholar PubMed PubMed Central
[69] Ryan, A. M., Blustein, J., Michelow, M. D., & Casalino, L. P. (2012). The effect of phase 2 of the premier hospital quality incentive demonstration on incentive payments to hospitals caring for disadvantaged patients. Health Services Research, 47(4), 1418–1436. 10.1111/j.1475-6773.2012.01393.xSearch in Google Scholar
[70] Ryan, A. M., Krinsky, S., Maurer, K. A., & Dimick, J. B. (2017). Changes in hospital quality associated with hospital value-based purchasing. New England Journal of Medicine, 376(24), 2358–2366. 10.1056/NEJMsa1613412Search in Google Scholar
[71] Santos, G. F. (2013). Fuel demand in Brazil in a dynamic panel data approach. Energy Economics, 36, 229–240. 10.1016/j.eneco.2012.08.012Search in Google Scholar
[72] Siciliani, L., Straume, O. R., & Cellini, R. (2013). Quality competition with motivated providers and sluggish demand. Journal of Economic Dynamics and Control, 37(10), 2041–2061. 10.1016/j.jedc.2013.05.002Search in Google Scholar
[73] Smith, L. R. (2017). A case study: The executive leadership response at a community hospital to the value-based purchasing requirements on the patient protection and affordable care act. (Ph.D. thesis). Jacksonville: University of North Florida, College of Education and Human Services. Search in Google Scholar
[74] Stock, J. H. (1991). Confidence intervals for the largest autoregressive root in us macroeconomic time series. Journal of Monetary Economics, 28(3), 435–459. 10.1016/0304-3932(91)90034-LSearch in Google Scholar
[75] Suwabe, A. (2004). Our efforts on DPC in Iwate Medical University hospital. Rinsho Byouri, 52, 1011–1014. Search in Google Scholar
[76] Vaghela, P., Ashworth, M., Schofield, P., & Gulliford, M. C. (2009). Population intermediate outcomes of diabetes under pay-for-performance incentives in England from 2004 to 2008. Diabetes Care, 32(3), 427–429. 10.2337/dc08-1999Search in Google Scholar PubMed PubMed Central
[77] Vina, E. R., Rhew, D. C., Weingarten, S. R., Weingarten, J. B., & Chang, J. T. (2009). Relationship between organizational factors and performance among pay-for-performance hospitals. Journal of General Internal Medicine, 24(7), 833. 10.1007/s11606-009-0997-6Search in Google Scholar PubMed PubMed Central
[78] Werner, R. M., Kolstad, J. T., Stuart, E. A., & Polsky, D. (2011). The effect of pay-for-performance in hospitals: Lessons for quality improvement. Health Affairs, 30(4), 690–698. 10.1377/hlthaff.2010.1277Search in Google Scholar PubMed
[79] Wilcox, M. A., Chang, A. M., & Johnson, I. R. (1996). The effects of parity on birthweight using successive pregnancies. Acta Obstetricia et Gynecologica Scandinavica, 75(5), 459–463. 10.3109/00016349609033354Search in Google Scholar PubMed
[80] Windmeijer, F. (2005). A finite sample correction for the variance of linear efficient two-step GMM estimators. Journal of Econometrics, 126, 25–51. 10.1016/j.jeconom.2004.02.005Search in Google Scholar
[81] Wu, J.-L. (2000). Mean reversion of the current account: Evidence from the panel data unit-root test. Economics Letters, 66(2), 215–222. 10.1016/S0165-1765(99)00198-6Search in Google Scholar
[82] Zhao, M., Bazzoli, G. J., Clement, J. P., Lindrooth, R. C., Nolin, J. M., & Chukmaitov, A. S. (2008). Hospital staffing decisions: Does financial performance matter? Inquiry, 45(3), 293–307. 10.5034/inquiryjrnl_45.03.293Search in Google Scholar PubMed
© 2022 Galina Besstremyannaya and Sergei Golovan, published by De Gruyter
This work is licensed under the Creative Commons Attribution 4.0 International License.