1,457
views
1
recommends
+1 Recommend
1 collections
    3
    shares

      Interested in becoming an AMM published author?

      • Platinum Open Access with no APCs.
      • Fast peer review/Fast publication online after article acceptance.

      Check out the call for papers on our website https://amm-journal.org/index.php/2023/04/26/acta-materia-medica-call-for-papers-2/

      scite_
       
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Statistical evaluation of absolute change versus responder analysis in clinical trials

      Published
      research-article
      Bookmark

            Abstract

            In clinical trials, the primary analysis is often either a test of absolute/relative change in a measured outcome or a corresponding responder analysis. Although each of these tests may be reasonable, determining which test is most suitable for a particular research study remains an open question. These tests may require different sample sizes or define different clinically meaningful differences; most importantly, they may lead to different study conclusions. The aim of this study was to compare a typical non-inferiority test using absolute change as the study endpoint to the corresponding responder analysis in terms of sample-size requirements, statistical power, and hypothesis-testing results. From numerical analysis, using absolute change as an endpoint generally requires a larger sample size; therefore, when the sample size is the same, the responder analysis has higher power. The cut-off value and non-inferiority margin are critical and can meaningfully affect whether the two types of endpoints yield conflicting conclusions. Specifically, extreme cut-off values are more likely to yield different conclusions. However, this influence decreases as population variance increases. One important reason for conflicting conclusions is a non-normal population distribution. To eliminate conflicting results, researchers should consider the population distribution and cut-off value selection.

            Main article text

            1. INTRODUCTION

            In clinical trials, primary study endpoints are often analyzed to determine whether the intended studies will achieve the study objectives with the desired statistical power. In practice, investigators can consider four different types of primary endpoints or outcomes according to a single study objective: (i) absolute change (i.e., the endpoint’s absolute change from baseline), (ii) relative change (e.g., the endpoint’s percentage change from baseline), (iii) responder analysis based on absolute change (i.e., an individual participant is defined as a responder if the absolute change in the primary endpoint exceeds a pre-specified threshold known as a clinically meaningful improvement), or (iv) responder analysis based on relative change. Although analyses based on these endpoints all appear reasonable, the following statements are often of great concern to principal investigators [1]. First, a clinically meaningful difference in one endpoint does not directly translate to a clinically meaningful difference in another endpoint. Second, these derived endpoints generally have different sample-size requirements. Third, and most importantly, these derived endpoints may not yield the same statistical conclusion (based on the same data set). Consequently, determining which type of primary endpoint is most appropriate and can best inform on disease status and treatment effects is of particular interest.

            Some researchers have criticized responder analysis because of a loss of information; i.e., the statistical power of a trial decreases if a continuous outcome is categorized into a binary variable [2, 3]. Although responder analysis may come at the expense of power, it still provides value. For example, the original scale (continuous) outcome is used to make binary decisions, such as whether a patient should be hospitalized. In a heterogeneous disease, a subset of patients may have more benefit than others, thus resulting in a non-normal distribution of the outcome [4]. If the trial is aimed at investigating an additional second agent, and the proportion of patients with more benefit is of interest, responder analysis is more suitable [4]. Because of its benefits and drawbacks, [3] have suggested that responder analysis be used as a secondary analysis to better interpret findings from the main analysis. However, analysis using absolute change as an endpoint and the corresponding responder analysis have different statistical properties. Hence, investigating their differences in terms of statistical power, sample size, and conclusions is highly important.

            To study the relative performance of these derived endpoints, in addition to mathematical derivations, we conducted a numerical study and real case study based on data on a recent rehabilitation program in lung transplant candidates and recipients [5]. The 6-minute walk distance (6MWD), a commonly used clinical indicator for patients with pulmonary disease, can be used as not only a prognostic factor but also as a health outcome variable [6]. For example, the 6MWD has been used to measure the functional status and exercise capacity of lung transplant candidates or recipients [7, 8]. Some studies have used the endpoint of the change in 6MWD from baseline (absolute change) as the outcome variable to evaluate the performance of pulmonary disease treatment [9], whereas others have performed responder analysis using 6MWD [10, 11]. Individuals can be defined as responders if they meet a pre-specified threshold of improvement in 6MWD and as non-responders otherwise. In one example, [11] have classified patient performance after rehabilitation according to the following criteria: good, 6MWD increase ≥50 m; moderate, ≥25 to <50 m; and non-response, <25 m. However, in another study [12], have reported that the minimum important change in 6MWD in chronic respiratory disease is 25 to 33 m. [13] have used 25 m as the threshold for equivalence in the change in 6MWD. Although the wider range of 6MWD is generally accepted as 25 to 30 m, the exact threshold of change in 6MWD that is clinically meaningful is under debate.

            In this case study, for simplicity, we focused on statistical evaluation of a rehabilitation program in lung transplant candidates and recipients in terms of absolute change in 6MWD and responder analysis based on a pre-specified threshold (improvement) of 6MWD on the basis of absolute change. A comparison between the absolute change and responder analysis with various pre-specified thresholds was performed in terms of sample-size requirement and statistical power. In the next section, we presented statistical methods for analysis using absolute change as the study endpoint as well as the corresponding responder analysis. Additionally, we compared the performance of these study endpoints in terms of statistical power, sample size and study results/conclusions. In Section 3, we discussed a numerical analysis of the comparison between absolute change and responder analysis and a case study of a rehabilitation program in lung transplant candidates and recipients. Brief concluding remarks and recommendations were given in the last section of this article.

            2. METHODS

            2.1 Hypothesis testing for efficacy

            Non-inferiority testing is commonly considered in randomized clinical trials evaluating the performance of a new drug or new treatment versus an active control (e.g., standard of care). The success of a non-inferiority trial depends on the selection of the study endpoint and the non-inferiority margin. As indicated earlier, for a given study endpoint, four types of primary endpoints exist: absolute change (e.g., endpoint change from baseline), relative change (e.g., endpoint percentage change from baseline), responder analysis based on a pre-specified improvement (threshold) in absolute change, and responder analysis based on a pre-specified improvement (threshold) in relative change. Consequently, the inference from responder analysis is very sensitive to the pre-specified threshold (cutoff) value [14]. For simplicity and illustration purposes, we examine the performance of the first two primary endpoints: absolute change and a corresponding responder analysis.

            We assume a two-arm parallel randomized clinical trial comparing a test treatment (T) and an active control (C) with a 1:1 treatment allocation ratio. Let W 1 ij and W 2 ij be the original response of the ith patient in the jth treatment group at baseline and post-treatment, where i = 1, …, nj and j = C, T, respectively. Furthermore, W 1 ij is assumed to follow a log-normal distribution

            and W 2 ij = W 1 ij (1 + Δ ij ), where
            Hence, the absolute change from baseline is as follows:

                            

                  (1)

            where W 1 ij and Δ ij are assumed to be independent. Let Xij  = log (W 2 ij  – W 1 ij ) represent the log absolute change; then

            Let xij denote the observations of random variable Xij . The reason for using W 1 ij and W 2 ij instead of directly using Xij , following a normal distribution, is that the same notation can be used to denote relative change. For example,
            can represent log relative change, which follows
            Although the relative-change endpoint is not the focus of this work, this notation will benefit future studies.

            The outcome variable for responder analysis based on a pre-specified absolute change is then given by

            where c 1 is the cutoff value. Then the endpoint for the responder analysis becomes
            For a sufficiently large sample size,
            can be verified to asymptotically follow
            [1]. According to the definition of

                            

                     (2)

            where Φ(·) is the cumulative distribution function (CDF) of a standard normal distribution. The hypotheses for non-inferority testing based on the derived endpoint of absolute change and the corresponding responder analysis can be set up as follows.

            1. Absolute change:

                              

                       (3)

              where δ 1 is the non-inferiority margin in hypothesis testing using absolute change.

            2. Responder analysis based on a pre-specified threshold (improvement) of absolute change:

                              

                       (4)

              where δ 2 is the non-inferiority margin in hypothesis testing using responder analysis.

            For a non-inferority test based on the derived endpoint of absolute difference, the Z test statistic under the null hypothesis in Equation (3) is given by

                            

                     (5)

            where

            and
            are the sample means of absolute change in the treatment and control groups, respectively, and n 1 is the sample size of the treatment or control group, assuming an allocation ratio of 1:1. Let δ 1 A denote the true sample mean difference. The corresponding statistical power can be written as follows:

                            

                            

                     (6)

            The sample-size requirement for the non-inferiority test using absolute difference can then be obtained as follows:

                            

                     (7)

            For a non-inferiority test of responder analysis based on a pre-specified threshold (improvement) of absolute difference, the Z test statistic under the null hypothesis in Equation (4) can be derived as follows:

                            

                     (8)

            where

            and
            are the sample proportions in the treatment and control groups, respectively, and n 2 is the sample size of the treatment or control group, assuming an allocation ratio of 1:1. Similarly, let δ 2 A denote the true proportion difference; then the corresponding statistical power is as follows:

                            

                            

                   (9)

            where the last approximate equation holds according to Slutsky’s theorem [1]. The sample-size requirement for a non-inferiority test for responder analysis based on a pre-specified threshold (improvement) of absolute difference is then given by the following:

                            

                   (10)

            2.2 Statistical power comparison in non-inferiority tests

            Many previous studies have suggested avoiding relative difference because of statistical inefficiency [15]. Extending this idea, we compared statistical power for non-inferiority testing using absolute change and a responder analysis using absolute change. The comparisons of required sample sizes and conclusion for non-inferiority tests are also shown in this section. Let AC denote absolute change and PAC denote responder analysis using absolute change.

            From the formula of statistical power of a non-inferiority test shown in Section 2, the power difference can be computed with the CDF of N(0,1). Through Taylor expansion, the CDF of N(0,1) Φ(·) can be written as follows:

                            

                     (11)

            Keeping the first term of the Taylor expansion in Equation (11), Φ(x 1) – Φ(x 2) can be simplified as follows:

                            

                     (12)

            To compare the statistical power of a non-inferiority test using the absolute-change endpoint with the statistical power for a responder analysis using the absolute-change endpoint, we first simplify

            On the basis of Equation (11),
            can be written as follows:

                            

                     (13)

            and

                            

                     (14)

            Hence,

                            

                     (15)

            and

                            

                 (16)

            If the sample sizes of a non-inferiority test using absolute change and the corresponding responder analysis are assumed to be the same, denoted n, on the basis of Equation (12), the difference between power 1 and power 2 can be written as follows:

                            

                     (17)

            2.3 Sample-size comparison in non-inferiority tests

            From Equation (15) and (16), the sample size for the responder analysis using the absolute-change endpoint in Equation (10) can be written as follows:

                            

                     (18)

            When the significance level and desired statistical power are the same, the necessary sample sizes for a responder analysis and a test from absolute change can be compared with the following ratio:

                            

                     (19)

            2.4 Conflict probability in non-inferiority tests

            In this section, we investigate the probabilities of a non-inferority test using absolute change as the endpoint and the corresponding responder analysis having similar or different conclusions. We assume that the samples used to conduct these two types of non-inferority test are the same. Thus, four types of events are possible:

            Both AC and PAC reject H 0

                            

                            

                            

                   (20)

            AC fails to reject H 0, whereas PAC rejects H 0

                            

                            

                            

                   (21)

            AC rejects H 0, whereas PAC does not reject H 0

                            

                            

                            

                   (22)

            Both AC and PAC do not reject H 0

                            

                            

                            

                   (23)

            3. RESULTS

            In this section, a numerical analysis using simulated data is conducted to investigate the difference between using absolute change as an endpoint and the corresponding responder analysis in terms of sample-size requirement, statistical power, and non-inferiority-test conclusions. Responses are assumed to follow a normal distribution. The allocation ratio is 1:1. The simulation is conducted 1000 times. Additionally, a case study is used to investigate the difference between a typical non-inferiority test and responder analysis by using real clinical data from [5]. Again, AC denotes a typical non-inferiority test using absolute change as the endpoint, and PAC denotes the corresponding responder analysis. The significance level is 0.05, and the desired power is 0.80.

            3.1 Numerical analysis

            According to Equation (7), the sample size of AC is associated with the population mean, population variance, and non-inferiority margin. The treatment-group population mean is set to 0.2 or 0.3, the control-group population mean is set to 0, and the population variance of both groups is 1.0, 2.0, or 3.0. Table 1 presents the required sample size of AC to achieve 80% statistical power. The sample size is associated with the effect size and the non-inferiority margin. When the effect size is fixed, a larger non-inferiority margin leads to a smaller sample size in AC; when the non-inferiority margin is fixed, a larger effect size leads to a smaller sample size in AC. Similarly, according to Equation (10), the sample size of PAC is additionally associated with the cut-off value (threshold) used to determine responders. As shown in Table 2 , the influences of effect size and non-inferiority margin on the sample size are the same as in Table 1 when the cut-off value is fixed. However, the influence of the cut-off value on the sample-size calculation is quite complex, because its effects are associated with not only its absolute value but also the population mean and variance.

            Table 1 |

            Sample sizes for non-inferiority tests using the absolute-change endpoint (AC).

            1.02.03.01.02.03.0
            1.02.03.01.02.03.01.02.03.01.02.03.01.02.03.01.02.03.0
            δ 1 = 0.25246368490368490612490612734164246328246328410328410492
            δ 1 = 0.30198298396298396496396496594138208276208276344276344414
            δ 1 = 0.35164246328246328410328410492118176236176236294236294352
            δ 1 = 0.40138208276208276344276344414102152202152202254202254304
            δ 1 = 0.4511817623617623629423629435288132176132176220176220264
            δ 1 = 0.5010215220215220225420225430478116156116156194156194232
            δ 1 = 0.558813217613217622017622026470104138104138172138172206
            δ 1 = 0.6078116156116156194156194232629212492124154124154184
            δ 1 = 0.6570104138104138172138172206568411084110138110138166
            δ 1 = 0.70629212492124154124154184507610076100124100124150
            Table 2 |

            Sample sizes for responder analysis using the absolute-change endpoint (PAC).

            1.02.03.01.02.03.0
            1.02.03.01.02.03.01.02.03.01.02.03.01.02.03.01.02.03.0
            Cut-off value = 0.1
            δ 2 = 0.251141221261221321361261361429038100104110114110118122
            δ 2 = 0.30869294929810094100104709676808486848892
            δ 2 = 0.35687274727678747880567460626668667072
            δ 2 = 0.40545858586062586264466050505454545656
            δ 2 = 0.45444648465050485052904840424444444646
            Cut-off value = 0.2
            δ 2 = 0.251141321421141321421141321429010411096110118100114122
            δ 2 = 0.30869810486981048698104708084748488768692
            δ 2 = 0.35687680687680687680566266606670606872
            δ 2 = 0.40546062546062546062465054485456505456
            δ 2 = 0.45444852444852444852384244404446404446
            Cut-off value = 0.3
            δ 2 = 0.25112142158104132146102126140901101229011012290110122
            δ 2 = 0.308410411480981067894104708492708492708492
            δ 2 = 0.35668086647482627480566670566670566670
            δ 2 = 0.40546268526064505862465456465456465456
            δ 2 = 0.45445054424852424850384446384446384446
            Cut-off value = 0.4
            δ 2 = 0.251101501769613015092122140881181348411012482106120
            δ 2 = 0.3084108124749610870901026888100668294648090
            δ 2 = 0.35648292587482567078566876526672526470
            δ 2 = 0.40526472485864465662465660445258425256
            δ 2 = 0.45425258404852384650384650364448364246

            A comparison of sample sizes in Tables 1 and 2 indicates that when the non-inferiority margin is fixed, the required sample size of AC is much larger than that of PAC. One important assumption is that the non-inferiority margins of the two tests are the same. The reason for this assumption is that many researchers have suggested using responder analysis as a secondary analysis [3]; i.e., the sample size is computed on the basis of the primary analysis, and statistical analysis of a typical non-inferiority test and the responder analysis are conducted on the same dataset. However, in practice, the non-inferiority margins in these two tests are likely to differ, because these two tests have different meanings. Therefore, when conducting responder analysis is the secondary analysis, the power of the secondary analysis might potentially be insufficient.

            Next, we compare the statistical power of AC and PAC by using Equations (6) and (9), when the sample size is fixed. In the simulation process, the sample size used to generate random samples is the minimum of all possible sample sizes, given the population mean and standard deviation. Here, the population means of the treatment and control groups are 0.2 and 0; the population variance of the treatment group is 2; the population variance of the control group ranges from 1 to 3; and the cut-off value ranges from 0.1 to 0.8. To make the power comparable, the non-inferiority margins of AC and PAC are assumed to be the same. As shown in Figure 1 , with a fixed sample size, the statistical power of AC is smallest, thus suggesting that a larger sample size is required to achieve the desired power with AC than PAC. This finding is consistent with the results in Tables 1 and 2 . Additionally, in PAC, the statistical power values when different cut-off values are used become closer to one another as the population variance increases. The statistical power of PAC is either slightly lower than 80% or above 80%, regardless of the cut-off value. The statistical power of AC is always below 60%. Hence, if researchers conduct a sample-size calculation based on responder analysis but ultimately use a typical non-inferiority test, they will not be able to achieve sufficient statistical power.

            Figure 1 |

            Statistical power comparison of non-inferiority tests using absolute change as an endpoint (AC) and corresponding responder analysis (PAC).

            To illustrate the relationships among required sample sizes, we assume that the non-inferiority margins of using the absolute-change endpoint and responder analysis are the same. The setting is the same as that in Figure 1 . In Figure 2 , the ratio of the PAC sample size to AC is used to represent the relationship between the AC and PAC sample size, where N 1 denotes the sample size of AC, and N 2 denotes the sample size of PAC. Under the settings used here, N 2/N 1 is always smaller than 0.35, thus suggesting that the sample size of AC is much larger than that of PAC. When the non-inferiority margin increases, ratios with different cut-off values become not only smaller but also closer to one another. Comparison of Figure 2a–c indicates that the ratio of sample sizes decreases, and the sample-size ratios with different cut-off values become closer to one another when the variance of the control group increases.

            Figure 2 |

            Sample-size comparison of non-inferiority tests using absolute change as an endpoint (AC) and corresponding responder analysis (PAC).

            Another essential parameter of interest in responder analysis is the cut-off value (threshold) to determine whether an observation indicates a responder. Let the population mean of the treatment group range from 0.10 to 0.30. To make the results comparable, the non-inferiority margins in AC and PAC are set to 0. The range of the cut-off value is set to be larger than before, from −3 to 3. The simulation process initially randomly generates continuous samples from a normal distribution, where the sample size is computed with the AC sample-size formula in Equation (7). Then, using the cut-off value, we label each participant as either a responder or a non-responder. As shown in Figure 3 , the cut-off value can indeed drive the conclusion in a different direction. In Figure 3a , a negative cut-off value provides conflicting results; in Figure 3b , a more extreme cut-off value provides conflicting results; the same findings are indicated in Figure 3c . Additionally, the influence of the cut-off value on the hypothesis-testing result is associated with the population mean and variance; however, the overall pattern is similar. Hence, a more extreme cut-off value, i.e., a cut-off farther away from the population mean, is more likely to lead to conflicting conclusions.

            Figure 3 |

            Comparison of non-inferiority-test results of typical tests using absolute change as an endpoint (AC) and corresponding responder analysis (PAC).

            3.2 Case study

            In Section 3.1, we studied the effects of essential parameters on sample-size requirements, statistical power, and test conclusions by using simulated data. To provide a clearer illustration of the influence of the cut-off value on non-inferiority-test results, we conducted a case study using real clinical data from an observational study on rehabilitation in patients who had received lung transplants [5]. That study’s primary aim was to compare the performance of individual rehabilitation and group rehabilitation in participants both pre-operatively and post-operatively, measured by a primary outcome variable of the change in 6MWD (detailed information in Table 3 ).

            Table 3 |

            Changes in 6MWD in pre-operative and post-operative participants in [5].

            Pre-operativePost-operative
            RehabilitationGroupIndividualGroupIndividual
            Sample size9381110105
            Mean (SD)51.6 (81.3)56.6 (62.9)174 (97.6)160 (89.4)
            Median [Q1, Q3]44.5 [6.40,102]59.7 [25.0,93.9]168 [106, 232]159 [104, 208]

            In this section, the non-inferiority test was used to study the circumstances in which AC and PAC may lead to different conclusions. According to previous studies [12, 13], a clinically meaningful change in 6MWD is between 25 m and 33 m. The cut-off value herein ranged from 20 m to 35 m, to provide more comprehensive understanding of the effects of cut-off value selection on study conclusions. The non-inferiority margin of AC ranged from −0.3 to 0.3, and the non-inferiority margin of PAC ranged from 0 to 0.03. As shown in Figure 4 , for pre-operative patients, some cut-off values may lead to different conclusions. For instance, in Figure 4a , a cut-off value larger than 27 yielded conflicting results. However, for post-operative patients, if the cut-off value was between 20 and 35, both AC and PAC always yielded consistent results. Closer examination of the data indicated that, for most post-operative patients, the change in 6MWD was either extremely large (above 35) or extremely small (below 20). That is, in this scenario, an extreme cut-off value (ranging from 20 to 35) did not significantly affect the proportion of responders among post-operative patients. These results suggest that cut-off value selection may cause responder analysis and typical non-inferiority tests to yield conflicting findings under certain circumstances.

            Figure 4 |

            Comparison of non-inferiority-test results of typical tests using absolute change as an endpoint (AC) and corresponding responder analysis (PAC) in a study examining a rehabilitation program after lung transplantation [5].

            As described in Section 1, a responder analysis answers a different question from a typical non-inferiority test. Specifically, if an extreme cut-off value is selected, responder analysis investigates whether the test treatment might provide substantial clinical benefits to patients. For example, in Figure 4a , AC yields an insignificant conclusion, i.e., individual rehabilitation is inferior to group rehabilitation, whereas PAC yields a significant conclusion when the cut-off value is large. These findings suggest that individual rehabilitation is non-inferior to group rehabilitation only for a small proportion of patients and provides a benefit of substantial improvement. The large cut-off value allowed us to focus on a smaller proportion of patients with substantial improvement, but this difference may not be detectable in typical non-inferiority tests, thus yielding conflicting findings.

            4. DISCUSSION

            One of the most important steps in any clinical trial is determining the primary study endpoint, which may influence aspects including establishment of hypotheses, selection of statistical models, and calculation of sample size. In general, four types of study endpoints exist: (i) absolute change, (ii) relative change, (iii) responder analysis using absolute change, and (iv) responder analysis using relative change. To demonstrate a comparison of different study endpoints, this work focused on the comparison of endpoints (i) and (iii) in non-inferiority tests, in terms of the sample-size requirement, statistical power, and whether different endpoints might lead to different conclusions. However, the comparison process in this study could also be generalized to compare any two study endpoints described above.

            In the numerical study section, both a simulation study and a case study using data from [5] were conducted. According to the simulation study, the required sample size of a non-inferiority test using AC is associated with the population mean, variance of the treatment and control groups, and non-inferiority margin. The sample size of the corresponding PAC is additionally associated with the cut-off value used to determine responders. After fixing all parameters, we observed that PAC requires a smaller sample size than AC. That is, for the same sample size, PAC will always have greater statistical power than AC, as shown in Figure 1 . When the desired statistical power is the same, the sample-size ratio of PAC to AC is always smaller than 1, an aspect also associated with the non-inferiority margin and cut-off value. However, the effects of these two parameters decrease with increasing population variance. As the cut-off value becomes more extreme, the likelihood of obtaining conflicting conclusions from a non-inferiority hypothesis test increases. This aspect was observed in both the simulation study and the case study. Our findings indicated that the selection of cut-off value selection is highly important, because it may lead to conflicting results when the mean and median in the treatment and control groups are close to the cut-off value.

            Without a loss of generalizability, similar conclusions may be found in superiority and equivalence tests. The fundamental reason why typical non-inferiority/superiority/equivalence tests using absolute change as an endpoint and corresponding responder analysis provide conflicting conclusions is the distribution of the target population. If the samples follow a normal distribution, typical tests and responder analysis are highly likely to yield the same conclusion when the cut-off value is close to the population mean. Otherwise, these two types of analysis would provide conflicting results, particularly when the cut-off value is far from the population mean.

            Because of the great importance of cut-off value selection and the possibility of obtaining conflict conclusions, we suggest determining cut-off values on the basis of existing knowledge in combination with statistical analysis of the collected sample in conducting responder analysis. Although the clinically important difference (MCID) is always used as the cut-off value [4], some studies have proposed some guidance or approaches for the selection of cut-off values [16, 17]. Additionally, because the sample-size requirements of AC and PAC are different, the sample size must be verified to be sufficiently large to achieve the desired statistical power. Notably, typical tests and responder analysis may require different sample sizes, differ in power, and yield different study conclusions; moreover, tests using absolute instead of relative change as a study endpoint may face the same challenges. Some studies have reported that absolute- and relative-change endpoints may lead to conflicting conclusions [1, 18]. In addition, these endpoints are viewed differently among drug approval administrations. According to the non-inferiority-test guidance from the US Food and Drug Administration (FDA), study constancy is expected to be based on the constancy of relative effects, not absolute effects [19]. However, the European Medicines Agency’s (EMA) guidance uses absolute difference to illustrate instructions for non-inferiority tests [20]. Hence, a drug approved by the FDA might not be approved by the EMA, or vice versa, because the required sample size and statistical power of using absolute change and relative change as a study endpoint differ [1]. Hence, providing the confidence intervals of cut-off values may be useful when a typical non-inferiority test and responder analysis might lead to consistent conclusions. Moreover, the circumstances in which both absolute- and relative-change endpoints provide the same non-inferiority-test results should be investigated.

            ACKNOWLEGEDMENTS

            This research was partly supported by grant UL1TR002553 from the National Center for Advancing Translational Sciences (NCATS) of the National Institutes of Health (NIH) and NIH Roadmap for Medical Research. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of NCATS or NIH.

            CONFLICTS OF INTEREST

            Shein-Chung Chow is Deputy Editor-in-Chief of Acta Materia Medica. He was not involved in the peer-review or handling of the manuscript. The other authors have no other competing interests to disclose.

            REFERENCES

            1. Chow CS. Controversial Statistical Issues in Clinical Trials. Boca Raton, FL, USA: CRC Press. 2011. p. 135–147

            2. Snapinn SM, Jiang Q. Responder Analyses and the Assessment of a Clinically Relevant Treatment Effec. Trials. 2007. Vol. 8:1–6

            3. Henschke N, van Enst A, Froud R, WG Ostelo R. Responder Analyses in Randomised Controlled Trials for Chronic Low Back Pain: An Overview of Currently Used Methods. European Spine Journal. 2014. Vol. 23:772–778

            4. Jones PW, Rennard S, Tabberer M, Riley JH, Vahdati-Bolouri M, Barnes NC. Interpreting Patient-Reported Outcomes from Clinical Trials in COPD: A Discussion. International Journal of Chronic Obstructive Pulmonary Disease. 2016. Vol. 11:3069

            5. Byrd R, Breslin R, Wang P, Peskoe S, Chow SC, Lowers S, et al.. Group versus Individual Rehabilitation in Lung Transplantation: A Retrospective Non-Inferiority Assessment. 2022.

            6. Tuppin MP, Paratz JD, Chang AT, Seale HE, Walsh JR, Kermeeen FD, et al.. Predictive Utility of the 6-Minute Walk Distance on Survival in Patients Awaiting Lung Transplantation. The Journal of Heart and Lung Transplantation. 2008. Vol. 27:729–734

            7. Martinu T, Babyak MA, O’Connell CF, Carney RM, Trulock EP, Davis RD, et al.. Baseline 6-Min Walk Distance Predicts Survival in Lung Transplant Candidates. American Journal of Transplantation. 2008. Vol. 8:1498–1505

            8. Munro PE, Holland AE, Bailey M, Button BM, Snell GI. Pulmonary Rehabilitation Following Lung Transplantation. Transplantation Proceedings. 2009. Vol. 41:292–295

            9. Ryerson CJ, Cayou C, Topp F, Hilling L, Camp PG, Wilcox PG, et al.. Pulmonary Rehabilitation Improves Long-Term Outcomes In Interstitial Lung Disease: a Prospective Cohort Study. Respiratory Medicine. 2014. Vol. 108:203–210

            10. Gilbert C, Brown MC, Cappelleri JC, Carlsson M, McKenna SP. Estimating a Minimally Important Difference in Pulmonary Arterial Hypertension Following Treatment with Sildenafil. Chest. 2009. Vol. 135:137–142

            11. Stoilkova-Hartmann A, Janssen DJ, Franssen FM, Wouters EF. Differences in Change in Coping Styles between Good Responders, Moderate Responders and Non-Responders to Pulmonary Rehabilitation. Respiratory Medicine. 2015. Vol. 109(12):1540–1545

            12. Holland AE, Spruit MA, Troosters T, Puhan MA, Pepin V, Saey D, et al.. An Official European Respiratory Society/American Thoracic Society Technical Standard: Field Walking Tests in Chronic Respiratory Disease. European Respiratory Journal. 2014. Vol. 44:1428–1446

            13. Holland AE, Mahal A, Hill CJ, Lee AL, Burge AT, Cox NS, et al.. Home-Based Rehabilitation for COPD Using Minimal Resources: A Randomised, Controlled Equivalence Trial. Thorax. 2017. Vol. 72:57–65

            14. Chow SC, Song F. On Controversial Statistical Issues in Clinical Research. Open Access Journal of Clinical Trials. 2015. Vol. 7:43–51

            15. Vickers AJ. The Use of Percentage Change from Baseline as an Outcome in a Controlled Trial is Statistically Inefficient: a Simulation Study. BMC Medical Research Methodology. 2001. Vol. 1:1–4

            16. Farrar JT, Dworkin RH, Max MB. Use of the Cumulative Proportion of Responders Analysis Graph to Present Pain Data Over a Range of Cut-Off Points: Making Clinical Trial Data More Understandable. Journal of Pain and Symptom Management. 2006. Vol. 31:369–377

            17. Harrell FE. Regression Modeling Strategies. Springer International Publishing.

            18. Curran-Everett D, Williams CL. Explorations in Statistics: The Analysis of Change. Advances in Physiology Education. 2015. Vol. 39:49–54

            19. FDA. Non-Inferiority Clinical Trials to Establish Effectiveness. 2016

            20. EMA. Guideline on the Choice of the Non-inferiority Margin. 2015

            Graphical abstract

            Highlights
            • In clinical trials, a non-inferiority test using absolute change as the study endpoint and a responder analysis are commonly considered.

            • When considering different derived primary endpoints, such as absolute change and the corresponding responder analysis, a clinically meaningful difference in one endpoint does not directly translate to a clinically meaningful difference in another endpoint.

            • A non-inferiority test using absolute change requires a larger sample size when compared with a corresponding responder analysis.

            • A non-inferiority test and a corresponding responder analysis may lead to different study conclusions.

            • Responder analysis is more favorable in terms of smaller sample size when the endpoint does not follow exact normal distribution.

            In brief

            In this article, the merits and disadvantages in terms of sample size, statistical power, and study conclusions between two derived endpoints (absolute change and the corresponding responder analysis) are studied.

            Author and article information

            Journal
            amm
            Acta Materia Medica
            Compuscript (Ireland )
            2737-7946
            26 August 2022
            : 1
            : 3
            : 320-332
            Affiliations
            [a ]Department of Biostatistics and Bioinformatics, Duke University School of Medicine, Durham, North Carolina, USA
            [b ]Department of Cardiology, Duke Health System, Durham, North Carolina, USA
            [c ]Department of Psychiatry and Behavioral Sciences, Duke Health System, Durham, North Carolina, USA
            Author notes
            *Correspondence: pj.wang@ 123456duke.edu (P. Wang)
            Article
            10.15212/AMM-2022-0020
            37274016
            87231813-5f94-4920-adf0-6d11a8519784
            Copyright © 2022 The Authors.

            Creative Commons Attribution 4.0 International License

            History
            : 20 June 2022
            : 25 July 2022
            : 10 August 2022
            Page count
            Figures: 4, Tables: 3, References: 20, Pages: 13
            Categories
            Research Article

            Toxicology,Pathology,Biochemistry,Clinical chemistry,Pharmaceutical chemistry,Pharmacology & Pharmaceutical medicine
            threshold selection,primary endpoints,responder analysis

            Comments

            Comment on this article