Treatment Effectiveness and the Russo–Williamson Thesis, EBM+, and Bradford Hill's Viewpoints

ABSTRACT Establishing the effectiveness of medical treatments is one of the most important aspects of medical practice. Bradford Hill's viewpoints play an important role in inferring causality in medicine, and EBM+ seeks to improve evidence-based medicine, which is influential in establishing treatment effectiveness. At EBM+'s foundations lies the Russo–Williamson thesis (RWT), which can be seen as providing a reduction of Hill's viewpoints into those involving difference-making and mechanistic evidence, both of which are claimed by the RWT's proponents to be typically required for establishing causal claims in medicine. Yet little has been written on whether and how the RWT, EBM+, and Hill's viewpoints establish treatment effectiveness. This could be because of uncertainty over what treatment effectiveness is. I provide an account of treatment effectiveness, analysing the role of the RWT, EBM+, and Hill's viewpoints in this regard. I argue that Hill's viewpoints and EBM+ can be useful in helping to evaluate evidence, but cannot directly establish or confirm treatment effectiveness. This is partly because effectiveness, I claim, is subject to inductive risk and therefore determined by non-epistemic values, neither of which Hill's viewpoints nor EBM+ offer guidance on. I conclude by reinterpreting Hill's viewpoints in light of establishing treatment effectiveness.


Introduction
In contemporary medical research and practice, much of the evidence regarding the putative effectiveness of medical treatments is based on randomised controlled trials (RCTs) and an evidence-based medicine (EBM)-inspired epistemic framework. Although EBM has been defined as 'the conscientious, explicit, and judicious use of current best evidence in making decisions about the care of individual patients' (Sackett et al. 1996), as Bluhm (2017a, 210) notes, '[t]he central idea underlying EBM is that research can be ranked hierarchically based on study design,' with RCTs and their meta-analyses sitting at the apex, and research on physiological mechanisms at the bottom. There are limitations with this approach (Cartwright and Stegenga 2011;Williamson (2021) provides a framework to help determine which and how many of Hill's viewpoints, when satisfied, can establish causality, and Luján and Todt (2020) characterise the RWT as a reduction of Hill's viewpoints to considerations of difference-making and mechanistic evidence. However, neither the RWT, EBM+, nor Hill's viewpoints-all of which are related to each other-are explicit as to how they can inform treatment effectiveness claims. Accordingly, my focus in this article is analysing whether, and if so, how, the RWT, EBM+, and Hill's viewpoints contribute to establishing or confirming treatment effectiveness, and then suggesting ways to build upon these approaches.
Immediately, at least with respect to the RWT ostensibly not being meant for this purpose, I could be charged with engaging a strawman. Illari and Russo (2014, 53, italics in original), for example, state, 'The point [of the RWT] is not to judge causal conclusions drawn by scientists either in the past or today, but to understand what evidential elements science uses in drawing causal conclusions.' Yet the RWT and its associated epistemological framework do apparently aim to establish sufficiency conditions for establishing causal claims, despite the RWT's avowed purpose of simply stating necessary conditions and the RWT's distinction between evidence that establishes causation and evidence that merely warrants action (Williamson 2019, 36). 4 For example, based on case studies advanced by its proponents, by 'establish' the RWT appears to also mean stating evidence sufficient to justify claims that could very well be considered a call to action for scientists, clinicians, or public policy makers (e.g. 'prescribe this treatment, because it is effective'). Ostensibly then, the RWT can be viewed as serving not merely as a test of whether a putative causal claim (e.g. 'a treatment is effective') has been legitimately established, but as a method for establishing it in the first place. Luján and Todt (2020), for example, apply the RWT to regulatory science, using it as a hypothetical test of whether an already regulated substance (such as a treatment) should continue to be regulated, or an unregulated substance should be regulated. If the RWT did not somehow indicate what the sufficient conditions for a causal claim in medicine are it might appear hollow in the face of researchers' and clinicians' needs for actionable information. Parkinnen and colleagues' (2018) framework of grading evidence indeed appears to provide a decision procedure by which the status of a causal claim can be assessed based on the quality of the evidence, such as done by the Grading of Recommendations, Assessment, Development and Evaluation (GRADE; Guyatt et al. 2008) approach, with GRADE being used to not only rule out claims but to rule them in in the first place.
What the sufficient conditions are for establishing or confirming treatment effectiveness may differ depending on the disease/condition being treated and the treatment in question. This appears to accord with Russo and Williamson (2011b, 63), who think that whether a plausible mechanism or confirmed mechanism is needed for a causal claim in medicine to be established 'depends on the epistemic context.' Since they also suggest (Russo and Williamson 2011b, 65) that purpose plays an important role in determining the contours of a causal epistemology, at heart this seems to countenance a view whereby sensitivity to actual medical practice is important to not only understand but to justify making effectiveness claims. Understanding how the RWT and EBM+ have been used to establish or confirm treatment effectiveness might thus help clarify the limits of the RWT and EBM+, improve understanding of treatment effectiveness itself, and elucidate their interrelationships. This is particularly important because what it means for a treatment to be effective in the first place is unclear. Accordingly, I first canvas several approaches in medicine to establishing or confirming treatment effectiveness, before explicating my own approach. I then use this as a way of understanding what role the RWT and EBM+ can and should play in helping to establish or confirm a treatment as being effective. Before concluding, I reinterpret Hill's viewpoints-which have been almost exclusively applied to disease aetiology-to make them more applicable for helping to establish or confirm treatment effectiveness.

Standard Approaches to Treatment Effectiveness
There are few philosophical accounts of treatment effectiveness (some include Ashcroft [2002] and Stegenga [2015]). 5 In medicine, EBM, through its hierarchies of methods of generating evidence, specifies the sorts of evidence that warrant a treatment as being considered effective, and how such claims should be established. One common method of ascertaining effectiveness-and what could be considered the standard EBM approach -is called simple extrapolation. This method of generalising RCT results involves transporting effect sizes from RCTs to a target population, unless there is a compelling reason not to (Post, de Beer, and Guyatt 2013). This approach, however, is problematic. Some problems include patients in clinical practice possibly being markedly different from the participants in RCTs, publication bias (meaning that the studies that are published may give a distorted view of a treatment's effects in comparison with any unpublished studies), inattention to the mechanisms by which treatments work, lack of mathematical justification, and the fact that some individuals respond differently than others to the same intervention, even if they are all representative of the study population in clinically relevant aspects (Stegenga 2015;Fuller 2021;Tresker forthcoming). Moreover, regardless of the type of study, study results do not represent or automatically determine the effectiveness of the treatment studied. Rather, as I have argued elsewhere, they serve as evidence for inferences to effectiveness claims (Tresker forthcoming). This distinction may seem obvious but it can be easily overlooked when studies, especially RCTs, are viewed as the means and the standard by which the effectiveness of treatments is established. Yet an account of effectiveness that merely or mostly looks at how a treatment performs in a given trial may not accurately reflect how the treatment will perform in a specific individual in clinical practice. 6 Even if RCT results could be simply transferred to a patient in clinical practice, effect sizes still need to be put in a form comprehensible to patients to serve as the basis for a treatment decision. One common measure of effectiveness used to counsel patients and communicate among clinicians the benefit of a treatment is the number needed to treat (NNT). The NNT can be seen as exemplifying a method of utilising effect sizes since it is the reciprocal of the absolute risk reduction, and therefore is derived directly from an RCT's effect size. The NNT is the number of patients that would be needed to be treated to avoid a given outcome, such as a heart attack. Prima facie, this appears like a useful metric to assess effectiveness and guide patients in making treatment decisions. However, the NNT is prone to being misunderstood, such as if it is not made clear to patients that it is not an objective measure of effectiveness but is instead specific to a particular comparison from an RCT. As Stang and colleagues (2010, 820) warn, 'Without stating the direction of the effect, the alternative treatment, the treatment period, and the follow-up period, information in terms of NNTs is uninterpretable.' Even with this information, however, NNTs may provide poor measures of effectiveness for patients in clinical practice because NNTs cannot easily be used to compare treatment options. To validly do this the different treatments would need to have been tested in similar populations, diseases (including similar disease stages), and timeframes, and use the same outcomes and comparator treatments (McAlister 2008). In the absence of this, decisions based on NNTs could be misleading.
Other approaches that rely on RCTs to assess the potential benefits of treatments include effectiveness scales, such as that used by the National Comprehensive Cancer Network (2021): (1) (Palliative only): Provides symptomatic benefit only (2) (Minimally effective): No, or unknown impact on survival, but sometimes provides control of disease (3) (Moderately effective): Modest impact on survival, but often provides control of disease (4) (Very effective): Cure unlikely but sometimes provides long-term survival advantage (5) (Highly effective): Cure likely and often provides long-term survival advantage While this scale could be useful under certain circumstances, for medical conditions in which outcomes are markedly different other scales would be needed. One general shortcoming of effectiveness scales like this is that in addition to mostly relying on RCTs at the expense of other types of evidence, such scales require agreement among its users in terms of which outcomes are most representative of effectiveness. Yet such scales are typically created and used by regulatory authorities or other organisations without incorporating the patient's voice or being able to be adjusted by a patient or clinician in clinical practice. They thus may offer a limited sense of effectiveness by not reflecting different standards of effectiveness.
Evidence interpreted in the light of tools such as effectiveness scales and the NNT may be easy to digest by policy makers, clinicians, or other stakeholders but offer rather blunt and sometimes fallacious inferences to-or (by omitting non-RCT evidence) impoverished views of-a treatment's effectiveness for a patient in clinical practice. Reliance on unitary, simplistic measures of effectiveness is therefore unrealistic, potentially misleading, and frustrates a comprehensive inferential account needed for a treatment effectiveness claim. This is not to impugn their value, however. Such approaches can serve important roles in understanding and simplifying medical evidence. But they are fundamentally simplifications that do not account for the complexity and context of effectiveness claims and the purposes for which they exist. It is doubtful that a single heuristic or tool could capture the widely different situations inherent in what role effectiveness plays for individual patients in clinical practice. Even a complex epistemological framework like EBM is unsuited for this end, partly because of its reliance on hierarchies of methods of generating evidence (Stegenga 2014).
Effectiveness may nonetheless seem to have a unitary meaning, especially for drugs, where the background assumptions underwriting effectiveness attributions are implicitly understood by stakeholders. This can happen for effectiveness claims arising in the context of study publications of RCTs, whereby an EBM framework is implied. If a doctor says that a Food and Drug Administration (FDA)-approved drug is effective for treating diabetes, it will be understood that-on the basis of FDA approval proceduresthe drug has shown superiority to placebo in typically at least two RCTs, with an acceptable adverse event profile. Yet what this conception of 'effectiveness' means can be highly subject to factors irrelevant to or even misleading with respect to what a particular patient in clinical practice might need to know. For example, special-interest groups campaigned for the approval of flibanserin (a drug used to treat low sexual desire in women), despite the FDA previously having deemed it to have an unfavourable benefit-risk ratio. The evidence of effectiveness of this drug thus took on different meanings if evaluated in a medical or policy context-here the medical context reflecting a weighing of clinical benefits with clinical risks, and the policy context incorporating economic and normative considerations. This illustrates how the standards for evidence for influencing policy can shift depending on context (La Caze and Colyvan 2017, 11), with possible distorting effects on effectiveness claims for specific patients.
Treatment effectiveness claims are often conveyed via trialists' statements in journal articles' discussion and conclusions sections. Such effectiveness statements reflect unspoken norms of effectiveness attribution among members of the clinical and research community, regulatory guidance on what constitutes a result of clinical relevance, and/or the personal views of the articles' authors. Such statements can be wrong, however. Relying on authors' declarations of effectiveness in trials' publications, or interpretations provided in secondary sources, to support effectiveness claims is therefore of dubious utility and can lead to faulty decisions if the standard of effectiveness used by any given patient is markedly different.
Many of the differences underlying various standards of effectiveness may derive from different assessments of the relative values of different clinical-trial outcomes. No clinical-trial outcome is universally applicable to all treatments and medical conditions; outcomes are domain-specific and depend on the goals, stakes, and features of the treatment in question. Although assessing whether an outcome has been obtained is usually relatively straightforward, diverse opinions exist as to which thresholds reflect clinical significance and which outcomes are important to patients in the first place (Nowinski, Miller, and Cella 2017;Dankers et al. 2021). For example, patients-who typically lie downstream of clinical trials-may only have their voices heard in retrospect, after trial results have already been reported and interpreted, and then only in an ancillary way, deciding whether to take a treatment that has already been deemed effective according to criteria and methods they had no input into and may not understand. If patients can only state their preferences 'after-the-fact' on predetermined outcomes, they are not being afforded the opportunity to more integrally contribute to effectiveness evaluations.
For at least these reasons, a better approach to effectiveness is needed.

An Account of Treatment Effectiveness
Effectiveness claims, I have previously suggested (Tresker forthcoming), are predictions based on arguments for how target patients/populations will fare after taking a treatment in a given set of circumstances, compared with the counterfactual of not taking that treatment. Effectiveness claims involve the specific situation of an individual (such as a patient in clinical practice deciding what treatment to take) or population (such as regarding an effectiveness claim meant to apply to a certain group of people). 7 For example, the claim that dapagliflozin is effective for the treatment of type 2 diabetes could mean several things: a patient taking the drug for 6 weeks will likely have a substantial reduction in glycated haemoglobin levels; taking the drug for a year could do the same and decrease the patient's risk of hospitalisation for heart failure; or any number of other predictions. On a population level it could mean that if certain types of patients (say, those with a glycated haemoglobin level of 7-10% and inadequate glycaemic control with metformin) take the drug for 24 weeks they can expect, say, a 0.7% decrease in glycated haemoglobin levels (as Bailey et al. [2010] report). What would make the effectiveness claim reasonable to accept in such cases is reason to think these standards are therapeutically relevant with respect to the target condition, plus evidence that the treatment yields improvement according to these standards. Importantly, there is no way to prove the truth of an effectiveness claim's prediction, since even a person taking a treatment and improving clinically does not vindicate the particular effectiveness claim for that treatment. If effectiveness claims could be vindicated by simply observing an individual's response to a treatment, then there would be no need for RCTs or large observational studies. Trying a treatment out is subject to far too many biases to in general be a reliable means of establishing valid treatment effectiveness claims. 8 A person getting better after taking a treatment deemed effective only vindicates the claim insofar as it can contribute as evidence to other treatment effectiveness claims for the same treatment, and then only when interpreted in light of the many biases present (such as those articulated by Lilienfeld et al. [2014]), whereby there could be multiple competing reasons besides the treatment itself for why a person taking the treatment could have improved. For example, their response could be due to factors independent of the treatment, such as regression to the mean or spontaneous fluctuations in their condition. RCTs are useful insofar as they can help eliminate these alternative explanations on a population level via the assumption that the average response of these factors in both the treatment and the comparison group is approximately equal. The existence of population-level and individual-level effectiveness claims raises the possibility that the two could conflict. For example, as indicated by the 'prevention paradox' (Rose 2008), whereby, as John (2011, 250) relates, 'the most effective public health strategies have little benefit for the vast majority of those they affect,' treatments such as statins could be considered effective on a population level but ineffective for most individual patients in clinical practice. Rather than undermine my account, I suggest that this possibility represents the fact that: (1) patients, even with similar clinically relevant characteristics (e.g. age, the presence of certain diseases), respond to treatments differently, and (2) the goals of effectiveness determinations on a population versus individual level often differ and should not be expected to coincide. Even solely on an individual level, though, a clinician could deem a treatment differentially effective for different patients. This reflects the heterogeneity of treatment effect commonly encountered in medicine. What then of different clinicians (ceteris paribus) deeming a treatment differentially effective for the same patient? Instead of rendering my account of effectiveness unacceptably fickle, this represents the common occurrence in medicine of different clinicians arriving at different conclusions (and why patients sometimes seek 'a second opinion'). This does not mean the different clinicians are all correct about the treatment's effectiveness. Some effectiveness claims can be wrong. Accordingly, understanding why clinicians reach different conclusions offers opportunities for elucidating the reasoning and argumentation that are important for justifying treatment effectiveness claims.
A good account of treatment effectiveness, I stipulate, should possess at least some of the following features: effectiveness determinations should be predictive; can benefit from, but need not always require, a plurality of evidence; should ideally reflect the relevant stakeholders' standard of effectiveness; and should facilitate the formulation of effectiveness claims that can be easily communicated and relied upon for treatment decisions. That effectiveness on my account inherently represents a prediction based on available evidence predisposes to being responsive to a plurality of evidence (including mechanistic evidence) and not simply the outcome of a clinical trial. By not privileging the results of RCTs as the arbiter of treatment effectiveness, my account of effectiveness as a prediction does not encounter some of the problems (as mentioned in the previous section) that beset the EBM approach that simply transports an effect size from a clinical trial to a patient in clinical practice.
An RCT's outcome measure, indeed, may be inappropriate as the standard of effectiveness for a patient in clinical practice (Tresker forthcoming). Patients themselves could help choose a standard of effectiveness consistent with their values, preferences, and goals. Effectiveness on my account is responsive to which outcomes patients consider to be indicative of effectiveness because such outcomes can be the patient-informed standard of effectiveness for the treatment. Effectiveness claims on my account can therefore possibly be easily communicated to patients in terms they understand and can take action on. My account consequently considers important epistemic constraints of explanations of treatment effectiveness claims, exemplified by challenges some people have with numeracy when interpreting health information (Peters et al. 2007), the limited ability of some people to understand personalised medical risk calculators (Damman et al. 2017), and the fact that absolute and relative risks are perceived differently, by both doctors and patients (Perneger and Agoritsas 2011). Since a treatment effectiveness claim can involve a qualitative prediction (such as, 'Will this treatment make me feel better?'), this potentially makes it easier than the EBM approach for patients to understand treatment effectiveness (though it should not downplay the usefulness of risk calculators and other ways of communicating risk to patients). The importance of this will hopefully become apparent in the context of inductive risk, which I discuss next.
The argument from inductive risk, early on developed by Rudner (1953) and Hempel (1965), and more recently by Douglas (2000Douglas ( , 2009, claims that non-epistemic values 'help determine how serious it would be to make a false positive or false negative error, and thus how much evidence should be demanded in order to accept a hypothesis' (Elliott and Richards 2017, 262). Inductive risk has mostly been discussed in the context of climatology and toxicology but, with the exception of its relationship to drug regulation and trial design (e.g. Plutynski 2017; Stegenga 2017; Bluhm 2017b), remains unexamined in the context of an account of treatment effectiveness. As considerations of science-policy communication differ among these fields, it is important to understand how inductive risk applies to treatment effectiveness, particularly regarding individuallevel effectiveness claims.
A common belief in the inductive risk literature, as nicely summarised by Resnik (2017, 64), is that 'scientists are morally obligated to consider the ethical or social consequences of accepting or rejecting hypotheses when they are asked to provide advice to policymakers and the public.' I, therefore, suggest that in the context of a clinical encounter patients fulfil the role analogous to the 'policymaker' and clinicians fulfil the role analogous to the 'scientist.' Treatment effectiveness claims, in my view, are not decisions by proxy as to whether patients should take treatments, but rather predictions as to what effects treatments might have if specific patients do take them. While it could be argued that the role of the clinician is to simply present information to patients about treatments, such as the results of clinical trials and how treatments work, patients also typically want to know whether a treatment will achieve a certain outcome for them -i.e. patients seek predictions to inform treatment decision-making.
The argument from inductive risk applied to treatment effectiveness could be viewed as the idea that non-epistemic (social, ethical) values determine how much and what type of evidence is needed for treatment effectiveness claims, which depend on the non-epistemic consequences of the risks of error-as Rudner (1953, 2) put it, 'how serious a mistake would be.' A treatment that were it to be deemed effective when really not (i.e. a false positive) might, for example, have serious consequences, such as patients being likely to take that treatment to the exclusion of other safer or less expensive alternatives. At the same time, the risk of false positives should be balanced against the risk of false negatives (missing out on calling a treatment effective that really is). To take an example I discuss briefly in the next section, art therapy could be deemed either effective or ineffective for the same patient depending on what the consequences of receiving art therapy would be were it in fact (in)effective. For example, assume the risks of adverse effects are low from a particular type of art therapy for a patient seeking to diminish their anxiety. Further assume the therapy is inexpensive, not inconvenient to the patient, and that there are no potentially more effective treatments available that would be foregone by engaging in the art therapy. The consequences of a false positive are therefore low-if the treatment doesn't work not much is lost. The patient's clinician, however, is faced with conflicting evidence on this art therapy's effectivenesswhile some RCTs indicate benefit, others do not, and the controls used in all the studies consisted of usual care or a no treatment wait-list (limitations of which are well known; see Tresker forthcoming). The clinician nonetheless determines, based also on mechanistic rationale, that if this patient engages in the therapy they will be likely to experience a moderate degree of diminished anxiety in comparison with the counterfactual of not engaging in the therapy. Alternatively, the clinician could make the less certain but equally warranted claim that the patient might experience the same outcome. The clinician could justifiably anticipate that the 'likely' claim will result in the patient electing to receive the treatment, whereas the 'might' claim will result in the patient refusing the treatment. There may be no way under the circumstances that the clinician can abdicate making any treatment effectiveness claim at all, because even if the clinician says nothing when asked if the treatment is effective, their reticence could still be interpreted by the patient as a 'no' or 'not likely.' The clinician making the more certain claim does so on the basis that the consequences of conveying that claim would be minor, a decision clearly influenced by non-epistemic values. Treatment effectiveness claims are therefore subject to inductive risk.
There are various points along the evidence chain for hypotheses (such as treatment effectiveness claims) at which inductive risk decisions are present. These include which level of statistical significance to choose, how evidence is characterised, and how data are interpreted (Douglas 2000); the choice of model organism (Wilholt 2009); and numerous methodological aspects of trials, such as the choice of comparator (placebo or active comparator), outcome measure, and trial length (Plutynski 2017;Stegenga 2017;Bluhm 2017b). Deciding on the evidence sufficient to establish a standard of effectiveness as being met is also subject to inductive risk, a point made starkly clear by Stegenga (2017) in discussing the effectiveness of antidepressants in the face of significant publication bias. In summary, that non-epistemic concerns play a defining role in determining whether effectiveness is established for a particular treatment can be illustrated by the fact that inductive risk is involved in: (1) some of the decisions made by researchers in generating evidence on the effectiveness of treatments, even before a stakeholder (such as a clinician) has to decide whether to accept or reject a treatment effectiveness claim, (2) the decision on the standard of effectiveness, (3) the decision on what evidence and how much evidence is sufficient to establish that standard as being met, and (4) the decision on how the effectiveness claim should be conveyed (in terms of its formulation and any of its nonverbal features, including the clinician's demeanour and aspects of the environment in which it is conveyed that might affect how the claim is interpreted).
One potential objection to applying the argument from inductive risk to treatment effectiveness could be that when researchers, clinicians, regulatory authorities, and other stakeholders accept or reject a hypothesis about whether a treatment is effective they need not do so tout court; instead, they can make the less-certain claim that a treatment is, for example, hardly/somewhat/moderately/very/extremely effective, and qualify any claims of effectiveness they make with varying degrees of probability-possibly, probably, definitely, etc. Indeed, as Steele (2012) points out, scientists' beliefs about uncertainty often need to be converted into predefined evidence scales (a conversion which at least implicitly involves value judgments). Conveying degrees and likelihoods of effectiveness are thus important communicative devices, reflections of the spectrum of effectiveness upon which treatments fall, and emblematic of the nuances captured by #4 of the previous paragraph. Thus, even if a hypothesis involves 'hedged' effectiveness claims it is still subject to inductive risk.
Because of the effects a communicated effectiveness claim can have on its recipients, such as licensing action (taking a treatment, regulating it, reimbursing it, etc.), I suggest that clinicians conveying treatment effectiveness claims to patients do more than merely accept hypotheses. Instead, treatment effectiveness claims should be viewed as assertions, a position that should not be too difficult to accept given that others (e.g. John 2015, 2019; Franco 2017) even hold this to be true of non-treatment-related hypotheses in general. Assertions better reflect the social and pragmatic dimensions of scientific practice (Franco 2017, 167-168). Accordingly, an important role for clinicians is thinking through how the treatment effectiveness claims they convey to patients will affect those patients.
Closely apposing effectiveness claims with likely actions (e.g. taking treatments 9 ) might encourage stakeholders to be more circumspect in their effectiveness claims and to be more explicit in articulating which non-epistemic values inform such claims. Douglas (2009, 155) suggests: Scientists should be clear about why they make the judgments they do, why they find evidence sufficiently convincing or not, and whether the reasons are based on perceived flaws in the evidence or concerns about the consequences of error.
As noted by Elliott and Richards (2017, 272), '[i]t is impractical, however, to think that scientists could elucidate all the scientific judgments associated with a particular line of inquiry and clarify the roles that values played in making each of them.' In the strict time limitations of a typical clinic visit, clinicians may only be able to convey to patients the most salient value considerations that influenced their inductive risk decisions in formulating a treatment effectiveness claim. Despite the infeasibility of such analyses and their disclosure to patients in clinical practice, such analyses could be valuable additions to medical education via case studies of individual-patient treatment effectiveness claims. On a population level, making explicit and defending the values involved in a treatment effectiveness claim could involve substantive argumentation that could be conveyed through narrative review articles or specialised investigations by regulatory authorities or other stakeholders. Such work could adduce a wide range of evidence in the context of an overriding argument in the specific situation in which the effectiveness determination needs to be made. Accordingly, I have refrained in this article from describing what is required for a treatment to be considered effective, and instead highlighted general constraints on effectiveness determinations. Towards this end, the close interconnection of effectiveness claims with how they are conveyed suggests that treatment effectiveness claims should be seen as part research and clinical science, and part medical communications. Plutynski (2017, 165) nicely summarises the issues at stake (with respect to screening, but also applicable to treatments): 'At issue here then are fundamentally philosophical disagreements about justice, harm, autonomy, and beneficence, and the role of the physician with respect to both individual patients and the patient population more generally.' Accordingly, the ways in which treatment effectiveness claims raise questions of harm and of paternalism (e.g. with respect to how clinicians modify their inductive risk calculus in relation to their presumptions about their patients' psychology and anticipated decisions) offer an opportunity for the exposition of such cases to expand beyond the ambit of medical expertise to also involve patients as well as experts in non-medical fields, including philosophy, sociology, psychology, and economics.

How do the RWT and EBM+ Establish Effectiveness?
To provide an example of the RWT/EBM+'s approach to establishing treatment effectiveness I next discuss Gillies's (2019a) analysis of the effectiveness of acupuncture. Relying on a single, secondary source from approximately two decades ago (Kaptchuk 2002), Gillies (2019a, 174-175) concludes: 'It seems that RCTs definitely show that acupuncture is effective for adult postoperative and chemotherapy nausea, and for acute dental pain.' This, for Gillies, provides the difference-making evidence. The mechanistic evidence he adduces not from Traditional Chinese Medicine (TCM), which he finds speculative and unscientific, but from contemporary scientific medicine. I concur with Gillies that plausible mechanisms exist for potential pain-relieving effects of acupuncture treatment. However, he does not cite or discuss any evidence that these exist for nausea. 10 Moreover, to say that enough (and presumably of high-enough quality) difference-making evidence exists for the effectiveness of acupuncture to be established (for the mentioned indications or any others) is perhaps premature, because it neglects discussion of the appropriate standard of effectiveness for acupuncture and a critical analysis of the available studies. For example, regarding acupuncture for postoperative and chemotherapy nausea, Kaptchuk (2002) cites a 1999 Cochrane review meta-analysis by Lee et al. which included 19 studies, 10 of which involved a no-treatment control group. This raises the question as to whether expectations, the treatment ritual, mere participation in a clinical trial, and/or response bias, as opposed to any effect of the needling, accounted for the clinical benefit. If perhaps Gillies means for the acupuncture treatment package (i.e. needling delivered by an empathetic practitioner with a plausible rationale and other trappings of a healing clinical environment) to be considered effective and not acupuncture itself (e.g. needling delivered to a person with postoperative nausea somehow unaware they are being given the treatment) then it might be fair to say that one of these treatments is effective and the other is not. However, to claim that the treatment package is effective would require a standard of effectiveness for such types of treatment packages. For effectiveness, superior performance to a no-treatment control group is a low bar to meet.
Perhaps then an updated, more rigorously controlled meta-analysis (Lee and Fan 2009) can provide the difference-making evidence to substantiate Gillies's assertion. This meta-analysis compared acupuncture in the TCM-sanctioned point to the illusion of this point or to a non-TCM-sanctioned point. There was a relative risk of 0.71 (95% confidence interval 0.61-0.83), with moderate heterogeneity (I 2 = 60%). Whether this should be considered effective has been challenged (Colquhoun and Novella 2013). Also, both meta-analyses included modalities other than acupuncture, such as acupressure, oppugning how representative the results are of acupuncture per se. An even morerecent updated meta-analysis (Lee, Chan, and Fan 2015) included the authors' conclusions as indicating there being low-quality evidence supporting the claim.
Since numerous studies have shown that the effects of acupuncture do not necessarily depend on the location of needle placement or whether they are inserted (reviewed in Dincer and Linde 2003), it is rebuttable whether the 'acupuncture' Gillies begins talking about (i.e. the one explained by TCM theory) can be considered the same treatment as 'acupuncture' explained according to contemporary scientific medicine. This is because the latter could involve the insertion of needles at non-TCM points, or the absence of insertion at all, whereas under TCM theory needles inserted at non-TCM points do not constitute acupuncture.
I am not saying that some form of acupuncture may not be effective for some conditions. I am saying, however, that too easily accepting the pronouncements of experts or the medical community (ignoring the existence of multiple communities, disputing voices, or countervailing rationales) can lead one to arrive at a simplistic, incorrect conclusion regarding effectiveness. 11 Moreover, it can lead to omitting key considerations that are important in evaluating the effectiveness of medical treatments, such as the legitimacy of the placebo comparison; also, focusing only on population-level effectiveness risks neglecting individual-level effectiveness.
Nonetheless, the RWT/EBM+ can be useful for disclaiming some effectiveness claims. For treatments such as homeopathy and retroactive intercessory prayer, for example, the RWT can serve as a disqualifier because of the absence of a plausible mechanism or the postulation of a highly implausible one, despite the presence of RCT difference-making evidence (Russo and Williamson 2011a). These are cases of what Jerkert (2015) calls 'negative mechanistic reasoning.' 12 In fact, Jerkert (2015) criticises the Lee and Fan (2009) Cochrane review for not taking into account any mechanistic considerations. The correct application of the RWT in such cases, however, does not necessarily extend support of it or EBM+ to more nuanced cases. These nuanced cases do not simply require a disqualification of an effectiveness claim but instead require positive warrant supporting one. Cases in which intuitions may clash, such as acupuncture, illustrate how simply knowing that there is some putative evidence of difference-making and mechanisms is insufficient to establish treatment effectiveness. This is because the roles such evidence plays can widely vary and be interpreted in multiple ways. The RWT does not specify what type or how much mechanistic evidence is required for establishing causal claims in medicine. Does mechanistic evidence always need to involve biological mechanisms, or would solely psychological or social mechanistic evidence count? Regarding the amount of evidence needed, Williamson (2021) states that the RWT does not require establishing the details of a mechanism, while other explanations of the RWT (e.g. Clarke et al. 2014, 357;Parkkinen et al. 2018, 94) emphasise the quality of the evidence. In this regard, Machamer, Darden, and Craver's (2000) influential distinction between mechanism sketches, mechanistic schemata, and full mechanistic models might be a useful way of distinguishing between the extent of needed mechanistic evidence. Still, a way to decide upon a choice would be needed. Adopting the RWT and EBM+ is not a way to solve how treatment effectiveness can be established, but might better be interpreted as a framework that emphasises the importance of differencemaking, and, in particular, mechanistic evidence, for establishing treatment effectiveness.
An ongoing theme underlying the RWT and EBM+ is engaging with the medical community. For example, the charts used by EBM+  to determine the status of mechanistic claims span the gamut from 'ruled out' to 'speculative' to 'established,' and therefore are judgments based on expert opinion. Illari (2011, 155) also thinks that 'a certain amount of consensus' is required for assessing the strength of certain forms of mechanistic evidence. The reliance by the RWT and EBM+'s proponents (e.g. Clarke et al. 2014) on the International Agency for Research on Cancer and other sources in medicine, such as the 'medical community' (e.g. Gillies 2019a, 148), to underwrite their application of the RWT are ways in which they have engaged with the medical community and apparently why they have accepted these agencies' claims as authoritative. Given the plethora of causal claims that could be made in medicine it would be unrealistic to specify criteria by which each could be validly satisfied, so EBM+'s reliance on community standards and expert judgment is sensible in this respect. It is important to not forget, however, the various ways in which a scientific community could be wrong or in epistemic flux. 13 This is illustrated by the acupuncture example, as well as the one discussed next.
Take an NPT-art therapy-as an example of a treatment heretofore unanalysed by the RWT or EBM+. The existence of some difference-making evidence (Uttley et al. 2015) combined with what might be considered good mechanistic evidence (Czamanski-Cohen and Weihs 2016) presents a situation whereby art therapy would appear to be effective (on a population level) according to the RWT. However, a pragmatic RCT (the MATISSE trial) showed art therapy, for one indication at least, to not be effective (Crawford et al. 2012). Which conclusion is to be believed? 14 Because of space, I will not perform a comprehensive analysis of the effectiveness of art therapy or how it can or should be evaluated, but as judged by critical editorials (Wood 2013;Holttum and Huet 2014) on whether the MATISSE trial really showed art therapy to not be effective, a valid and comprehensive assessment of art therapy is far more complex and nuanced than anything the RWT and EBM+ appear capable of. For example, a thorough evaluation that goes beyond the RWT might look to the legitimacy of the placebo controls employed in the existing RCTs, 15 and a deep investigation of all the biases inherent in the existing clinical trials, analysing such evidence in light of the existing mechanistic evidence and skilfully synthesised in a way that a good narrative review article does. It is uncertain that the Is your policy really evidence-based? tool offered by EBM+ , is capable of this sort of analysis, even when combined with expert judgment and GRADE (as EBM+ [Parkkinen et al. 2018, 27] suggests could be done to assess a correlation claim). For example, neither EBM+ nor GRADE facilitates an in-depth analysis of the legitimacy of placebo groups in terms of whether they offer valid comparisons to the experimental treatment. This, however, is crucial considering how widely a treatment's effectiveness can be underestimated or overestimated depending upon the comparison (Howick and Hoffmann 2018). GRADE is also beset with other limitations (Rehfuess and Akl 2013) that it would be helpful for EBM+ to clarify if they expect GRADE to be used alongside evaluation of mechanistic evidence. Moreover, as alluded to in the previous section, a richer understanding of art therapy's effectiveness should consider how its effectiveness differs on a population versus individual level, and how it is affected by decisions made under inductive risk.
According to Holman (2019, 4379), EBM+ is an example of friction-free epistemology in the sense that it does not take economic forces closely into consideration when evaluating evidence; he writes, 'What such an approach is not well-equipped to do is make any recommendations about what inferences should be made given the current state of evidence or to make any policy recommendations about how the medical community should reform its institutions or epistemological priorities.' Although Gillies (2019b) responds to some of Holman's criticisms of EBM+, regardless of how well EBM+ takes economic forces into consideration EBM+ is useful insofar as it helps evaluate evidence. However, to help establish effectiveness EBM+ should be viewed in a framework of inductive risk and supplemented with guidance on navigating and establishing appropriate standards of effectiveness.
Overall, using the RWT and EBM+ as a general means of establishing treatment effectiveness stretches them beyond what they were formulated for or are capable of. Cases (i.e. acupuncture) where some of their proponents reach incorrect conclusions about effectiveness claims that have already been putatively established illustrate this. Granted, EBM+ is positioned as an aid to judgment, not a replacement (Parkkinen et al. 2018, 7). EBM+ is also an improvement over EBM in that it abandons hierarchies of methods of generating evidence, and lends more evidential weight to mechanistic evidence. Nonetheless, it is not entirely clear which other of EBM's commitments EBM+ accepts and rejects. 16 Rejection of EBM's hierarchies and increased consideration of mechanistic evidence, while welcome improvements, do not necessarily produce a system or means by which the effectiveness of treatments can be established or confirmed. Like EBM, EBM+ still involves rule-based reasoning, supplemented with guidance for mechanistic evidence and how to integrate this with difference-making evidence. Essentially, EBM+ is akin to an evidence assessment tool and is thus better seen as a method of assessing evidence as opposed to determining or confirming whether a treatment's effectiveness should be considered established. Accordingly, the tools it proposes should be evaluated empirically and assessed for inter-rater and inter-tool reliability, as well as their validity and usability in terms of the applicability of items and how items should be weighted. 17 If EBM+ and the RWT cannot establish or confirm effectiveness, this raises the question whether Hill's viewpoints can. I address this next. Russo and Williamson (2011a, 572) claim that their interpretation of causality 'provides a conceptual framework to underpin causal assessment as advised by Bradford Hill.' Hill's nine viewpoints (which are best seen not as causal criteria but as considerations to inform evaluation of causal claims) remain influential in philosophy and medicine. For example, Reiss (2015), in his pragmatist theory of evidence, implicitly seems to utilise at least some of Hill's viewpoints. Stegenga (2011, 497) writes (in comparison with meta-analysis), '-the plurality of reasoning strategies appealed to by the epidemiologist Sir Bradford Hill-is a superior strategy for assessing a large volume and diversity of evidence. ' Howick and colleagues (2009) offer a revision and categorisation of Hill's viewpoints and then apply them to examples of putative causation when RCT evidence is unavailable. Application of Hill's viewpoints to diverse problems in medicine by clinicians and researchers is also popular (e.g. Kotsovilis and Slim 2012;Armon 2018;Smith et al. 2018). Even the GRADE approach to evidence evaluation, which is firmly rooted in EBM and is often used to assess treatment effectiveness by relying on a hierarchy of methods of generating evidence and not a plurality of evidence as championed by Hill, has been argued as reflecting Hill's viewpoints (Schünemann et al. 2011). Indeed, common approaches to causal inference in philosophy and medicine, including GRADE, have been shown to overlap with Hill's viewpoints (Shimonovich et al. 2021).

Reinterpretation of Hill's Viewpoints for Treatment Effectiveness
Hill's viewpoints were developed with disease aetiology in mind (Hill 1965) and have mainly been applied to the establishment of disease causation. 18 Based on an informal search on PubMed and PhilPapers.org, they appear to have been only infrequently applied to medical treatment, and there to the effectiveness of drugs-less so (if at all as far as I can tell) to the effectiveness of NPTs. Hill's viewpoints have also typically been applied to situations in which evidence from observational studies, but not RCTs, is available. How they apply to evaluating a treatment for effectiveness, therefore, does not appear to be well known. I address this lacuna by charitably reinterpreting Hill's viewpoints to see how they can help evaluate evidence for treatment effectiveness claims. Because Hill's viewpoints were not designed for treatment effectiveness, my approach is revisionary (and preliminary), with no pretence at capturing Hill's original meanings. My descriptions of some viewpoints may appear applicable to other viewpoints. This could reflect that there are no clear boundaries between some of the viewpoints.

Strength of Association
A strong association between a treatment and an outcome could, in combination with Hill's other viewpoints, indicate causality. Williamson (2019) rightly notes that large effect sizes observed in observational studies combined with excellent evidence of mechanisms offer sufficient evidence for effectiveness to be established, even in the absence of RCTs. However, strong causal associations can also entail weak effects, and strong correlations can be spurious. Knowing the causal structure of a treatment can make a strong observational association between that treatment and an effect more plausible as an indicator of causality. Here mechanistic knowledge can be useful. An example would be a drug that suddenly shrunk a person's tumour far beyond what could be expected by chance occurrence or the natural history of the disease (see also the 'coherence' and 'plausibility' viewpoints).

Consistency
According to Hill (1965), consistency involves consistent findings or replication across different locations, populations, and methods. While this could be a useful indicator of a treatment being effective, this could also reflect an artefact in how treatments are studied or administered. Placebo effects, for example, are commonly encountered with most treatments, such as via conditioned responses and by being secondary to an empathetic clinician who provides support, encouragement, and expectation that a treatment will work. A benefit of the consistency viewpoint might instead rest with how it brings attention to the importance of treatments being defined consistently (i.e. between study and clinical settings). For example, a treatment should be consistently described as containing, or not containing, whatever aspects of the treatment produces placebo effects, depending upon whether the treatment's standard of effectiveness includes that it should or should not work by placebo effects. I discuss this further under the 'plausibility' viewpoint.

Specificity
Instead of meaning (as originally formulated by Hill) that an exposure causes only one disease, specificity could be reinterpreted to mean whether a treatment works only for a specific disease and not others. Some treatments for inborn errors of metabolism or rare diseases, for example, could be seen as exemplifying this viewpoint. However, there are many treatments, such as broad-spectrum antibiotics, that are effective for many conditions.
The viewpoint could also be reinterpreted in terms of causal specificity as explicated by Woodward (2010), whereby specificity represents the degree to which causes exert fine-grained influence over their effects. However, a relative lack of causal specificity can actually be a boon to successful intervention, with some treatments (such as HIV drugs) involving less specific causes being more highly valued clinically (Neal 2019).
How useful this viewpoint is to establishing or confirming treatment effectiveness therefore currently seems unclear.

Temporality
Since for medical treatment it is typically clear which comes first-the treatment or the effects-rather than serving any useful purpose for establishing or informing a claim of effectiveness, this viewpoint merely serves as the precondition for a causal attribution. To be useful, temporality can be reinterpreted as whether the time course of a treatment's effects is consistent with mechanistic knowledge, or can be used to establish what the putative mechanisms might be. For example, in the EMPA-REG OUTCOME trial of empagliflozin treatment in patients with type 2 diabetes (Zinman et al. 2015), the benefit on cardiovascular mortality was evident as early as 3 months, whereas previous large trials of glucose-lowering drugs had not shown a reduction in cardiovascular events until about 10 years of treatment, which argues against glycaemic control as being a significant mechanism by which empagliflozin benefits the heart.

Biological Gradient, or Dose-response
This may apply to some drugs but for many does not. For example, there was no difference between empagliflozin 10 mg and 25 mg on outcome measures in EMPA-REG OUTCOME (Zinman et al. 2015). For some NPTs, such as surgery, this viewpoint is largely irrelevant. Similarly, for other NPTs, such as psychotherapy, the phenomenon of responsiveness, whereby the therapist and client adjust the number of sessions according to the client's need (Stiles et al. 2008;Stiles 2009), makes it very difficult to determine a consistent dose-response relationship.

Plausibility
This viewpoint can be seen to reflect whether plausible mechanistic evidence exists. Although this has traditionally been considered to involve mechanisms explainable on a biological level, given the diverse levels on which treatments work, it could also involve psychological and social mechanisms.
In terms of my account of treatment effectiveness, mechanisms are important for establishing effectiveness because of how they can contribute to valid treatment effectiveness claims by supporting the argument for a prediction about a treatment's effects; moreover, they can serve an important communicative function when clinicians explain treatments to patients, thus contributing to how a patient understands a treatment effectiveness claim.
When interpreting epidemiological studies, Evidence of mechanisms can help rule in or out various explanations of a correlation. For example, it can help to determine the direction of causation, which variables are potential confounders, whether a treatment regime [sic] is likely to lead to performance bias, and whether measured variables are likely to exhibit temporal trends. (Parkkinen et al. 2018, 16) Mechanisms can also facilitate the synthesis of diverse kinds of evidence, generalisation of study results, what outcome measures to use in clinical trials, and other trial-design decisions (Clarke et al. 2014;La Caze and Colyvan 2017;Aronson et al. 2018). As mentioned for the 'consistency' viewpoint, when stakeholders view certain mechanisms as being crucial to how a treatment works, mechanistic knowledge can aid in formulating an unambiguous description of the treatment-a description that is used to define that treatment with respect to how the standard of effectiveness for it can be met, such as by informing, for example, what the control group in an RCT should comprise.
The reason a treatment should be clearly and unambiguously characterised is because only then can tests of a treatment be made whose results can be validly used as evidence for effectiveness claims for that same treatment, and not slightly different ones. When 'covert treatment substitutions' happen, the helping or hindering factors that are present/absent in a trial are not present/absent in the same way in the clinical setting in which the treatment is administered. While this could be viewed as the treatment's effectiveness being contextual, it can also be viewed as the description of a treatment in clinical practice subsuming contextual factors, but differing from those factors subsumed during the trial. An unambiguous description of a treatment helps eliminate alternative explanations for the treatment's effects and helps inform the standard of effectiveness. It does this by indicating what alternative explanations are deemed ineligible for effectiveness. This could guide the appropriate control group in an RCT. For example, a good standard of effectiveness for a treatment might involve it not working through placebo effects if stakeholders deem those effects not integral to how the treatment should work.

Coherence
This viewpoint was used by Hill to mean lack of conflict with the natural history and biology of the disease. When such information is known, this can be useful for ruling out alternative explanations for a treatment's effects. 'Coherence' thus underscores the importance of mechanistic knowledge of diseases' natural histories.

Experiment
How do experiments offer evidence for treatment effectiveness? Unlike disease aetiology, for which conducting RCTs is usually unethical or infeasible, the effectiveness of treatments (particularly pharmacological treatments) can be and typically are evaluated by RCTs. This seems like it would obviate the need for evidence from observational studies and hence the need to apply Hill's viewpoints. Yet observational studies play important roles in evaluating treatment effectiveness (Black 1996), particularly for NPTs. The 'experiment' viewpoint then, at least with respect to treatment effectiveness, should not be seen as something separate from the other viewpoints, as if it could clinch treatment effectiveness and therefore diminish or eliminate the need for evidence from observational studies. Indeed, responses to some treatments have been shown to differ depending on if studied in RCTs or observational studies (Naudat, Maria, and Falissard 2011), and such average treatment responses may also differ from what could be expected with an individual patient in clinical practice. For both observational studies and RCTs the target population/individual in clinical practice is different than the study population. As discussed in the 'An account of treatment effectiveness' section, clinical trial results therefore need to be generalised.
A clinical trial (of any type; see note 2) I have proposed can be seen as being generalisable to the extent to which its results can contribute to valid inferences to treatment effectiveness claims (Tresker forthcoming). Although generalisability has traditionally been considered to be a reflection of a trial's external validity in terms of how representative the trial subjects are to those in the target population, I have also argued that representativeness, though important, is not a suitable feature for determining how well a trial contributes to valid treatment effectiveness claims (Tresker forthcoming). RCTs, though often very helpful for obtaining a treatment effect in a specific population, hold no privileged position in my account of treatment effectiveness when it comes to determining treatment effectiveness for a specific patient in clinical practice. The 'experiment' viewpoint should therefore be viewed as underscoring one of the many ways by which inferences to treatment effectiveness claims can be made. Looking at effectiveness this way could result in some nonintuitive conclusions, though. For example, Wilde and Parkkinen (2019) argue that when difference-making evidence in humans is absent, mechanism-based extrapolation from animal models (combined with appropriate mechanistic evidence) is sufficient to establish an exposure as a cause of disease. Applied to medical treatment, this could entail that the effectiveness of a treatment could be established in the absence of human studies, using extrapolated difference-making evidence from animal studies. The view of effectiveness I have articulated is consistent with this possibility.

Analogy
Analogy can be reinterpreted to apply to medical treatment by considering whether similar treatments work in similar patients. For example, drugs of the same class often produce similar effects. But because they can also produce very different effects this viewpoint should be applied with caution.
The foregoing reinterpretations preliminarily indicate that Hill's viewpoints could be useful for helping to establish or confirm a treatment as being effective. Indeed, approximately 14 years before his President's Address to the Section of Occupational Medicine of the Royal Society of Medicine in which he expanded upon and refined the viewpoints first promulgated by the Surgeon General's Advisory Committee on Smoking and Health (US Public Health Service 1964), Hill had alighted upon some principles that could be used to assess the effectiveness of treatments. These included replication of the result in similar patients and, 'whether the result was merely due to the natural history of the disease or in other words to the lapse of time, or whether it was due to some other factor which was necessarily associated with the therapeutic measure in question' (Hill 1951, 278;quoting Pickering 1949, 231). Hill thus adumbrated the 'coherence' viewpoint, and in doing so identified an important feature for helping to establish the effectiveness of treatments. Nonetheless, for Hill's viewpoints to be more useful for establishing or confirming treatment effectiveness, a tenth viewpoint that reflects the dependence of effectiveness claims on non-epistemic considerations might be helpful. However, since demarcating the specific criteria that can identify the legitimate uses of values (and which ones) in science is unsettled (Holman and Wilholt 2022), demarcating such criteria for treatment effectiveness is not something I can attempt here. Similarly, although I have indicated that an application of Hill's reinterpreted viewpoints might correctly judge an intervention to be effective, such an investigation is only a beginning and also exposes an unresolved tension in just how the RWT expresses Hill's original viewpoints. 19 Further work can help clarify the application of the reinterpreted viewpoints to unequivocally effective treatments and those whose effectiveness is contested, to ultimately determine the usefulness of Hill's reinterpreted viewpoints to establishing or confirming treatment effectiveness.

Conclusions
In this article, I have articulated an account of treatment effectiveness and described how the RWT, EBM+, and Hill's viewpoints contribute to understanding and establishing treatment effectiveness. The RWT, when not stretched beyond what it was formulated for, is useful by highlighting the importance of mechanistic evidence and its typically joint necessity with difference-making evidence for contributing to good treatment effectiveness predictions. How much and what type of evidence is required for establishing treatment effectiveness, however, is constrained by the idiosyncrasies of specific treatments and the purposes and contexts in which they are used. Accordingly, the RWT and EBM+ should be seen as pieces of a much larger puzzle for establishing or confirming treatment effectiveness. My reinterpretation of Hill's viewpoints is a modest attempt at expanding the repertoire of tools for establishing or confirming treatment effectiveness, especially when applied with an understanding of how effectiveness claims are fundamentally infused with values, some of whose influence should be addressed so that they can be managed appropriately. Areas I have suggested that could build on EBM+ include a greater sensitivity to inductive risk in formulating and communicating treatment effectiveness claims, and better procedures by which the medical community can assess standards of effectiveness and what amount, quality, and combination of difference-making and mechanistic evidence are sufficient to establish treatment effectiveness. Consistent with these suggestions, I have described an account of treatment effectiveness that emphasises non-epistemic features of treatment effectiveness claims, the importance of defending such claims and subjecting them to normative scrutiny, the desirability of establishing a clinically relevant and (particularly in the case of individual patients) patient-endorsed standard of effectiveness, and the predictive and rhetorical aspects of treatment effectiveness claims. Notes 1. According to Williamson (2019, 36) '"appropriately correlated" just means probabilistically dependent conditional on potential confounders, where the probability distribution in question is relative to a specified population or reference class of individuals.' 2. Clinical trials include, inter alia, RCTs; observational trials whereby the investigator observes outcomes without providing an intervention, such as case-control, cross-sectional, and cohort studies; case series; and N-of-1 trials.
3. Though as Leuridan and Weber (2011) note, examples attempting to show the RWT to be undermined could also be interpreted as calls for the associated claim of effectiveness to be amended. 4. I thank an anonymous reviewer for this journal for the lattermost point. 5. In the interest of space, I only discuss my own, in the next section. 6. One might nonetheless try to rely on the efficacy-effectiveness distinction to argue that RCTs (or, specifically, explanatory RCTs as opposed to pragmatic RCTs) are used to assess the efficacy of treatments whereas observational studies (and pragmatic RCTs) assess effectiveness. This distinction, however, is problematic and typically either just indicates the method by which a treatment is studied, or the population/setting in question (i.e. a study/research setting, or target population/clinical practice), but does not indicate anything intrinsic about a treatment's effectiveness (Tresker forthcoming). In the next section I offer an account of effectiveness that, in the case of an individual patient in clinical practice, looks at how a treatment is likely to affect that patient. In this regard, both so-called efficacy and effectiveness trials can produce evidence of effectiveness, since in both cases inferences will need to be made regarding how the treatment will affect a given patient in clinical practice. 7. For reasons of space I restrict my discussion of treatment effectiveness claims to the individual case and leave a discussion of treatment effectiveness claims for populations (such as might be made by researchers, regulatory authorities, health insurers, etc.) to potential future work. 8. Nonetheless, N-of-1 trials, although primarily just suitable for chronic, symptomatic conditions with easily measurable clinical outcomes, could provide useful information about treatment effectiveness, though typically only for the patient subject to the trial (Lillie et al. 2011;Duan, Kravitz, and Schmid 2013). 9. While a patient's treatment decision can be considered in the context of a benefit-cost analysis, there may be good reasons for an effectiveness claim to be seen as an input to this analysis and not as equivalent to the outcome. In work in progress, I articulate how better understanding and facilitating treatment decision-making in the context of a clinical encounter, particularly from a shared decision-making standpoint, can help to better understand the contours of how treatment effectiveness claims can be integrated into decision theory. 10. This is not to say that plausible mechanisms of the effects of acupuncture on nausea do not exist; see Chen et al. (2014), for example, for some possibilities, although some acupuncture researchers claim that the mechanisms remain undiscovered (Shi et al. 2019). 11. Although Gillies's account of acupuncture is problematic this should not be seen as impugning EBM+ as a whole, since alternative EBM+-inspired analyses might reach different conclusions. Moreover, Gillies should not be faulted, because rather than provide a rigorous substantive analysis of the effectiveness of acupuncture his chief aim appears to have been to illustrate the importance of mechanisms for establishing causal claims in medicine, for which I think he is successful. 12. 'NegC' reasoning in his typology whereby meta-mechanistic reasoning is used to argue that there is no plausible mechanistic chain between the putative cause and effect. 13. Illustrated by clinical practice guidelines (CPGs) when some experts hold diametrically opposed viewpoints. For example, CPGs promulgated by the American Academy of Orthopaedic Surgeons (2013) strongly recommended against viscosupplementation for the treatment of osteoarthritis (OA) of the knee, despite statistically significant results versus placebo, based on what they saw as clinical insignificance for some patients. This was met with strong resistance by physicians and industry members (Peer Review & Public Comments and AAOS Responses 2012). As Lubowitz, Provencher, and Poehling (2014, 4, italics in original) opined, 'In our opinion, it is not in the interests of all patients to recommend against a treatment that is of significant benefit for some patients, especially when that treatment is for a disease (knee OA) that is not preventable, and for which there is no cure.' This example also illustrates how non-epistemic values can influence effectiveness claims, and the importance of not conflating effectiveness determinations with treatment recommendations or decisions.
14. EBM+ proponents might argue that both sets of evidence should be evaluated and a conclusion drawn from that. However, it is unclear how such conclusions can be drawn using an EBM+ framework. Parkinnen and colleagues (2018, 74), for example, in offering an example of an intervention to which an EBM+ analysis could apply, subsequently state, 'We did not undertake a systematic review of the evidence on how probiotics might work.' This implies an actual substantive investigation is needed, possibly rendering the philosophical work otiose. 15. In this regard, use of the Template for Intervention Description and Replication (TIDieR-Placebo) checklist (Howick et al. 2020) could be used alongside EBM+'s checklists, given that it allows key features of both active interventions and placebo or sham controls to be concisely summarised by researchers, and also whether blinding was successful. Since this can easily be integrated with EBM, this seems like something that could also be integrated with EBM+. 16. As Howick (2011b) indicates, EBM's view of evidence is fundamentally based on hierarchies of methods of generating evidence. If this is true then it is not clear how EBM+ merits the EBM appellation given its apparent abandonment of evidence hierarchies by viewing mechanistic evidence on the same footing as difference-making evidence. 17. Some of these may very well be concerns shared by members of the EBM+ community, so my arguments should be construed less as criticism and more as spotlighting areas where EBM+ can be built upon. 18. Despite their popularity, Hill's viewpoints-even with respect to disease causation-have received their fair share of criticism (e.g. Rothman and Greenland [2005]). Hill's viewpoints are not well-defined conditions and it is not clear what is required for any given viewpoint to be satisfied, much less how they can be weighted and integrated to establish causation (Phillips and Goodman 2006). According to Ward (2009), Hill's viewpoints do not justify inferring that a statistical association is causal. 19. For example, Dammann (2018, 3) raises the point that since Hill didn't require all of the viewpoints be fulfilled for a causal claim to be supported and no single one (with the exception of temporality) to be necessary, claiming as Russo and Williamson do that Hill's guidelines express principles of the RWT elides over the fact that the RWT requires at least some of the viewpoints (namely, those expressing mechanistic and difference-making evidence) to be necessary.