Your Hypothesis or Mine? Terminological and Conceptual Variation Across Disciplines

Cross-disciplinary research (CDR) is a necessary response to many current pressing problems, yet CDR practitioners face diverse research challenges. Communication challenges can limit a CDR team’s ability to collaborate effectively, including differing use of scientific terms among teammates. To illustrate this, we examine the conceptual complexity and cross-disciplinary ambiguity of the term hypothesis as it is used by researchers participating in 16 team building workshops. These workshops assist CDR teams in finding common ground about fundamental research assumptions through philosophically structured dialogue. Our results show that team members often have very different perceptions about the nature of hypotheses, the role of hypotheses in science, and the use of hypotheses within different disciplines. Furthermore, we find that such assumptions can be rooted in disciplinary-based training. These data indicate that potentially problematic terminological differences exist within CDR teams, and exercises that reveal this early in the collaborative process may be beneficial.

Variation in technical terminology is a communication issue that consistently emerges across teams of disparate disciplinary composition, career maturity, and team objectives (Bracken & Oughton, 2006;Endter-Wada, Blahna, Krannich, & Brunson, 1998;Jeffrey, 2003;Thompson, 2009). There is some consensus that technical language is an area where cross-disciplinary literacy is critical (Marzano, Carrs, & Bell, 2006;Morse et al., 2007;Thompson, 2009;Wear, 1999). This is especially evident in science and academia, where words can have more precise and constrained disciplinary meanings than they do in general language. For an example of this, consider the word significance. In lay language, this word simply means important or meaningful, whereas in statistics, the term is used to indicate that there is a low probability that observed measurements or values of interest occurred by chance. Although scientists may be more closely aligned in how they define the term significance, there is still a great deal of variability and potential confusion. For example, Heath (1995) explained that statistical significance is not always representative of biological significance, because an experiment may show statistical significance but have results that are not biologically meaningful. Other words that represent basic, foundational concepts-such as hypothesis, science, research, empirical, qualitative, quantitative, basic, applied, experimental, replication, and validation-are used by CDR teams and may also have disparate meanings among team members across disciplinary contexts (Bracken & Oughton, 2006;Bruce, Lyall, Tait, & Williams, 2004;Thompson, 2009;Wear, 1999).
Semantic variation across technical vocabularies challenges successful CDR in a number of ways. Cross-disciplinary communication requires practitioners to (a) understand their own disciplinary perspective well enough to convey the meanings to others and (b) develop a shared understanding of commonly used terms (Jeffrey, 2003). The understanding one has of one's own disciplinary perspective, though, will often be implicit or tacit, acquired over time based on one's experiences with words (Nagy & Scott, 2000;Polanyi, 1966Polanyi, /2009). In the event that obstacles to communication arise, explicit understanding of this perspective that can be readily conveyed to one's collaborators could be of great value. Mutual understanding of how other disciplines use their technical terms is needed to support the desired level of integration and build common ground across disciplines (McKibbon et al., 2010). Once achieved, this mutual understanding can enable collaborators to avoid unwittingly talking past one another because they assume that the terms they use are univocal. Appreciation for these disciplinary nuances allows for greater word consciousness, which promotes more effective word choice and word use (Graves, 2006).
In this article, we examine the conceptual complexity and cross-disciplinary ambiguity of technical terminology from the perspective of a central scientific technical term, "hypothesis." In some disciplines, the term hypothesis may simply refer to an assumption about a topic or process; in other disciplines, by contrast, the term calls to mind statistical significance and criteria for rejecting a given assumption. It is striking that a word fundamental to scientific endeavors can be used so differently across disciplines. We begin by supplying context for understanding the hypothesis concept, rooted in the philosophy of science. We then describe our methods that involve isolating and analyzing transcript data concerning the hypothesis concept collected through the Toolbox Project, a workshop-based initiative that focuses on enhancing cross-disciplinary communication through systematic, philosophical dialogue about research assumptions. Working, cross-disciplinary scientists participate in the Toolbox dialogues, expressing their own views about the conceptual assumptions underpinning scientific practice and negotiating collective views that can guide their collaboration going forward. Evidence from these cross-disciplinary workshops (a) highlights the range and diversity of ways that scientists use the term hypothesis and its cognates and (b) illustrates how the different uses of core scientific technical terms can create the possibility of confusion within CDR teams. Finally, we argue that close attention to the term hypothesis in a cross-disciplinary context can highlight the critical yet elusive nature of cross-disciplinary understanding, and that engaging in conversations explicitly about scientific jargon and paradigms reveals differences within teams that participants may never have considered.

Theme and Variation in the Hypothesis Concept
Few concepts are as central to scientific research as the concept hypothesis. "Hypothesis" comes from the Greek hupo, meaning under, and thesis, meaning placing; together, they yield a word close in meaning to the English word foundation. Although it could certainly be said that the concept hypothesis plays a foundational role in contemporary conceptions of the scientific process, the term hypothesis has taken on a different sense over the intervening millennia. Specifically, hypotheses have become the statements that express "alternative possible explanations" (Martin, 2000, p. 161) that are placed "under test" (Hempel, 1966, p. 19) in scientific research. Glass and Hall (2008) supplied a brief account of how the hypothesis concept underwent this transition, tracing it from Galileo through Bacon, Newton, Hume, and into the 20th century with Popper and Kuhn. Not all these figures were equally sanguine about the importance of hypotheses to science-as Glass and Hall noted, Newton makes no room for hypotheses in his "experimental philosophy"-but the concept received a decided boost from Popper, whose work is the "basis for much of the framing terminology used by practicing scientists today" (Glass & Hall, 2008, p. 379).
According to what we will call the "Standard View" of the scientific method, science begins with observation that leads to the formation and then the testing of hypotheses, followed by interpretation of the results (McErlean, 2000, p. 8). An important method of forming and testing hypotheses is the hypothetico-deductive method, in which a hypothesis is tested against an observational statement that it deductively implies (Hempel, 1966). The Standard View gives privilege of place to hypotheses as the "tentative answers" (Hempel, 1966, p. 10) that would serve to explain the phenomena under investigation. Specifically, on this approach, hypotheses are statements of relationships between variables that are formed prior to derivation of the data and, if valid, suggest explanations for the effects under investigation. As such, they are different from both research questions, which pose but do not explain, and models, which are built after data have been derived (Krebs, 2000). Hypotheses in this view are regarded as necessary parts of the scientific process (Martin, 2000), without which "analysis and classification are blind" (Hempel, 1966, p. 13).
On the Standard View, hypotheses are "invented" to explain "observed facts" (Hempel, 1966, p. 15). They are "happy guesses" produced by creative applications of "inventive talent" (Whewell, 1984, p. 211). These suppositions play a key functional role in science, guiding scientific inquiry and constraining scientific experimentation. Historically, though, there has been concern about giving hypotheses such a formative influence over scientific inquiry; in fact, it is this that turned Newton against them, since hypotheses would "frame" the inquiry with an "unproven premise" (Glass & Hall, 2008, p. 379). Here is where Popper's contribution to the Standard View is most keenly felt. By introducing the idea of falsification, Popper turned the scientific attitude toward hypotheses on its head. If it were the case that science aimed to verify a priori speculations, then it would be legitimate to worry that science was not explaining the world but rather recasting it in our image. However, Popper argued, hypotheses are not to be verified, but rather falsified. It is for this reason that they remain tentative-"provisional conjectures" that can only be definitively falsified and never confirmed (Popper, 1968, p. 265).
Although the Standard View dominates textbook introductions to the scientific method, the reality of scientific practice reveals wide variability in the hypothesis concept. This can be seen in disciplines such as sociology and ecology, which are fluid in how or even whether a hypothesis is incorporated in research methodology. One might define the term hypothesis in these fields as a tentative answer to a research question that is sometimes a testable statement of relationships (Quinn & Dunham 1983;Vogt, 1999). But others define "hypothesis" using the stricter, statistical definition that limits the application of hypotheses to theoretically based, quantitative research questions about a given relationship and requires statistical data analysis to interpret (Teddlie & Tashakkori, 2009). Although social science and ecological researchers tend to use quantitative approaches when employing hypotheses, qualitatively based research is also importantly associated with hypotheses. In particular, qualitative research plays a central role in generating hypotheses. According to Auerbach and Silverstein (2003), although the traditional quantitative approach to research most often includes the testing of hypotheses, the focus of qualitatively driven research may be to develop hypotheses that can later be tested. Furthermore, some researchers propose that qualitative data can be used to test hypotheses, so long as other deductive assumptions are maintained (Flyvbjerg, 2006;Hyde, 2000).
There is no single way to think about the nature of hypotheses or their role in science. Variability in this most basic of scientific research concepts suggests that different disciplines embed different epistemological perspectives, that is, different ways of looking at the nature of research and the business of producing new knowledge (Miller et al., 2008). 2 The fact that conceptual variability manifests at such a foundational level suggests the need to communicate about the different epistemological perspectives in play in collaborative CDR. Failure to do this can leave these variations lying in wait, threatening to give rise to unreasonable disagreement or, perhaps worse, unreasonable agreement (e.g., Bracken & Oughton, 2006;Slatin, Galizzi, Melillo, & Mawn, 2004;Thompson, 2009). In short, revealing assumptions and misconceptions around fundamental terms such as "hypothesis" can create opportunities for CDR practitioners to recognize potential communication hurdles that could hinder team progress, development, and efficiency.

Method
The "Toolbox" approach described by Eigenbrode et al. (2007) is designed to create common ground through conceptual dialogue about fundamental, philosophically identified research assumptions. Specifically, the Toolbox approach focuses members of interdisciplinary research groups on epistemic and metaphysical dimensions of scientific research, structuring a dialogue among them about assumptions that frame and inform their research approaches. By limning research orientations in these fundamental terms, collaborators are afforded the opportunity to develop their disciplinary self-understanding as well as their mutual understanding of the various perspectives represented in the group.
These dialogues take place in workshops organized primarily for groups engaged in cross-disciplinary projects. Toolbox workshop groups have comprised ad hoc collections of researchers, groups of students, and functioning research teams (the majority). Most groups had already coalesced around or were about to begin work on a CDR question. The focus and composition of these groups has varied widely, addressing research questions ranging from public health to clean energy systems to evolutionary biology. Before the workshop, participants are asked to self-report their disciplinary background on their workshop sheets. Participants can record up to six selfidentified disciplines in the demographic section of their Toolbox instruments. Most participants list several disciplines, typically variations on a more general discipline (e.g., microbiology, microbial evolution, and microbial ecology).
Participants in the workshops base their discussion around the Toolbox instrument, a structured set of prompts that draws out their views on issues concerning epistemic dimensions (viz., motivation, methodology, confirmation) and metaphysical dimensions (viz., reality, values, reductionism) of the research process. (The Toolbox instrument can be downloaded from www.sagepub.com/orourke, under Chapter 11 student resources.) Immediately prior to the dialogue, participants are provided with a paper copy of the instrument, asked to read the prompts and associated questions, and respond to the prompts by scoring a Likert-type scale and/or making notes about their impressions. In responding to the Toolbox prompts, participants are asked to be mindful of both their own practice and their disciplinary perspective(s); as a result of this prompting, both their individual scientific assumptions and a representation of the assumptions underpinning their discipline(s) emerge in the dialogue. The prompts express foundational claims about scientific research that support different and incompatible perspectives, and as a result motivate dialogue that can reveal unexpected disagreements and surprising similarities. This enables the group to develop an enhanced, negotiated, mutual understanding of their respective approaches to science that is conducive to more effective and efficient research communication.
The subsequent dialogue session typically lasts about 2 hr. Discussion can begin around any point and anywhere within the instrument, and is usually initiated by a participant who found a particular prompt intriguing. While the group dialogue is intended to unfold in an organic, participant-led fashion, each session is moderated by a facilitator who is part of the Toolbox research personnel. The facilitator's role is to prompt participants to explore other areas of the Toolbox instrument if discussion becomes "bogged down," or simply comes to a halt, without being intrusive or guiding the dialogue. Toolbox sessions are audio-recorded and subsequently transcribed. Each session also includes an observer drawn from the Toolbox project personnel, whose role is to take notes that ensure the accurate attribution of speaking turns to distinct participants and to capture non-verbal data.
Our research has been exploratory, and our aim has been to run workshops with as many different types of cross-disciplinary groups as possible. Initial recruitment focused on cross-disciplinary groups associated with the National Science Foundation's Integrative Graduate Education and Research Traineeship (NSF-IGERT) projects; subsequent groups were recruited by targeted invitation or when they approached us about participating in a workshop.
For this article, we analyzed how 16 interdisciplinary groups used the term hypothesis and its cognates within their Toolbox workshop dialogues. Workshop dialogue about hypotheses was generated primarily by the following Toolbox prompts in the "Epistemology" section-the first in the "Methodology" sub-section, and the second in the "Confirmation" sub-section: • • Scientific research (applied or basic) must be hypothesis driven.
• • There are strict requirements for determining when empirical data confirm a tested hypothesis.
Transcripts of these dialogues were coded and analyzed in the qualitative software Nvivo 8 by Toolbox researchers using narrative analysis that involved, first, developing an iterative coding scheme (Miles & Huberman, 1994), and second, using that scheme to reveal patterns related to the use of the term hypothesis in four coding stages: preliminary coding, peer-debriefing, recoding, and code analysis. Use of an iterative coding process and peer-debriefing helped ensure the trustworthiness of our data. The only dialogue considered for this analysis was limited to paragraphs of conversation that directly referenced the term hypothesis or its cognates. The results of these discussions are described below.

Results
We analyzed dialogue about hypotheses from 16 Toolbox workshops, involving a total of 127 researchers (mean group size = 7.8) from multiple disciplines (Table 1). For clarity, we have characterized the participants based on the first of the several disciplines listed. We categorized these disciplines into research areas, after Morillo, Bordons, and Gomez (2001), to provide a simplified look at the makeup of participating interdisciplinary teams. These data are not analyzed further. For the purpose of this article, we focused our analysis on participant discussion of the meaning and use of "hypotheses" and its cognates in science. Three themes about hypotheses emerged across workshops. Here we arrange those themes conceptually, moving from basic discussion of  what a hypothesis is to the critical question of whether hypotheses must be employed to perform science, and concluding with evaluation of when and how hypotheses are employed across and within disciplines.

Theme 1: Definitions of Hypotheses
Questions about how to define and employ the word hypothesis were raised by participants during each Toolbox session. It was apparent in most of these discussions that the term was employed differently, and the conceptual and linguistic polysemy of "hypothesis" in a scientific context was not lost on the participants. As one ecologist noted, in talking about his team, (1) [The problem] is that different definitions of "hypothesis" are in use. (Session 10, Ecology) Some participants approached negotiating this problem by reflecting on the intrinsic nature of the hypothesis concept, that is, on the characteristics that constitute the meaning of this concept. It was in this spirit that participants offered characterizations of the term hypothesis, which varied considerably across the sessions. For example, a few participants offered rather general characterizations, such as Comments 2 and 3: (2) And so now I've been going back to my students, and they've been telling me they've done all these great things, and I said, well, what is your hypothesis? What did you think you were going to see? (Session 16, Computer Science) (3) I teach that a hypothesis is our best guess to a question. (Session 10, Plant Pathology) These perspectives align well with the closely associated conceptions of hypotheses as tentative answers (Hempel, 1966), happy guesses (Whewell, 1984), or alternative possible explanations (Martin, 2000). The second is also consistent with the idea that hypotheses can be distinct from and posterior to research questions. Such a clean distinction between hypotheses and questions is not universally maintained, though, as evinced by Comment 4: (4) I view that hypotheses are directive questions . . . (Session 2,Geochemistry) This statement (Comment 4) conforms to the idea presented in Hempel (1966) and Martin (2000) that hypotheses guide scientific inquiry. One specific mode of guidance supplied by hypotheses, so understood, involves highlighting what counts as evidence, as expressed by an ecologist in Comment 5: (5) Without a hypothesis, even if it's just something you think may happen, you need to know what evidence you need to take. You need to know what to look for and you don't know what you're looking for without a hypothesis of some sort. (Session 7,Zoology) Two additional modes of guidance-identifying what is to be tested (Hempel, 1966;McErlean, 2000) and providing the foundation for experimental design-are expressed in Comment 6: (6) I'm a big proponent of at least knowing what your question is. I'm a little less concerned about whether you frame it as a null [hypothesis] or not. But you ought to have some sort of hypothesis that you're going to try to test or use to design your project. (Session 3, Natural Resources) Of course, guidance need not always be positive, especially if it involves a "rigid" hypothesis: (7) . . . where you might have a really "rigid" hypothesis . . . that might in fact constrain your thinking, constrain your creative juices . . . (Session 16,Computer Science) Other Toolbox participants offered more narrowly defined conceptions of hypotheses. Common among our data are expressions of statistical conceptions. One participant stated, (8) I think that a hypothesis has to do with using statistical tests to prove or disprove a bunch of numbers (Session 2, Marine Ecology), while another offered, (9) I'm really stuck in that biostatistics hypothesis box where a hypothesis has to be quantitatively proven or disproven. And maybe that, that's just kind of the box I have around the word "hypothesis." (Session 2, Geochemistry) Not all participants endorsed the standard, statistical approach, however; for example, one conservation biologist noted, (10) The big trend now is away from [using] the standard [null hypothesis]. Instead you have a list of ten, twenty hypotheses and you're gathering evidence and seeing which of those fits rather than forcing null and alternative hypotheses. (Session 2, Ecology) A second way of getting at the meaning of hypothesis involves locating it extrinsically in the context of the practice of scientific research. Instead of seeking to define the hypothesis concept by identifying conditions that constitute it, this approach takes it to play an important input/output role in the scientific process. As we saw above in Comment 3, one way to understand the generation of hypotheses is as tentative answers to interesting research questions. This view is championed by a philosopher for his collaborators in Comment 11: (11) . . . and what we asked people to do . . . was not necessarily to come up with hypotheses but come up with questions, right, interesting questions, right? And that's the start of a process that might lead to interesting hypotheses. . . . (Session 16,Philosophy) In responding to this claim, one of the philosopher's collaborators questioned whether research questions precede hypotheses: (12) But to know even how to begin with an exploration of those questions, you have to have some idea in mind of what the answer could be or you won't have any idea where to look. (Session 16,Computer Science) This sentiment that hypotheses are guiding ideas about "what the answer could be" is widely shared, but others take the "idea in mind" that frames pursuit of research questions to be a vague sense that something scientifically interesting is afoot. As one social scientist explained to her teammates, (13) I think of the value qualitative research is that it could generate hypotheses; I don't know the degree to which that's true within most of the disciplines represented here. Take a Margaret Mead classic case of understanding the situation in a very holistic way, and in doing that, calling out things that strike you or that seem curious or ought to be understood more, or could be considered. And in the describe-predictcontrol sequence of scientific research, then that description phase can [be] generating [a] hypothesis about that. (Session 10,Education) In Comment 13, hypotheses are understood to be an output of the descriptive phase of scientific research, serving then as an input into predict-control phase. In other workshops, we find participants articulating how hypotheses contribute to the predict-control phase: (14) The data we collect are empirical, because we're out there collecting them according to some design. Then we test them statistically, and set criteria for determining whether we're going to reject or accept the hypothesis. (Session 3, Civil Engineering) (15) Every time you build a model, it is like you're testing a hypothesis because you build models to collect data. (Session 2,Economics) In Comment 14, as in Comment 5 above, the hypothesis is described as giving shape to the experimental design that yields data that determine the resulting attitude toward the hypothesis; Comment 15 attributes the same functional role to hypotheses in the context of model-based science. In both these cases, the role ascribed to hypotheses is consistent with the hypothetico-deductive method, described above (Hempel, 1966).
Emphasis on hypotheses understood extrinsically suggests that hypotheses are necessary components of scientific research, but this is a controversial topic that has been debated in many of our workshops, with participants expressing strong opinions about the centrality of hypotheses to valid research. We turn to this topic in the next theme. The idea that scientific research must be hypothesis driven, or must at least make room for hypotheses, is expressed in Comments 5 and 6 above-hypotheses are critical design elements that identify both what is to be tested and what is to count as evidence for those tests. This attitude is also reflected in the following pair of quotes: Both of these reinforce the picture that scientific research requires systematic tests, and that these depend on hypotheses for their content and focus. In Comment 19, we find this view expressed in connection with graduate education: (19) . . . if you don't formulate a hypothesis, you tend to just drift with the wind. So if you're training students, it's good to be able to teach them how to formulate a specific hypothesis at some point. So for me it's like uh, I think it was Emerson said about consistency, a foolish consistency is the hobgoblin of small minds. A foolish addiction to hypotheses is not productive, but you've got to have it sometimes. (Session 16, Microbial Ecology) These three comments were comments from participants who consider themselves scientific researchers, but an essential connection between scientific research and hypotheses was also expressed by those who did not think of themselves in this way. For example, one engineer announced, (20) I don't really use hypotheses in my disciplines, but I don't really consider myself a scientist either. I guess I had thought if you considered yourself a scientist that meant you used hypotheses . . . (Session 15, Engineering) (21) . . . the first thing that came to my mind is wait, you know, I'm doing design. That's not hypothesis driven. I mean that's problem solving and, and has nothing to do with hypotheses basically, but then I went and said well wait a second. That's not scientific research. The question says scientific research, applied or basic. It's certainly scientific work, but I wouldn't call it scientific research. It's design. That's why it has its own name. . . . (Session 16,Computer Science) Both Comments 20 and 21 cast scientific research as research that involves hypotheses by definition. Others take hypotheses, understood as antecedent guides to scientific research, to be necessary but only as part of the rhetoric of science. This view is expressed in the following comment from a computer scientist: (22) It's just that a lot of times what happens is that you do the research and you formulate the hypothesis after you have the results . . . A lot of really great new discoveries happen that way. You come up with something you did not expect, you did not anticipate. You have to write it up, you have to fill out the formula, and you write it up . . . misleadingly that way. (Session 16,Computer Science) According to this view, hypotheses are retroactively crafted to bring scientific discovery in line with accepted scientific rhetoric, that is, with community expectations that research generates conclusions that are anticipated and not accidental. But so understood, this necessity is less a conceptual requirement of scientific research than a sociological one.
Other participants were less adamant that hypotheses testing per se is required to conduct science: (23) Not all the research is based on experimentation. So [research] doesn't necessarily have to be hypothesis-driven. (Session 7, Biology) By distinguishing non-hypothetical research from experimentation, Comment 23 points us in the direction of more observational or descriptive science. Another biologist offered an example of this sort of project as an instance of scientific research that does not involve hypotheses: These comments suggest that scientific research that does not involve hypotheses is distinct from and perhaps independent of hypothesis-driven research; however, scientific research is more complex and intertwined than that, as revealed in this exchange between a biologist and a computer scientist: Observational and qualitative studies can stand alone as rigorous contributions to scientific knowledge, but they can feed into experimental work when they are exploratory, a description that emphasizes the search for insight that can frame an experimental study. This resonates with comments made above by those who sought to understand hypotheses extrinsically, locating them in a process that involves a "descriptive phase" in advance of hypothesis formulation. It is worth noting, though, that not everyone looks upon the work conducted prior to the formation of hypotheses as scientific research. Consider the following exchange between an ecologist (P01) and an economist (P02): (27) P01: [On] intuition and predispositions? (Session 2, Ecology) P02: Yeah, that's how I build my hypothesis.

P02:
But then I develop a hypothesis and I test a hypothesis and it's done [using] quantitative methods.

P01:
But that [i.e., exploratory work involving intuition] is not research, that's just freeing stuff . . . boiling down to a hypothesis.
"Freeing stuff" is not research, from this perspective, even if it is necessary to "boil down" to a hypothesis. Conversations like (27) highlight the common view that regardless of whether or not a discipline is traditionally hypothesis-based, research must involve hypotheses from the start to be considered scientific.
Thus, understanding perceptions concerning hypotheses at the level of scientific practice is critical to revealing varied expectations about whether and when a CDR team is conducting research that can be properly classified as scientific. There are also similar, important variations at the disciplinary level that are manifest in the workshop data. Disciplinary deployment of hypotheses is our focus in the third theme.

Theme 3: How Do Disciplines Use Hypotheses?
As we noted above, emphasis on how hypotheses figure into the practice of disciplinary research is motivated by the two statements in the Toolbox, one focusing on hypotheses as drivers of scientific research and the other on what is involved in confirming hypotheses. Because participants are asked in part to be mindful of their disciplines in dialogue, they typically respond to these prompts by embedding the hypothesis concept into their disciplinary practice.
As we have seen in Comments 14 and 15, certain participants report that hypotheses are indispensible to conducting research in their disciplines. The engineer who produced Comment 14 went on to say that, whatever anyone might think, their discipline demanded that research products include testable hypotheses to meet publication standards: (28) . . . so I was just thinking methodologically, if you don't follow those protocols, you don't get published in my [discipline's] journals. (Session 3,Engineering) Others express more ambivalence about how hypotheses are relevant to the practice of a particular discipline. In Comment 21, an engineer distinguished between scientific research and scientific work, reporting that the latter did not involve hypotheses. Another engineer describes the role of hypotheses in engineering this way: Although Comment 29 does not explicitly claim that engineering counts as scientific research, the view expressed conforms with a conception of scientific research described by Glass and Hall (2008), namely, question-driven investigation that leads to the formation of models, which are then verified based on their ability to predict outcomes.
Engineering was not the only area described as operating without hypotheses. Another discipline mentioned by name was oceanography: (30) A lot of times in oceanography, you can't really have a hypothesis . . . because oceanography is about deploying instruments, seeing what those instruments say, and deploying more instruments. And it can be completely un-hypothesis driven. (Session 2, Marine Ecology) While Comment 30 met with criticism in the workshopin fact, Comment 17 was a response to this comment-it reflects a perspective according to which scientific research can consist in data collection that is not hypothesis driven.
Another discipline that has a complicated relationship with hypotheses, as we have seen, is social science. In one workshop, a biologist proclaimed, (31) Social scientists don't use hypotheses; they only use observations. (Session 13,Biology) To this, a public health scientist responded, (32) I am a social scientist and I use hypotheses all the time. There are misconceptions that all social scientists do qualitative research, but that is not true. The kind of research we do depends on the questions we are trying to answer and I often want to answer more quantitatively-based questions. (Session 13,Public Health) The social scientist who reported a similarly complex relationship between social science and hypotheses in Comment 25 prefaced that remark by speaking to these misconceptions, saying, (33) Many people on my team assumed I did not use hypotheses because I am a social scientist. (Session 14,Sociology) Similarly the engineer responsible for Comment 20, in learning more about the work of social scientists, continued, (34) . . . I guess I did not realize that sociologists are scientists. So I had not really considered whether or not they might use hypotheses. It is very interesting for me to find out that [sociologists] sometimes do use hypotheses! (Session 15,Engineering) This statement illustrates how members of CDR teams can have limited knowledge of other approaches to science. Perhaps one of the most critical aspects of team communication is how members of CDR teams apply their disciplinary expectations to the contributions made by their collaborators. A fundamental contribution of the Toolbox workshops is helping promote self-and mutual understanding regarding communication styles and philosophical differences between team members. The quotes weaved throughout this section and the dialogue captured in Comments 33 and 34 provide just a few examples of workshop participants learning more about how and why other disciplines chose to use or not use hypotheses in their research.

Discussion
Terminological variation across disciplines is a well-known challenge for those who engage in CDR. However, although it has been often noted (e.g., Eigenbrode et al., 2007;Galison, 1997;Klein, 2000;Naiman, 1999;NAS, 2004;Schoenberger, 2001), the point has not often been explored in detail with respect to a particular term. Our attempt to do this highlights the need to avoid the assumption that one's collaborators will agree with how fundamental concepts like hypothesis are interpreted across a CDR project. It also supplies a map of the space of alternative interpretations of "hypothesis" that is in itself useful, given the centrality of this concept in science. Our specific goals in this article are to (a) supply data that demonstrate variable use of the term hypothesis and its cognates across disciplines, (b) analyze these data to reveal insights into the hypothesis concept as it is understood by practicing scientists, and (c) use these insights to illuminate communicative and conceptual obstacles that confront crossdisciplinary collaborators. The first of these goals was addressed in the previous section, and in this section we address the second and third. We speak initially to a methodological concern, and then examine the data from the perspectives of the three themes introduced above.
Our approach has been to collect comments and insights from scientists about the hypothesis concept and use those as the foundation for conceptual analysis. One might critique this approach by asking why we put so much stock into what scientists think about hypotheses in general, given that many scientists have not had much training in the relevant conceptual methods or philosophical literatures. If, as Glass and Hall (2008) asserted, "many scientists use the term 'hypothesis' when they mean 'model'" (p. 378), why should we rely on their views in analyzing the hypothesis concept? There are two reasons. First, the conceptions expressed in these workshop dialogues are what guide these scientists as they conduct their research business. They have ample practical experience in forming and testing hypotheses; furthermore, their intuitions and interactions with hypotheses are the ones that matter to scientific practice. Although some may be philosophically untutored, the evidence from the dialogues is that when given the opportunity, many are quite capable of articulating how they understand the hypothesis concept. Second, one of our goals is to highlight the fact that you cannot presume your collaborators will have the same understanding of this core concept as you work toward achieving project objectives. The differences between interpretations can be subtle and insidious, making it difficult to appreciate when they are there and when they are not. Failure to recognize that even fundamental concepts like hypothesis can be variously interpreted can lead to disagreements about project decisions concerning research direction and publication (Comments 28 and 29; see also Bracken & Oughton, 2006;Thompson, 2009;Wear, 1999). Thus, we are interested in what the scientists think because they are the ones who do the science and they are the ones who collaborate.
Turning to Theme 1 and the nature of hypotheses, it is worth repeating the two statements in the Toolbox instrument that have prompted the dialogue we sample above: • • Scientific research (applied or basic) must be hypothesis driven. • • There are strict requirements for determining when empirical data confirm a tested hypothesis.
These supply the conceptual frame within which the comments supplied above were made. The first of these presents hypotheses as what guide scientific research, whereas the second highlights the role that hypothesis testing plays in advancing scientific understanding. These aspects resonate with the historical, philosophical development of hypotheses as an element of the scientific method, and they are reflected in our data. For example, Comments 4 to 6 and 14 highlight the guidance role, whereas Comments 6 and 15 emphasize a different but not necessarily incompatible role as what is tested. In its tested role, the hypothesis is what one expects to see (Comment 2), one's "best guess" (Comment 3), or a possible "answer" (Comment 12).
What emerges from these comments is a picture of hypotheses as representations of empirical possibilities whose realization is expected under certain conditions. Hypothesis-driven science, then, involves producing these conditions to evaluate these expectations. As possibilities, hypotheses are akin to questions, sharing in their uncertainty and in their ability to direct investigation in the pursuit of a more settled epistemic position. As representations, they are also akin to models-they figure into hypothetical worldviews that guide scientists in the development of experimental designs and data analysis. Interestingly, and pace Popper, many of the scientists who participate in our workshops accept the characterization that hypothesis-driven science aims to confirm hypotheses, and that the goal of scientific research is to augment our understanding. We acknowledge that the second prompt presupposes the legitimacy of hypothesis confirmation; however, one reason for disagreeing with the prompt is the rejection of hypothesis confirmation as a possibility, and this perspective is not articulated in the transcripts we considered and is only rarely articulated in Toolbox workshops. As Glass and Hall (2008) noted, falsification is favored philosophically over confirmation in part because it changes the status of the hypothesis understood as an unproven premise; however, the scientists who have participated in Toolbox workshops tend to see hypotheses as part of a process, emerging from prior research in an ongoing effort to examine and explain the phenomena that interest them. Their hypotheses exist as embedded premises, dependent in many complex ways on substantial scientific work.
As we have noted, Glass and Hall (2008) distinguished hypotheses from questions and models, and suggest that scientists are not always so careful. Our data would appear to bear them out, but may point to a different possibility. Glass and Hall's view entails a relatively restrictive sense of the concept according to which a hypothesis is "an idea or postulate that must be phrased as a statement of fact, so that it can be subjected to falsification" (Glass & Hall, 2008, p. 378). But our data suggest that there is a range of hypothesis concepts that vary in terms of how restrictive they are. Comments 7 to 10 and 14 describe "rigid" hypotheses that meet specific criteria, rendering them subject to test and falsification, consistent with the view of Glass and Hall. By contrast, though, consider Comments 2 to 6 and 12-these characterize hypotheses not in terms of their status as statements of fact, but rather in terms of their potential to motivate and frame broader scientific inquiry. These comments describe less restrictive, more general versions of the hypothesis concept. One does not falsify questions, and hunches about what you might see are not precise enough to be subject to classical falsification. Interestingly, this difference could account for why so many participants in these dialogues are happy to think of their work as involving confirmation of hypotheses, pace Popper, Glass and Hall, and many other philosophers of science. If the more expansive version of the concept is in play, then establishing the legitimacy of the inquiry framed by such a hypothesis could well be taken to confirm the hypothesis in that framing role. Participants across workshops seemed sensitive to gradations in the hypothesis concept and how such variability reflected functionally different categories in research, for example, testable hypotheses versus general hypotheses. This variation in the concept could explain some of the disagreements expressed in Toolbox workshops as merely apparent, for example, the disagreement found in the Exchange 11 and 12, where it appears that the restrictive concept informs Comment 11 whereas a more expansive concept informs Comment 12. Another point to note is that the variability did not necessarily follow preconceptions of "hard" and "soft" disciplines; for example, oceanography, a very quantitative physical science, was described as less hypothetico-deductive by its practitioners than what the profuse instrumentation and precision might lead one to believe.
We turn now to the second theme and the question of whether hypotheses are necessary to science. The first point to make in this connection is that there is no one conception of science in play in these dialogues. This is evident in the comments that dispute the scientific standing of work that is not hypothesis driven, for example, Comments 17, 18, 21, and 27. Nevertheless, the statements that introduce the hypothesis concept serve to specify a common conception of science for the purposes of the discourse, even if there is disagreement about whether that specification is apt. Hypotheses are introduced as drivers of scientific research which, when successful, issues in confirmation of those hypotheses. Here hypotheses are characterized extrinsically, with the scientific process built around them. This is the conception of science to which participants in the workshop react when they remark on the necessity of hypotheses.
Although our results do not reveal unanimity across the disciplinary spectrum concerning the nature of science, upon analysis, they deliver a broad model of science. Beginning with the observation made in Comment 21, we can distinguish between scientific work, such as design, and scientific research, which has something to do with hypotheses. Although there is disagreement among the participants in our workshops, many will allow that scientific research can involve hypotheses even if it is not itself experimental or hypothesis driven. Thus, the data support the distinction between experimental and non-experimental scientific research, with the latter variously described in our data as observational (Comment 31), descriptive (Comment 24), and exploratory (Comment 26). Experimental research so distinguished involves use of the more restrictive hypothesis concept, whereas non-experimental research involves work that can be productive of hypotheses (but see Flyvbjerg, 2006;Hyde, 2000). As we saw in Comments 23 and 24, though, non-experimental research can also be regarded as fully formed even if it is not experimental and it does not issue in more restrictive hypotheses. Furthermore, model building and testing could be included in the picture in several places: scientific work that does not qualify as research, 3 non-experimental research conducted without hypotheses, or part of hypothesis-driven experimentation.
Theme 3 concerns the disciplinary deployment of hypotheses, and it is here that we address the third goal mentioned at the top of this section, namely, the goal of evaluating communication and collaboration in CDR. In the workshops, there was ample discussion about how different disciplines did-or did not-employ hypotheses. Our results demonstrate that these discussions were important for better understanding the research approaches of fellow participants, and in some instances (e.g., Comment 21), perhaps better understanding one's own discipline. It is important to emphasize that the goal of the dialogue is not to standardize the understanding of the hypothesis concept in a project, because differences about it are an important part of the conceptual diversity that recommend CDR as a mode of research in the first place; rather, the goal should be to discover those differences so as to enhance the collaborators' collective ability to look at their common problem through each other's eyes. Thus, the dialogues emphasize that one's research approachbroadly embodied here under the hypothesis concept-is an important topic for discussion among CDR collaborators (Eigenbrode et al., 2007;Khagram et al., 2010;Miller et al., 2008).
These discussions also revealed erroneous preconceptions participants had about one another's disciplines, often in debate about the use of hypotheses by a particular discipline. Of course, debate among scientists about the use of hypotheses in research is not new. Within ecology, for example, there is a history of adapting increasingly quantitative methods, at least in part to counter the notion that ecology was not a hard science (Simberloff, 1983). The fallout of this "physics envy" has been a trend away from inductive approaches in ecology and the dominance in the field's flagship journals by "the scientific method" in the spirit of Platt (1964), despite recognition that there are numerous ways ecologists actually practice science (Forbes et al., 2004). One important and reoccurring example of these erroneous preconceptions in our workshops was the statement by natural and physical scientists that social scientists did not employ hypotheses at all (e.g., Comments 31 and 34). This evinces a lack of understanding about the social sciences as a whole, as reflected in reactions by social scientists (e.g., Comments 32 and 33), and the use of complementary quantitative and qualitative approaches within social science disciplines. Given our data, it is tempting to ascribe this knowledge gap to persistent disciplinary chauvinism within the natural and physical sciences (Giri, 2002). Qualitative, observational approaches that do not employ statistically testable hypotheses certainly do look non-scientific to some quantitative scientists. However, it is worth noting that debates over research methodology continue within social sciences, as is suggested by Comment 25. In fact, many social science disciplines have struggled internally with methodological conflicts for quite some time. More than 70 years ago, for example, White (1943) argued that quantitative methodology, although philosophically attractive, could not answer many of the most intriguing questions in sociology, and injudicious use of mathematical language to explore inherently non-mathematical concepts actually reduces understanding.
The importance of how researchers use hypotheses in discourse and practice is not merely of abstract or academic interest, but a critical aspect for evaluating CDR efforts (Klein, 2006). Narrow understanding of scientific processes and variation in methodology across disciplines can derail developing careers (NAS, 2004). Basic conflicts in methodology can reduce success in interdisciplinary education programs ; indeed, one of us participated in an NSF-IGERT project where the interdisciplinary component of the dissertation was nearly rejected because of its non-hypothesis-driven and non-statistical approach. Whether or not research is hypothesis driven can even constrain funding opportunities (NAS, 2004)-an obstacle addressed by a participant during this study: (35) Well, you see, in our case, with any kind of expensive research, you have to formulate some kind of hypothesis to get funding. (Session 1, Plant Science) Beyond the immediate and personal, restrictive concepts and definitions limit science as a way of knowing and discovering, constraining exploration of both basic and applied problems (Gallopin, Funtowicz, O'Connor, & Ravetz, 2001;Pickett, Burch, & Grove, 1999). Miller et al. (2008) invoked "epistemological pluralism" as a guiding principle for CDR efforts. This principle requires disciplinary participants to recognize the multiple ways of investigating a CDR problem, and requires research plans to accommodate multiple ways of knowing (Khagram et al., 2010).
One way of cultivating epistemological pluralism and reconciling epistemological difference is through the use of dialogue methods (Bammer, 2013), an approach instantiated by the Toolbox workshops. Although not an essential exercise-CDR is practiced by teams that are epistemically synchronized without formal discussion of these differences-the data we have collected in these workshops suggest that conversation that explicitly focuses on primary epistemological concepts, such as the hypothesis concept, could lead to valuable gains in understanding within a team, both of oneself and others. This nuanced form of scientific communication may be of special concern to research teams focused on producing professionally articulate and integrated products, particularly those hoping to generate novel answers to the type of complex problems interdisciplinary research is expected to address (Bammer, 2013;Daily & Ehrlich, 1999;Kinzig, 2001;Klein, 1990;NAS, 2004). Differences in philosophical assumptions may result in the collection of incompatible data or fundamental disagreement about data integration among collaborators who each believe they are employing valid approaches that will enable easy future analysis.
3. Engineers who share the opinion expressed in Comment 21 might be inclined to look at it this way, given the prevalence of model-based research in engineering, as indicated by Comment 29.