Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Success Factors of European Syndromic Surveillance Systems: A Worked Example of Applying Qualitative Comparative Analysis

  • Alexandra Ziemann ,

    alexandra.ziemann@maastrichtuniversity.nl

    Affiliation Department of International Health, CAPHRI School of Public Health and Primary Care, Faculty of Health, Medicine and Life Sciences, Maastricht University, Maastricht, The Netherlands

  • Anne Fouillet,

    Affiliation Institute de Veille Sanitaire, Saint Maurice cedex, France

  • Helmut Brand,

    Affiliation Department of International Health, CAPHRI School of Public Health and Primary Care, Faculty of Health, Medicine and Life Sciences, Maastricht University, Maastricht, The Netherlands

  • Thomas Krafft

    Affiliation Department of Health, Ethics and Society, CAPHRI School of Public Health and Primary Care, Faculty of Health, Medicine and Life Sciences, Maastricht University, Maastricht, The Netherlands

Abstract

Introduction

Syndromic surveillance aims at augmenting traditional public health surveillance with timely information. To gain a head start, it mainly analyses existing data such as from web searches or patient records. Despite the setup of many syndromic surveillance systems, there is still much doubt about the benefit of the approach. There are diverse interactions between performance indicators such as timeliness and various system characteristics. This makes the performance assessment of syndromic surveillance systems a complex endeavour. We assessed if the comparison of several syndromic surveillance systems through Qualitative Comparative Analysis helps to evaluate performance and identify key success factors.

Materials and Methods

We compiled case-based, mixed data on performance and characteristics of 19 syndromic surveillance systems in Europe from scientific and grey literature and from site visits. We identified success factors by applying crisp-set Qualitative Comparative Analysis. We focused on two main areas of syndromic surveillance application: seasonal influenza surveillance and situational awareness during different types of potentially health threatening events.

Results

We found that syndromic surveillance systems might detect the onset or peak of seasonal influenza earlier if they analyse non-clinical data sources. Timely situational awareness during different types of events is supported by an automated syndromic surveillance system capable of analysing multiple syndromes. To our surprise, the analysis of multiple data sources was no key success factor for situational awareness.

Conclusions

We suggest to consider these key success factors when designing or further developing syndromic surveillance systems. Qualitative Comparative Analysis helped interpreting complex, mixed data on small-N cases and resulted in concrete and practically relevant findings.

Introduction

Syndromic surveillance aims at augmenting traditional public health surveillance systems by providing (near) real-time information on the public health impact of events. To gain a head start, syndromic surveillance mainly analyses syndromes generated from existing data that originally were not collected for surveillance purposes [1]. Such data can come from web query logs, telephone helpline registries, or patient or veterinary records [2]. Syndromic surveillance is used for two main purposes. The first is timely infectious disease outbreak detection. The second purpose is near real-time situational awareness during events such as mass gatherings or extreme weather events [3]. Over the past 15 years, an increasing number of syndromic surveillance systems were set up in Europe and elsewhere. There is still doubt about the benefit of the approach [4]. A lack of clinical specificity, which can lead to false alerts and undetected events is considered the major weakness of syndromic surveillance [5]. The major advantages are timeliness, flexibility in using the approach for different types of health threats, and cost-effectiveness, because no additional data have to be collected [6, 7].

There are frameworks for evaluating syndromic surveillance systems, which propose a range of quantitative performance indicators such as timeliness and validity, and more qualitative indicators such as flexibility and acceptability [8, 9]. Evaluations to date mostly focus only on single, usually quantifiable performance indicators [1013]. This is likely due to lack of data or the effort to collect data for all the different performance indicators. Further, there are many case reports assessing one syndromic surveillance system at a time. There are only few comparative analyses or syntheses of several syndromic surveillance systems allowing for a generalizable assessment [7, 11, 14]. Such comparisons could unveil differences in performance and the impact of certain system characteristics on performance. Decision makers could use this information to design syndromic surveillance systems and improve their performance. Such comparative analyses can become a complex endeavour because different characteristics of a syndromic surveillance system can affect performance. These can be the analysed data source, the data collection, analysis and reporting process, and the purpose for and context in which the system is set up. The social science method Qualitative Comparative Analysis (QCA) offers the opportunity to analyse a combination of different system characteristics related to certain performance measures by analysing several case reports at the same time. We aimed at applying QCA for evaluating performance and identifying general success factors of syndromic surveillance systems in Europe.

Materials and Methods

Qualitative Comparative Analysis

The aim of QCA is to identify if certain conditions or combinations of conditions (so called configurations, in our case characteristics of syndromic surveillance systems) are part of an outcome set (in our case defined by the performance of syndromic surveillance systems). QCA is an approach and analysis method based on Boolean algebra. It allows for a systematic comparative analysis of small-N and especially case-based data which are not suited for statistical analysis such as regression analysis [15]. Ragin gives a good introduction to QCA in a short online presentation [16]. Until now, QCA was mainly applied in the social sciences, especially the political sciences. It is relatively new to the health sciences. Few studies used it to identify determinants of health policies or interventions but not for analysing topics in public health surveillance [1719].

QCA allows for analysing quantitative, qualitative or mixed data across several cases (in our case reports of syndromic surveillance system applications for a specific purpose). Depending on the characteristics of the data input, one can choose between two main QCA variants. Crisp set QCA (cs/QSA) analyses dichotomous data while fuzzy set QCA analyses ordinal or interval data or ratios. QCA is an iterative process, characterised by a dialogue between the results of the different steps of the QCA and the researcher’s case knowledge and expertise. The data input for the QCA is not fixed a priori but is likely to be adjusted in the process to increase validity of the results. Nevertheless, QCA follows a structured approach of predefined steps: (1) building a data table consisting of outcome indicators and conditions, (2) constructing a “truth table” consisting of configurations (combinations of the conditions and the outcome), (3) Boolean minimisation to reduce the complexity of the configurations to necessary or sufficient configurations known as solution terms, and (4) interpretation of the solution terms [20]. A key aspect of QCA is the inclusion of configurations without an observed outcome, so called logical remainders [20]. Thus, QCA does not only analyse the observed cases. It includes all logical possible configurations in the minimisation process to reduce the complexity of the solution terms. We found that cs/QCA is the suitable analysis method for our study because of our small-N, case-based and mixed data basis reflecting absence or presence of case characteristics. In the following, we present our setup of the QCA steps 1 to 3.

Data table: conceptual model

There are two main advantages of syndromic surveillance: it provides timelier information and information on events for which no other public health information is available. These virtues are mainly applied for two purposes: (1) timelier detection of seasonal infectious disease outbreaks, mainly influenza, and (2) real-time situational awareness during events with potential public health impact such as environmental threats or mass gatherings. The latter also includes reassurance that an event has no public health impact. We performed two cs/QCA to identify success factors of syndromic surveillance systems for these two major areas of application.

Data table: data basis

The data for this study was collected in the framework of the European project Triple S-AGE, which aimed at supporting a harmonised setup of syndromic surveillance systems across Europe [21]. Based on a literature review and an inventory, 60 European syndromic surveillance systems were identified [2, 22]. 36 systems in eight countries and one European consortium were selected for a site visit based on the extent of syndromic surveillance experience. Active, pilot, planned and expired syndromic surveillance systems were visited during nine site visits between June 2011 and June 2012. Data on the visited syndromic surveillance systems were collected through presentations and transcripts of discussions [23]. The 18 syndromic surveillance systems initially selected for this study were chosen based on two aspects: (1) the status of the system, that is if it is or was active and is not just a pilot or planned system AND (2) sufficient availability of data for the study, defined as results published in peer-reviewed journals and coverage during the site visits (Table 1). In February 2015, we updated the literature review purposefully on the selected systems searching in Google Scholar and PubMed and by hand-searching references (list of publications included in data collection available in S1 Appendix). Case-based data for the QCA were compiled from scientific and grey literature and the site visits. The visits provided additional information for the QCA beyond what was reported in scientific publications. In the course of the QCA, we chose to add another case from Germany to increase the quality of the QCA. This ad-hoc system was not considered for a site visit because the plan for the site visits was already fixed at the time the system was set up. All necessary data on this system were collected from scientific and grey literature [24, 25].

thumbnail
Table 1. Key characteristics of syndromic surveillance systems selected for QCA.

https://doi.org/10.1371/journal.pone.0155535.t001

Data table: outcome indicators

Syndromic influenza surveillance.

We anticipate a syndromic surveillance system to detect cases earlier if it analyses data that is collected earlier in the course of illness compared to data analysed by traditional surveillance systems. Further, we anticipate that a syndromic surveillance system detects cases earlier if it is based on pre-diagnostic clinical data such as chief complaints, in comparison to confirmed diagnoses [2, 11, 26]. The data basis was largest for seasonal influenza surveillance. Therefore, we focused on this area of application in the QCA. We had to define a cut-off for differentiating a successful from an unsuccessful case in the QCA. We chose this based on the data reported in the cases and based on the following theoretic consideration. Traditional influenza surveillance information based on sentinel general practitioner reports or laboratory confirmations is usually available on a weekly basis. We defined a successful case as a syndromic surveillance system indicating the onset or peak of the influenza season the week before a traditional surveillance system first indicated the same. We used the average timeliness per system if results were reported for several influenza seasons. The outcome indicator for the QCA (QCA coding: OUTCOME) was coded as 1 for successful cases if timeliness of case detection was less than 0 weeks. The outcome was coded as 0 for unsuccessful cases if the timeliness of case detection was 0 or more weeks. Data on syndromic influenza surveillance were available for nine syndromic surveillance systems in six countries (Table 2).

thumbnail
Table 2. Raw data for cs/QCA of seasonal syndromic influenza surveillance.

https://doi.org/10.1371/journal.pone.0155535.t002

Situational awareness.

We consider syndromic situational awareness as beneficial if a syndromic surveillance system can be applied to different potential public health threats (applicability) and if it provides rapid information (timeliness of reporting) [7, 27, 28]. We defined the cut-off for a successful outcome for applicability as prospective surveillance during two or more types of events. Four different event types can be differentiated: (1) environmental threats such as the volcanic ash plume 2010, heat waves, or floods, (2) the A/H1N1 pandemic 2009, (3) mass gatherings such as political summits or sporting events, and (4) industrial accidents. We chose the definition of this outcome indicator and its cut-off based on the reported applications in the cases. We combined the applicability outcome indicator with an indicator for timeliness. Here, we defined successful outcome as reporting of syndromic surveillance results of under 3 days. Reporting time referred to the onset of the health impact of an event or to the frequency of reporting. We used the average timeliness if data were reported for several events. Also this cut-off was derived based on the reported applications in the cases. If the two outcome indicators contradicted each other, we decided to weigh applicability higher than timeliness. As can be seen from the cases, most syndromic surveillance systems provide timely reports but it seems more difficult to apply a system to more than one type of event. Therefore, the outcome indicator for the QCA (OUTCOME) was coded as 1 for successful cases if the timeliness of reporting was under 3 days and/or the system was applied during two or more types of events. The outcome was coded as 0 for unsuccessful cases if timeliness was 3 or more days and/or the system was applied during only one type of event. Data on situational awareness were available on syndromic surveillance systems in nine countries (Table 3).

thumbnail
Table 3. Raw data for cs/QCA of syndromic situational awareness.

https://doi.org/10.1371/journal.pone.0155535.t003

Data table: conditions

Syndromic influenza surveillance.

Timeliness of seasonal influenza case detection can be influenced by the analysed data source. We expect non-clinical data sources (NONCLIN) and information collected prior to confirmed diagnoses from acute care data sources (ACUTE) to support timelier case detection. Non-clinical data sources comprised web searches and telephone helplines. Clinical sources referred to primary and acute care data sources. Acute care data sources were referring to emergency departments or out-of-ours general practitioner services. After a first cs/QCA analysis of these two conditions, a contradictory configuration occurred. That means that the same combination of conditions resulted in a positive and a negative outcome in the observed cases. The most frequently chosen action to resolve this contradiction is to add conditions [20]. We chose the analysis of population subgroups as additional condition. The analysis of different age groups can support timeliness of detecting cases (AGE). Furthermore, the analysis of small-scale, subnational syndromic surveillance data can support timeliness in comparison to traditional surveillance data referring to a higher administrative level (SUBNAT).

Situational awareness.

Factors influencing applicability of syndromic surveillance systems for various events are flexibility and acceptance of the system. A common experience shared during the site visits was that flexibility is supported by the analysis of multiple data sources (MULTDATA). We added the ability to generate multiple syndromes as another condition (MULTSYND). This was a logical deduction from the analysis of multiple data sources. Furthermore, free text in comparison to coded diagnostic information was reported as supporting factor (FREETEXT) [24]. Acceptability and timeliness of the system is supported by the collection of electronic compared to paper-based data (ELEC). We chose automation of data transfer and possibly data analysis and reporting as additional condition (AUTOM). Furthermore, the existence of a system before the occurrence of the event compared to systems that are set up ad-hoc during an event can support acceptability and timeliness of reporting (EXIST).

Quality of the data table

A limited variety and consistency can reduce the quality and informational value of the cs/QCA results [20]. In order to check the quality of the data table for the minimisation process, we analysed the variety of values across outcome, conditions and cases. Further, we analysed the consistency of the conditions for explaining a positive outcome. As a rule, there should be a mix of cases with a negative and a positive outcome. Furthermore, at least one third of the cases should represent a certain condition value. And, conditions should not have the same values across the cases. There are no rules for defining appropriate levels of consistency; this depends on the study. But, there is a general agreement about the lower level of consistency of 0.75 [29]. To increase quality of the data table, the selection of cases and conditions or the definition of conditions can be reconsidered [20].

Syndromic influenza surveillance.

The condition AGE did not show enough variety across all cases and we excluded it from the minimisation process. The consistency levels of the three conditions left for the minimisation process were between 0.43 and 0.57 (Table 4). There is a limitation to take into account when analysing the condition ACUTE. Only clinical data sources can also be acute data sources. The consistency level for ACUTE is rising to 0.75 when referring only to cases using clinical data sources instead of all cases. Despite the low consistency levels for the conditions NONCLIN and SUBNAT, we included the conditions in the analysis. We took the low consistency for these conditions into account in the interpretation of the results.

thumbnail
Table 4. Truth table for cs/QCA of seasonal syndromic influenza surveillance.

https://doi.org/10.1371/journal.pone.0155535.t004

Situational awareness.

For three conditions, MULTSYND, EXIST and ELEC, there was not enough variety over all cases and the conditions ELEC and EXIST were showing the same value pattern over all cases. We decided to exclude the condition ELEC from the minimisation process. Data are more and more electronically available. Thus, this factor will be of lesser relevance in the future compared to the factor of existence before an event. For increasing the variety of values across cases, we had no theoretic justification to exclude any of the cases or conditions or to reconsider the coding of the conditions. Instead, we decided to add another case. We chose the system from Germany, which was set up to monitor the O104:H4 outbreak in 2011. We have analysed this syndromic surveillance system in the framework of another study [28]. We knew that it was qualifying for the QCA and that we could retrieve all necessary data from the broad publication coverage of this case. The consistency levels of the conditions were between 0.75 and 1.00, except for FREETEXT for which it was only 0.5. Thus, we decided to exclude FREETEXT from the minimisation process (Table 5).

thumbnail
Table 5. Truth table for cs/QCA of syndromic situational awareness.

https://doi.org/10.1371/journal.pone.0155535.t005

Truth table and Boolean minimisation

We used the software TOSMANA Version 1.302 for accomplishing the cs/QCA steps of constructing the truth table and Boolean minimisation [30].

Syndromic influenza surveillance.

The truth table contained five configurations of the three conditions (Table 4). Six cases were combined into two groups of configurations while the other three had individual configurations. The first round of Boolean minimisation including logical remainders was based on contradictory simplifying assumptions. That means that a logical remainder was used to explain both positive and negative outcome at the same time. This was resolved by applying the following procedure suggested by Delreux and Hesters [31]. The contradictory simplifying assumption was assigned the likelier outcome of 1 according to case and theoretic knowledge. It was included as logical remainder in the minimisation process for successful cases. The contradictory simplifying assumption was excluded as logical remainder from the minimisation process of the less likely outcome of 0. Instead, it was included as additional case.

Situational awareness.

The truth table contained six configurations of the five conditions (Table 5). Three cases were combined in one configuration group while the other cases had individual configurations. The inclusion of logical remainders in the minimisation process did not result in any contradictions.

Results and Discussion

Results of the Qualitative Comparative Analysis

Syndromic influenza surveillance.

In more than 75% (n = 7) of the cases, syndromic surveillance systems detected the onset or peak of an influenza season earlier compared to traditional surveillance systems (Table 2). Timeliness ranged from -2.5 weeks in the SOS Medecins system to 2.0 weeks in the QSurveillance system with an average of -0.75 weeks and a median of -0.9 weeks. The minimisation process resulted in two different solution terms explaining each 57% (n = 4) and 43% (n = 3) of the successful cases. Further, it resulted in two solution terms explaining the unsuccessful cases, each explaining 50% (n = 1) of the cases (Table 6). According to these solutions, successful cases of syndromic surveillance systems for timely influenza case detection are analysing non-clinical data sources. Alternatively, they are analysing acute data sources in combination with applying subnational data analysis. In unsuccessful cases, systems are analysing clinical data sources in combination with either analysing non-acute care data or without applying subnational analysis. No solutions leading to successful cases are necessary as the outcome also occurred in the absence of the solutions. The solutions are sufficient as the outcome always occurred when the solutions are present. But, there are also other solutions leading to the outcome.

thumbnail
Table 6. Solution terms for explaining successful and unsuccessful cases of seasonal syndromic influenza surveillance.

https://doi.org/10.1371/journal.pone.0155535.t006

Situational awareness.

The systems covered different types of events ranging from one event in four cases to four events in two cases (Table 3). Reporting time for the six successful cases was around one day. The timeliness over all cases was very similar. For the two unsuccessful cases, reporting time was around 4 and 14 days. The minimisation process resulted in one solution term explaining all successful cases. Two solution terms explained 50% (n = 2) and 75% (n = 3) of the unsuccessful cases respectively (Table 7). According to the solutions, in successful cases of situational awareness, syndromic surveillance systems are analysing multiple syndromes and are automated. In unsuccessful cases, systems are analysing single syndromes and/or are not automated. The combination of analysing multiple syndromes in an automated system is a necessary condition as it was always present when the outcome occurred. Further, the outcome did not occur when this configuration was absent. The analysis of multiple data sources and the existence of the system before the event occurred were not identified as key influencing factors.

thumbnail
Table 7. Solution terms for explaining successful and unsuccessful cases of syndromic situational awareness.

https://doi.org/10.1371/journal.pone.0155535.t007

Interpretation

Syndromic influenza surveillance.

The low coverage levels of the QCA solution terms reflect the low consistency levels of the data input. They call for careful interpretation of the findings and indicate a limitation of their practical relevance. The solution term for successful cases confirmed our hypothesis that the analysis of non-clinical data can support timely syndromic influenza surveillance. The analysis of acute care data sources seems to be of lesser importance and only in combination with another condition such as subnational analysis.

Only cases featuring clinical data based systems were unsuccessful. In one of the two unsuccessful cases, QSurveillance, the system is based on data from general practitioners. In the course of illness or patient treatment, this data source is positioned closer to the traditional data sources of confirmed diagnoses from sentinel general practitioners or laboratories. Thus, compared to the non-clinical data sources, timeliness of general practitioner data sources must be lower. The timeliness of the second system, Oscour, which is based on emergency department data, was referring to two influenza seasons in 2005 and 2006. Since then, the system was used for influenza surveillance in France on a regular basis. It might very well have shown timeliness values in other seasons that were comparable to the other acute care data based systems that we analysed. We expected subnational data analysis to have an impact on the performance. According to the solution term, the analysis of age groups does not seem to influence timely case detection. That does not mean that age-adjusted analysis does not have an impact on timelier case detection. Such analysis might just not be applied in many systems. QCA says more about conditions in relation to each other than about the impact of single conditions [29].

When looking at applications of the analysed syndromic surveillance systems for other infectious disease outbreaks, the two QCA solutions for successful outcome also apply to those cases. For detection of seasonal gastrointestinal outbreaks, the NHS Direct system in England achieved a timeliness of -5.68 weeks on average using non-clinical data [32]. No results from subnational analysis were reported, but from analysis by age group. The syndromic surveillance system in Genoa monitored two measles outbreaks with a timeliness of -3.57 weeks on average based on acute care data [33, 34]. In this case, subnational surveillance was applied in one study and was connected to a 1.43 week advance detection [34]. No results from age-adjusted analysis were reported. We could only collect data on syndromic surveillance of other infectious disease outbreaks for these two systems. Therefore, these findings can only indicate a confirmation of our results.

Looking into other systems for syndromic seasonal influenza surveillance outside Europe, we can best look into a review by Dailey and colleagues [11]. They compared different systems and data sources for timeliness of syndromic seasonal influenza surveillance. The authors also found a higher timeliness of systems analysing non-clinical and acute care data sources. When looking into the single studies included in the review, there were no comparable non-clinical data based systems. Six systems were based on emergency department data. These are comparable to the acute care data based systems included in our study. One of these systems was associated with a negative outcome according to our definition. This system applied subnational analysis [35]. Another system applied subnational analysis and had a positive outcome [36]. Three other systems associated with positive outcome according to our definition did not report application of subnational analysis [3739]. One applied spatial cluster analysis but did not find any clusters associated with influenza [40]. One also applied age-adjusted analysis [39]. From the six systems, five only reported results for one influenza season, which reduces the representativeness of the results [3537, 39, 40]. We included cases of systems analysing emergency department data, which were all based on more than one season. The sixth study did not compare syndromic with traditional surveillance results, which limits the comparability to our results [38].

The comparison with other systems indicates that the two conditions concerning data sources might influence timeliness of syndromic influenza surveillance. The analysis of non-clinical data sources might have a greater positive influence than the analysis of acute data sources. Subnational analysis in combination with acute care data sources seems to be of lesser influence.

Situational awareness.

The possibility to analyse multiple syndromes and the use of automated systems were identified as key success factors, which is in line with our hypothesis. Even more, the combination of these factors turned out to be a necessary configuration for successful cases. This solution term explained all observed successful cases. The absence of one or both of these conditions was connected with unsuccessful cases of syndromic situational awareness. Against the expectations formulated during the site visits, the condition of analysing multiple data sources turned out not to be a success factor for flexible situational awareness. Multiple data sources were analysed in successful and unsuccessful cases. Further, in successful and unsuccessful cases, systems existed before the monitored event. However, in all successful cases, systems were established before the event. This indicates that this aspect could support syndromic situational awareness.

Systems outside Europe, which were only applied during one event, shared similar characteristics with the systems labelled as unsuccessful cases in our study. One example are systems that were set up in 2005 in the aftermath of hurricane Katrina in different cities in the United States of America. They analysed multiple databases and multiple syndromes on a daily basis. But, they were not existing before the event, and were based on manual data collection and analysis [41, 42]. Another example is the system set up for monitoring the Kentucky Derby in Louisville from 2002 until 2004. The system was neither automated nor established prior to the event (and terminated between events). But, it was monitoring multiple syndromes on a daily basis [43]. Also in these examples, the combination of multiple syndrome analysis and automation was absent as suggested by the QCA solution.

Turning to successful cases of systems outside Europe, we can look at the system in New York City, which analysed different types of events. Next to multiple other data sources, especially emergency department data were used for situational awareness [44]. The emergency department data based system part was set up ad-hoc as non-automated system in the days after the September 2011 terrorist attacks [45, 46]. The system was maintained by changing it into an automated system. The automated emergency department data based system was used to monitor multiple syndromes during the blackout in 2003 [47], the A/H1N1 pandemic 2009 [48, 49] and the effects of hurricane Sandy in 2012 [50]. Another system part, which was analysing multiple other data sources, was only used for monitoring one syndrome during the blackout in 2003 [47]. This system part seemed to be automated for certain data sources, but the information we could obtain from the literature was unambiguous. For at least one part of the New York City system, the combination of multiple syndrome analysis in an automated system supported situational awareness. This example also confirms that the analysis of multiple data sources is not a key success factor.

Limitations

The low quality of the data table and solution terms of the QCA regarding syndromic influenza surveillance limits the quality and practical relevance of the findings. We tried to include representative data providing results for more than one influenza season. We decided to include cases of two systems despite results referred only to one season: NHS24 in Scotland and QSurveillance in England. NHS24 was applied in more than one influenza season and reported timeliness was positive but not explicitly quantified [51]. As QSurveillance was the only case representing a primary care data based system, we decided to include it to add variety to the QCA. The QSurveillance system was reportedly changed, so performance could also have changed [52]. Nevertheless, the QCA results might not be representative for other syndromic surveillance systems based on general practitioner data.

The conditions analysed in the QCA were chosen based on theoretic considerations but also based on the available data. There are other conditions such as validity or representativeness, which might alter our QCA results. But, we could not include them due to limited data availability, or in the case of validity, because of the large diversity in measuring this indicator in different surveillance systems.

QCA makes it necessary to differentiate successful from unsuccessful cases. We defined the cut-off points for both analyses based on the results reported for the cases and theoretic considerations. We could only include few unsuccessful cases with a limited representativeness in our QCA on syndromic influenza surveillance. This might have limited our options to define a representative cut-off. It is important to highlight that the distinction of successful or unsuccessful cases should not be generalised to the syndromic surveillance systems behind these cases. The syndromic surveillance systems might perform differently when they are monitoring other syndromes, for example.

Finally, the data in our study were not collected with QCA in mind. We chose the method afterwards to add value to our complex case-based data. With QCA in mind, we might have collected different or additional data, which could have improved quality of the input and output of the QCA.

Practical relevance

For the first time, QCA allowed a structured, comparative analysis of complex, small-N case reports of syndromic surveillance system performance. The identified success factors can inform the process of designing or further developing a syndromic surveillance system for situational awareness or seasonal influenza surveillance. The success factors could be used to prioritise certain system characteristics over others for improving performance. This especially applies to the system characteristics included in our analysis. For example, a syndromic influenza surveillance system might detect cases earlier rather by analysing non-clinical data than clinical data. Further, the choice of data source seems to influence timely influenza surveillance more than the analysis of subnational level data or age groups. Timely syndromic situational awareness during different types of events is supported by analysing multiple syndromes in an automated system. This combination of system characteristics is of greater influence than the analysis of multiple data sources or the existence of the syndromic surveillance system before the occurrence of an event.

The relatively low coverage levels of the QCA results suggest that the success factors regarding syndromic influenza surveillance should be applied with caution. The relatively high coverage level of the QCA results for syndromic situational awareness suggests a higher reliability of these findings. Any changes in specific syndromic surveillance systems should be based on comprehensive test runs using historical data.

Conclusions

We identified key success factors for the two main areas of syndromic surveillance system application using cs/QCA. For syndromic influenza surveillance, a system might be timelier if analysing non-clinical data sources. Syndromic situational awareness is supported by automated syndromic surveillance systems capable of analysing multiple syndromes. Analysing multiple data sources is no pre-requisite for flexible situational awareness. We suggest to consider the success factors when designing or further developing a syndromic surveillance system.

We showed that the social science analysis method QCA can add value for the interpretation of case-based, small-N and mixed data in the area of public health surveillance. QCA can only produce valid results if guided by good case and subject expertise and if based on varied and representative data. We propose to apply QCA to other case-based and small-N data next to more traditional analysis methods. It might yield meaningful and relevant results for policy and practice that would otherwise remain undiscovered.

Supporting Information

S1 Appendix. List of publications included in data collection.

https://doi.org/10.1371/journal.pone.0155535.s001

(PDF)

Acknowledgments

We would like to thank the anonymous reviewer for the valuable comments on an earlier version of this article.

Author Contributions

Conceived and designed the experiments: AZ. Performed the experiments: AZ. Analyzed the data: AZ AF HB TK. Wrote the paper: AZ AF HB TK.

References

  1. 1. Triple S Project. Assessment of syndromic surveillance in Europe. Lancet. 2011;378(9806):1833–4. pmid:22118433
  2. 2. Ziemann A, Krafft T. Guidelines for Assessment of Data Sources. 2013. Available: http://www.syndromicsurveillance.eu/triple-s_guidelines_datasources.pdf. Accessed 28 May 2015.
  3. 3. Fouillet A, Sala-Soler M, Conti S, Kanieff M, Rago G, Perrin JB, et al. Inventory of syndromic surveillance systems in Europe by the Triple-S project. Emerg Health Threats J. 2011;4:10946.
  4. 4. Koopmans M. Surveillance strategy for early detection of unusual infectious disease events. Curr Opin Virol. 2013;3(2):185–91. pmid:23612329
  5. 5. Kaydos-Daniels SC, Rojas Smith L, Farris TR. Biosurveillance in outbreak investigations. Biosecur Bioterror. 2013;11(1):20–8. pmid:23448272
  6. 6. Buehler JW, Whitney EA, Smith D, Prietula MJ, Stanton SH, Isakov AP. Situational uses of syndromic surveillance. Biosecur Bioterror. 2009;7(2):165–77. pmid:19635001
  7. 7. Paterson BJ, Durrheim DN. The remarkable adaptability of syndromic surveillance to meet public health needs. J Epidemiol Glob Health. 2013;3(1):41–7. pmid:23856537
  8. 8. Buehler JW, Hopkins RS, Overhage JM, Sosin DM, Tong V. Framework for evaluating public health surveillance systems for early detection of outbreaks: recommendations from the CDC Working Group. MMWR Recomm Rep. 2004;53(RR-5):1–11. pmid:15129191
  9. 9. Triple S Project. Guidelines for designing and implementing a syndromic surveillance system. Triple S Project; 2013. Available: http://www.syndromicsurveillance.eu/Triple-S_guidelines.pdf. Accessed: 28 May 2015.
  10. 10. Rosenkotter N, Ziemann A, Riesgo LG, Gillet JB, Vergeiner G, Krafft T, et al. Validity and timeliness of syndromic influenza surveillance during the autumn/winter wave of A(H1N1) influenza 2009: results of emergency medical dispatch, ambulance and emergency department data from three European regions. BMC Public Health. 2013;13(1):905.
  11. 11. Dailey L, Watkins RE, Plant AJ. Timeliness of data sources used for influenza surveillance. J Am Med Inform Assoc. 2007;14(5):626–31. pmid:17600101
  12. 12. Doroshenko A, Cooper D, Smith G, Gerard E, Chinemana F, Verlander N, et al. Evaluation of syndromic surveillance based on National Health Service Direct derived data—England and Wales. MMWR Morb Mortal Wkly Rep. 2005;54 Suppl:117–22.
  13. 13. Jefferson H, Dupuy B, Chaudet H, Texier G, Green A, Barnish G, et al. Evaluation of a syndromic surveillance for the early detection of outbreaks among military personnel in a tropical country. J Public Health (Oxf). 2008;30(4):375–83.
  14. 14. Buckeridge DL. Outbreak detection through automated surveillance: a review of the determinants of detection. J Biomed Inform. 2007;40(4):370–9. pmid:17095301
  15. 15. Ragin CC. Using qualitative comparative analysis to study causal complexity. Health Serv Res. 1999;34(5 Pt 2):1225–39. pmid:10591281
  16. 16. Ragin CC. What is Qualitative Comparative Analysis (QCA)? 2008. Available: http://eprints.ncrm.ac.uk/250/1/What_is_QCA.pdf. Accessed 28 May 2015.
  17. 17. Blackman T, Wistow J, Byrne D. A Qualitative Comparative Analysis of factors associated with trends in narrowing health inequalities in England. Soc Sci Med. 2011;72(12):1965–74. pmid:21640455
  18. 18. Thomas J, O'Mara-Eves A, Brunton G. Using qualitative comparative analysis (QCA) in systematic reviews of complex interventions: a worked example. Systematic reviews. 2014;3:67. pmid:24950727
  19. 19. Warren J, Wistow J, Bambra C. Applying qualitative comparative analysis (QCA) in public health: a case study of a health improvement service for long-term incapacity benefit recipients. J Public Health (Oxf). 2014;36(1):126–33.
  20. 20. Rihoux B, De Meur G. Crisp-Set Qualitative Comparative Analysis (csQCA). In: Rihoux B, Ragin CC, editors. Configurational Comparative Methods: Qualitative Comparative Analysis (QCA) and Related Techniques. Thousand Oaks: SAGE; 2009. p. 33–68.
  21. 21. Triple S-AGE [Internet]; 2015. Available: www.syndromicsurveillance.eu. Accessed: 28 May 2015.
  22. 22. Conti S, Kanieff M, Rago G, on behalf of the Triple-S project. Inventory of Syndromic Surveillance Systems in Europe. Triple S-AGE project; 2012. Available: http://www.syndromicsurveillance.eu/images/stories/Final_material/triple-S_country_visits.pdf. Accessed: 28 May 2015.
  23. 23. Ziemann A, Krafft T, Sala Soler M, Sypniewska P. Country visits. Triple S-AGE project; 2013. Available: http://syndromicsurveillance.eu/images/stories/Final_material/triple-S_country_visits.pdf. Accessed: 28 May 2015.
  24. 24. Robert Koch Institute. Final presentation and evaluation of epidemiological findings in the EHEC O104:H4 Outbreak, Germany 2011. Berlin: Robert Koch-Institute; 2011.
  25. 25. Wadl M, Rieck T, Nachtnebel M, Greutelaers B, an der Heiden M, Altmann D, et al. Enhanced surveillance during a large outbreak of bloody diarrhoea and haemolytic uraemic syndrome caused by Shiga toxin/verotoxin-producing Escherichia coli in Germany, May to June 2011. Euro Surveill. 2011;16(24). pii: 19893.
  26. 26. Mandl KD, Overhage JM, Wagner MM, Lober WB, Sebastiani P, Mostashari F, et al. Implementing syndromic surveillance: a practical guide informed by the early experience. J Am Med Inform Assoc. 2004;11(2):141–50. pmid:14633933
  27. 27. Rosenkötter N, Ziemann A, Krafft T, Riesgo LG, Vergeiner G, Brand H. Non-infectious events under the International Health Regulations (2005) in Europe—a case for syndromic surveillance. J Public Health Policy. 2014;35(3):311–26. pmid:24804952
  28. 28. Ziemann A, Rosenkötter N, Riesgo L Garcia-Castrillo, Fischer M, Krämer A, Lippert FK, et al. Meeting the International Health Regulations (2005) surveillance core capacity requirements at the subnational level in Europe: the added value of syndromic surveillance. BMC Public Health. 2015 Feb 7;15:107. pmid:25879869
  29. 29. Schneider C, Wagemann C. Standards of Good Practice in Qualitative Comparative Analysis (QCA) and Fuzzy-Sets. Comparative Sociology. 2010;9:397–418.
  30. 30. Cronqvist L. Tool for Small-N Analysis (Version 1.302). Trier: Universität Trier; 2011. Available: http://www.compasss.org/software.htm#tosmana.
  31. 31. Delreux T, Hesters D. Solving contradictory simplifying assumptions in QCA: presentation of a new best practice. 2010;58:1–27. Available: http://www.compasss.org/wpseries/DelreuxHesters2010.pdf. Accessed 28 May 2015.
  32. 32. Loveridge P, Cooper D, Elliot AJ, Harris J, Gray J, Large S, et al. Vomiting calls to NHS Direct provide an early warning of norovirus outbreaks in hospitals. J Hosp Infect. 2010;74(4):385–93. pmid:20172625
  33. 33. Ansaldi F, Orsi A, Altomonte F, Bertone G, Parodi V, Carloni R, et al. Syndrome surveillance and molecular epidemiology for early detection and tracing of an outbreak of measles in Liguria, Italy. J Med Virol. 2009;81(10):1807–13. pmid:19697412
  34. 34. Orsi A, Alicino C, Patria AG, Parodi V, Carloni R, Turello V, et al. Epidemiological and molecular approaches for management of a measles outbreak in Liguria, Italy. J Prev Med Hyg. 2010;51(2):67–72. pmid:21155408
  35. 35. Ritzwoller DP, Kleinman K, Palen T, Abrams A, Kaferly J, Yih W, et al. Comparison of syndromic surveillance and a sentinel provider system in detecting an influenza outbreak—Denver, Colorado, 2003. MMWR Morb Mortal Wkly Rep. 2005;54 Suppl:151–6.
  36. 36. Yuan CM, Love S, Wilson M. Syndromic surveillance at hospital emergency departments—southeastern Virginia. MMWR Morb Mortal Wkly Rep. 2004;53 Suppl:56–8.
  37. 37. Irvin CB, Nouhan PP, Rice K. Syndromic analysis of computerized emergency department patients' chief complaints: an opportunity for bioterrorism and influenza surveillance. Ann Emerg Med. 2003;41(4):447–52. pmid:12658241
  38. 38. Ivanov O, Gesteland PH, Hogan W, Mundorff MB, Wagner MM. Detection of pediatric respiratory and gastrointestinal outbreaks from free-text chief complaints. AMIA Annual Symposium proceedings / AMIA Symposium AMIA Symposium. 2003:318–22. pmid:14728186
  39. 39. Miller B, Kassenborg H, Dunsmuir W, Griffith J, Hadidi M, Nordin JD, et al. Syndromic surveillance for influenzalike illness in ambulatory care network. Emerg Infect Dis. 2004;10(10):1806–11. pmid:15504267
  40. 40. Heffernan R, Mostashari F, Das D, Karpati A, Kulldorff M, Weiss D. Syndromic surveillance in public health practice, New York City. Emerg Infect Dis. 2004;10(5):858–64. pmid:15200820
  41. 41. Murray KO, Kilborn C, DesVignes-Kendrick M, Koers E, Page V, Selwyn BJ, et al. Emerging disease syndromic surveillance for Hurricane Katrina evacuees seeking shelter in Houston's Astrodome and Reliant Park Complex. Public Health Rep. 2009;124(3):364–71. pmid:19445411
  42. 42. Centers for Disease Control and Prevention. Injury and illness surveillance in hospitals and acute-care facilities after Hurricanes Katrina And Rita—New Orleans area, Louisiana, September 25-October 15, 2005. MMWR Morb Mortal Wkly Rep. 2006;55(2):35–8. pmid:16424856
  43. 43. Carrico R, Goss L. Syndromic surveillance: hospital emergency department participation during the Kentucky Derby Festival. Disaster Manag Response. 2005;3(3):73–9. pmid:15986027
  44. 44. Heffernan R, Mostashari F, Das D, Besculides M, Rodriguez C, Greenko J, et al. New York City syndromic surveillance systems. MMWR Morb Mortal Wkly Rep. 2004;53 Suppl:23–7.
  45. 45. Centers for Disease Control and Prevention. Syndromic surveillance for bioterrorism following the attacks on the World Trade Center—New York City, 2001. MMWR Morb Mortal Wkly Rep. 2002;51 Spec No:13–5. pmid:12238536
  46. 46. Das D, Weiss D, Mostashari F, Treadwell T, McQuiston J, Hutwagner L, et al. Enhanced drop-in syndromic surveillance in New York City following September 11, 2001. J Urban Health. 2003;80(2 Suppl 1):i76–88. pmid:12791782
  47. 47. Marx MA, Rodriguez CV, Greenko J, Das D, Heffernan R, Karpati AM, et al. Diarrheal illness detected through syndromic surveillance after a massive power outage: New York City, August 2003. Am J Public Health. 2006;96(3):547–53. pmid:16380562
  48. 48. Plagianos MG, Wu WY, McCullough C, Paladini M, Lurio J, Buck MD, et al. Syndromic surveillance during pandemic (H1N1) 2009 outbreak, New York, New York, USA. Emerg Infect Dis. 2011;17(9):1724–6. pmid:21888804
  49. 49. Hadler JL, Konty K, McVeigh KH, Fine A, Eisenhower D, Kerker B, et al. Case fatality rates based on population estimates of influenza-like illness due to novel H1N1 influenza: New York City, May-June 2009. PLoS One. 2010;5(7):e11677. pmid:20657738
  50. 50. Chen BC, Shawn LK, Connors NJ, Wheeler K, Williams N, Hoffman RS, et al. Carbon monoxide exposures in New York City following Hurricane Sandy in 2012. Clin Toxicol (Phila). 2013;51(9):879–85.
  51. 51. Kavanagh K, Robertson C, Murdoch H, Crooks G, McMenamin J. Syndromic surveillance of influenza-like illness in Scotland during the influenza A H1N1v pandemic and beyond. J R Stat Soc a Stat. 2012;175:939–58.
  52. 52. Green HK, Zhao H, Boddington NL, Andrews N, Durnall H, Elliot AJ, et al. Detection of varying influenza circulation within England in 2012/13: informing antiviral prescription and public health response. J Public Health (Oxf). 2015;37(2):295–304.