Introduction

Constitutional genetic information is constant throughout life; patients are typically tested once in their lifetime and the test may never be repeated or confirmed. Consequently, incorrect results may not become apparent for years and can lead to harmful outcomes for patients and family members, such as unnecessary surgery or prenatal diagnosis, or erroneous carrier testing. Given the complex and sensitive nature of genetic information, testing services have the responsibility to accurately test and interpret genetic results in the context of the clinical indication and family history, while taking into account that the recipient of the laboratory report may not have genetic expertise.

In the 1990s, a system known as external quality assessment (EQA, in Europe) or proficiency testing (PT, in the United States) was introduced in molecular genetic diagnostic laboratories to evaluate the ability to accurately determine genotypes and to assess the accuracy and completeness of the interpretation of the test results.1 Independent assessment of test performance is a recommendation of the Organisation for Economic Cooperation and Development, and a requirement for laboratory accreditation to the International Organisation for Standardisation (ISO) 15 189 and 17 025 standards and for complying with the US Clinical Laboratory Improvement Amendments (CLIA).2, 3, 4, 5 One of the main European EQA providers in molecular genetics is the Cystic Fibrosis Network, which has been organising an EQA scheme for cystic fibrosis (CF; OMIM no. 219700) since 1996.6

Assessment of reports through the Cystic Fibrosis Network and other disease-specific schemes organised by the European Molecular Genetics Quality Network revealed that there is room for improvement of the interpretation in laboratory reports. For example, the EQA scheme for molecular diagnosis of hereditary recurrent fevers described that only 8–47% of the reports mentioned interpretation of testing results (15–50 participants, 2006–2008).7 Similarly, the spinocerebellar ataxias EQA scheme reported average interpretation scores of 1.53/2.00 and 1.76/2.00 (26–37 participants, 2005–2006) and the scheme for mutation detection in the breast cancer genes revealed average interpretation scores from 1.46/2.00 to 1.78/2.00 (25–41 participants, 2000–2002).8, 9 The maximum interpretation score in these schemes was 2.0 per laboratory; the average interpretation scores represented the average of interpretation scores of all participating laboratories in a certain year.

Although EQA is accepted as a corner stone of quality assurance, direct evidence that participation in EQA leads to improved interpretation is lacking because measuring improvement of interpretation is challenging.10, 11 First, interpretation depends heavily on the complexity of the mock clinical cases, which change every year.12 Second, published interpretation scores represent the mean annual score for the whole group of participants, whose members vary from year to year, with a mix of regular, irregular and new participants.

This large retrospective longitudinal study was performed through the Cystic Fibrosis Network over a 6-year period among more than 300 laboratories. We examined and analysed changes in the completeness of interpretation in clinical laboratory reports when similar versions of mock clinical cases were distributed and thus when the influence of the complexity of the clinical question was eliminated. In addition, we investigated the effect of the number of previous EQA participations on the completeness of the interpretation and thus evaluated the efficacy of EQA. Monitoring and improving the completeness of interpretation in clinical reports is essential as it will lead to better quality of care for patients and families.

Materials and methods

Context and setting of the study

In a standard CF EQA round (which takes 1 year, starting from autumn until summer the year after), a set of samples (purified DNA) and mock cases for closely related clinical situations are sent for analysis to voluntarily registered laboratories. Laboratory reports submitted by the participants to the EQA provider should reflect routine reports issued to requesting physicians. The assessors determine the key interpretation elements that should be present in participants’ laboratory reports before dispatching the samples. Their decisions are based on expert consensus and best practice guidelines, and depend on the clinical indication in the case.13, 14, 15, 16 Two assessors independently evaluate genotypes and interpretation in the submitted reports. Results are also discussed during an assessment meeting. Apart from laboratory reports, the CF Network requests information on the accreditation and certification status of the laboratory, the laboratory setting (eg, hospital, university, industry) and the number of CF samples tested per year (laboratory test volume). At the end of the EQA round, participants receive feedback in the form of laboratory-specific individual comments and a general report addressed to all participants. Laboratory information and performance data have been stored systematically in a database since 2004, using a unique identification code for each laboratory.

Study design

Over a 6-year period, we have intentionally included similar versions of two mock cases, presenting clinical situations requiring essentially the same interpretation, masked by changes to the case descriptions (patients’ names, genders, ethnic origins and dates of birth) and the genotypes: three versions of a patient with CF in 2004, 2008 and 2009, named case 1, and three versions of a CF carrier in 2006, 2008 and 2009, named case 2 (Table 1). The genotypes were chosen not to affect the evaluation. The laboratories were not informed about the similarity of the cases, nor that we were conducting this study.

Table 1 Overview of the similar versions of two cases, sent to the laboratories that participated in the CF EQA scheme in 2004 or 2006, in 2008 and in 2009

The interpretation elements evaluated in the clinical laboratory reports for both cases in the study are outlined in Table 2. A laboratory received marks when an interpretation element was present and correct in the report, or when an element was present but not clearly defined (eg, a laboratory recommends genetic counselling without specifying whom it may concern). No marks were awarded when the element was absent or if the element was present, but wrong (eg, a wrong risk figure). The sum of the marks (further referred to as the ‘interpretation score’) is maximum 2.0 per case.

Table 2 Overview of the interpretation elements evaluated in the clinical laboratory reports for both cases in the study

We collected the stored performance data for these specific similar cases only at the end of the 6-year period and then conducted a longitudinal retrospective analysis. If participants provided laboratory information (accreditation/certification status, setting and test volume) in different years, only the most recently submitted information was used in this study.

Laboratories that registered for the EQA scheme but did not submit laboratory reports were not taken into account in the study (eight laboratories in 2004, five in 2006, four in 2008 and six in 2009). We also excluded those reports in which a genotyping error was made, because interpretation was not assessed the same way (one laboratory in 2004, four in 2008, five in 2009 for case 1, and six in 2006, three in 2008, seven in 2009 for case 2). Furthermore, we excluded reports from laboratories that did not report a certain mutation because they had not tested for it (two laboratories in 2006, three in 2008, eight in 2009 for case 1, and one in 2008, one in 2009 for case 2).

Statistical analysis

The interpretation score for a laboratory comprised values between 0.0 and 2.0 with nine possible levels. Given the very skew distribution of the score (higher scores are more frequent), we adopted a proportional odds model (odds ratios (OR) and 95% confidence intervals (CI)) to analyse the effect of the number of previous EQA participations, the effect of the year, the laboratory accreditation/certification status, setting and test volume. The proportional odds model considered the interpretation score as an ordered category response and modelled the probability that the response fell within the higher categories (equivalent to a higher score) versus the lower categories (lower score). A P-value <0.05 was considered significant. Generalised estimating equations were used for estimation to account for clustering of observations coming from the same laboratory.

The specific interpretation elements on which the laboratory reports were evaluated were given a binary score (0/1; element absent or element wrong/element present and correct or present but not clearly defined). The effect of the number of previous participations on the success probability on each aspect were analysed using a logistic regression model and using generalised estimating equations to take into account the clustering of the scores due to the longitudinal data structure.

All analyses have been performed using SAS software, version 9.2 of the SAS System for Windows (SAS Institute Inc., Cary, NC, USA).

Results

Study population

Data regarding the characteristics of the 311 laboratories (272 European), from 39 different countries included in the study are described in Table 3.

Table 3 Characteristics of 311 laboratories, from 39 countries included in the study

All these laboratories offered testing for cystic fibrosis transmembrane conductance regulator (CFTR) gene mutations as a clinical test. We assessed the interpretation in clinical reports from 298 different laboratories for case 1 (n=232 in 2004, n=202 in 2008, n=203 in 2009) and from 263 different laboratories for case 2 (n=200 in 2006, n=205 in 2008, n=208 in 2009). Roughly half of the laboratories provided reports that could be analysed in each of the 3 investigated years: 44% (132/298) for case 1 and 56% (147/263) for case 2. About one-third of the laboratories provided reports in only one of the three investigated years: 31% (91/298) for case 1 and 23% (60/263) for case 2.

Trend of the mean annual interpretation score

First, we calculated the mean annual interpretation score for the whole group of participants for both cases (black rounds in Figure 1). In general, scores improved over the years. A significant difference was found between the two cases with better interpretation scores for case 1 (P=0.0007; OR: 1.41; CI: 1.16–1.72). For case 1 in 2004, only 20% (46/232) of the laboratories got a maximum interpretation score of 2.0, compared with 66% (134/203) in 2009. A similar trend was observed for case 2: 31% (61/200) got a maximum score of 2.0 in 2006 and 61% (126/208) in 2009. We then analysed whether the laboratories that participated more frequently in the CF EQA scheme had better scores than those that participated less frequently. The score for the laboratories that participated three times or only once during the investigated period are indicated with grey symbols in Figure 1. We observed a significant positive effect of the number of previous participations on the interpretation score, both for case 1 (P<0.0001; OR: 2.34; CI: 1.60–3.44) and for case 2 (P<0.0001; OR: 2.64; CI: 1.85–3.76).

Figure 1
figure 1

Trend of the interpretation scores for different groups for case 1 (2004, 2008 and 2009) and case 2 (2006, 2008 and 2009).

We analysed the effect of the year on the interpretation score and corrected for the number of previous participations. A lower score was observed for 2004 compared with 2008 in case 1 (P<0.0001; OR: 0.43; CI: 0.28–0.65), and a lower score was observed in 2009 compared with 2008 in case 2 (P=0.017; OR: 1.67; CI: 1.10–2.54). No effect of the year was observed between 2008 and 2009 in case 1, and between 2006 and 2008 in case 2.

Trend of the presence of specific interpretation elements

We visualised the trend of the presence or absence of the specific interpretation elements for both cases between 2004 and 2009 (Figure 2). The presence of all elements increased steadily over the years, except ‘confirmation of CF carrier status’, which decreased in 2009 (86%, 178/208) compared with 2008 (97%, 199/205) and ‘genetic counselling for the couple’ for case 2, which decreased slightly in 2008 (85%, 175/205) compared with 2006 (88%, 175/200). Note that we accepted both ‘elements present’ and ‘elements present, but not clearly identified’ as ‘present’ in this analysis. The presence of interpretation element ‘need for qualification of the genotype’ (that is, study the parents to confirm homozygosity or compound heterozygosity) increased most, from 49% (113/232) in 2004 to 93% (189/203) in 2009. ‘Cascade screening for relatives’ increased also more than 30% (from 48% (111/232) to 84% (171/203) in case 1 and from 46% (91/200) to 78% (162/208) in case 2). At the end of the study, two elements were still relatively absent in the clinical reports: the element ‘cascade screening in relatives’, as mentioned just above, and the element ‘offer or suggest prenatal diagnosis or preimplantation genetic diagnosis for next pregnancy’, which was omitted from 24% (49/203) of the reports in 2009 (46% (107/232) in 2004). As with the interpretation scores, we observed a positive effect of the number of previous participations on the presence of all specific interpretation elements, except for ‘confirmation of the CF carrier status’ (P=0.47) in case 2 (Table 4).

Figure 2
figure 2

Presence or absence of specific interpretation elements, in the laboratory reports for case 1 (2004, 2008 and 2009) and case 2 (2006, 2008 and 2009).

Table 4 Effect of the number of previous EQA participations on the presence of specific interpretation elements

Effect of the laboratory accreditation or certification status, setting and test volume on the interpretation

We evaluated whether laboratories that were accredited or certified had more complete interpretation than those that were not accredited or certified. In addition, we evaluated whether a certain setting was associated with better interpretation and whether laboratories that test more CF samples per year had better scores than those that test fewer samples per year. No effect was observed of the laboratories’ accreditation/certification status on the interpretation scores (P=0.18 for case 1, P=0.27 for case 2). However, within the group of accredited and certified laboratories, there was an effect of the number of years since the accreditation or certification. We found lower scores for laboratories that had been accredited or certified for a longer time for case 1 (P=0.0004; OR: 0.90; CI: 0.85–0.95) and case 2 (P=0.005; OR: 0.93; CI: 0.89–0.98). For example, the 2009 mean annual interpretation score for case 1 was 1.77/2.00 for the laboratories that had achieved accreditation before 2004, 1.83/2.00 for the laboratories accredited between 2004 and 2007 and 1.90/2.00 for the laboratories accredited since 2008. The outcome was similar when the analysis was restricted to accredited laboratories. No differences in interpretation score were found between the three different categories of laboratory settings (P=0.99 for case 1, P=0.63 for case 2) and between the four different categories of laboratory test volumes (P=0.71 for case 1, P=0.2 for case 2).

Discussion

Principal findings

First, this study with similar mock clinical cases revealed that regular participation in EQA leads to a more complete interpretation in laboratory reports. Our observations support the view that EQA participation is a form of continuous education, whereby a comprehensive EQA general report and individual comments contribute to the improvement observed. Interpretation could additionally be influenced by external factors such as the availability of best practice guidelines and publications.13, 14, 15

Second, the presence in clinical reports of all investigated interpretation elements improved over the years. Two elements increased remarkably more (>30%) than others: stating the need to test the parents of a CF patient to confirm homozygosity or compound heterozygosity (need for qualification of the genotype) and mentioning the possibility of testing the relatives. The fact that most laboratories currently mention that genotypes must be confirmed represents an important improvement in laboratory practice and a reduced probability of error or misinterpretation, and shows that laboratories better understand the limitations of genetic analysis, especially PCR-based tests.

Third, although all investigated interpretation elements were included in the majority (>80%) of reports by the end of the study, there was room for improvement for those aspects that concern offering testing for familial mutations in relatives or future pregnancies. The possibility of testing relatives should be mentioned in all cases in which an individual is positive for a CF-causing mutation. It is also recommended that laboratories mention the possibility of offering prenatal diagnosis for future pregnancies in all cases where a child was born with CF, as the risk of a future child with CF is one-fourth in each pregnancy. Opportunities for improvement were noticed more for case 2 than for case 1. This may be because laboratories interpret test results of patients with CF more assiduously than results of CF carriers. Nevertheless, in either case, genetic test reports can be referred to at a later date for relatives and future generations, and both situations must therefore be treated with the same assiduity.

Finally, no effect was observed of the laboratory accreditation/certification status on the completeness of the interpretation. Laboratories that are not accredited, but that perform EQA, reach an equal quality of their clinical reports compared with accredited or certified laboratories, confirming EQA participation as a fundamental element of laboratory quality assurance. The reason why recently accredited/certified (accredited) laboratories performed significantly better compared with those with more years of accreditation/certification (accreditation) is unclear. There might be a chance that recently accredited laboratories pay more attention to the improvement of their reports as they still have regular audits, whereas laboratories that have their quality management system in place for a long time, may become complacent. They perhaps just maintain their systems, with less attention to major adaptations.

Strengths and limitations

Our study has several important strengths. First, this is the largest longitudinal expert peer review of interpretation in genetic reporting for a single disorder. We were able to take advantage of the unique opportunity of having a large number of laboratories involved, from Europe, the United States and Australasia, principally because CF is a very commonly tested disorder compared with other genetic diseases. To evaluate the proportion of CFTR-testing laboratories in this study, we consulted Orphanet and estimated that 75% (272/361) of listed CFTR-testing laboratories were included in our study.17 Second, this is the first study that intended to overcome the influence of the difficulty of the clinical case on the interpretation score by including similar cases over a relatively long period. Further, difference in interpretation scores were analysed statistically. Published results of other EQA schemes have mainly included data from EQA pilot years, without any statistical evidence that the interpretation improved.7, 8, 9 It is known that assessment criteria and procedures are being set up in the pilot years and this could have a considerable impact on the trend of the interpretation score. In comparison, the CF EQA scheme was set up 8 years before this study and with a fixed group of assessors over the years.

Although this study provides important data on the impact of the number of previous participations on the interpretation in laboratory reports, it has potential limitations. Firstly, only reports for the EQA scheme were evaluated, and it is not possible to judge whether the laboratories’ real practice has improved, or if tailored reports are prepared to ‘satisfy’ the EQA provider’. Secondly, although we have eliminated the influence of the variety of the clinical question, there might be a variety in the strictness of assessment. Even though the assessors group did not change over the years, it is likely that assessment was stricter year after year, thus reducing interpretation scores at later years. There might be a chance that the group of laboratories that participated only one time during the investigated period experienced a greater effect of the increase of strictness, compared with those that participated more times (cfr. decrease of the mean annual interpretation score in Figure 1 for the group that participated one time). Further, the data on laboratory accreditation/certification status were provided by the laboratories themselves. We know from experience within the European project EuroGentest that misunderstanding about the exact meaning of accreditation is common and we cannot rule out that some laboratories claimed to be accredited, but meant licensed, or mistook hospital accreditation for laboratory accreditation and thus overestimated their status.18, 19 The outcome with regard to the accreditation and certification status could have been different if the information given by the laboratories was 100% correct. Seen the fact that the accurateness of the data could not be guaranteed, we decided not to perform detailed research on the accreditation/certification status data. We thus did not analyse for instance whether or not the effect of accreditation/certification status on the interpretation score is generalisable to all accreditation and certification programs.

Conclusion and implications

A genetic test result differs from other diagnostic results in that it can have far-reaching effects on patients’ and their relatives’ life plans; furthermore, genetic tests are usually performed only once in a lifetime. Consequently it is of utmost importance that the result is correct and accompanied with an unambiguous, accurate and complete interpretation. We demonstrate that regular participation in EQA contributes to a more complete interpretation in laboratory test reports. Unfortunately, 50% of the CF EQA participants still do not participate regularly in the CF EQA scheme. In addition, based on data from Orphanet, at least 25% of the European CFTR-testing laboratories do not participate in the CF EQA scheme at all. Some of these are likely to participate in national CF EQA schemes organised by UKNEQAS, IEQA-ISS, Afssaps or INSTAND (only for c.1521_1523delCTT).20, 21, 22, 23 We expect similar percentages in schemes for other disorders or methods. It seems likely that laboratories that do not participate in EQA and whose reports have never been objectively assessed by peers and against expert opinions would have relatively low interpretation scores, comparable to those that have participated in only 1 of the 3 investigated years (between 1.2/2.0 and 1.7/2.0). Therefore, participation of laboratories in EQA on a regular basis, for all disorders they test for, needs to be continuously encouraged. The adoption of ISO 15 189, for implementing a quality management system and assessing technical competence, in national legislations could accelerate the implementation of accreditation, and thus assure EQA participation in genetic testing laboratories. Today, this is the case in only a few European countries, for example in France where accreditation will be mandatory by 2016 for medical biological laboratories, and in Switzerland where EQA participation for all genetic tests is becoming mandatory.24, 25 In the United States, the situation is different and all laboratories conducting moderate and/or high complexity testing are required by law (CLIA) to participate in PT for certain tests they perform. However, the CLIA regulations do not have PT requirements specific for molecular genetic tests. Therefore, laboratories that perform genetic tests must comply with the general requirements for alternative performance assessment for any test or analyte not specified as a regulated analyte to verify the accuracy of any genetic test or procedure they perform.26 Unfortunately in the United States, focus is mainly on the assessment of the accuracy of the test result (genotyping) and not on interpretation assessment, which is sufficient to comply with CLIA.5 We conclude that EQA contributes to improved interpretation of sensitive genetic test results and improved interpretation will lead to better quality of care for patients and their families by all healthcare professionals involved in and collaborating with genetic testing services.