Scolaris Content Display Scolaris Content Display

AD‐8 for detection of dementia across a variety of healthcare settings

Collapse all Expand all

Background

Dementia assessment often involves initial screening, using a brief tool, followed by more detailed assessment where required. The AD‐8 is a short questionnaire, completed by a suitable 'informant' who knows the person well. AD‐8 is designed to assess change in functional performance secondary to cognitive change.

Objectives

To determine the diagnostic accuracy of the informant‐based AD‐8 questionnaire, in detection of all‐cause (undifferentiated) dementia in adults. Where data were available, we described the following: the diagnostic accuracy of the AD‐8 at various predefined threshold scores; the diagnostic accuracy of the AD‐8 for each healthcare setting and the effects of heterogeneity on the reported diagnostic accuracy of the AD‐8.

Search methods

We searched the following sources on 27 May 2014, with an update to 7 June 2018: ALOIS (Cochrane Dementia and Cognitive Improvement Group), MEDLINE (Ovid SP), Embase (Ovid SP), PsycINFO (Ovid SP), BIOSIS Previews (Thomson Reuters Web of Science), Web of Science Core Collection (includes Conference Proceedings Citation Index) (Thomson Reuters Web of Science), CINAHL (EBSCOhost) and LILACS (BIREME). We checked reference lists of relevant studies and reviews, used searches of known relevant studies in PubMed to track related articles, and contacted research groups conducting work on the AD‐8 to try to find additional studies. We developed a sensitive search strategy and used standardised database subject headings as appropriate. Foreign language publications were translated.

Selection criteria

We selected those studies which included the AD‐8 to assess for the presence of dementia and where dementia diagnosis was confirmed with clinical assessment. We only included those studies where the AD‐8 was used as an informant assessment. We made no exclusions in relation to healthcare setting, language of AD‐8 or the AD‐8 score used to define a 'test positive' case.

Data collection and analysis

We screened all titles generated by electronic database searches, and reviewed abstracts of potentially relevant studies. Two independent assessors checked full papers for eligibility and extracted data. We extracted data into two‐by‐two tables to allow calculation of accuracy metrics for individual studies. We then created summary estimates of sensitivity, specificity and likelihood ratios using the bivariate approach and plotting results in receiver operating characteristic (ROC) space. We determined quality assessment (risk of bias and applicability) using the QUADAS‐2 tool.

Main results

From 36 papers describing AD‐8 test accuracy, we included 10 papers. We utilised data from nine papers with 4045 individuals, 1107 of whom (27%) had a clinical diagnosis of dementia. Pooled analysis of seven studies, using an AD‐8 informant cut‐off score of two, indicated that sensitivity was 0.92 (95% confidence interval (CI) 0.86 to 0.96); specificity was 0.64 (95% CI 0.39 to 0.82); the positive likelihood ratio was 2.53 (95% CI 1.38 to 4.64); and the negative likelihood ratio was 0.12 (95% CI 0.07 to 0.21). Pooled analysis of five studies, using an AD‐8 informant cut‐off score of three, indicated that sensitivity was 0.91 (95% CI 0.80 to 0.96); specificity was 0.76 (95% CI 0.57 to 0.89); the positive likelihood ratio was 3.86 (95% CI 2.03 to 7.34); and the negative likelihood ratio was 0.12 (95% CI 0.06 to 0.24).

Four studies were conducted in community settings; four were in secondary care (one in the acute hospital); and one study was in primary care. The AD‐8 has a higher relative sensitivity (1.11, 95% CI 1.02 to 1.21), but lower relative specificity (0.51, 95% CI 0.23 to 1.09) in secondary care compared to community care settings.

There was heterogeneity across the included studies. Dementia prevalence rate varied from 12% to 90% of included participants. The tool was also used in various different languages. Among all the included studies there was evidence of risk of bias. Issues included the selection of participants, conduct of index test, and flow of assessment procedures.

Authors' conclusions

The high sensitivity of the AD‐8 suggests it can be used to identify adults who may benefit from further specialist assessment and diagnosis, but is not a diagnostic test in itself. This pattern of high sensitivity and lower specificity is often suited to a screening test. Test accuracy varies by setting, however data in primary care and acute hospital settings are limited. This review identified significant heterogeneity and risk of bias, which may affect the validity of its summary findings.

How accurate is the AD‐8 informant questionnaire for diagnosing dementia in all healthcare settings?

Why is recognising dementia important?

Many people are living with dementia but have never had the condition diagnosed. Not recognising dementia when it is present (a false negative result) may deny people access to social support, medications, and financial assistance. It also prevents the individual and their family from planning for the future. However, incorrectly diagnosing dementia when it is not present (a false positive result) can cause distress or fear and lead to additional investigations which can waste resources.

What is the aim of the review?

The aim of this Cochrane Review was to find out how accurate the AD‐8 informant questionnaire is for detecting dementia in all healthcare settings. The researchers included 10 studies to answer this question, nine of which reported numerical information that could be used.

What was studied in the review?

The AD‐8 questionnaire includes eight ‘yes or no’ questions, to be answered by someone who knows the person under investigation; for example, a relative, carer or close friend (sometimes described as an informant). The questions ask about whether the informant has noticed a change in the individual’s memory and thinking abilities over the past years. A point is given for every item where they think the person’s abilities have changed. Higher scores occur when more changes are noted by the informant. The AD‐8 would not usually be used to make a final diagnosis of dementia, but it may help identify those who require further assessment.

What are the main results of the review?

The review included data from nine relevant studies, with a total of 4045 participants.

Seven of the studies used a score of two or more to indicate dementia. A score of two or more is the cut‐off recommended for the AD‐8. The results of these studies indicate that, in theory, if the AD‐8 were to be used to diagnose dementia in a group of 1000 people across all healthcare settings, of whom 280 (28%) have dementia, an estimated 517 would have an AD‐8 result indicating dementia is present and of these 259 (50%) will not have dementia. Of the 483 people with a result indicating that dementia is not present, 22 (5%) would be incorrectly classified as not having dementia

It is possible that the AD‐8 may work differently in different settings, for example in hospital or General practice. In secondary care, the AD‐8 produces more false positive results, but fewer false negative results, than when it is used in community settings. We could not directly compare the performance of the AD‐8 in primary care as there was only one study from this setting.

How reliable are the results of the studies in this review?

In the included studies, the diagnosis of dementia was made by assessing all patients with a detailed clinical assessment. (In these studies, detailed clinical assessment was the gold standard we compared the AD‐8 to.) This is likely to have been a reliable method for deciding whether patients really had dementia. However, there were some problems with how the studies were conducted. This may have resulted in the AD‐8 appearing more accurate than it really is. The numbers described are an average across studies in the review. However, as estimates from individual studies varied we cannot be sure that the AD‐8 will always produce these results.

Who do the results of this review apply to?

Studies included in the review were conducted in Brazil, China, Japan, Singapore, Taiwan, the UK and USA. Studies included those attending primary and secondary care services, and populations of older adults living in the community. Five studies used the English‐language version of the AD‐8. The percentage of people with a final diagnosis of dementia was between 12% and 90% (an average of 38%).

What are the implications of this review?

The studies included in this review suggest the AD‐8 can identify adults who may have a diagnosis of dementia, who would benefit from specialist assessment and diagnosis. If the AD‐8 was used alone to diagnose dementia, the chance of wrongly diagnosing someone with dementia when they do not actually have it is high (50% of those whose AD‐8 score suggests they have dementia). This makes the AD‐8 an unsuitable single diagnostic test as it would potentially create anxiety and distress. The chance of missing a diagnosis of dementia is much lower (5% of those whose AD‐8 score suggests they do not have dementia when they actually have it). This group will miss out on opportunities to plan their future care and would not be eligible to be assessed for treatment with medicines. These findings should be considered when deciding whether or not to use the AD‐8 to test for dementia.

How up‐to‐date is this review?

The review authors searched for and used studies published up to June 2018.

Authors' conclusions

available in

Implications for practice

The AD‐8 was developed as a tool to help inform the assessment of an individual, rather than being a diagnostic test for dementia in itself. The data from this review support the use of the AD‐8 (informant cut‐off score of two) as a tool to help identify those who may benefit from further specialist assessment. The setting where the test is administered affects the diagnostic accuracy, though we could not evaluate estimates of performance in primary care using the available data (albeit this may be the setting where the tool is used most often).

Test accuracy is only one metric to consider when using a test in clinical practice. For a test that is likely to be used predominantly in older adults, consideration must be given to feasibility, acceptability and test burden (Lees 2017). The feasibility of using the AD‐8 in the range of settings reported cannot be fully determined from the included studies, as there was a lack of data on the handling of indeterminate results or methods of assessing those whose informant was not present. Lack of an available informant was a common reason for exclusion, and one study excluded those whose informant was considered 'unreliable' due to age or perceived knowledge of the individual (Larner 2015). As such, there are insufficient data on the acceptability and feasibility of AD‐8 use in practice to offer any useful comment on these properties.

In the context of a large and increasing variety of available cognitive assessment tools (Harrison 2016), the important clinical question is around which test to use for a particular patient. Our review only goes some way to answering this question. We offer data on a single test, rather than formal comparative analyses with other tests. Techniques to allow comparisons across a network of differing tests with a common reference standard are emerging and may be a useful approach for future reviews (Owen 2018).

There is no ideal value for sensitivity or specificity in clinical practice. The optimal 'trade‐off' between sensitivity and specificity will depend on the purpose of testing and the implications of erroneous test results. With a pattern of high sensitivity and lower specificity, the AD‐8 is best suited for use as an initial screening or triage tool, with a view to selecting those who need more detailed assessment. In this scenario, the higher 'false positives' will go on to have additional testing and an incorrect label of dementia should be avoided. This form of cognitive screening is the usual purpose for which the AD‐8 is employed in clinical practice and our review data would support this approach. It is interesting that the pattern of high sensitivity and lower specificity seems more pronounced when the AD‐8 is used in a secondary care (specialist memory service) setting. In this setting, where there is a high prevalence of cognitive impairment and the population is likely to have already had some form of cognitive assessment before referral, the utility of a high‐sensitivity screening tool is questionable.

Implications for research

In this review, we observed differences in test accuracy when the AD‐8 was used in different healthcare settings. It is therefore important that future studies stratify results by the setting of recruitment, and that researchers are aware of the differences in dementia prevalence which may affect test performance. We only identified one study that was conducted in each of the primary care and acute hospital settings. These settings may provide opportunities for the use of brief cognitive tests such as the AD‐8, and thus diagnostic accuracy should be established.

This review considered only the use of the AD‐8 as a single test in isolation. In practice, it is likely that testing is done sequentially, informed by results. This review is one in a series examining the test accuracy of other dementia tests, including the Informant Questionnaire on Cognitive Decline in the Elderly (Harrison 2015), and Montreal Cognitive Assessment (Davis 2015). It would be helpful to consider the testing pathway and determine which tests have greatest diagnostic utility at which stage. Tests may not be used sequentially and informant assessments such as the AD‐8 are often completed at the same time as direct‐to‐patient cognitive tests. How these results should be combined to give the greatest accuracy is a topic that requires further research.

In applying our assessments of internal and external validity, use of the QUADAS‐2 tool was complicated by poor reporting within the included studies. Poor reporting is a recognised issue in test accuracy studies (Davis 2013). Specific guidance for reporting test accuracy with a dementia focus is available, and we would recommend that future primary test accuracy research makes use of this resource (Noel‐Storr 2014).

The ultimate goal of interventions to improve rates of dementia diagnosis is to enable individuals to access specialist services, resources and support. Historical modelling suggested early diagnosis could improve the quality of life of those diagnosed, and reduce the need for care‐home admission (Banerjee 2009). Longitudinal data from the UK identified evidence of increased rates of dementia diagnosis and reductions in antipsychotic prescribing following introduction of the National Dementia Strategy (Donegan 2017). A challenge is therefore for researchers to identify if those identified as having possible dementia and referred for formal diagnosis derive benefit from the impact of this diagnosis through the care and treatment they subsequently receive. Ultimately, the design of studies looking at assessment tools such as the AD‐8 may move away from the classical index test versus gold‐standard paradigm to a model that looks at test‐treatment‐outcome pathways (Takwoingi 2018).

Summary of findings

Open in table viewer
Summary of findings 1. Summary of findings table

Study ID

Country

Subjects (n)

Age (Years) (Mean, SD)

AD‐8 language

Cut‐off scores reported

Reference standard

Chan 2016

Singapore

309

Malay

3

Dementia

(CDR and DSM‐IV)

Chio 2017

Taiwan

153

76.9

(8.6)

Unclear

1, 2, 3, 4

Early cognitive impairment

(CDR)

Correia 2011

Brazil

109

Portuguese

2, 3

Dementia

(DSM‐IV)

Galvin 2006

US

255

73.3

(11.3)

English

2, 3

Various criteria for dementia

Galvin 2007

US

325

76.8

(8.9)

English

1, 2, 3

Dementia

(CDR)

Jackson 2016

UK

125

84.4

English

3, 7

Pre‐delirium dementia

(DSM‐IV)

Larner 2015

UK

212

Median 64.5

(IQR 12.7)

English

2

Dementia

(DSM‐IV)

Meguro 2015

Japan

572

Japanese

2

Dementia

(CDR)

Razavi 2014

US

186

77.8
(8.2)

English

2

Various criteria for dementia

Yang 2016

China

2063

79.5 (7.6)

Unclear

2

Dementia

(NIA‐AA criteria)

CDR: Clinical Dementia Rating scale
DSM‐IV: Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition
NIA‐AA: National Institute on Aging and Alzheimer's Association

Open in table viewer
Summary of findings 2. Summary of findings table

What is the accuracy of the AD‐8 test for the detection of dementia when differing thresholds are used to define AD‐8 positive cases?

Population

Adults whose informant completed an AD‐8 questionnaire, with no restrictions on the case mix of recruited participants

Setting

Primary care, community and secondary care

Index test

AD‐8 informant‐based questionnaire

Reference standard

Clinical diagnosis of dementia made using any recognised classification system

Studies

We included cross‐sectional but not case‐control studies

Test

Summary accuracy (95% CI)

No. of participants (studies)

Dementia prevalence

Implications, quality and comments

AD‐8

cut‐off score 1

Sensitivity: 0.90 (0.81 to 0.96)

Specificity: 0.68 (0.62 to 0.74)

n = 324

(1 study)

n = 73

(60%)

The majority of studies used the informant cut‐off score of 2 to define AD‐8 test positivity.

As the cut‐off score is raised, there is an increase in specificity. Sensitivity is similar at a cut‐off score of 2 or 3, but falls as the cut‐off score to define test positivity rises.

False positives (individuals wrongly identified as having possible dementia and referred for further investigations) and false negatives (individuals with dementia who are not identified and not referred for specialist assessment and management) can both be associated with harms. If used as a screening test (i.e. an initial assessment, which will be followed by further testing) then a pattern of higher sensitivity may be preferable.

The dementia prevalence within the included studies was highly varied (12% to 90%) and may reflect that for each cut‐off score, data were included from all settings in which the AD‐8 has been used. These analyses need to be interpreted with caution as the properties of the test may vary by setting and the desired balance of sensitivity and specificity will vary by purpose of testing.

AD‐8 cut‐off score 2

Sensitivity: 0.92

(0.86 to 0.96)

Specificity: 0.64

(0.39 to 0.82)

+LR: 2.53

(1.38 to 4.64)

‐ve LR: 0.12

(0.07 to 0.21)

n = 3659

(7 studies)

n = 1016 (28%

range:

12% to 90%)

AD‐8 cut‐off score 3

Sensitivity: 0.91

(0.80 to 0.96)

Specificity: 0.76

(0.57 to 0.89)

+LR: 3.86

(2.03 to 7.34)

‐ve LR: 0.12

(0.06 to 0.24)

n = 1060

(5 studies)

n = 396

(37%

range:

14% to 90%)

AD‐8 cut‐off score 7

Sensitivity: 0.83

(0.69 to 0.92)

Specificity: 0.90

(0.73 to 0.98)

n = 77

(1 study)

n = 47 (61%)

95% CI: 95% confidence interval
+LR: positive likelihood ratio
‐LR: negative likelihood ratio

Open in table viewer
Summary of findings 3. Summary of findings table

What is the accuracy of the AD‐8 test for detection of dementia in different settings and using different languages of administration?

Population

Adults whose informant completed an AD‐8 questionnaire, with no restrictions on the case mix of recruited participants

Setting

Primary care, community and secondary care

Index test

AD‐8 informant‐based questionnaire using a cut‐off score of 2 to define positivity

Reference standard

Clinical diagnosis of dementia made using any recognised classification system

Studies

We included cross‐sectional but not case‐control studies

Comparative analyses

Test

No. of participants (studies)

Dementia prevalence

Findings

Implications

Community versus primary care versus secondary care setting

Total n = 4045 (9)

n = 2937 (4) Community

n= 309 (1) Primary care

n = 716 (4) Secondary care

Total n = 1107 (27%)

Community n = 601 (20%)

Primary n = 44 (14%)

Secondary care n = 462 (65%)

Using a cut‐off score of 2, the AD‐8 has a higher relative sensitivity (1.11, 95% CI 1.02 to 1.21) in secondary care compared to community care settings.

Using a cut‐off score of 2, the AD‐8 has a lower relative specificity (0.51, 95% CI 0.23 to 1.09) in secondary care compared to community care settings.

Only one study used the AD‐8 in primary care settings; data were at a cut‐off score of 3.

Two studies used AD‐8 in community and two studies used AD‐8 in secondary care settings, with a cut‐off score of 3. The numbers were insufficient to allow formal quantitative synthesis.

Dementia prevalence in secondary care is more than double that seen in the community.

The AD‐8 has a higher sensitivity, but poorer specificity, in secondary care settings.

English language

Total: n = 1040 (5) English language

Total n = 535 (51%)

The test accuracy of English language administration of the AD‐8, at a cut‐off score of 2, has a higher sensitivity and lower specificity than summary estimates for all studies at a cut‐off score of 2. Estimates for specificity were imprecise due to the small number of studies included.

Sensitivity: 0.95 (95% CI 0.87 to 0.98)

Specificity: 0.59 (95% CI 0.21 to 0.89)

+LR: 2.30 (95% CI 0.87 to 6.07)

‐LR: 0.08 (95% CI 0.03 to 0.22)

We were unable to assess English language at a cut‐off score of 3 due to lack of available data at this cut‐off score.

We were unable to assess the non‐English‐language group due to the small numbers reporting language of administration clearly.

There is a need to improve clarity of reporting around language of test administration to evaluate diagnostic accuracy of non‐English‐language versions of the AD‐8

95% CI: 95% confidence interval
+LR: positive likelihood ratio
‐LR: negative likelihood ratio

Background

available in

Dementia is a substantial and growing public health concern (Livingston 2017; Prince 2013). As an example, depending on case definition employed, contemporary estimates of dementia prevalence in the United States are in the range 2.5 to 4.5 million individuals (up to 1.6% of the population; 6.5% of those aged over 65) (Hebert 2013). Changes in population demographics will be accompanied by increases in global dementia incidence and prevalence (Ferri 2005). Although the magnitude of the increase in prevalent dementia may have been overestimated in previous prediction models (Matthews 2013), there is no doubt that absolute numbers of older adults with dementia will increase substantially in the short‐ to medium‐term future (Ferri 2005).

A key element of effective management in dementia is a firm diagnosis. Recent guidelines place emphasis on early diagnosis to facilitate improved management and enable informed discussions and planning with patients and carers. The benefits of screening for cognitive decline are debated (Brunet 2012); however, in certain healthcare systems, screening or case‐finding has already been introduced for certain groups, e.g. unscheduled hospital admissions of older adults (Burn 2018; Robinson 2015).

Given the projected global increase in dementia prevalence, there is potential tension between the clinical requirements for robust diagnosis at the individual patient level and the need for equitable, easy access to diagnosis at a population level. The ideal would be expert, multidisciplinary assessment informed by various supplementary investigations (neuropsychology; neuroimaging or other biomarkers). This approach is only really feasible in a specialist memory service and is not suited to population screening or case‐finding at scale.

In practice, a two‐stage process is often employed. This involves initial 'triage' assessments, suitable for use by non‐specialists, to select those who require second‐stage, further detailed assessment (Boustani 2003).

Various tools for initial cognitive screening have been described (Brodaty 2002; Folstein 1975). Regardless of the methods employed, there is scope for improvement, with observational work suggesting that many people with dementia are not diagnosed (Bradford 2009; Chodosh 2004; Lang 2017; Valcour 2000).

Screening assessment often takes the form of brief, direct cognitive testing. Such an approach will only provide a 'snapshot' of cognitive function. However, a defining feature of dementia is cognitive or neuropsychological change over time. People with cognitive problems themselves may struggle to make an objective assessment of personal change and so an attractive approach is to question collateral sources with sufficient knowledge of the person. These informant‐based interviews aim to assess change in function retrospectively.

An instrument prevalent in research and clinical practice, particularly in North America, is the AD‐8 informant questionnaire (Galvin 2005). The authors of the tool describe it as the AD‐8, without further definition of the acronym (Galvin 2005). However, it is also described as the Ascertaining Dementia interview (Correia 2011), Ascertain Dementia 8 questionnaire (Chen 2018) and the Ascertain Dementia 8‐item Informant Questionnaire (Chen 2017) in the literature. This screening/triage tool will be the focus of this review.

A number of properties can be described for a clinical assessment (reliability, responsiveness, feasibility). For our purposes, the test property of greatest interest is diagnostic test accuracy (DTA) (Cordell 2013).

Although we describe test accuracy of the AD‐8 for dementia diagnosis, the AD‐8 is not suitable for establishing a clinical diagnosis of dementia when used in isolation. The AD‐8 can be described in different ways dependent on the setting and purpose of testing.

Target condition being diagnosed

The target condition for this DTA review is all‐cause dementia (clinical diagnosis).

Dementia is a syndrome characterised by cognitive or neuropsychological decline sufficient to interfere with usual functioning. The neurodegeneration and clinical manifestations of dementia are progressive and at present there is no 'cure', although numerous interventions to slow or arrest cognitive decline have been described, including pharmacological (Birks 2018; McShane 2006), and non‐pharmacological (Bahar‐Fuchs 2013).

Dementia remains a clinical diagnosis, based on history from the person and suitable collateral sources and direct examination, including cognitive assessment. There is no universally‐accepted, ante‐mortem, gold‐standard diagnostic strategy. We have chosen expert clinical diagnosis as our gold standard (reference standard) for describing AD‐8 test properties, as we believe this is most in keeping with current diagnostic criteria and best practice.

Dementia diagnosis can be made according to various internationally‐accepted diagnostic criteria, with exemplars being the World Health Organization, International Classification of Diseases (ICD) (ICD‐10), and the American Psychiatric Association Diagnostic and Statistical Manual of Mental Disorders (DSM) (DSM‐IV) for all‐cause dementia and subtypes (Appendix 1). The label of dementia encompasses varying pathologies, of which Alzheimer’s disease is the most common. Diagnostic criteria are available for specific dementia subtypes, e.g. NINCDS‐ADRDA criteria for Alzheimer’s dementia (McKhann 1984; McKhann 2011); McKeith criteria for Lewy body dementia (McKeith 2005); Lund criteria for frontotemporal dementias (McKhann 2001); and the NINDS‐AIREN criteria for vascular dementia (Román 1993).

Index test(s)

Our index test was the AD‐8 informant questionnaire (Galvin 2005) (Appendix 2; Appendix 3).

First published in 2005, the AD‐8 is a screening tool which has been used to distinguish individuals with dementia or mild cognitive impairment (MCI) from those with normal cognitive function. It is designed to be administered to a relevant proxy, usually a relative or carer, in questionnaire form. The AD‐8 is a brief screening tool. With only eight questions, it takes less than three minutes to complete and was developed to replace other lengthy informant questionnaires (Galvin 2006). The AD‐8 was originally developed for administration in the English language but has been reproduced in other languages including Brazilian Portuguese (Correia 2011), Taiwanese (Yang 2011), and Korean (Ryu 2009).

The AD‐8 items cover domains of judgement, hobby/activity level, repetitive conversations, learning ability, memory in relation to date/appointments, finances and daily thought processes. Informants indicate presence of change 'over several years' using responses of ‘Yes, a change’, ‘No, no change’ or ‘NA, don’t know’. Each 'yes' answer is scored one point. Final scores range from zero, where no change has been noticed by the informant, to eight, where change has been noted across all domains. The commonly employed threshold score for the AD‐8 to differentiate cognitive from no cognitive impairment is greater than or equal to two out of eight (i.e. a 'yes' response for two or more items) (Galvin 2005).

The AD‐8 has a number of features that potentially make it attractive for clinical and research use. The questions used have an immediacy and relevance that is likely to appeal to users. Assessment and (informant) scoring is brief, and as the scale is not typically interviewer‐administered it requires minimal training in application and scoring. There are data to suggest that, compared to standard direct assessments, informant interviews may be less prone to bias from cultural norms and previous level of education (Jorm 2004). As diagnostic criteria for dementia make explicit reference to documenting decline and involving collateral informants, the potential utility of an informant interview tool such as the AD‐8 is clear.

Clinical pathway

Dementia develops over a trajectory of several years, and tests may be performed at different stages in the dementia pathway. In this review we considered any use of the AD‐8 as an initial assessment for cognitive decline, and did not limit our inclusion criteria to a particular healthcare setting. We have operationalised the various settings where the AD‐8 may be used as secondary care, primary care, and community.

In secondary‐care settings, patients will have been referred for expert input but not exclusively due to memory complaints. Cognitive testing in secondary care involves two main groups: opportunistic screening of adults presenting as unscheduled admissions to hospitals, and those people referred to specific services for dementia, memory or psychiatry of older age. Both populations will have a high prevalence of cognitive disorders and mimics. In‐hospital estimates of dementia prevalence vary from 15% to 42%, dependent on methodology used (Jackson 2017). Secondary‐care patients are more likely to have had a degree of prior cognitive assessment than those in other settings, although we recognise that cognitive screening prior to referral to specialist services is neither consistent nor guaranteed (Menon 2011).

The prevalence of dementia in primary care is estimated to be 1.4% (Donegan 2017), and it has been projected that most General practitioners could expect one to two new cases each year, per physician (Iliffe 2009). However, cognitive testing in the General practice/primary‐care setting is likely to be conducted in response to an individual person self‐presenting because of subjective memory complaints, a common finding in older adults, estimated to be reported by 25% to 50% of those aged 65 years and over (Iliffe 2010). Using the AD‐8 in this setting could be described as 'triage' or 'case finding' to determine individuals with objective concerns requiring further investigation.

In the community setting, the cohort is largely unselected.The estimated prevalence of dementia in those aged 65 years and over is 6.5% (Matthews 2013) and this age group accounts for ˜18% of the population (Office for National Statistics 2018). This means dementia prevalence will be low compared to other settings. In the community setting, use of the AD‐8 as a cognitive testing approach may be described as 'population screening'.

Most studies of test accuracy compare the test against contemporaneous reference standards (in this case, clinical dementia diagnosis). An alternative is to describe the test properties for detection of early, 'pre‐clinical' problems that are formally diagnosed during prospective, longitudinal follow‐up. This delayed verification approach is commonly employed in studies describing properties of dementia biomarkers, but may have utility for other test strategies such as informant interview.

Rationale

There is no consensus on the optimal initial assessment for dementia, and choice is currently dictated by experience with a particular instrument, time constraints and training. A better understanding of the diagnostic properties of various strategies would allow for an informed approach to testing. Critical evaluation of the evidence base for screening tests or other diagnostic markers is of major importance. Without a robust synthesis of the available information there is the risk that future research, clinical practice and policy will be built on erroneous assumptions about diagnostic validity.

The AD‐8 is commonly used in practice and research; it is used internationally and is one of only a few validated informant‐based screening/diagnostic tools. A body of literature describing the test accuracy of the AD‐8 in different settings is available, although some of these studies have been modest in size. Thus, systematic review and, if possible, meta‐analysis of the diagnostic properties of the AD‐8 is warranted.

This review forms part of a body of work describing the diagnostic properties of commonly‐used dementia tools (Appendix 4). At present we are conducting single‐test reviews and meta‐analyses. However, the intention is then to collate these data and perform an overview, to enable comparison of various test strategies.

Objectives

available in

To determine the diagnostic accuracy of the informant‐based AD‐8 questionnaire, in detection of all‐cause (undifferentiated) dementia in adults.

Secondary objectives

Where data were available, we described the following.

  1. The diagnostic accuracy of the AD‐8 at various predefined threshold scores (one, two, three). These thresholds have been chosen to represent the range of cut‐off scores that are commonly used in practice and research; we have been inclusive in our choice of cut‐off score to maximise available data for review.

  2. The diagnostic accuracy of the AD‐8 for each healthcare setting (community; primary care; secondary care).

  3. The effects of heterogeneity on the reported diagnostic accuracy of the AD‐8. Potential sources of heterogeneity that we intended to explore were: case mix of cohort; method of dementia diagnosis; method (language) of AD‐8 assessment.

Methods

available in

Criteria for considering studies for this review

Types of studies

We anticipated that the majority of the studies would be of AD‐8 test properties compared against a contemporaneous clinical diagnosis of dementia in secondary‐care settings. We included test studies performed in other healthcare settings, and classified settings as secondary care, primary care, or community.

Case‐control studies are known to potentially overestimate properties of a test and we did not include such studies in the review (Davis 2013).

We also did not include case studies or samples with very small numbers (chosen as 10 participants, for the purposes of this review).

Where settings were mixed, for example, a population study 'enriched' with additional cases from secondary care, we did not consider such studies unless separate data were presented for participants from each setting. This design can be affected by similar biases to a case‐control design.

Participants

All adults (aged over 18 years) were eligible.

We did not predefine exclusion criteria relating to the 'case mix' of the population studied, but assessed this aspect of the study as part of our assessment of heterogeneity.

Index tests

Studies must have included (not necessarily exclusively) the AD‐8, used as an informant questionnaire.

The AD‐8 has been translated into various languages to enable international administration. The properties of a translated AD‐8 in a cohort of non‐English speakers may differ from the properties of the original English‐language questionnaire. We collected data on the principal language used for AD‐8 assessment.

For this review we did not consider other cognitive screening/assessment tools. Where a paper described the AD‐8 with in‐study comparison against another screening tool, we included the AD‐8 data only. Where the AD‐8 was used in combination with another cognitive screening tool, we included the AD‐8 data only.

Target conditions

We included any clinical diagnosis of all‐cause (unspecified) dementia. We did not require a definition of a particular dementia subtype, although we recorded this information where available.

Reference standards

Our reference standard is clinical diagnosis of dementia. We recognise that clinical diagnosis itself has a degree of variability but this is not unique to dementia studies and does not invalidate the basic diagnostic test accuracy (DTA) approach.

The primary analysis for clinical diagnosis included all‐cause (unspecified) dementia, using any recognised diagnostic criteria (for example, DSM‐IV; ICD‐10). We included use of the full Clinical Dementia Rating (CDR) Scale (Morris 1993) as an acceptable method for making a dementia diagnosis, with a CDR score greater than or equal to one for a clinical diagnosis of dementia. Dementia diagnosis may have specified a pathological subtype and we included all common dementia subtypes, e.g. NINCDS‐ADRDA (Alzheimer's disease), Lund‐Manchester (frontotemporal dementia), McKeith (Dementia with Lewy bodies), NINCDS‐AIREN (vascular dementia) (McKeith 2005; McKhann 1984; McKhann 2001; Román 1993). We did not define preferred diagnostic criteria for rarer forms of dementia (e.g. alcohol‐related; HIV‐related; prion disease‐related). These were considered under our rubric of 'all‐cause' dementia, and not considered separately.

The label 'dementia' can span a range of disease severities, from mild disease to 'end stage'. The diagnostic properties of a tool will vary depending on disease stage, for example, true positives are more likely when disease is advanced and diagnosis is clear. For our primary analysis we included any dementia diagnosis, at any stage of disease.

Clinicians may use imaging, pathology or other data to aid diagnosis; however, we did not include diagnosis based only on these data without corresponding clinical assessment. We recognise that different iterations of diagnostic criteria may not be directly comparable, and that diagnosis may vary with the degree or manner in which the criteria have been operationalised (e.g. individual clinician versus algorithm versus consensus determination). We collected data on method and application of dementia diagnosis for each study and explored the validity of the dementia diagnosis as part of our 'Risk of Bias' assessment. We did not accept use of other (brief) direct performance tests in isolation as a basis for diagnosis.

We recognise that dementia diagnosis often comprises a degree of informant assessment. Thus there is potential for incorporation bias. We explored the potential effects of this bias through our 'Risk of bias' assessment.

Search methods for identification of studies

We used a variety of information sources to ensure all relevant studies were included. We devised terms for electronic database searching in conjunction with the team at the Cochrane Dementia and Cognitive Improvement Group. As this AD‐8 review forms part of a suite of reviews looking at informant scales, we have created a comprehensive search strategy designed to pick up all cognitive assessment scales (Quinn 2014); we complemented this generic search with searches specific to AD‐8 terminology.

Electronic searches

We searched the specialised register of the Cochrane Dementia and Cognitive Improvement Group, ALOIS (which includes both intervention and diagnostic accuracy studies), MEDLINE (Ovid SP), Embase (Ovid SP), BIOSIS (ISI Web of Science), Web of Science Core Collection (ISI Web of Science), PsycINFO (Ovid SP), CINAHL (EBSCOhost) and LILACS (Bireme). We designed similarly‐structured search strategies and used search terms appropriate to each database. We used MeSH words and other controlled vocabulary where appropriate.

We also searched sources specific to diagnostic accuracy or systematic review:

· MEDION database (Meta‐analyses van Diagnostisch Onderzoek www.mediondatabase.nl);

· DARE (Database of Abstracts of Reviews of Effects) and HTA Database (Health Technology Assessments Database), both The Cochrane Library;

· ARIF database (Aggressive Research Intelligence Facility; www.arif.bham.ac.uk).

See Appendix 5 and Appendix 6 for the search strategies run.

We did not apply language or date restrictions to the electronic searches, nor did we apply restrictions regarding publication status when assessing records for potential inclusion, including abstracts, conference proceedings and unpublished data. We used translation services where necessary (see: Acknowledgements), and used a translation proforma for data extraction (Appendix 7).

The Cochrane Dementia and Cognitive Impairment Group Information Specialist (ANS) ran the initial searches.The most recent search for this review was performed on 7 June 2018.

Searching other resources

Grey literature and proceedings: we searched our chosen electronic databases and included relevant assessments of conference proceedings. Where possible we accessed theses or PhD abstracts from institutions known to be involved in prospective dementia studies.

Handsearching: we did not perform handsearching as there is little published evidence of the benefits of handsearching for diagnostic studies (Glanville 2010).

Reference lists: we checked the reference lists of all relevant studies and reviews in the field for further possible titles, and repeated the process until no new titles were found (Greenhalgh 2005).

Correspondence: We contacted research groups who have published or are conducting work on AD‐8 for dementia diagnosis, informed by results of initial search (see: Acknowledgements).

We used relevant studies in PubMed to search for additional studies using the 'related article' feature. We examined key studies in citation databases, such as Science Citation Index and Scopus, to ascertain details of any further relevant studies.

Data collection and analysis

Selection of studies

One review author (ANS) screened all titles generated by initial electronic database searches for relevance. The initial search was a sensitive, generic search, designed to include all potential dementia screening tools. Two review authors (KH, CG) then independently screened all remaining titles for relevance. The two review authors inspected abstracts of selected titles and selected all potentially eligible studies for full‐text review. They independently assessed full manuscripts against the inclusion criteria, and resolved disagreements by discussion or by involving an arbitrator where necessary.

Where a study included useable data but did not present these in the published manuscript, we contacted the authors directly to request further information. If the same data were presented in more than one paper we included the primary paper only.

We have detailed the study selection process in a PRISMA flow diagram.

Data extraction and management

We extracted data to a study‐specific pro forma that included clinical/demographic details of the participants; details of setting; details of AD‐8 administration and details of the dementia diagnosis process.

Where AD‐8 data were presented for differing threshold scores, we extracted data for each threshold and collated these separately. We extracted test accuracy data to a standard two‐by‐two table.

Two review authors (KH, CG), blinded to study identifiers, performed data extraction independently. They resolved disagreements by discussion, with the use of an arbitrator if necessary. As a further check of validity of search and data extraction, a third reviewer (SA) assessed all titles and papers from the most recent updated literature search.

For each included paper, we detailed the flow of participants (numbers recruited, included, assessed) in a flow diagram.

Assessment of methodological quality

We assessed the methodological quality of each study using the QUADAS‐2 tool (www.bris.ac.uk/quadas/quadas‐2) (Appendix 8). This tool incorporates domains specific to participant selection; index test; reference standard; and participant flow. Each domain is assessed for risk of bias and the first three domains are also assessed for applicability. Certain key areas important for quality assessment are participant selection, blinding, and missing data. Following a group meeting of review authors, we created guidance for the application of QUADAS‐2 to dementia screening assessments, specifically developing anchoring statements for QUADAS‐based assessment that are suited to dementia test accuracy studies. This QUADAS‐2 guidance was created through a multidisciplinary working group and has been extensively piloted. The process and resulting statements for assessment are described in Appendix 9.

We did not use QUADAS‐2 data to form a summary quality score; we produced a narrative summary describing numbers of studies for which we found high/low/unclear risk of bias, or concerns regarding applicability with corresponding tabular and graphical displays.

Paired independent raters (KH, CG), blinded to each other’s scores, performed both assessments. They resolved disagreements by further review and discussion, with recourse to a third‐party arbitrator where necessary. A third reviewer assessed all titles selected from the most recent literature search.

Statistical analysis and data synthesis

We were interested in the test accuracy of the AD‐8 for the dichotomous variable 'dementia'/'no dementia'. Thus, we applied the current Cochrane DTA framework for analysis of a single test. We extracted data from included papers to allow creation of a standard two‐by‐two data table showing dichotomised AD‐8 test results (AD‐8 positive or AD‐8 negative), cross‐classified with binary reference standard (dementia or no dementia).

We used Review Manager 5 software (RevMan 2012) to calculate sensitivity, specificity and 95% confidence intervals (CIs) from the two‐by‐two tables abstracted from the included studies. We used a threshold score of two or more on the AD‐8 for primary analyses. If data at other thresholds were presented, we examined these in separate analyses (we were able to describe accuracy at cut‐off scores of one, two, three, and seven). We presented individual study results graphically by plotting estimates of sensitivities and specificities as forest plots.

To allow for summary analysis, we used SAS release 9.4, in addition to Review Manager 5. Using the bivariate approach we described metrics of pooled sensitivity, specificity, positive and negative likelihood ratios, all with corresponding 95% confidence intervals. We plotted summary data in receiver operating characteristic (ROC) space, including 95% confidence regions. Where we were interested in comparative accuracy and where data allowed, for example differential accuracy by setting, we plotted summary accuracy data for each variable in shared ROC space and described relative sensitivity and specificity of one factor to the other.

We suspected papers would use the classical cross‐sectional test accuracy study design. An alternative is the 'delayed verification' study design, i.e. where the AD‐8 is performed at baseline and those without disease are prospectively followed up for development of incident dementia. No included studies took this approach, but if they had, we planned to use baseline (contemporaneous testing) data for our primary analysis.

Investigations of heterogeneity

Heterogeneity is expected in DTA reviews, and 'traditional' measures of heterogeneity used in meta‐analysis are not appropriate to DTA reviews.

The properties of a tool describe behaviour of the instrument under particular circumstances. We included all AD‐8 studies in narrative review. We prespecified particular areas of potential heterogeneity, as follows.

AD‐8 threshold score

We included data from all AD‐8 threshold scores described in the included studies. Where data allowed, we collated scores at each cut‐off score of test positivity, using meta‐analytical techniques to create summary estimates of sensitivity and specificity.

Healthcare setting

We suspected that healthcare setting would impact on properties of the test. Our primary analysis was across all settings; where data allowed, we performed sensitivity analysis comparing hospital‐based (secondary care) settings and primary care or community settings.

Case mix

In the first instance we explored age, taking age over 65 years as a reference point. We anticipated that the majority of included participants in eligible studies would be aged over 65 years. The AD‐8 may have different properties in younger cohorts and so we looked at age ranges within studies. We planned to grade studies with greater than 20% of included participants younger than 65 years as potentially unrepresentative. Where studies offered test accuracy at different age ranges, we chose the range that best represented our primary population of interest, namely older adults (aged over 65 years).

We anticipated that most studies would be of unselected adults, however we included studies that limited inclusion to a specific population, for example stroke survivors. For studies of selected populations, we assessed the validity of inclusion in a summary analysis with other included studies on a case‐by‐case basis.

We suspected that the majority of studies would focus on all‐cause (undifferentiated) dementia or Alzheimer's disease, but included any dementia pathology in the review. We assessed generalisability according to age, comorbidity and dementia pathology, as part of our 'Risk of Bias' assessment. If data allowed, we planned to run sensitivity analyses removing studies with unrepresentative populations.

Criteria used to reach dementia diagnosis

We recorded the classification used (for example, DSM‐IV; ICD‐10) and the methodology used to reach dementia diagnosis (for example, individual assessment; group (consensus) assessment). We assessed the 'quality' of diagnosis at study level using the QUADAS‐2 tool. If data allowed, we planned subgroup analyses comparing differing approaches to diagnosis.

Technical features of the testing strategy

Our focus was on language of assessment. We classified this as either English language or non‐English‐language tests. We performed subgroup analyses comparing the English‐language AD‐8 versus non‐English‐language AD‐8.

Sensitivity analyses

Where appropriate (i.e. if not already explored in our analyses of heterogeneity), and as data allowed, we explored the sensitivity of any summary accuracy estimates to other methodological aspects of the included studies. We performed a sensitivity analysis excluding a paper with a disease‐specific patient group (those with delirium). We also performed a post‐hoc sensitivity analysis excluding a paper whose test accuracy estimates were based on a more inclusive definition of dementia than was used in the other papers. We had planned to perform sensitivity analyses based on key aspects that could indicate risk of bias, such as nature of blinding and loss to follow‐up, guided by the anchoring statements developed in our QUADAS‐2 exercise. Due to the modest number of included studies, and the lack of papers deemed to have low risk of bias on all domains, we were unable to perform these analyses.

Assessment of reporting bias

We did not investigate reporting bias because of current uncertainty about how it operates in test accuracy studies, and about the interpretation of existing analytical tools such as the funnel plot (van Enst 2014).

Results

Results of the search

Our search resulted in 3661 records, from which we identified 36 full‐text papers for eligibility. We included 10 studies, with a total of 4309 participants (summary of findings Table 1).

We identified two studies which required translation (Munoz 2010; Pardo 2013). We contacted three authors to provide additional data, and one author provided this for incorporation in the review (see: Acknowledgements).

We excluded 26 papers. The most frequent reasons for exclusion were: case‐control data (n = 10); unsuitable reference standard (n = 7) and repeat data set (n = 2) (see Characteristics of excluded studies). We excluded one study following repeated review and consultation with the review group Editor (Malmstrom 2009). The study subgroup, which could have been eligible, did not include any participants with a cognitive diagnosis that was compatible with our prespecified criteria for dementia (clinical diagnosis using a recognised clinical classification system and/or Clinical Dementia Rating (CDR) Scale score of one or more) As our review focuses on dementia, we chose to exclude this study. For a PRISMA flow diagram depiciting study selection, see Figure 1.


Study flow diagram.

Study flow diagram.

Methodological quality of included studies

We described the risk of bias and applicability using the QUADAS‐2 methodology (Appendix 8); our anchoring statements for the AD‐8 are summarised in Appendix 9.

We did not rate any included study as being at low risk of bias for all the categories of QUADAS‐2. For all domains there were papers where risk of bias was high or where reporting was insufficient, which necessitated an assessment of unclear risk of bias. There were particular issues around inappropriate exclusions of participants (for example, exclusion those who had dementia) and around patient flow (for example, substantial dropouts due to inability to complete test). We had fewer concerns around applicability. See: Figure 2.


Risk of bias and applicability concerns summary: review authors' judgements about each domain for each included study

Risk of bias and applicability concerns summary: review authors' judgements about each domain for each included study

Findings

We have described the individual included studies in Characteristics of included studies and Table 1. We have also presented tabulated data for test accuracy by AD‐8 threshold (summary of findings Table 2), and by setting and language of administration (summary of findings Table 3).

Open in table viewer
Table 1. Summary of test accuracy at study level by dementia prevalence

Study

Included participants

(n)

Dementia

n (%)

AD‐8 Threshold

Sensitivity

(%)

Specificity

(%)

Galvin 2006

241

217 (90)

2

92

46

Galvin 2006

241

217 (90)

3

89

67

Razavi 2014

186

129 (69)

2

99

74

Jackson 2016

77

47 (61)

3

98

40

Larner 2015

212

69 (33)

2

97

11

Galvin 2007

324

73 (23)

2

84

93

Galvin 2007

324

73 (23)

3

75

90

Yang 2016

2015

444 (22)

2

90

78

Correia 2011

109

15 (14)

2

73

61

Correia 2011

109

15 (14)

3

100

67

Chan 2016

309

44 (14)

3

91

91

Meguro 2015

572

69 (12)

2

88

68

Where multiple thresholds were reported in the primary paper, we present the data for both a cut‐off score of 2 and 3.

The total number of participants is adjusted to reflect the data used in quantitative synthesis.

The total number of participants across the studies was 4309 (range: 109 to 2063). One study (n = 153 participants) identified cases of dementia (those with a CDR score of one or more), but did not provide the test accuracy data specific to this (Chio 2017). The focus of the paper was on identifying early cognitive impairment and so, while the study was eligible for inclusion in review, we could not include its data in the quantitative synthesis. We performed quantitative synthesis for nine studies, including a total of 4045 participants of whom 1107 (27%) had a clinical dementia diagnosis. This excludes 14 participants from Galvin 2006, who were missing a clinical diagnosis; one participant from Galvin 2007, who had no reference standard data; and 48 participants from Jackson 2016, plus 48 participants from Yang 2016, who were lacking complete follow‐up assessment data. Dementia prevalence in the included studies varied from 12% to 90%.

Nine of the ten studies are included in the quantitative synthesis. Seven studies used a cut‐off score of two to differentiate dementia from no dementia, and five reported data using a cut‐off of three. We did not consider it was appropriate to combine data using different cut‐off scores of test positivity. The cut‐off scores used to define a 'positive' AD‐8 included scores of one, two, three, and seven.

The included studies are international, including datasets from seven countries (Brazil, China, Japan, Singapore, Taiwan, the UK, the USA). Four different language versions of the AD‐8 were used, although in two papers it was not clear if the authors had used a translated version.

Using a cut‐off score of two, there was a spread of sensitivity and specificity (sensitivity range: 73% to 99%; specificity range: 11% to 93%). Using a cut‐off score of three, there was a spread of sensitivity and specificity (sensitivity range: 75% to 100%; specificity range: 40% to 91%). Table 1 provides a summary of the test accuracy for each study, presenting results at cut‐off scores of two and three where available, ordered by dementia prevalence.

AD‐8 using an informant cut‐off score of two

Seven studies (n = 3659) reported data using an AD‐8 informant cut‐off score of two (Correia 2011; Galvin 2006; Galvin 2007; Larner 2015; Meguro 2015; Razavi 2014; Yang 2016). Sensitivity was 0.92 (95% confidence interval (CI) 0.86 to 0.96); specificity was 0.64 (95% CI 0.39 to 0.82). The overall positive likelihood ratio was 2.53 (95% CI 1.38 to 4.64) and the negative likelihood ratio was 0.12 (95% CI 0.07 to 0.21). The summary describing test accuracy across the included studies in receiver operating characteristic (ROC) space is presented in Figure 3.


Summary ROC plot of AD‐8 informant cut‐off score 2. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Summary ROC plot of AD‐8 informant cut‐off score 2. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

AD‐8 using an informant cut‐off score of three

Five studies (n = 1060) reported data using an AD‐8 informant cut‐off score of three (Chan 2016; Correia 2011; Galvin 2006; Galvin 2007; Jackson 2016). These gave a sensitivity of 0.91 (95% CI 0.80 to 0.96) and specificity of 0.76 (95% CI 0.57 to 0.89). The overall positive likelihood ratio was 3.86 (95% CI 2.03 to 7.34) and the negative likelihood ratio was 0.12 (95% CI 0.06 to 0.24). The summary describing test accuracy across the included studies in ROC space is presented in Figure 4 .


Summary ROC plot of AD‐8 informant cut‐off score 3. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Summary ROC plot of AD‐8 informant cut‐off score 3. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

AD‐8 test accuracy at other diagnostic thresholds for dementia

AD‐8 informant cut‐off score of one

One study (n = 324) reported data using an AD‐8 informant cut‐off score of one (Galvin 2007). This had a sensitivity of 0.90 (95% CI 0.81 to 0.96) and specificity of 0.68 (95% CI 0.62 to 0.74).

AD‐8 informant cut‐off score of seven

One study (n = 77) reported data using an AD‐8 informant cut‐off score of seven (Jackson 2016). This had a sensitivity of 0.83 (95% CI 0.69 to 0.92) and specificity of 0.90 (95% CI 0.73 to 0.98).

Heterogeneity relating to setting

Primary care

One study (n = 309) was conducted in a primary care setting, specifically the waiting areas of healthcare centres in Singapore (Chan 2016). Data were only reported using an AD‐8 cut‐off score of three. The sensitivity was 0.91 (95% CI 0.78 to 0.97) and specificity was 0.91 (95% CI 0.87 to 0.94).

Community

Four studies (n = 2937) were conducted in community settings (Correia 2011; Galvin 2007; Meguro 2015; Yang 2016).

Data were available at a cut‐off score of two for all four studies. The sensitivity was 0.86 (95% CI 0.78 to 0.92) and specificity was 0.78 (95% CI 0.57 to 0.90). The overall positive likelihood ratio was 3.89 (95% CI 1.76 to 8.59) and the negative likelihood ratio was 0.18 (95% CI 0.10 to 0.31). The dementia prevalence ranged from 12% to 23%.

Secondary care

Four studies (n = 716) were conducted in secondary care settings. This included three in outpatient neurology, cognitive function or memory clinics (Galvin 2006; Larner 2015; Razavi 2014), and one study of hospitalised individuals with delirium (Jackson 2016).

Data were available for three of the studies at a cut‐off score of two (excluding Jackson 2016, which only reported data using a cut‐off score of three). The sensitivity was 0.96 (95% CI 0.92 to 0.98) and specificity was 0.39 (95% CI 0.16 to 0.69). The overall positive likelihood ratio was 1.58 (95% CI 0.98 to 2.57) and the negative likelihood ratio was 0.10 (95% CI 0.03 to 0.33). The dementia prevalence ranged from 33% to 90%.

Comparing pooled test accuracy between the two settings, the AD‐8 has a higher relative sensitivity (1.11, 95% CI 1.02 to 1.21), but lower relative specificity (0.51, 95% CI 0.23 to 1.09), in secondary care compared to community care settings (Figure 5).


Summary ROC plot of AD‐8 informant cut‐off score 2 comparing test accuracy in community versus secondary care settings. Red diamonds represent community studies and black circles those in secondary care. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region. No confidence region calculable for secondary care due to lack of available data.

Summary ROC plot of AD‐8 informant cut‐off score 2 comparing test accuracy in community versus secondary care settings. Red diamonds represent community studies and black circles those in secondary care. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region. No confidence region calculable for secondary care due to lack of available data.

Other sources of heterogeneity and sensitivity analyses

Age

We performed a sensitivity analysis removing the study which included participants with a low mean or median age (less than 65 years) (Larner 2015). Using a cut‐off score of two, test accuracy was similar after exclusion of this study, with an improvement in the specficity of the AD‐8: sensitivity was 0.91 (95% CI 0.83 to 0.95); specificity was 0.73 (95% CI 0.59 to 0.84); the positive likelihood ratio was 3.42 (95% CI 2.12 to 5.49); and the negative likelihood ratio was 0.12 (95% CI 0.07 to 0.23) (Figure 6).


Summary ROC plot of sensitivity analysis removing low average age. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Summary ROC plot of sensitivity analysis removing low average age. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Disease‐specific populations

One study included a disease‐specific population, individuals with a diagnosis of delirium (Jackson 2016). We performed a sensitivity analysis, removing this study from all the others using a cut‐off score of three. Results showed a lower sensitivity and higher specificity than the pooled analysis result. Sensitivity was 0.88 (95% CI 0.78 to 0.94) and specificity was 0.83 (95% CI 0.69 to 0.91). The overall positive likelihood ratio was 5.04 (95% CI 2.82 to 9.01) and the negative likelihood ratio was 0.15 (95% CI 0.08 to 0.26).

Reference standard dementia diagnostic method

Three studies used the CDR as their reference standard (Chan 2016; Galvin 2007; Meguro 2015); three used the DSM‐IV criteria (Correia 2011; Jackson 2016; Larner 2015); two studies used various criteria for dementia diagnosis (Galvin 2006; Razavi 2014); and one used the National Institute on Aging and Alzheimer's Association (NIA‐AA) criteria (Yang 2016). Quantitative comparisons based on the diagnostic criteria used were not possible due to the small numbers of studies using each method.

One study took a more inclusive definition of dementia, including those with a CDR score of 0.5 in their summary test accuracy data (Galvin 2007). For the purposes of analysis, we used the total numbers with a clinical diagnosis of dementia (CDR score greater than or equal to one), in common with the approach used in all other studies and the CDR definitions (Morris 1993). We used the summary sensitivity and specificity data presented in the paper, however the prevalence of dementia was lower than that quoted due to our exclusion of those 102 participants scoring 0.5 from the disease‐positive group (54% quoted in study reduced to 23%). As such, we performed a post‐hoc sensitivity analysis to check the effects of including the data from this study on our summary estimates at cut‐off scores of two and three.

Using a cut‐off score of two, removing data from Galvin 2007 resulted in similar sensitivity (0.93, 95% CI 0.86 to 0.97); but lower specificity (0.56, 95% CI 0.34 to 0.76); a positive likelihood ratio of 2.10 (95% CI 1.28 to 3.43) and negative likelihood ratio of 0.12 (95% CI 0.06 to 0.25). Using a cut‐off score of three, removing data from Galvin 2007 resulted in similar sensitivity (0.92, 95% CI 0.80 to 97); but lower specificity (0.71, 95% CI 0.48 to 0.87), with wider confidence intervals. The overall positive likelihood ratio was 3.19 (95% CI 1.61 to 6.33) and the negative likelihood ratio was 0.11 (95% CI 0.04 to 0.28).

Disease subtype

No studies considered a specific dementia subtype to facilitate sensitivity analysis based on this.

Language of AD‐8 administration

The English‐language version of the AD‐8 was used in five studies (Galvin 2006; Galvin 2007; Jackson 2016; Larner 2015; Razavi 2014). Three studies named which specific non‐English‐language versions were used (Chan 2016; Correia 2011; Meguro 2015). In two studies the language of AD‐8 administration was unclear (Chio 2017; Yang 2016).

The summary test accuracy of the English‐language version, used at a cut‐off score of two (n = 963) was calculated, resulting in sensitivity of 0.95 (95% CI 0.87 to 0.98); and specificity of 0.59 (95% CI 0.21 to 0.89). The overall positive likelihood ratio was 2.30 (95% CI 0.87 to 6.07) and the negative likelihood ratio was 0.08 (95% CI 0.03 to 0.22). Only three studies using a cut‐off score of three used the English‐language version, and thus we could not pool the data.

Discussion

available in

Summary of main results

This review summarises the test accuracy of the AD‐8 for dementia, across the range of settings in which it may be applied. Quantitative synthesis from seven of the identified studies suggests a summary sensitivity of 0.92 and specificity of 0.64 when an AD‐8 cut‐off score of two is used for the diagnosis of dementia. These results are international and include a population of 3659 participants. Using an AD‐8 cut‐off score of three to diagnose dementia achieves a similar sensitivity of 0.91, and higher specificity of 0.76, based on data from five studies and 1060 participants. There was significant clinical and methodological heterogeneity between the included studies, regarding their study populations, and dementia prevalence (which varied from 12% to 90%).

We found the usual pattern of compromise between sensitivity and specificity at extremes of scoring (for example, an AD‐8 score of one versus an AD‐8 score of seven). At the thresholds commonly used in practice — cut‐off scores of two or three — differences were less obvious. Our summary estimates of test properties at these scores have relatively large confidence intervals, particularly for specificity; this reflects the heterogeneity in the included studies. We can conclude that at usual threshold scores, the AD‐8 favours sensitivity over specificity. A cut‐off of two was used most commonly, which is consistent with the original development of the tool (Galvin 2005).

The setting of test administration affected the test accuracy estimates produced. Using a cut‐off score of two, the AD‐8 has a higher relative sensitivity (1.11), but lower relative specificity (0.51) in secondary care compared to community care settings. This suggests use of the AD‐8 in secondary care will produce more false positive results in those referred for specialist assessment, but fewer false negatives than when used in community cohorts. Arguably, in those referred for a specialist assessment and having already had some form of cognitive testing, specificity is important. In this case, our data would suggest that the AD‐8 is less suited as a diagnostic aid in a secondary‐care memory‐clinic setting. The data at a cut‐off score of three could not be pooled due to the small number of studies in each setting group; similarly, only one study examined the AD‐8 in primary care.

Removal of the study with a low average age (prespecified as less than 65 years) resulted in similar estimates of sensitivity (0.91), but improved specificity (0.73). The excluded study still had a mean age of greater than 60 years.

One study included only adults diagnosed as having delirium at the time of presentation (Jackson 2016). Delirium is more likely to occur in adults with a diagnosis of dementia (Inouye 2014), so we performed a sensitivity analysis removing this from the pooled estimate at a cut‐off score of three. Removing this study resulted in lower sensitivity and higher specificity than the summary estimate.

Quantitative analysis comparing the accuracy of different reference standard classifications was not possible due to the range of different approaches used.

Analysis comparing the accuracy of different language versions of the AD‐8 was limited due to the unclear description of the version used in two studies, and the lack of data available at a cut‐off score of three. The summary test accuracy of the English‐language version was comparable with the overall pooled estimate.

Strengths and weaknesses of the review

Strengths and weaknesses of included studies

Quality assessment using the QUADAS‐2 tool identified risks of bias and applicability concerns across all the included studies; we assessed no study as having low risk on all domains. Specific concerns are discussed in Characteristics of included studies and are graphically presented in Figure 2. These included the potential for selection bias introduced by exclusion criteria around comorbidities or unclear exclusion of participants; and issues around the conduct of index tests and suitability of informants, based on length of relationship, age, etc. There was a lack of data from primary care settings and acute hospital inpatient settings, which may be targets for opportunistic assessment using brief instruments.

Strengths and weaknesses of review process

This review has been conducted using a rigorous search strategy, developed and managed by an experienced Information Specialist in Cochrane. The search was determined using a structured approach. No restrictions were placed based on date or language, and translation services were used where necessary to facilitate the evaluation and inclusion of non‐English‐language publications. Our correspondence with study authors was successful; we obtained additional data, clarification around aspects that were unclear in the original publication, and additional papers evaluating the AD‐8 (Acknowledgements). Quality assessment was informed by use of dementia‐specific QUADAS‐2 anchoring statements, which were developed for this suite of reviews by experts in the field (Davis 2013).

Our review question was focused on evaluating the diagnostic accuracy of the AD‐8 for dementia, thus we excluded data evaluating the AD‐8 for diagnosing mild cognitive impairment from our quantitative analyses. We conducted prespecified sensitivity analyses around setting, studies of low average age, disease‐specific populations, and language of administration.

Comparisons with previous research

The AD‐8 is one of two informant‐based questionnaires whose test accuracy for dementia has been assessed by the Cochrane Dementia and Cognitive Improvement Group; the other is the Informant Questionnaire on Cognitive Decline in the Elderly (IQCODE). The IQCODE was found to have a similar test accuracy to the AD‐8 in community settings (sensitivity 0.80; specificity 0.84; 10 studies; 2644 individuals) (Quinn 2014). In secondary care settings, data were further divided into those evaluating the IQCODE in specialist memory settings and non‐specialist settings (Harrison 2015). In keeping with the current review, the majority of the data in the IQCODE review were from specialist settings and test accuracy estimates showed a higher specificity but comparable sensitivity (sensitivity 0.90; specificity 0.54; six studies; 1352 individuals) to the overall secondary care estimate for the AD‐8. The relative lack of AD‐8 data from primary care settings is consistent with the review of the IQCODE in primary care, which identified only one relevant study in this setting (Harrison 2014).

The AD‐8 was designed to differentiate cognitively normal individuals from those with mild dementia as part of clinical assessment in a primary care setting, derived using data from a longitudinal community cohort study (Galvin 2005). Prevalence of dementia was recognised to affect the test accuracy of the AD‐8, even at its inception (Galvin 2005). In the original validation study, the inclusion of 56 individuals with mild to severe dementia raised prevalence from 38% to 53%, improved sensitivity to 0.85 from 0.74 without a change to specificity of 0.86 (Galvin 2005). The pooled estimates reported in this review of all‐cause/all‐severity dementia have a higher sensitivity, but poorer specificity, than described in the first AD‐8 report.

During the production of this review, another author group published a systematic review of AD‐8 test accuracy (Chen 2018). Although the inclusion and exclusion criteria were similar, the review did not include some studies that we included in our review. It is not clear if these titles were considered and then rejected, or not considered. For using the AD‐8 to differentiate dementia from normal cognition, the summary test accuracy data presented in the review by Chen and colleagues were similar to our results, with high sensitivity (0.91, 95% CI 0.89 to 0.92) and lower specificity (0.78, 95% CI 0.76 to 0.80) (Chen 2018). No assessment of internal or external validity was presented.

We excluded case‐control designs, in common with the other reviews in the suite, due to the potential to over‐estimate test accuracy (Rutjes 2006; Kohn 2013). The case‐control studies excluded from this review have been analysed, and a summary sensitivity of 0.88 and specificity of 0.76 was reported (Anwer 2017). The effect of combining these studies with the studies included in our review has not been evaluated, however meta‐regression analysis did not show evidence of a systematic difference in the accuracy reported for the AD‐8 between non‐case‐control and case‐control designs (Anwer 2017).

Applicability of findings to the review question

This review focused on collating all available cross‐sectional data on the diagnostic accuracy of the AD‐8 to identify dementia. Our search strategy was robust and we believe we have evaluated all the evidence relevant to this question. Summary estimates of AD‐8 test accuracy are provided where sufficient appropriate data allowed, and findings are reported with consideration of study quality, risks of bias and applicability associated with study conduct.

Study flow diagram.
Figures and Tables -
Figure 1

Study flow diagram.

Risk of bias and applicability concerns summary: review authors' judgements about each domain for each included study
Figures and Tables -
Figure 2

Risk of bias and applicability concerns summary: review authors' judgements about each domain for each included study

Summary ROC plot of AD‐8 informant cut‐off score 2. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.
Figures and Tables -
Figure 3

Summary ROC plot of AD‐8 informant cut‐off score 2. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Summary ROC plot of AD‐8 informant cut‐off score 3. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.
Figures and Tables -
Figure 4

Summary ROC plot of AD‐8 informant cut‐off score 3. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

Summary ROC plot of AD‐8 informant cut‐off score 2 comparing test accuracy in community versus secondary care settings. Red diamonds represent community studies and black circles those in secondary care. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region. No confidence region calculable for secondary care due to lack of available data.
Figures and Tables -
Figure 5

Summary ROC plot of AD‐8 informant cut‐off score 2 comparing test accuracy in community versus secondary care settings. Red diamonds represent community studies and black circles those in secondary care. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region. No confidence region calculable for secondary care due to lack of available data.

Summary ROC plot of sensitivity analysis removing low average age. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.
Figures and Tables -
Figure 6

Summary ROC plot of sensitivity analysis removing low average age. The dark point is a summary point, the other points individual studies; the broken line represents 95% confidence region.

AD‐8 informant cutpoint 1.
Figures and Tables -
Test 1

AD‐8 informant cutpoint 1.

AD‐8 informant cutpoint 2.
Figures and Tables -
Test 2

AD‐8 informant cutpoint 2.

AD‐8 informant cutpoint 3.
Figures and Tables -
Test 3

AD‐8 informant cutpoint 3.

AD‐8 informant cutpoint 7.
Figures and Tables -
Test 4

AD‐8 informant cutpoint 7.

Sensitivity analysis removing low average age.
Figures and Tables -
Test 5

Sensitivity analysis removing low average age.

Sensitivity analysis removing disease‐specific study.
Figures and Tables -
Test 6

Sensitivity analysis removing disease‐specific study.

English language studies ‐ cutpoint 2.
Figures and Tables -
Test 7

English language studies ‐ cutpoint 2.

English language studies ‐ cutpoint 3.
Figures and Tables -
Test 8

English language studies ‐ cutpoint 3.

Sensitivity analysis removing CDR 0.5 ‐ cutpoint 2.
Figures and Tables -
Test 9

Sensitivity analysis removing CDR 0.5 ‐ cutpoint 2.

Sensitivity analysis removing CDR 0.5 ‐ cutpoint 3.
Figures and Tables -
Test 10

Sensitivity analysis removing CDR 0.5 ‐ cutpoint 3.

Summary of findings 1. Summary of findings table

Study ID

Country

Subjects (n)

Age (Years) (Mean, SD)

AD‐8 language

Cut‐off scores reported

Reference standard

Chan 2016

Singapore

309

Malay

3

Dementia

(CDR and DSM‐IV)

Chio 2017

Taiwan

153

76.9

(8.6)

Unclear

1, 2, 3, 4

Early cognitive impairment

(CDR)

Correia 2011

Brazil

109

Portuguese

2, 3

Dementia

(DSM‐IV)

Galvin 2006

US

255

73.3

(11.3)

English

2, 3

Various criteria for dementia

Galvin 2007

US

325

76.8

(8.9)

English

1, 2, 3

Dementia

(CDR)

Jackson 2016

UK

125

84.4

English

3, 7

Pre‐delirium dementia

(DSM‐IV)

Larner 2015

UK

212

Median 64.5

(IQR 12.7)

English

2

Dementia

(DSM‐IV)

Meguro 2015

Japan

572

Japanese

2

Dementia

(CDR)

Razavi 2014

US

186

77.8
(8.2)

English

2

Various criteria for dementia

Yang 2016

China

2063

79.5 (7.6)

Unclear

2

Dementia

(NIA‐AA criteria)

CDR: Clinical Dementia Rating scale
DSM‐IV: Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition
NIA‐AA: National Institute on Aging and Alzheimer's Association

Figures and Tables -
Summary of findings 1. Summary of findings table
Summary of findings 2. Summary of findings table

What is the accuracy of the AD‐8 test for the detection of dementia when differing thresholds are used to define AD‐8 positive cases?

Population

Adults whose informant completed an AD‐8 questionnaire, with no restrictions on the case mix of recruited participants

Setting

Primary care, community and secondary care

Index test

AD‐8 informant‐based questionnaire

Reference standard

Clinical diagnosis of dementia made using any recognised classification system

Studies

We included cross‐sectional but not case‐control studies

Test

Summary accuracy (95% CI)

No. of participants (studies)

Dementia prevalence

Implications, quality and comments

AD‐8

cut‐off score 1

Sensitivity: 0.90 (0.81 to 0.96)

Specificity: 0.68 (0.62 to 0.74)

n = 324

(1 study)

n = 73

(60%)

The majority of studies used the informant cut‐off score of 2 to define AD‐8 test positivity.

As the cut‐off score is raised, there is an increase in specificity. Sensitivity is similar at a cut‐off score of 2 or 3, but falls as the cut‐off score to define test positivity rises.

False positives (individuals wrongly identified as having possible dementia and referred for further investigations) and false negatives (individuals with dementia who are not identified and not referred for specialist assessment and management) can both be associated with harms. If used as a screening test (i.e. an initial assessment, which will be followed by further testing) then a pattern of higher sensitivity may be preferable.

The dementia prevalence within the included studies was highly varied (12% to 90%) and may reflect that for each cut‐off score, data were included from all settings in which the AD‐8 has been used. These analyses need to be interpreted with caution as the properties of the test may vary by setting and the desired balance of sensitivity and specificity will vary by purpose of testing.

AD‐8 cut‐off score 2

Sensitivity: 0.92

(0.86 to 0.96)

Specificity: 0.64

(0.39 to 0.82)

+LR: 2.53

(1.38 to 4.64)

‐ve LR: 0.12

(0.07 to 0.21)

n = 3659

(7 studies)

n = 1016 (28%

range:

12% to 90%)

AD‐8 cut‐off score 3

Sensitivity: 0.91

(0.80 to 0.96)

Specificity: 0.76

(0.57 to 0.89)

+LR: 3.86

(2.03 to 7.34)

‐ve LR: 0.12

(0.06 to 0.24)

n = 1060

(5 studies)

n = 396

(37%

range:

14% to 90%)

AD‐8 cut‐off score 7

Sensitivity: 0.83

(0.69 to 0.92)

Specificity: 0.90

(0.73 to 0.98)

n = 77

(1 study)

n = 47 (61%)

95% CI: 95% confidence interval
+LR: positive likelihood ratio
‐LR: negative likelihood ratio

Figures and Tables -
Summary of findings 2. Summary of findings table
Summary of findings 3. Summary of findings table

What is the accuracy of the AD‐8 test for detection of dementia in different settings and using different languages of administration?

Population

Adults whose informant completed an AD‐8 questionnaire, with no restrictions on the case mix of recruited participants

Setting

Primary care, community and secondary care

Index test

AD‐8 informant‐based questionnaire using a cut‐off score of 2 to define positivity

Reference standard

Clinical diagnosis of dementia made using any recognised classification system

Studies

We included cross‐sectional but not case‐control studies

Comparative analyses

Test

No. of participants (studies)

Dementia prevalence

Findings

Implications

Community versus primary care versus secondary care setting

Total n = 4045 (9)

n = 2937 (4) Community

n= 309 (1) Primary care

n = 716 (4) Secondary care

Total n = 1107 (27%)

Community n = 601 (20%)

Primary n = 44 (14%)

Secondary care n = 462 (65%)

Using a cut‐off score of 2, the AD‐8 has a higher relative sensitivity (1.11, 95% CI 1.02 to 1.21) in secondary care compared to community care settings.

Using a cut‐off score of 2, the AD‐8 has a lower relative specificity (0.51, 95% CI 0.23 to 1.09) in secondary care compared to community care settings.

Only one study used the AD‐8 in primary care settings; data were at a cut‐off score of 3.

Two studies used AD‐8 in community and two studies used AD‐8 in secondary care settings, with a cut‐off score of 3. The numbers were insufficient to allow formal quantitative synthesis.

Dementia prevalence in secondary care is more than double that seen in the community.

The AD‐8 has a higher sensitivity, but poorer specificity, in secondary care settings.

English language

Total: n = 1040 (5) English language

Total n = 535 (51%)

The test accuracy of English language administration of the AD‐8, at a cut‐off score of 2, has a higher sensitivity and lower specificity than summary estimates for all studies at a cut‐off score of 2. Estimates for specificity were imprecise due to the small number of studies included.

Sensitivity: 0.95 (95% CI 0.87 to 0.98)

Specificity: 0.59 (95% CI 0.21 to 0.89)

+LR: 2.30 (95% CI 0.87 to 6.07)

‐LR: 0.08 (95% CI 0.03 to 0.22)

We were unable to assess English language at a cut‐off score of 3 due to lack of available data at this cut‐off score.

We were unable to assess the non‐English‐language group due to the small numbers reporting language of administration clearly.

There is a need to improve clarity of reporting around language of test administration to evaluate diagnostic accuracy of non‐English‐language versions of the AD‐8

95% CI: 95% confidence interval
+LR: positive likelihood ratio
‐LR: negative likelihood ratio

Figures and Tables -
Summary of findings 3. Summary of findings table
Table 1. Summary of test accuracy at study level by dementia prevalence

Study

Included participants

(n)

Dementia

n (%)

AD‐8 Threshold

Sensitivity

(%)

Specificity

(%)

Galvin 2006

241

217 (90)

2

92

46

Galvin 2006

241

217 (90)

3

89

67

Razavi 2014

186

129 (69)

2

99

74

Jackson 2016

77

47 (61)

3

98

40

Larner 2015

212

69 (33)

2

97

11

Galvin 2007

324

73 (23)

2

84

93

Galvin 2007

324

73 (23)

3

75

90

Yang 2016

2015

444 (22)

2

90

78

Correia 2011

109

15 (14)

2

73

61

Correia 2011

109

15 (14)

3

100

67

Chan 2016

309

44 (14)

3

91

91

Meguro 2015

572

69 (12)

2

88

68

Where multiple thresholds were reported in the primary paper, we present the data for both a cut‐off score of 2 and 3.

The total number of participants is adjusted to reflect the data used in quantitative synthesis.

Figures and Tables -
Table 1. Summary of test accuracy at study level by dementia prevalence
Table Tests. Data tables by test

Test

No. of studies

No. of participants

1 AD‐8 informant cutpoint 1 Show forest plot

1

324

2 AD‐8 informant cutpoint 2 Show forest plot

7

3659

3 AD‐8 informant cutpoint 3 Show forest plot

5

1060

4 AD‐8 informant cutpoint 7 Show forest plot

1

77

5 Sensitivity analysis removing low average age Show forest plot

6

3447

6 Sensitivity analysis removing disease‐specific study Show forest plot

4

983

7 English language studies ‐ cutpoint 2 Show forest plot

4

963

8 English language studies ‐ cutpoint 3 Show forest plot

3

642

9 Sensitivity analysis removing CDR 0.5 ‐ cutpoint 2 Show forest plot

6

3335

10 Sensitivity analysis removing CDR 0.5 ‐ cutpoint 3 Show forest plot

4

736

Figures and Tables -
Table Tests. Data tables by test