Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

Test Accuracy of Cognitive Screening Tests for Diagnosis of

Dementia and Multidomain Cognitive Impairment in Stroke


Rosalind Lees, MA; Johann Selvarajah, PhD; Candida Fenton, MSc; Sarah T. Pendlebury, DPhil;
Peter Langhorne, PhD; David J. Stott, MD; Terence J. Quinn, MD

Background and Purpose—Guidelines recommend screening stroke-survivors for cognitive impairments. We sought to
collate published data on test accuracy of cognitive screening tools.
Methods—Index test was any direct, cognitive screening assessment compared against reference standard diagnosis of
(undifferentiated) multidomain cognitive impairment/dementia. We used a sensitive search statement to search multiple, cross-
disciplinary databases from inception to January 2014. Titles, abstracts, and articles were screened by independent researchers.
We described risk of bias using Quality Assessment of Diagnostic Accuracy Studies tool and reporting quality using Standards
for Reporting of Diagnostic Accuracy guidance. Where data allowed, we pooled test accuracy using bivariate methods.
Results—From 19 182 titles, we reviewed 241 articles, 35 suitable for inclusion. There was substantial heterogeneity:
25 differing screening tests; differing stroke settings (acute stroke, n=11 articles), and reference standards used
(neuropsychological battery, n=21 articles). One article was graded low risk of bias; common issues were case–control
methodology (n=7 articles) and missing data (n=22). We pooled data for 4 tests at various screen positive thresholds:
Addenbrooke’s Cognitive Examination-Revised (<88/100): sensitivity 0.96, specificity 0.70 (2 studies); Mini Mental
State Examination (<27/30): sensitivity 0.71, specificity 0.85 (12 studies); Montreal Cognitive Assessment (<26/30):
sensitivity 0.95, specificity 0.45 (4 studies); MoCA (<22/30): sensitivity 0.84, specificity 0.78 (6 studies); Rotterdam-
CAMCOG (<33/49): sensitivity 0.57, specificity 0.92 (2 studies).
Conclusions—Commonly used cognitive screening tools have similar accuracy for detection of dementia/multidomain
impairment with no clearly superior test and no evidence that screening tools with longer administration times perform
better. MoCA at usual threshold offers short assessment time with high sensitivity but at cost of specificity; adapted
cutoffs have improved specificity without sacrificing sensitivity. Our results must be interpreted in the context of modest
Downloaded from http://ahajournals.org by on November 28, 2022

study numbers: heterogeneity and potential bias.   (Stroke. 2014;45:3008-3018.)


Key Words: cognitive impairment ◼ dementia ◼ sensitivity ◼ specificity ◼ stroke

S troke-survivors are at particular risk of cognitive decline.


Three month dementia prevalence is ≥30%, and even
minor stroke events have cognitive sequel.1,2 Poststroke cog-
for screening or triage and specialist assessment to define the
cognitive problem offered depending on the results.
Although there is general agreement on the merits of post-
nitive impairment is associated with increased mortality, dis- stroke cognitive assessment, there is no consensus on a pre-
ability, and institutionalization.3 The importance of cognitive ferred testing strategy.6–8 Various cognitive screening tools are
change is highlighted by stroke-survivors themselves. In a available with substantial variation in test used.9,10
national priority setting exercise, cognitive impairment was The clinical meaning of cognitive problems after stroke will
voted the single most important topic for stroke research.4 vary according to test context. Cognitive impairment diagnosed
A first step in management of cognitive problems is recog- in the first days post stroke may reflect a mix of delirium, stroke-
nition and diagnosis. Informal clinician assessment will miss specific impairments, and prestroke cognitive decline.2,11,12 In
important cognitive problems,5 and formal cognitive testing is the longer term, assessments aim to make or refute a demen-
recommended.6–8 The ideal would be expert, multidisciplinary tia diagnosis. Common to all test situations is a final diagno-
assessment informed by comprehensive investigations. This sis of presence/absence of clinically important impairments. A
approach is not feasible at a population level. In practice, a screening assessment should detect this clinical syndrome of
2-step system is adopted, with baseline cognitive testing used all-cause, poststroke multidomain cognitive impairment.

Received April 16, 2014; accepted July 31, 2014.


From the Institute of Cardiovascular and Medical Sciences (R.L., P.L., D.J.S., T.J.Q.), Institute of Neuroscience and Psychology (J.S.), MRC/CSO Social
and Public Health Sciences Unit (C.F.), University of Glasgow, UK; and NIHR Biomedical Research Centre and Stroke Prevention Research Unit, John
Radcliffe Hospital, Oxford, UK (S.T.P.).
Presented in part at the European Stroke Conference, Nice, France, May 6–9, 2014.
The online-only Data Supplement is available with this article at http://stroke.ahajournals.org/lookup/suppl/doi:10.1161/STROKEAHA.114.
005842/-/DC1.
Correspondence to Terence J. Quinn, MD, Institute of Cardiovascular and Medical Sciences, University of Glasgow, Room 2.44, New Lister Building,
Glasgow Royal Infirmary, 10–16 Alexandra Parade, Glasgow, UK. E-mail terry.quinn@glasgow.ac.uk
© 2014 American Heart Association, Inc.
Stroke is available at http://stroke.ahajournals.org DOI: 10.1161/STROKEAHA.114.005842

3008
Lees et al   Cognitive Screening in Stroke   3009

Collation and synthesis of the evidence around cognitive blinded to each others’ results. On review of paired data, disagree-
screening tools will help guide policy and practice and high- ment was resolved by group discussion.
We developed a sensitive search strategy in collaboration with an
light knowledge gaps. We sought to perform systematic review Information Scientist (CF) and with assistance from the Cochrane
and meta-analysis, describing accuracy of screening tools for Dementia and Cognitive Improvement Group. Search terms were
assessing dementia and multidomain cognitive impairment in developed using a concepts-based approach, using Medical Subject
stroke-survivors. Heading terms and other controlled vocabulary. Concepts of inter-
est were stroke, dementia, and cognitive assessment. Our cognitive
assessment concept included terms relating to cognitive tests used
Methods in stroke, based on previous survey data.9,10 We searched multiple,
We used systematic literature review and meta-analysis techniques international, cross-disciplinary electronic databases from inception
specific to test accuracy13 and followed best practice in reporting.14 to January 2014. (Full search strategy is detailed in the online-only
Data Supplement I–II.) We checked reference lists of relevant studies
Aims and reviews for further titles, repeating the process until no new titles
Coprimary aims were to describe the following: were found.
We screened all titles generated by initial searches for relevance, and
1. Accuracy of cognitive screening tests for clinical diagnosis of corresponding abstracts were assessed and potentially eligible studies
multidomain, cognitive impairment/dementia in stroke-survivors. reviewed as full manuscripts against inclusion criteria. As a check of in-
2. Accuracy of brief (<5 minutes completion time), cognitive screen- ternal validity, a random selection of 2000 titles from the original search
ing tests against more detailed neuropsychological assessments. was reassessed by a third author (TQ). We tested external validity of our
If data allowed, secondary objectives were to compare differing search by sharing lists of included articles with independent researchers
cutpoint scores used to define a threshold of test positivity and to (Acknowledgments). One author (TQ) identified exemplar articles rel-
compare the effects of heterogeneity with specific reference to test evant to the review question (online-only Data Supplement II), and we
context and diagnostic reference standard. assessed whether these titles were detected by search strategy.

Data Extraction and Management


Index Test We extracted data to a study-specific proforma piloted against 2 ar-
Index tests of interest were any direct-to-patient, cognitive screening ticles (available from contact author).
tests. We included any screening test where the authors described it as For screening tests that give an ordinal summary score, various cut-
such. We excluded informant-based assessments and tests that require points can be used to define test positive cases. Where data were given
testing equipment considered nonstandard for a stroke service. We rec- for several thresholds, we extracted separate data for each. Where a
ognize that language and visuospatial function are important compo- study may have included useable data, but these were not presented
nents of cognition, but did not include assessments of tools designed in the published article, we contacted the authors directly. If the same
to exclusively test these domains. We did not include studies that com- data set was presented in >1 publication, we included the primary
pared screening tools with no reference to diagnostic gold standard. article.
Downloaded from http://ahajournals.org by on November 28, 2022

Target Condition and Reference Standard Quality Assessment


Our target condition of interest was all-cause (undifferentiated) multido- We assessed quality of study reporting using the dementia-specific ex-
main cognitive impairment post stroke. This rubric recognizes that a di- tension to the Standards for Reporting of Diagnostic Accuracy check-
agnosis of important poststroke cognitive problems can be made without list.16 We assessed methodological quality using the Quality Assessment
necessarily assigning a dementia label or specifying a pathological subtype. of Diagnostic Accuracy Studies tool (QUADAS-2).17 We have previ-
As reference standard, we included clinical diagnosis of dementia ously developed QUADAS-2 anchoring statements specific to cognitive
made using any recognized classification system. We also included assessment.18
multidomain, cognitive problems as detailed on a neuropsychologi-
cal assessment, provided the test battery was comprehensive and the
results were interpreted to give a diagnostic formulation. We did not Statistical Analysis
include single domain cognitive impairment or cognitive impairment We assessed accuracy of screening tests against a dichotomous vari-
no dementia because of inconsistency in operationalization of these able cognitive impairment/no cognitive impairment. We created stan-
syndromes.15 We prespecified subgroup analyses comparing demen- dard 2-by-2 data tables describing binary test results cross-classified
tia diagnosis and neuropsychological battery-based diagnosis. with binary reference standard. We did not include case–control stud-
For assessment of brief tests, we accepted results from a more ies in pooled analyses. We used accepted cut-offs for multidomain
detailed multidomain, screening assessment as reference standard, impairment/dementia published in the literature: Addenbrooke’s
for example, comparison of Hodgkinson’s abbreviated mental test Cognitive Examination-Revised <88, Rotterdam-CAMCOG<33,
against Folstein’s Mini Mental State Examination (MMSE). MMSE<27, and <25. For Montreal Cognitive Assessment (MoCA),
a cut-off of <26 was used together with the lower cut-off of <22 be-
Participants and Setting cause the former was developed to detect single-domain impairment.
Where data allowed, we calculated sensitivity, specificity, and corre-
Our focus was stroke-survivors. Where study population was a mix sponding 95% confidence intervals (95% CI) and created test accuracy
of stroke survivors and other patients, we included if proportion of forest plots (RevMan 5.1, Cochrane Collaboration). We pooled test ac-
stroke-survivors was >75%. We made no distinction between stroke curacy data using the bivariate approach (Statistical Analysis Software
subtypes, but excluded studies of traumatic intracerebral hemorrhage v9.1; SAS Institute Inc, USA). We used a bespoke macro developed with
and subarachnoid hemorrhage. We did not include case studies (de- assistance of a statistical team with an interest in test accuracy.19 Summary
fined as having fewer than 10 participants). metrics of interest were sensitivity/specificity and positive/negative like-
We included studies conducted in any clinical setting and at any lihood ratios and clinical utility index.20 We created summary curves in
time post stroke. We operationalized time since stroke as hyperacute received operating characteristic space with corresponding 95% predic-
(first 7 days); acute (8–14 days); post acute (15 days to 3 months); tion intervals.
medium term (3–12 months), and longer term (post 1 year). We assessed potential heterogeneity through visual inspection of for-
est plots. We prespecified 2 factors that may contribute to heterogeneity,
Search Strategy timing of assessment and reference standard. We dichotomized studies
All aspects of searching, data extraction, and study assessment were into acute (classified as hyperacute or acute) or nonacute and described
performed by 2 reviewers (RL, JS) based in different centers and reference standard as clinical (clinical diagnosis of dementia) and
3010  Stroke  October 2014

neuropsychological battery (multidomain cognitive impairment). We as- Results


sessed effect by plotting summary received operating characteristic curves
From 19 182 titles, we reviewed 241 full articles of which 35 (34
by covariate. Taking timing of assessment as example, we calculated
sensitivity/specificity restricted to the single cognitive assessment with data sets)21–56 were suitable for inclusion. Scope of included litera-
greatest number of studies. We described a relative sensitivity/specific- ture was international, with articles from 16 different countries.
ity comparing acute and nonacute studies, where a result of unity sug- We detailed the study selection process in a Preferred
gests no difference in that diagnostic property between settings. We did Reporting Items for Systematic Review and Meta-Analysis
not quantify publication bias because there is no assessment applicable
to test accuracy.
(PRISMA) flow diagram (Figure 1).
Where articles fulfilled inclusion criteria but did not have data suitable Our validation check suggested the initial search was appro-
for this method of analysis, we offer a tabulated/narrative description. priate because all prespecified articles were found on first search.
Downloaded from http://ahajournals.org by on November 28, 2022

Figure 1. Preferred Reporting Items for Systematic Review and Meta-Analysis (PRISMA) flow diagram. ACE-R indicates Addenbrooke’s
cognitive examination-revised; AMT, Abbreviated Mental Test; FN, false-negative; FP, false-positive; HSROC, Hierarchical Summary
ROC; MMSE, mini-mental state examination; MoCA, Montreal cognitive assessment; R-CAMCOG, Rotterdam-CAMCOG; ROC, received
operating characteristic; TN, true-negative; and TP, true-positive.
Lees et al   Cognitive Screening in Stroke   3011

Table 1. Test Accuracy Data for Studies Where Reference Standard was
Clinical Diagnosis of Dementia
n n (%) With Index Test Summary
Study Included Dementia (Threshold) Results
Brookes34* 152 N/A BMET Discriminates AD/SVD
Cumming 201022 149 42 (28%) Cog4 Sens: 53% Spec: 84%
de Koning 200037 300 55 (19%) R-CAMCOG (33) Sens: 91% Spec: 90%
de Koning 200538 121 35 (29%) R-CAMCOG (33) Sens: 66% Spec: 94%
de Koning 52
300 55 (19%) CAMCOG (81) Sens: 91% Spec: 80%
MMSE (25) Sens: 80% Spec: 76%
Dong 201239 300 60 (25%) MoCA (21) Sens: 88% Spec: 64%
MMSE (24) Sens: 88% Spec: 67%
Hershey 42
63 13 (21%) CSCE (20) Sens: 85% Spec: 87%
Hobson43* 149 N/A PNB (55) Sens: 71% Spec: 93%
Srikanth24 67 8 (12%) MMSE (23) Sens: 14% Spec: 100%
Tang47 142 10 (12%) MDRS (22) Sens: 82% Spec: 90%
MMSE (18) Sens: 80% Spec: 88%
Wu49 206 95 (46%) MoCA (23) Sens: 65% Spec: 79%
Where >1 test threshold was described, we present the primary data. AD indicates Alzheimer’s
disease; BMET, brief memory and executive test; CSCSE, cognitive capacity screening examination;
MDRS, Mattis dementia rating scale; MMSE, mini-mental state examination; MoCA, Montreal cognitive
assessment; N/A, not applicable; PNB, preliminary neuropsychological battery; R-CAMCOG, Rotterdam-
CAMCOG; and SVD, small vessel disease.
*Case–control study, data not included in pooled analyses.

Accuracy of Screening Tools for Diagnosis of descriptions of training and expertise of assessors (n=25 arti-
Cognitive Impairment cles; online-only Data Supplement V)
Downloaded from http://ahajournals.org by on November 28, 2022

In total n=32 articles (n=3562 participants) were eligible.21–53


We tabulated summary descriptors for studies using clinical Meta-Analyses
diagnosis reference standard (n=11 articles)22,24,34,37–39,43,47,49,51,52 We were able to pool test accuracy data for 4 tests:
and those using detailed neuropsychological assessment Addenbrooke’s Cognitive Examination-Revised (threshold
(n=21; Tables 1–2).21,23,25-33,35,36,40–42,44–46,48,50 score ≤88); MMSE (thresholds ≤24 and ≤26), MoCA (thresh-
There was considerable heterogeneity in study population: olds <26 and <22), and Rotterdam-CAMCOG (threshold
setting and test strategy. Twenty-three different tests were <33). No test had sensitivity and specificity that were signifi-
described, commonest MMSE (n=16 articles) and MoCA cantly different from others. MoCA at traditional threshold
(n=8; Tables 1–2; online-only Data Supplement III). was sensitive at cost of specificity; specificity improved if
Screening tests gave a spread of test accuracy, sensitivity thresholds were adjusted (Table 3; Figure 3; online-only Data
range 14% to 100%, and specificity range 0% to 100% with Supplement VI)
trade-off between sensitivity and specificity. Where authors
compared >1 test in the same population, the comparator was Heterogeneity
usually MMSE, and for most articles, MMSE was more spe- We assessed effect of timing and reference standard using
cific and less sensitive than other tests. Both the Executive MMSE data. Comparing 6 acute studies20,21,31,37,38,43 and 6
Function Performance Test and the Repeatable Battery for nonacute studies22,30,34,39,46,50; gave a relative sensitivity, 0.73
Assessment of Neuropsychological Status showed strong (95% CI, 0.58–0.93) and relative specificity, 1.12 (95% CI,
correlations with scores on neuropsychological batteries, but 1.01–1.25). These data suggest that accuracy varies with
data were not suitable for pooled analysis.25–28 assessment timing and context, with acute testing yielding
higher sensitivity and lower specificity. From analysis of
Quality and Reporting paired summary received operating characteristic curves, the
One article was graded low risk for all QUADAS2 domains.21 overall accuracy was better for those studies where assess-
Common issues of concern were use of case–control methodol- ment was performed within the acute period (online-only
ogy (n=7 articles) and potential lack of blinding (n=12; Figure 2; Data Supplement VII). Comparing clinical dementia ref-
online-only Data Supplement IV). Five articles attempted to erence standard21,24,40,47,52 against neuropsychological bat-
include patients with moderate to severe aphasia.21–25 tery23,32,33,36,39,41,45,46 suggested no difference in test properties
Standards for Reporting of Diagnostic Accuracy assess- dependent on the reference standard used, with relative sensi-
ment suggested consistent areas of poor reporting, particu- tivity of 0.86 (95% CI, 0.67–1.11) and relative specificity of
larly around the handling of missing data (n=22 articles) and 1.05 (95% CI, 0.95–1.16).
3012  Stroke  October 2014

Table 2. Reference Standard of Cognitive Impairment Based on Multidomain


Neuropsychological Battery
n n (%) With Index Test Summary
Study Included Dementia (Threshold) Results
Agrell31* 64 N/A MMSE (23) Sens: 56% Spec: 24%
Baum26* 73 N/A EPFT Correlates with NPB
Blake32 112 31 (28%) MMSE (24) Sens: 62% Spec: 88%
Bour33 194 22 (11%) MMSE (23) Sens: 96% Spec: 83%
Cartoni35 30 27 (90%) MEAMS (3) Sens: 52% Spec: 100%
Cumming 201336 60 39 (65%) MMSE (24) Sens: 54% Spec: 81%
MoCA (24) Sens: 97% Spec: 52%
Desmond29* 72 6 (8%) TICS (25) Sens: 100% Spec: 83%
MMSE (24) Sens: 83% Spec: 87%
Dong 201040 100 60 (60%) MMSE (24) Sens: 86% Spec: 82%
MoCA (21) Sens: 90% Spec: 77%
Godefroy23 95 64 (67%) MMSE (24) Sens: 70% Spec: 94%
MoCA (24) Sens: 92% Spec: 58%
Grace41 70 32 (46%) MMS (79) Sens: 69% Spec: 79%
MMSE (24) Sens: 44% Spec: 84%
Green27 60 N/A RBANS (3) Sens: 84% Spec: 90%
Jodzio44 44 24 (55%) WCST PPV: 21 NPV: 75
Larson28 158 N/A RBANS Correlates with NPB
Morris45 101 51 (84%) MMSE (24) Sens: 58% Spec: 77%
ACE-R (88) Sens: 94% Spec: 0%
Nokleby 51
49 28 (60%) COGNISTAT (59) Sens: 59% Spec: 67%
SINS (2) Sens: 71% Spec: 67%
Downloaded from http://ahajournals.org by on November 28, 2022

CDT 9 Sens: 63% Spec: 67%


Nys50* 72 N/A MMSE (28) Sens: 100% Spec: 40%
Pendlebury 30
91 19 (21%) MoCA (25) Sens: 100% Spec: 46%
MMSE (24) Sens: 56% Spec: 96%
ACE-R (88) Sens: 84% Spec: 85%
Pendlebury46(telephone) 91 19 (21%) T’phone MoCA (19) Sens: 89% Spec: 46%
TICSm (25) Sens: 85% Spec: 56%
Salvadori21 80 47 (59%) MoCA (21) Sens: 91% Spec: 76%
Wolf25 20 N/A EPFT Correlate with NPB
Wong * 48
68 N/A MoCA (21) Sens: 73% Spec: 75%
Where >1 test threshold was described, we present the primary data. ACE-R indicates Addenbrooke’s cognitive
examination-revised; CDT, clock drawing test; EPFT, executive performance function test; MEAMS, middlesex elderly
assessment mental state; MMSE, mini-mental state examination; MoCA, Montreal cognitive assessment; N/A, not
applicable; NPB, neuropsychological battery; RBANS, repeatable battery for assessment of neuropsychological status;
SINS, screening instrument neuropsychological impairments in stroke; TICS, telephone interview cognitive status; and
WCST, Wisconsin Card Sorting Test.
*Case–control study, data not included in pooled analyses.

Accuracy of Brief Tools were partially successful in this aim. Although there is an
We found 3 suitable articles (n=294 participants; online-only extensive, cross-disciplinary literature describing cogni-
Data Supplement VIII).53–55 Two articles were graded high risk tive testing in stroke, number of articles using the classical
of bias and applicability concerns because of case–control test accuracy paradigm of index test versus reference (gold)
methodology and patient inclusion.53,55 standard was limited. Eligible articles were characterized by
substantial heterogeneity and risk of bias, and the potential to
Discussion describe summary analyses at individual test level was limited.
Previous systematic reviews of cognitive test accuracy have Accepting these caveats, we can still offer conclusions from
focused on older adults with narrative results only. Our aim our pooled analysis. There was no clearly superior cognitive
was to provide a stroke-specific, literature synthesis to allow screening test. Given the relative consistency in accuracy,
evidence-based recommendations on cognitive testing. We choice of test strategy should be informed by other factors,
Lees et al   Cognitive Screening in Stroke   3013

Figure 2. Assessment of risk of bias and


applicability concerns using the revised
Quality Assessment of Diagnostic Accuracy
Studies tool (QUADAS-2). Assessment at
individual study level is available in online-
only Data Supplement.

such as purpose of testing, feasibility; acceptability, and There is no universally accepted gold standard for demen-
opportunity cost. Recent guidance and practice has tended tia. Rather than restricting to a particular pathological subtype,
to favor novel tests compared with the traditionally popular our focus was all cause dementia post stroke. We think this
MMSE.6,7,56 Although there may be good reasons to favor approach is in keeping with current clinical practice, where
other tests, our data do not suggest that MMSE is inferior for initial screening highlights potential cognitive issues that can
the diagnosis of multidomain impairment, albeit MMSE may subsequently be characterized with more detailed assess-
lack sensitivity for single domain impairment.15,46,56 There was ments. Literature describing screening tests for diagnosis of
a trend toward better clinical utility for Rotterdam-CAMCOG, specific dementia subtypes was limited, and we were only
but the small number of studies and corresponding wide confi- able to describe test accuracy for all cause (undifferentiated)
dence intervals precludes definitive recommendation. dementia. A priori, we decided to include multidomain impair-
Preferred test properties will depend on test purpose, par- ment based on neuropsychological assessment in our reference
ticularly the level of cognitive impairment to be detected. It standard, and this approach maximized potential for pooled
could be argued that for initial screening, sensitivity is pre- analysis and recognizes that clinical diagnosis in certain stroke
ferred compared with specificity. In this case, MoCA and situations is not always feasible or appropriate. We used mul-
Addenbrooke’s Cognitive Examination-Revised may be pre- tidomain rather than single domain impairment because this is
ferred. The high false-positive rate for multidomain impair- closest to current operational classifications of dementia. We
ment obtained using MoCA <26 is to be expected because this did not specify content of the neuropsychological assessment
threshold was chosen for detection of single domain impair- and recognize the substantial heterogeneity in batteries used.15
ment/mild cognitive impairment.15,46 Our findings suggest that There were many potential sources of heterogeneity (sample
an adjusted cut-off <22 has improved overall test properties for size, case–control methodology, investigator training); indeed
multidomain impairment in stroke as suggested previously.15,46 a degree of heterogeneity is expected in test accuracy meta-
Downloaded from http://ahajournals.org by on November 28, 2022

We were pragmatic in our choice of index test and refer- analyses. We were not able to assess all potentially contributory
ence standard. Our screening test rubric included relatively factors across the modest number of included studies and accept
short assessments (MoCA) through to fairly lengthy batteries that this limits our findings. We chose to focus on reference stan-
(Repeatable Battery for Assessment of Neuropsychological dard and timing of assessment as the most important possible
Status). We did not find significant improvements in sensitiv- sources of heterogeneity. We found that test properties varied
ity/specificity comparing shorter and longer screens (MoCA with timing and seemed to favor earlier assessment, albeit our
and Addenbrooke’s Cognitive Examination-Revised). This rubric of acute assessment included studies from first hours ≤14
may suggest that increasing the length of test batteries does days post ictus. Comparative analysis of test properties require
not necessarily improve the test accuracy. careful interpretation; a conservative interpretation would be

Table 3. Pooled Test Accuracy for 4 Cognitive Screening Tests Against a Reference Standard of Cognitive Impairment
Articles Cognitive Sensitivity Specificity Positive Likelihood Negative Likelihood Clinical
Test (Threshold) (Patients) Impairment, n (%) (95% CI) (95% CI) Ratio (95% CI) Ratio (95% CI) Utility Index
ACE-R (<88/100) 2 (192) 52 (27%) 96.2 (0.90–1.0) 0.70 (0.59–0.80) 3.19 (2.24–4.54)) 0.06 (0.01–0.22) +0.67
−0.67
MMSE (<25/30) 12 (1639) 483 (30%) 0.71 (0.60–0.80) 0.85 (0.80–0.89) 4.73 (3.63–6.17) 0.34 (0.25–0.47) +0.46
−0.70
MMSE (<27/30) 5 (445) 195 (44%) 0.88 (0.82–0.92) 0.62 (0.50–0.73) 2.33 (1.72–3.17) 0.19 (0.13–0.29) +0.65
−0.54
MoCA (<22/30) 6 (726) 289 (39%) 0.84 (0.76–0.89) 0.78 (0.69–0.84) 3.75 (2.77–5.08) 0.20 (0.15–0.29) +0.59
−0.63
MoCA (<26/30) 4 (326) 131 (40%) 0.95 (0.89–0.98) 0.45 (0.34–0.57) 1.73 (1.43–2.10) 0.10 (0.04–0.23) +0.60
−0.36
R-CAMCOG (<33/49) 2 (421) 90 (21%) 0.81 (0.57–0.93) 0.92 (0.87–0.95) 10.18 (6.41–16.18) 0.20 (0.07–0.52) +0.59
−0.86
For clinical utility index, +, positive; −, negative. ACE-R indicates Addenbrooke’s cognitive examination-revised; MMSE, mini-mental state examination; MoCA,
Montreal cognitive assessment; and R-CAMCOG, Rotterdam-CAMCOG.
3014  Stroke  October 2014

B
Downloaded from http://ahajournals.org by on November 28, 2022

Figure 3. Summary received operating characteristic (ROC) curve and forest plot describing test accuracy studies of (A) Folstein’s mini-
mental state examination (MMSE) at threshold of <25/30; and (B) MMSE at threshold of <27/30. (Continued)
Lees et al   Cognitive Screening in Stroke   3015

MoCA threshold <22 for diagnosis cognitive impairment

D
Downloaded from http://ahajournals.org by on November 28, 2022

MoCA threshold <26 for diagnosis cognitive impairment

Figure 3 (Continued). Summary ROC curve and forest plot describing test accuracy studies of (C) Montreal cognitive assessment (MoCA) at
threshold of <26/30; and (D) MoCA at threshold of <22/30. The filled circle represents the summary (pooled) test accuracy; unfilled circles repre-
sent individual articles. The solid line is the summary ROC curve; the broken line is the 95% confidence interval (CI) around the summary point.
3016  Stroke  October 2014

that a test suitable for longer-term poststroke assessment may missing data, future studies may wish to describe feasibility
not be suitable for early assessment and that cognitive testing and acceptability of testing, as well as classical test accuracy
can (and should) be performed in the acute stroke unit. metrics.
Brief assessment tools are attractive for use in busy stroke
settings but only if they have acceptable accuracy. In this Conclusions
regard, the limited number of studies looking at brief assess- All our results must be interpreted with caution as included
ments is unfortunate. Based on data available, screening using studies had substantial heterogeneity and potential for bias.
clock drawing test or abbreviated mental test may have a role However, we can offer some practical advice to clinicians.
for initial assessment; but they should not be considered a sub- There is no clearly superior cognitive screening test approach
stitute for subsequent multidomain testing. and no evidence that newer or longer screening tests neces-
sarily offer better test accuracy than established screens with
Limitations of Included Studies shorter administration time. The strategy for cognitive testing
Our review highlights issues in the design and reporting of in stroke must be informed by the purpose of the test and by
cognitive test accuracy studies. There was little reporting of other metrics, such as feasibility, acceptability, and opportu-
missing or indeterminate data; however, we know that sub- nity cost. If the purpose of the initial screening is to pick-up
stantial numbers of stroke patients are unable to complete all potential cases to allow further assessment, then MoCA
multidomain screening tools. The same concern holds for a may be the preferable test because it offers high sensitivity
reference standard based on extensive neuropsychological with shorter administration time than other equally sensitive
testing. Limiting test accuracy data to those able to complete tests. For MoCA use in stroke, screen positive thresholds may
testing will bias results, tending to inflate test accuracy. We need to be adapted if the aim is to assess for dementia/mul-
would encourage use of the intention to diagnose approach, tidomain impairment, and test accuracy may vary with time
where the traditional 2×2 test accuracy table is expanded with since stroke event.
cells representing those unable to complete index test and ref-
erence standard.57 Acknowledgments
A related issue is around generalizability of the included We are grateful for expert assistance from Cochrane Dementia and
subjects. In our quality assessment, we scored several articles Cognitive Improvement Group to Cochrane authors who shared rel-
evant papers, specifically Dr Ingrid Arevalo Rodriguez, Dr Sarah
as inappropriate exclusion, including the studies that excluded Cullum/Dr Daniel Davis, and Professor Nadina Lincoln. We are
patients with moderate to severe aphasia or inability to consent. grateful to all study authors who responded to queries or shared data.
We recognize the challenges of cognitive testing in this group,
Downloaded from http://ahajournals.org by on November 28, 2022

who by definition have at least single domain impairment; Sources of Funding


however, excluding patients with communication problems or Project supported by a Chest Heart and Stroke Scotland research
frank confusion from test accuracy research will limit external award. J. Selvarajah is supported by a National Health Service
validity. Research Scotland fellowship, Chief Scientists Office Scotland. D.J.
Stott has received support toward cognitive assessment research from
Chief Scientists Office and the Orr Bequest. T.J. Quinn received trav-
Strengths and Limitations of Review el support from Royal College of Physicians and Surgeons Glasgow
Our review provides a contemporary synthesis of the rapidly Scholarship and Graham Wilson Travel Fund.
evolving field of cognitive screening in stroke. We offer a sen-
sitive search strategy following best practice in conduct and Disclosures
reporting58 and with multiple embedded internal and external None.
validity checks. The review has limitations; our focused ques-
tion excluded potentially informative articles, for example References
where screening tests are compared against each other.59 We did 1. Pendlebury ST, Wadling S, Silver LE, Mehta Z, Rothwell PM. Transient
cognitive impairment in TIA and minor stroke. Stroke. 2011;42:3116–3121.
not assess other important test metrics, such as responsiveness 2. Pendlebury ST, Rothwell PM. Prevalence, incidence, and factors associ-
to change or reliability, or other cognitive states, such as (single ated with pre-stroke and post-stroke dementia: a systematic review and
domain) mild cognitive impairment. These questions would meta-analysis. Lancet Neurol. 2009;8:1006–1018.
need their own review. We recognize that cognitive assessment 3. Leys D, Hénon H, Mackowiak-Cordoliani MA, Pasquier F. Poststroke
dementia. Lancet Neurol. 2005;4:752–759.
in stroke is evolving; 15 (47%) of included articles in our pri- 4. Pollock A, St George B, Fenton M, Firkins L. Top ten research priorities
mary review were published since 2010. Novel, stroke cognition relating to life after stroke. Lancet Neurol. 2012;11:209.
assessments are emerging, but no test accuracy data were avail- 5. Chodosh J, Petitti DB, Elliott M, Hays RD, Crooks VC, Reuben DB, et
al. Physician recognition of cognitive impairment: evaluating the need
able at time of review. for improvement. J Am Geriatr Soc. 2004;52:1051–1059.
6. Hachinski V, Iadecola C, Petersen RC, Breteler MM, Nyenhuis DL,
Future Research Black SE, et al. National Institute of Neurological Disorders and Stroke-
Canadian Stroke Network vascular cognitive impairment harmonization
Our subgroup analysis suggests that test accuracy will vary standards. Stroke. 2006;37:2220–2241.
depending on time since stroke, but could not suggest an 7. Royal College of Physicians. National Clinical Guidelines for Stroke
optimal time for assessment. Early assessment in the Acute Clinical. Effectiveness and Evaluation Unit. London Royal College of
Physicians; 2012.
Stroke Unit has practical advantages and could allow for
8. SIGN Guideline 118. Management of patients with stroke. Rehabilitation,
timely intervention; however, few studies assessed patients prevention and management of complications and discharge planning
in the acute period. Given the issues with generalizability and Scottish Intercollegiate Guidelines Network (SIGN); 2010
Lees et al   Cognitive Screening in Stroke   3017

9. Lees R, Fearon P, Harrison JK, Broomfield NM, Quinn TJ. Cognitive and and relationships to somatic and mental variables. Aging (Milano).
mood assessment in stroke research: focused review of contemporary 2000;12:439–444.
studies. Stroke. 2012;43:1678–1680. 32. Blake H, McKinney M, Treece K, Lee E, Lincoln NB. An evaluation of
10. Lees RA, Broomfield NM, Quinn TJ. Questionnaire assessment of usual screening measures for cognitive impairment after stroke. Age Ageing.
practice in mood and cognitive assessment in Scottish stroke units. 2002;31:451–456.
Disabil Rehabil. 2014;36:339–343. 33. Bour A, Rasquin S, Boreas A, Limburg M, Verhey F. How predic-
11. Hénon H, Lebert F, Durieu I, Godefroy O, Lucas C, Pasquier F, et al. tive is the MMSE for cognitive performance after stroke? J Neurol.
Confusional state in stroke: relation to preexisting dementia, patient 2010;257:630–637.
characteristics, and outcome. Stroke. 1999;30:773–779. 34. Brookes RL, Hannesdottir K, Lawrence R, Morris RG, Markus HS. Brief
12. Shi Q, Presutti R, Selchen D, Saposnik G. Delirium in acute stroke: a Memory and Executive Test: evaluation of a new screening test for cogni-
systematic review and meta-analysis. Stroke. 2012;43:645–649. tive impairment due to small vessel disease. Age Ageing. 2012;41:212–218.
13. Higgins JPT, Green S, eds. Cochrane Handbook for Systematic Reviews 35. Cartoni A, Lincoln NB. The sensitivity and specificity of the middlesex
of Interventions Version 5.0.2. The Cochrane Collaboration; 2009. elderly assessment of mental state (MEAMS) for detecting cognitive
14. Stroup DF, Berlin JA, Morton SC, Olkin I, Williamson GD, Rennie D, et impairment after stroke. Neuropsychol Rehabil. 2005;15:55–67.
al. Meta-analysis of observational studies in epidemiology. JAMA. 2000; 36. Cumming TB, Churilov L, Linden T, Bernhardt J. Montreal cognitive
283:2008–2012. assessment and mini-mental state examination are both valid cognitive
15. Pendlebury ST, Mariz J, Bull L, Mehta Z, Rothwell PM. Impact of dif- tools in stroke. Acta Neurol Scand. 2013;128:122–129.
ferent operational definitions on mild cognitive impairment rate and 37. de Koning I, Dippel DW, van Kooten F, Koudstaal PJ. A short screen-
MMSE and MoCA performance in transient ischaemic attack and stroke. ing instrument for poststroke dementia: the R-CAMCOG. Stroke.
Cerebrovasc Dis. 2013;36:355–362. 2000;31:1502–1508.
16. Noel-Storr AH, McCleery JM, Richard E, Ritchie CW, Flicker L, Cullum 38. de Koning I, van Kooten F, Koudstaal PJ, Dippel DW. Diagnostic value of
S, et al. Reporting standards for studies of diagnostic test accuracy in the Rotterdam-CAMCOG in post-stroke dementia. J Neurol Neurosurg
dementia. Neurology. 2014;83:364–373. Psychiatry. 2005;76:263–265.
17. Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ, Reitsma 39. Dong Y, Venketasubramanian N, Chan BP, Sharma VK, Slavin MJ,
JB, et al; QUADAS-2 Group. QUADAS-2: a revised tool for the Collinson SL, et al. Brief screening tests during acute admission in patients
quality assessment of diagnostic accuracy studies. Ann Intern Med. with mild stroke are predictive of vascular cognitive impairment 3-6
2011;155:529–536. months after stroke. J Neurol Neurosurg Psychiatry. 2012;83:580–585.
18. Davis DHJ, Creavin ST, Noel-Storr A, Quinn TJ, Smailagic N, Hyde 40. Dong YH, Sharma VK, Chan BPL, Venketasubramanian N, Teoh HL,
C, et al. Neuropsychological tests for the diagnosis of Alzheimer’s Seet RC, et al. The Montreal cognitive assessment (MoCA) is superior to
disease dementia and other dementias: a generic protocol for cross- the mini-mental state examination (MMSE) for the detection of vascular
sectional and delayed-verification studies. Cochrane Database Syst Rev. cognitive impairment after acute stroke. J Neurolog Sci. 2010;299:15–18.
2013;3:CD010460. 41. Grace J, Nadler JD, White DA, Guilmette TJ, Giuliano AJ, Monsch AU, et
19. Takwoingi Y, Deeks JJ. MetaDAS: A SAS macro for meta-analysis al. Folstein vs modified Mini-Mental State Examination in geriatric stroke.
of diagnostic accuracy studies. User Guide Version 1.3. http://srdta. Stability, validity, and screening utility. Arch Neurol. 1995;52:477–484.
cochrane.org/. Accessed January 2014. 42. Hershey LA, Jaffe DF, Greenough PG, Yang SL. Validation of cogni-
20. Mitchell AJ. SensitivityxPPV is a recognized test called the clinical util- tive and functional assessment instruments in vascular dementia. Int J
ity index. Eur J Epidemiol. 2011;26:251–252. Psychiatry Med. 1987;17:183–192.
Downloaded from http://ahajournals.org by on November 28, 2022

21. Salvadori E, Pasi M, Poggesi A, Chiti G, Inzitari D, Pantoni L. Predictive 43. Hobson JP, Leeds L, Meara RJ. The feasibility of cognitive screening of
value of MoCA in the acute phase of stroke on the diagnosis of mid-term patients with ischaemic stroke using the preliminary neuropsychological
cognitive impairment. J Neurol. 2013;260:2220–2227. battery. Psychol Health. 2003;18:655–665.
22. Cumming TB, Blomstrand C, Bernhardt J, Linden T. The NIH stroke 44. Jodzio K, Biechowska D. Wisconsin card sorting test as a measure of
scale can establish cognitive function after stroke. Cerebrovasc Dis. executive function impairments in stroke patients. Appl Neuropsychol.
2010;30:7–14. 2010;17:267–277.
23. Godefroy O, Fickl A, Roussel M, Auribault C, Burnicourt JM, Lamy 45. Morris K, Hacker V, Lincoln NB. The validity of the Addenbrooke’s
C, et al. Is the Montreal cognitive assessment superior to the mini- Cognitive Examination-Revised (ACE-R) in acute stroke. Disabil
mental state examination to detect poststroke cognitive impairment? Rehabil. 2012;34:189–195.
Stroke. 2011;42:1712–1716. 46. Pendlebury ST, Mariz J, Bull L, Mehta Z, Rothwell PM. MoCA, ACE-
24. Srikanth V, Thrift AG, Fryer JL, Saling MM, Dewey HM, Sturm JW, R, and MMSE versus the National Institute of Neurological Disorders
et al. The validity of brief screening cognitive instruments in the diag- and Stroke-Canadian Stroke Network Vascular Cognitive Impairment
nosis of cognitive impairment and dementia after first-ever stroke. Int Harmonization Standards Neuropsychological Battery after TIA and
Psychogeriatr. 2006;18:295–305. stroke. Stroke. 2012;43:464–469.
25. Wolf TJ, Stift S, Connor LT, Baum C. Feasibility of using the EFPT 47. Tang WK, Mok V, Chan SS, Chiu HF, Wong KS, Kwok TC, et al.
to detect executive function deficits at the acute stage of stroke. Screening of dementia in stroke patients with lacunar infarcts: compari-
Stroke. 2010;36:405–412. son of the mattis dementia rating scale and the mini-mental state exami-
26. Baum CM, Connor LT, Morrison T, Hahn M, Dromerick AW, Edwards nation. J Geriatr Psychiatry Neurol. 2005;18:3–7.
DF. Reliability, validity, and clinical utility of the Executive Function 48. Wong A, Xiong YY, Kwan PW, Chan AY, Lam WW, Wang K, et al.
Performance Test: a measure of executive function in a sample of people The validity, reliability and clinical utility of the Hong Kong Montreal
with stroke. Am J Occup Ther. 2008;62:446–455. Cognitive Assessment (HK-MoCA) in patients with cerebral small ves-
27. Green S, Sinclair E, Rodgers E, Birks E, Lincoln N. The Repeatable sel disease. Dement Geriatr Cogn Disord. 2009;28:81–87.
Battery for the Assessment of Neuropsychological Status (RBANS) for 49. Wu Y, Wang M, Ren M, Xu W. The effects of educational background
post stroke cognitive impairment screening. Int J Ther Rehab. 2013;20:7. on Montreal Cognitive Assessment screening for vascular cognitive
28. Larson E, Kirschner K, Bode R, Heinemann A, Goodman R. Construct impairment, no dementia, caused by ischemic stroke. J Clin Neurosci.
and predictive validity of the repeatable battery for the assessment of 2013;20:1406–1410.
neuropsychological status in the evaluation of stroke patients. J Clin Exp 50. Nys GM, van Zandvoort MJ, de Kort PL, Jansen BP, Kappelle LJ, de
Neuropsychol. 2005;27:16–32. Haan EH. Restrictions of the Mini-Mental State Examination in acute
29. Desmond DW, Tatemichi TK, Hanzawa L. The telephone interview for stroke. Arch Clin Neuropsychol. 2005;20:623–629.
cognitive status. Reliability and validity in a stroke sample. Int J Geriat 51. Nøkleby K, Boland E, Bergersen H, Schanke AK, Farner L, Wagle J, et
Psychiat. 1994;9:803–807. al. Screening for cognitive deficits after stroke: a comparison of three
30. Pendlebury ST, Welch SJ, Cuthbertson FC, Mariz J, Mehta Z, Rothwell screening tools. Clin Rehabil. 2008;22:1095–1104.
PM. Telephone assessment of cognition after transient ischemic attack 52. de Koning I, van Kooten F, Dippel DWJ, van Harskamp F, Grobbee DE,
and stroke: modified telephone interview of cognitive status and telephone Kluft C, Koudstaal PJ. Stroke. The CAMCOG: A useful screening instru-
Montreal Cognitive Assessment versus face-to-face Montreal Cognitive ment for dementia in stroke patients. Stroke. 1998:29:2080–2086.
Assessment and neuropsychological battery. Stroke. 2013;44:227–229. 53. Johnson-Greene D, Touradji P, Emmerson LC. The Three Cities Test:
31. Agrell B, Dehlin O. Mini mental state examination in geriatric preliminary validation of a short bedside memory test in persons with
stroke patients. Validity, differences between subgroups of patients, acute stroke. Top Stroke Rehabil. 2009;16:321–329.
3018  Stroke  October 2014

54. Lees R, Corbet S, Johnston C, Moffitt E, Shaw G, Quinn TJ. Test transient ischemic attack and stroke: a population-based study. Stroke.
accuracy of short screening tests for diagnosis of delirium or 2010;41:1290–1293.
cognitive impairment in an acute stroke unit setting. Stroke. 57. Schuetz GM, Schlattmann P, Dewey M. Use of 3x2 tables with an inten-
2013;44:3078–3083. tion to diagnose approach to assess clinical performance of diagnostic
55. Wong A, Mok VC, Yim P, Fu M, Lam WW, Yau C, et al. The executive tests. BMJ. 2012;345:e6717
clock drawing task (CLOX) is a poor screening test for executive dys- 58. Quinn TJ, Fearon P, Noel-Storr AH, Young C, McShane R, Stott DJ.
function in Chinese elderly patients with subcortical ischemic vascular Informant Questionnaire on Cognitive Decline in the Elderly (IQCODE)
disease. J Clin Neurosci. 2004;11:493–497. for the diagnosis of dementia within community dwelling populations.
56. Pendlebury ST, Cuthbertson FC, Welch SJ, Mehta Z, Rothwell PM. Cochrane Database Syst Rev. 2014;4:CD010079.
Underestimation of cognitive impairment by Mini-Mental State 59. Blackburn DJ, Bafadhel L, Randall M, Harkness KA. Cognitive screen-
Examination versus the Montreal Cognitive Assessment in patients with ing in the acute stroke setting. Age Ageing. 2013;42:113–116.
Downloaded from http://ahajournals.org by on November 28, 2022

You might also like