Introduction

Clinical assessment of cognition and functional abilities plays an important role in identifying, diagnosing, and monitoring individuals at risk for cognitive decline, mild cognitive impairment (MCI), and dementia due to Alzheimer’s disease (AD). Problems in everyday function help predict those who will more rapidly decline and convert to dementia (1, 2). The Everyday Cognition Scale (ECog) is a clinical assessment of decline in instrumental activities of daily living that map to six cognitive domains (3). ECog correlates well with functional and cognitive status, and is associated with clinical diagnosis and AD biomarker status (4, 5). However, many traditional cognitive assessments for clinical research are time consuming and expensive (6) because they are paper and pencil-based and are optimally conducted in a supervised clinical setting with an experienced and certified rater (7), which generates logistical challenges and financial barriers. In-person clinical assessments can deter broad participation in clinical studies and pose special barriers for individuals who do not live near a research clinic or have access to transportation (8). These barriers reduce diversity in AD research.

One strategy to overcome these barriers is to leverage digital technologies to facilitate screening, assessment, and enrollment (9), such as through establishing online registries (1012). The Brain Health Registry (BHR) is an online registry at the University of California San Francisco (UCSF) that captures data remotely using online questionnaires and neuropsychological tests adapted from traditional clinical assessments to longitudinally monitor cognition and functional status (13). There is emerging evidence for validity of BHR measures (5, 14, 15). However, the validity of unsupervised measures obtained remotely online compared to traditional measures obtained in a supervised clinical setting has not been fully established.

In this study, we tested the hypothesis that online ECog closely corresponded with in-clinic ECog in 94 participants enrolled in both the Alzheimer’s Disease Neuroimaging Initiative (ADNI) and Brain Health Registry (BHR). The direct comparison of the traditional in-clinic measure of ECog and the adapted online measure of ECog is a novel step in assessing the validity of online measures of cognition and functional ability. We assessed how well the measure of self-reported ECog collected in BHR, an unsupervised online setting, agrees with and predicts the same measure of self-reported ECog collected in ADNI, a supervised clinical setting. Since the correspondence of in-clinic and online ECog measures could not logically exceed the test-retest reliability of the in-clinic ECog measure, we used this as a benchmark for validity of the online measure by comparing the associations between online ECog and in-clinic ECog to the test-retest reliability of in-clinic ECog.

Methods

This study compared data collected in a supervised clinical setting from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) to data collected remotely online from the Brain Health Registry (BHR). Data were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). The ADNI was launched in 2003 as a public-private partnership, led by Principal Investigator Michael W. Weiner, MD. The primary goal of ADNI has been to test whether serial magnetic resonance imaging (MRI), positron emission tomography (PET), other biological markers, and clinical and neuropsychological assessment can be combined to measure the progression of mild cognitive impairment (MCI) and early Alzheimer’s disease (AD). During the four phases of the ADNI study (ADNI-1, ADNI-GO, ADNI-2, ADNI-3), participants were rolled over from previous phases for continued monitoring, while new participants were added with each phase (16). Participants were diagnosed as cognitively normal, MCI, or with dementia due to AD according to ADNI inclusion criteria (17).

Participants

Participants enrolled in both ADNI and the Brain Health Registry to complete the self-reported online Everyday Cognition (n=94)

Existing ADNI-3 participants were invited to enroll in BHR to complete supplemental online assessments. Of the 603 ADNI-3 participants invited to join BHR, 110 consented and enrolled in BHR (referred to as ADNI-BHR participants), and 94 of those completed ECog both in-clinic and online and have evaluable linked data (Figure 1). All 94 participants who completed ECog both in-clinic and online, regardless of cognitive status, were included in this study due to the limited sample size.

Figure 1
figure 1

Enrollment Flow Chart for ADNI and BHR

Approximate (approx.) ADNI enrollment numbers were provided from http://adni.loni.usc.edu/study-design (16). Bold boxes represent study samples used in this analysis.

Participants enrolled in ADNI to estimate test-retest reliability of the self-reported in-clinic Everyday Cognition (n=472)

Out of all of the participants enrolled in ADNI-2 [16], 472 completed two separate in-clinic ECog assessments about six months apart (Figure 1). The average time between assessments was 189 (interquartile range (IQR) 106 to 352) days. An estimate of the test-retest reliability of self-reported in-clinic ECog was obtained from these 472 ADNI participants, of which 250 were diagnosed as stable cognitively normal, 208 were diagnosed as stable MCI, 11 converted from cognitively normal to MCI, and 3 reverted from MCI to cognitively normal according to ADNI inclusion criteria (17).

Measurements

Measurements collected in-clinic in ADNI

Participants in ADNI were asked to complete the Everyday Cognition Scale (ECog) in-clinic. The ECog includes 39 items assessing an individual’s capability to perform everyday tasks in comparison to 10 years prior (3). Ratings use a four-point Likert scale ranging from “better or no change” to “consistently much worse” to report changes in ability to perform tasks in six cognitively-relevant domains: Memory, Language, Visuospatial Abilities, and Planning, Organization, and Divided Attention (18). A total ECog score, which was the sum of the ratings from all completed items divided by the number of items completed, ranged from 1 to 4 (19).

Participant age, gender (categorized as male or female), years of education (ranging from 0 to 20 years), self-reported race (categorized as American Indian or Alaskan Native, Asian, Native Hawaiian or other Pacific Islander, Black or African American, White, more than one race, or unknown) and ethnicity (categorized as Hispanic or Latino, not Hispanic or Latino, or unknown) were collected as part of their regular, ADNI study visits.

Measurements collected online in the Brain Health Registry

Participants in Brain Health Registry (BHR) were asked to complete an online adaptation of the ECog Scale in an unsupervised setting, such as a quiet space at home with an internet connection. All 39 of the in-clinic ECog items were used verbatim in the online ECog (14). The online ECog was scored in the same way as the in-clinic ECog, resulting in a total online ECog score ranging from 1 to 4.

Statistical Analysis

Associations between self-reported online Everyday Cognition and in-clinic Everyday Cognition in ADNI-BHR participants (n=94)

We tested the hypothesis that online self-reported total ECog score corresponded well with in-clinic self-reported total ECog score using a Bland-Altman plot and linear regression. The nearest date of in-clinic ECog completion in ADNI was matched to the date the online ECog was completed at time of enrollment into BHR. The Bland-Altman plot was constructed and the mean difference was calculated to evaluate the method agreement between online ECog and in-clinic ECog. 95% limits of agreement were calculated as ±1.96 standard deviations of the mean difference. We would expect 95% of differences between measurements by two methods to lie between these limits (20).

To evaluate how well online ECog could predict in-clinic ECog, a linear regression model was fit with online ECog total score as the sole predictor and in-clinic ECog total score as the outcome. Linearity of the regression model was assessed using diagnostics including a Locally Weighted Scatterplot Smoother (LOWESS) and component plus residual plots (21). The fit of the linear model was confirmed by fitting a linear spline as well as the restricted cubic spline to test for departure from linearity. Normality of the error term was assessed using diagnostic graphical methods, such as the kdensity nonparametric estimate of the normal distribution of the residuals and the curvature of the normal Q—Q plot. Constant variance was assessed using the shape of the residual versus fitted (RVF) plots. Influential points were identified from a boxplot using an absolute DFBETA statistic cutoff of 0.5. Prediction error was corrected for optimism with 10-fold cross-validation using STATA’s crossfold utility to avoid overfitting.

Test-retest reliability of self-reported in-clinic Everyday Cognition in ADNI participants (n=472)

We assessed the test-retest reliability of self-reported Everyday Cognition collected in a clinic from ADNI participants who completed two separate repeating in-clinic ECog assessments about six months apart using intraclass correlation coefficient (ICC), a Bland-Altman plot, and linear regression. The Bland-Altman plot and 95% limits of agreement, calculated as ±1.96 standard deviations of the mean difference, were constructed to evaluate the agreement between in-clinic ECog at time of enrollment into ADNI and in-clinic ECog at six months. The linear regression model was fit with in-clinic ECog total score at time of enrollment into ADNI as the sole predictor and in-clinic ECog total score at six months as the outcome. All of the same methods used to evaluate the first model for linearity, normality, constant variance, and influential points were used. Prediction error was corrected for optimism with 10-fold cross-validation using STATA’s crossfold utility to avoid overfitting.

We compared the Bland-Altman results assessing the method agreement between online and in-clinic ECog to the Bland-Altman results assessing the repeatability of in-clinic ECog. We evaluated prediction performance of the linear regression models using R2 and compared the R2 result from the estimated test-retest reliability of in-clinic ECog to the R2 result from the prediction of in-clinic ECog using online ECog. Statistical analyses were performed using STATA (version 16.1).

Results

Associations between self-reported online Everyday Cognition and in-clinic Everyday Cognition in ADNI-BHR participants (n=94)

94 participants were enrolled in ADNI-BHR and completed self-reported ECog both online and in-clinic. The median and interquartile range (IQR) of age for this sample was 75 (71 to 80) years, 61% of the sample was female, and 95% identified as white (Table 1). 77% of this sample was diagnosed as cognitively normal and 22% was diagnosed with MCI.

Table 1 Characteristics of participants

Out of the 94 participants enrolled in ADNI-BHR and included in this analysis, 67 of the participants completed the in-clinic ECog first and 27 of the participants completed the online ECog first. The average absolute time between the online ECog completion and the in-clinic ECog completion was 24.5 (IQR 0 to 112) days. A scatterplot showed a strong association between self-reported online ECog total scores and self-reported in-clinic ECog total scores (Figure 2A). The Bland-Altman plot (Figure 3A) showed that average in-clinic ECog scores were a little higher than online ECog scores (mean difference: 0.11 (95% CI 0.06 to 0.17)) and the 95% limits of agreement were −0.41 (95% CI −0.47 to −0.36) to 0.64 (95% CI 0.58 to 0.69).

Figure 2
figure 2

Scatterplots and Component Plus Residual Plots of Everyday Cognition Total Scores

(A) Self-report online Everyday Cognition and in-clinic Everyday Cognition (n=94) (B) LOWESS curve from the CPR plot for linear regression to predict in-clinic Everyday Cognition using self-report online Everyday Cognition (n=94) (C) Repeating self-report in-clinic Everyday Cognition at time of enrollment into ADNI and in-clinic Everyday Cognition at six months (n=472) (D) LOWESS curve from the CPR plot for in-clinic Everyday Cognition linear regression (n=472)

Figure 3
figure 3

Bland-Altman Plots and 95% Limits of Agreement

(A) Differences between measures of in-clinic Everyday Cognition and online Everyday Cognition against mean (Bland-Altman) Plot, with the limits of agreement (red dashed line), the mean difference (black dashed line), shaded confidence intervals. (B) Differences between repeated measures of in-clinic Everyday Cognition against mean (Bland-Altman) Plot, with the limits of agreement (red dashed line), the mean difference (black dashed line), shaded confidence intervals

In linear regression analysis, we found that the mean online ECog total score increased by 0.81 (95% CI 0.65 to 0.97) for each unit increase in in-clinic ECog total score. The optimism-corrected R2 using 10-fold cross-validation was estimated as 0.60 (95% CI 0.41 to 0.78) (Table 2). The LOWESS curve from the component plus residual (CPR) plot agreed with the linear fit (Figure 2B), as did the restricted cubic spline model. When two influential points were omitted from the regression model, we found that the mean online ECog total score increased by 0.89 (95% CI 0.76 to 1.03) for each unit increase in in-clinic ECog total score. The adjusted R2 was 0.65.

Table 2 Estimated regression coefficients of linear regression models

Test-retest reliability of self-reported in-clinic Everyday Cognition in ADNI participants (n=472)

We estimated the test-retest reliability of self-reported in-clinic ECog using the sample of 472 ADNI participants who completed two separate in-clinic ECog assessments approximately six months apart. The median and interquartile range (IQR) of age for this sample was 80 (64 to 97) years, 49% of the sample was female, and 92% identified as white (Table 1). A scatterplot showed a strong association between self-reported in-clinic ECog total scores at time of enrollment into ADNI-2 and six months (Figure 2C) and the intraclass correlation (ICC) was estimated to be 0.71 (95% CI 0.62 to 0.79). The Bland-Altman plot (Figure 3B) showed that average in-clinic ECog scores at enrollment were slightly higher than in-clinic ECog scores at six months (mean difference: 0.01 (95% CI −0.02 to 0.03)) and the 95% limits of agreement were −0.61 (95% CI −0.64 to −0.58) to 0.62 (95% CI 0.59 to 0.65). This indicated good test-retest reliability of the self-reported in-clinic ECog. 54% of this sample was diagnosed as cognitively normal and 46% was diagnosed with MCI.

In linear regression analysis, we found that the mean in-clinic ECog total score collected at time of enrollment into ADNI increased by 0.79 (95% CI 0.73 to 0.85) for each unit increase in in-clinic ECog total score at six months. The optimism-corrected R2 using 10-fold cross-validation was estimated as 0.61 (95% CI 0.53 to 0.69) (Table 2). The LOWESS curve from the component plus residual (CPR) plot indicated slight departure from linearity starting around an in-clinic ECog score at enrollment of 2.5 (Figure 2D). This departure from linearity was confirmed by fitting a linear spline with a knot placed at an in-clinic ECog score of 2.5, as well as by fitting the restricted cubic spline. Using a linear spline regression model with a knot placed at an in-clinic ECog score of 2.5, we found that the mean in-clinic ECog total score collected at time of enrollment increased by 0.85 (95% CI 0.78 to 0.92) for each unit increase in in-clinic ECog total score at six months. The adjusted R2 was 0.61. When one influential point was omitted from the regression model, we found that the mean in-clinic ECog total score at enrollment increased by 0.82 (95% CI 0.77 to 0.88) for each unit increase in in-clinic ECog total score. The adjusted R2 was 0.63.

Discussion

The major finding of this study was that the online ECog closely corresponded with in-clinic ECog and may provide as much information as repeating in-clinic ECog. This is an important finding because it demonstrates that the remotely administered online ECog may be just as useful as the traditional in-clinic ECog. In the Bland-Altman plots, we found that the mean difference of 0.01 between in-clinic ECog scores completed approximately six months apart indicated better agreement compared to the mean difference of 0.11 between online and in-clinic ECog scores. However, we found that the 95% limits of agreement closely corresponded. For 95% of individuals, a measurement by online ECog would be between 0.41 units less and 0.64 units greater than a measurement by in-clinic ECog. Similarly, for 95% of individuals, a measurement by in-clinic ECog at six months would be between 0.61 units less and 0.62 units greater than a measurement by in-clinic ECog at time of enrollment.

The intraclass correlation (ICC) of the repeating in-clinic ECog was 0.71 (95% CI 0.62 to 0.79), meaning that 71% of the total variability in ECog was due to differences between subjects (true between-subject variability) and 29% was due to within-subject random measurement error or changes over the six month follow up. Our results from the prediction of in-clinic ECog using online ECog suggested that the ECog collected in an unsupervised online setting could predict the ECog collected in a supervised clinic setting. We used 10-fold cross-validation to avoid overfitting and give a realistic estimate of the usefulness of the online ECog. The optimism-corrected R2 of 0.60 (95% CI 0.41 to 0.78) showed that online ECog scores explained a moderate proportion, 60%, of the total variability of in-clinic ECog scores. The optimism-corrected R2 of the model used to predict in-clinic ECog using online ECog (R2= 0.60, 95% CI 0.41 to 0.78) was similar to the optimism-corrected R2 of the model used to estimate the test-retest reliability of self-reported in-clinic ECog (R2=0.61 (95% CI 0.53 to 0.69). Based on R2 comparison, this indicated that online ECog corresponded well with in-clinic ECog and provided almost as much information as repeating the in-clinic ECog. These results demonstrate the realistic usefulness and clinical validity of online measures of cognition and functional ability in clinical research.

There was evidence of cognitive impairment in both the ADNI sample used to estimate the test-retest reliability of self-reported in-clinic ECog (46% (219/472) MCI) and the ADNI-BHR sample used for the prediction of in-clinic ECog using online ECog (22% (22/94) MCI or dementia). So the results from this study extend previous findings that unsupervised online measures (5, 14, 15) were associated with participant cognition and diagnosis along the CN to MCI continuum. The ADNI-BHR cohort used in this analysis is particularly valuable because their online data is linked to confirmed clinical data, including neuropsychological tests, clinical diagnosis, and AD biomarkers. In the future, the ADNI-BHR cohort can be used to study relationships between additional online and in-clinic variables. Since the cohort is followed longitudinally both in-clinic and online, it provides a unique opportunity to investigate online variables associated with disease progression and cognitive decline.

This study has limitations. First, in the ADNI-BHR cohort there is a selection bias for participants with computer and internet access and literacy, and the cohort overrepresents those identified as White and those with high educational attainment. 110 of the 603 invited ADNI participants consented to enroll in online BHR. These factors will impact the generalizability of the results. Although these cohort characteristics will limit the external generalizability of these findings, the validation of online assessments will nonetheless greatly expand the accessibility to alternative screening measures of cognitive and functional status at a much lower cost and encourage broad diverse participation in AD clinical research studies, such as from those who do not have access to a memory clinic but have access to a computer. It will be crucial to validate online functional measures in a more diverse and highly characterized cohort in the future. Second, the test-retest reliability of in-clinic ECog relied on administrations given approximately six months apart. Since ECog assesses an individual’s capability to perform everyday tasks in comparison to activity levels ten years prior [3], it may be reasonable to estimate the reliability with a six month interval. Third, the average time interval between ECog administrations was different for the test-retest analysis (time interval of six months) and the online versus in-clinic analysis (time interval of 24.5 days apart). In addition, the sequence of ECog administrations was not taken into account. We assumed that practice effects, if any, were the same. However, it may be possible that a person completing two assessments 24.5 days apart could perform differently compared to a person completing two assessments six months apart. In the future, a better estimate of test-retest reliability could be achieved in which the sequence of online and in-clinic assessments is randomly assigned and the timing of assessments is kept consistent.

In conclusion, this study suggests that assessment of cognition and functional abilities collected in an unsupervised online setting provided as much information as repeating the assessment of cognition and functional abilities collected in a supervised clinical setting. We suggest that these results support the validity of leveraging a digital approach to overcome barriers in AD clinical research, by more effectively and broadly facilitating recruitment, screening, and monitoring activities in clinical trials aimed to develop new treatments and prevention strategies for AD and related dementias. The validation of remotely collected online cognitive functional assessments indicative of meaningful clinical measures could expand the methods used to identify, diagnose, and monitor individuals at risk for cognitive decline and have impactful and scalable applications in brain aging and AD clinical research.