|
Neurocognitive SuperAging (SA) in Older Adults Living With HIV: Demographic, Neuromedical and Everyday Functioning Correlates (CHARTER)
|
|
|
Download the PDF here
.....younger age, less depression, better reading skills + more cannabis use all were predictors of "super aging in HIV+" with undetectable viral load. In all others including those with detectable & undetectable viral load diabetes was a predictor of not being a "super ager".
March 2019 - advance publication - Journal of the International Neuropsychological Society - Rowan Saloner,1,2 Laura M. Campbell,1,2 Vanessa Serrano,2 Jessica L. Montoya,2 Elizabeth Pasipanodya,2 Emily W. Paolillo,1,2 Donald Franklin,2 Ronald J. Ellis,2 Scott L. Letendre,3 Ann C. Collier,4 David B. Clifford,5 Benjamin B. Gelman,6 Christina M. Marra,7 J. Allen McCutchan,3 Susan Morgello,8 Ned Sacktor,9 Dilip V. Jeste,2,10 Igor Grant,2 Robert K. Heaton,2 David J. Moore,2
AND the CHARTER and HNRP Groups
1San Diego State University/University of California, San Diego Joint Doctoral Program in Clinical Psychology, San Diego, California 2Department of Psychiatry, University of California, San Diego, San Diego, California 3Department of Medicine, University of California, San Diego, San Diego, California 4Department of Medicine, University of Washington, Seattle, Washington 5Department of Neurology, Washington University, St. Louis, Missouri 6Department of Pathology, University of Texas Medical Branch, Galveston, Texas 7Department of Neurology, University of Washington, Seattle, Washington 8Department of Neurology, Icahn School of Medicine of Mount Sinai, New York, New York 9Department of Neurology, Johns Hopkins University School of Medicine, Baltimore, Maryland 10Stein Institute for Research on Aging, University of California, San Diego, San Diego, California
734 PLWH and 123 HIV-uninfected participants between 50 and 64 years of age underwent neuropsychological and neuromedical evaluations. SA was defined as demographically corrected (i.e., sex, race/ethnicity, education) global neurocognitive performance within normal range for 25-year-olds. Remaining participants were labeled cognitively normal (CN) or impaired (CI) based on actual age. ......The observation that SA prevalence was twice as high in HIV-uninfected comparison participants as compared to PLWH provides important context to our findings. This difference, in addition to the higher prevalence of CN and lower prevalence of CI in HIV-uninfected controls, aligns with the known independent neurotoxic effects of HIV and potential synergistic effects of aging with HIV. Compared to their seronegative counterparts, older PLWH must withstand a greater amount of exposure to neural insults to sustain an elite level of neurocognitive performance. It is important to note that the HIV-uninfected group was demographically distinct from the PLWH group, as indicated by a higher prevalence of non-Hispanic whites, more years of education, and better WRAT Reading scores. Thus, the estimated two-fold difference in SA prevalence may be partially confounded by potential socio-demographic advantages of the HIV-uninfected group.
1. Average age was 55 for both HIV+ & HIV-neg.
2. 17% of HIV+ met criteria for "super agers" vs 35% for HIV-neg.
3. 45% of HIV+ were Cognitively Impaired (CI) and 38% Cognitively Normal.
4. Younger age, higher verbal IQ, absence of diabetes, fewer depressive symptoms, and lifetime cannabis use disorder increased likelihood of SA [super aging]. SA reported increased independence in everyday functioning, employment, and health-related quality of life than non-SA.
5. To focus on a clinically relevant subgroup, we reran the multinomial logistic regression among participants with undetectable levels of HIV plasma RNA. Of the 535 participants with an undetectable viral load, 97 (18%) were SA, 208 (39%) were CN, and 230 (43%) were CI. Age, WRAT [reading skills], BDI-II [depression], and diagnosis of lifetime cannabis use disorder remained significant predictors of neurocognitive status in this virally suppressed subgroup. Although diabetes increased likelihood of CN (odds ratio [OR]=1.74; p=.13) or CI (OR=1.63; p=.19) compared to SA, these associations were no longer statistically significant.
6. Furthermore, a lifetime diagnosis of cannabis use disorder decreased the likelihood of classification as CI compared to SA.
SA displayed greater rates of lifetime cannabis use disorder in comparison to CI, and this pattern also remained significant in the multinomial logistic regression. This result is supported by evidence suggesting neuroprotective effects of cannabis use through activation of cannabinoid receptors (i.e., CB1 and CB2) in the central nervous system (Sanchez & Garcia-Merino, 2012). Specifically, CB1 agonists reduce excitotoxity in post-synaptic neurons (Marsicano et al., 2003) while CB2 agonists promote anti-inflammatory and immunomodulatory actions (Rom & Persidsky, 2013).
Nevertheless, the relationship between cannabis use and brain integrity among PLWH and HIV-uninfected adults remains a controversial matter. While chronic cannabis use has been associated with neurometabolic abnormalities, reduced gray matter volumes, and memory deficits in cohorts comprised of PLWH and seronegative controls (Battistella et al., 2014; Chang, Cloak, Yakupov, & Ernst, 2006; Cristiani, Pukay-Martin, & Bornstein, 2004; Thames et al., 2017), emerging evidence suggests that active cannabis use may limit HIV viral replication and attenuate HIV-related immunosuppression, inflammation, and cerebral glutamate depletion (Chang et al., 2006; Rizzo et al., 2018; Thames, Mahmood, Burggren, Karimian, & Kuhn, 2016). These neuroprotective properties of the cannabinoid system are not referenced in the context of a cannabis use disorder, which may reflect problematic use or heavy exposure that could exceed therapeutic levels.
Moreover, prior studies examining elite neurocognition in healthy elders have excluded participants with substance use histories that could influence neurocognition. Thus, our cannabis-related findings cannot be compared to prior SA studies and the relationship between cannabis use disorder and neuroprotection in HIV remains poorly characterized. Future research is needed to explore therapeutic levels of cannabis use and identify potential benefits of cannabinoid receptor activation on neurocognition among PLWH.
7. SA had lower rates of unemployment and IADL dependence than the other neurocognitive status groups and higher self-reported physical and mental HRQoL.
Among HIV-uninfected individuals, diabetes is also strongly associated with neurocognitive impairment and is considered to be a predisposing factor for later development of vascular dementia and Alzheimer's disease (Cheng, Huang, Deng, & Wang, 2012; Taguchi, 2009). Insulin resistance and diabetes are associated with MRI structural abnormalities and functional alterations of the blood brain barrier, resulting in processes that facilitate the pathogenesis and progression of neurocognitive impairment (Archibald et al., 2014; Mogi & Horiuchi, 2011; Prasad, Sajja, Naik, & Cucullo, 2014). We found a stair-step effect for the influence of diabetes on neurocognitive status such that CI individuals were characterized by the highest rates of diabetes, followed by CN, and then SA participants; associations between diabetes and neurocognitive status remained in multivariable analyses.....Other studies have found similar increases in risk for HAND among HIV-infected persons with self-reported diabetes or elevated fasting insulin levels (McCutchan et al., 2012; Valcour et al., 2006, 2005; Vance et al., 2014). Thus, for SA participants, their relatively low incidence of diabetes likely contributed to better neurocognitive functioning. However, the effect of diabetes was not significant when restricting our multinomial regression analysis to virally suppressed participants, underscoring the importance of other contributing factors to SA status.
Consistent with prior research, the WRAT reading subtest - reading skill- , an estimate of premorbid verbal IQ that is relatively resistant to HIV-associated neurocognitive decline (Casaletto et al., 2014), was higher in SA and predicted SA status. Moore et al. (2014) demonstrated a positive correlation between a composite measure of cognitive reserve, including verbal IQ, and successful cognitive aging in older PLWH. The theory of cognitive reserve postulates that effects of neural insults, such as age and comorbidities, are buffered by robust brain networks (Stern, 2002). The Wide Range Achievement Test 4 (WRAT4) is an achievement test which measures an individual's ability to read words, comprehend sentences, spell, and compute solutions to math problems.
Although SA displayed higher premorbid functioning on the WRAT, neurocognitive status groups did not differ on years of education. Thus, neuroprotective benefits measured by higher WRAT performance may be better explained by factors other than education, such as genetically driven neurocognitive resilience. More granular methods of quantifying both the genetic (e.g., polygenic risk scores) and environmental (e.g., educational quality, socioeconomic factors) loadings of cognitive reserve are needed to thoroughly address questions regarding premorbid functioning and age-related neuroprotection.
We compared neurocognitive functioning of our sample to normative standards for age 25 when neurocognitive functioning is maximal (Salthouse, 2009). The concept of SA (Rogalski et al., 2013) posits that, within an individual's adult life, aging does not necessitate neurocognitive decline. Rather, aging increases the likelihood of encountering adverse events that contribute to neuronal damage and decline in neurocognition. Defining SA in this way may facilitate understanding of the kinds of events or experiences that either support, or damage, neurocognitive functioning.
ABSTRACT
Objectives: Studies of neurocognitively elite older adults, termed SuperAgers, have identified clinical predictors and neurobiological indicators of resilience against age-related neurocognitive decline. Despite rising rates of older persons living with HIV (PLWH), SuperAging (SA) in PLWH remains undefined. We aimed to establish neuropsychological criteria for SA in PLWH and examined clinically relevant correlates of SA.
Methods: 734 PLWH and 123 HIV-uninfected participants between 50 and 64 years of age underwent neuropsychological and neuromedical evaluations. SA was defined as demographically corrected (i.e., sex, race/ethnicity, education) global neurocognitive performance within normal range for 25-year-olds. Remaining participants were labeled cognitively normal (CN) or impaired (CI) based on actual age. Chi-square and analysis of variance tests examined HIV group differences on neurocognitive status and demographics. Within PLWH, neurocognitive status differences were tested on HIV disease characteristics, medical comorbidities, and everyday functioning. Multinomial logistic regression explored independent predictors of neurocognitive status.
Results: Neurocognitive status rates and demographic characteristics differed between PLWH (SA=17%; CN=38%; CI=45%) and HIV-uninfected participants (SA=35%; CN=55%; CI=11%). In PLWH, neurocognitive groups were comparable on demographic and HIV disease characteristics. Younger age, higher verbal IQ, absence of diabetes, fewer depressive symptoms, and lifetime cannabis use disorder increased likelihood of SA. SA reported increased independence in everyday functioning, employment, and health-related quality of life than non-SA.
Conclusions: Despite combined neurological risk of aging and HIV, youthful neurocognitive performance is possible for older PLWH. SA relates to improved real-world functioning and may be better explained by cognitive reserve and maintenance of cardiometabolic and mental health than HIV disease severity. Future research investigating biomarker and lifestyle (e.g., physical activity) correlates of SA may help identify modifiable neuroprotective factors against HIV-related neurobiological aging. (JINS, 2019, 00, 1–13)
DISCUSSION
The emerging concept of neurocognitive SA has produced invaluable insights into age-related neurocognitive phenotypes and has undermined the widely-held assumption that age-related neurocognitive deterioration is inevitable. However, the prospect of maintaining intact neurocognitive capacities throughout the lifespan is highly daunting for PLWH. In our study sample with 17% meeting criteria for SA, we demonstrate that youthful neurocognitive performance is possible for older PLWH. Our findings suggest that SA status is independently related to diverse factors that reflect current physical and mental health as well as premorbid neurocognitive functioning. Furthermore, SA status is associated with better every day functioning, supporting the ecological validity of distinguishing SA from CN and CI individuals.
Given the marked difference in average age between our cohort of older PLWH and previous SA cohorts of healthy elders, our SA criteria and study results cannot be directly linked to the extant SA literature. However, there are several strengths of our peak-age approach to defining neurocognitive SA in the context of HIV infection. First, we do not focus on one specific domain of neurocognitive functioning. Instead, our SA criteria are defined by absence of peak-age impairment in global neurocognitive functioning and absence of actual-age impairment in all domains assessed. PLWH are a heterogeneous group whose neurocognition may be impacted by HIV and demographic and clinical confounds, contributing to a neurocognitive profile that is not defined by deficits in any one neurocognitive domain. Thus, we demonstrate merit in defining SA by global performance to match what is known about neurocognitive functioning among PLWH.
An important feature of our global estimates of neurocognitive functioning is that they are adjusted for practice effects, as some study participants had prior exposure to the neurocognitive testing battery. Practice, or learning, effects complicate assessment of SA because seemingly elite neurocognition can be an artifact of prior testing experience. By correcting for normal test–retest fluctuations, we reduce the likelihood of overestimating neurocognitive ability and enhance the stringency of our SA criteria.
We compared neurocognitive functioning of our sample to normative standards for age 25 when neurocognitive functioning is maximal (Salthouse, 2009). The concept of SA (Rogalski et al., 2013) posits that, within an individual's adult life, aging does not necessitate neurocognitive decline. Rather, aging increases the likelihood of encountering adverse events that contribute to neuronal damage and decline in neurocognition. Defining SA in this way may facilitate understanding of the kinds of events or experiences that either support, or damage, neurocognitive functioning.
SA had lower rates of unemployment and IADL dependence than the other neurocognitive status groups and higher self-reported physical and mental HRQoL. Thus, our method for defining SA appears to be concurrently valid with measures of everyday functioning and HRQoL. Importantly, CN and SA groups differ in real-world outcomes, indicating heterogeneity among neurocognitively unimpaired individuals. Unlike prior investigations of SA, our definition of SA did not require self-reported IADL independence as a criterion. Despite performance-based data indicating SA, a small proportion of the SA group endorsed IADL dependence. Among our SA group, self-reported declines in IADL may represent actual decline, such that SA individuals may have started at higher levels of functioning and experienced a decline that is not necessarily at an impaired level.
To this point, our measure of IADL dependence may be overly sensitive in detecting decline and not specific in detecting whether this decline represents a shift from within normal functioning to impairment status. Given that most other studies rely on absence of IADL dependence or decline when defining SA, these studies may be potentially misidentifying SA individuals who perform at peak-age levels on neurocognitive tests. Thus, future investigations need to consider the appropriate use of performance-based versus self-reported deficits when classifying individuals as SA versus CN.
Consistent with prior research, the WRAT reading subtest, an estimate of premorbid verbal IQ that is relatively resistant to HIV-associated neurocognitive decline (Casaletto et al., 2014), was higher in SA and predicted SA status. Moore et al. (2014) demonstrated a positive correlation between a composite measure of cognitive reserve, including verbal IQ, and successful cognitive aging in older PLWH. The theory of cognitive reserve postulates that effects of neural insults, such as age and comorbidities, are buffered by robust brain networks (Stern, 2002). Although operational definitions and methods of quantifying cognitive reserve may vary across studies (Moore et al., 2014; Nucci, Mapelli, & Mondini, 2012; Reed et al., 2010; Selzam et al., 2017), cognitive reserve is considered to reflect a combination of genetically-driven intellectual capacity and cognitively stimulating life experiences that promote resilience against age-related neurocognitive decline (Daffner, 2010; Stern, 2012).
Although SA displayed higher premorbid functioning on the WRAT, neurocognitive status groups did not differ on years of education. Thus, neuroprotective benefits measured by higher WRAT performance may be better explained by factors other than education, such as genetically driven neurocognitive resilience. More granular methods of quantifying both the genetic (e.g., polygenic risk scores) and environmental (e.g., educational quality, socioeconomic factors) loadings of cognitive reserve are needed to thoroughly address questions regarding premorbid functioning and age-related neuroprotection.
Among HIV-uninfected individuals, diabetes is also strongly associated with neurocognitive impairment and is considered to be a predisposing factor for later development of vascular dementia and Alzheimer's disease (Cheng, Huang, Deng, & Wang, 2012; Taguchi, 2009). Insulin resistance and diabetes are associated with MRI structural abnormalities and functional alterations of the blood brain barrier, resulting in processes that facilitate the pathogenesis and progression of neurocognitive impairment (Archibald et al., 2014; Mogi & Horiuchi, 2011; Prasad, Sajja, Naik, & Cucullo, 2014). We found a stair-step effect for the influence of diabetes on neurocognitive status such that CI individuals were characterized by the highest rates of diabetes, followed by CN, and then SA participants; associations between diabetes and neurocognitive status remained in multivariable analyses.
Other studies have found similar increases in risk for HAND among HIV-infected persons with self-reported diabetes or elevated fasting insulin levels (McCutchan et al., 2012; Valcour et al., 2006, 2005; Vance et al., 2014). Thus, for SA participants, their relatively low incidence of diabetes likely contributed to better neurocognitive functioning. However, the effect of diabetes was not significant when restricting our multinomial regression analysis to virally suppressed participants, underscoring the importance of other contributing factors to SA status.
SA had lower BDI-II scores than both CN and CI univariately and in the multinomial logistic regression. In contrast, rates of current and lifetime MDD diagnoses did not significantly differ by neurocognitive status group, indicating that among older PLWH, current subclinical depressive symptoms are associated with neurocognitive functioning more closely than active or remote clinical depression. This relationship may reflect known neurological consequences of depression, including neuroinflammation and associated neuronal damage, apoptosis, and reduced neurogenesis (Kubera, Obuchowicz, Goehler, Brzeszcz, & Maes, 2011; Maes et al., 2009). Behavioral mechanisms may also underlie the relationship between depression and neurocognition, as depressive symptoms (even those that are subclinical) negatively impact engagement in activities known to promote neurocognitive health, including exercise, healthy nutrition, and social activity (Jeste, Depp, & Vahia, 2010; Moore et al., 2018; Vahia et al., 2010).
SA displayed greater rates of lifetime cannabis use disorder in comparison to CI, and this pattern also remained significant in the multinomial logistic regression. This result is supported by evidence suggesting neuroprotective effects of cannabis use through activation of cannabinoid receptors (i.e., CB1 and CB2) in the central nervous system (Sanchez & Garcia-Merino, 2012). Specifically, CB1 agonists reduce excitotoxity in post-synaptic neurons (Marsicano et al., 2003) while CB2 agonists promote anti-inflammatory and immunomodulatory actions (Rom & Persidsky, 2013).
Nevertheless, the relationship between cannabis use and brain integrity among PLWH and HIV-uninfected adults remains a controversial matter. While chronic cannabis use has been associated with neurometabolic abnormalities, reduced gray matter volumes, and memory deficits in cohorts comprised of PLWH and seronegative controls (Battistella et al., 2014; Chang, Cloak, Yakupov, & Ernst, 2006; Cristiani, Pukay-Martin, & Bornstein, 2004; Thames et al., 2017), emerging evidence suggests that active cannabis use may limit HIV viral replication and attenuate HIV-related immunosuppression, inflammation, and cerebral glutamate depletion (Chang et al., 2006; Rizzo et al., 2018; Thames, Mahmood, Burggren, Karimian, & Kuhn, 2016). These neuroprotective properties of the cannabinoid system are not referenced in the context of a cannabis use disorder, which may reflect problematic use or heavy exposure that could exceed therapeutic levels.
Moreover, prior studies examining elite neurocognition in healthy elders have excluded participants with substance use histories that could influence neurocognition. Thus, our cannabis-related findings cannot be compared to prior SA studies and the relationship between cannabis use disorder and neuroprotection in HIV remains poorly characterized. Future research is needed to explore therapeutic levels of cannabis use and identify potential benefits of cannabinoid receptor activation on neurocognition among PLWH.
Despite stair-step patterns for HIV disease characteristics in SA individuals compared to CN and CI participants, only the proportion of participants with current CD4 counts below 200 was statistically significantly different among the neurocognitive status groups. Specifically, the SA group had a lower proportion with current CD4 counts below 200 than the CI group. However, this difference was not statistically significant when controlling for other clinical and demographic variables (e.g., age, WRAT, and depressive symptoms). Unexpectedly, the SA and CN groups had low nadir CD4 counts comparable to the CI group, possibly reflecting underlying resilience to the "legacy" effects of advanced immunosuppression.
In a comparison of predictors of HAND before and during the era of ART, only low nadir CD4 was found to increase risk of neurocognitive impairment in both treatment eras (Heaton et al., 2011). However, when examining factors associated with decline to symptomatic HAND, current CD4 also predicted decline to symptomatic status (Grant et al., 2014). SA with current CD4 counts below 200 were more likely to be off ART. Furthermore, the higher proportion of participants with CD4 counts below 200 in the CI group may result from poorer ART adherence that is a consequence of their cognitive impairment. Given that the majority of participants were likely to begin ART after having advanced HIV, it is unclear whether similar relationships between HIV disease severity and neurocognitive status exist for modern era patients who typically start treatment at earlier stages.
The observation that SA prevalence was twice as high in HIV-uninfected comparison participants as compared to PLWH provides important context to our findings. This difference, in addition to the higher prevalence of CN and lower prevalence of CI in HIV-uninfected controls, aligns with the known independent neurotoxic effects of HIV and potential synergistic effects of aging with HIV. Compared to their seronegative counterparts, older PLWH must withstand a greater amount of exposure to neural insults to sustain an elite level of neurocognitive performance. It is important to note that the HIV-uninfected group was demographically distinct from the PLWH group, as indicated by a higher prevalence of non-Hispanic whites, more years of education, and better WRAT Reading scores. Thus, the estimated two-fold difference in SA prevalence may be partially confounded by potential socio-demographic advantages of the HIV-uninfected group.
Several limitations to the present study warrant discussion. Our peak-age corrected neurocognitive scores, based on a normative sample of 25-year-olds, serve as proxy measures for neurocognitive resilience and do not directly capture the true within-subject change in neurocognitive performance since age 25. Because our data are cross-sectional, we cannot rule out the possibility that members of the SA group have experienced considerable lifetime neurocognitive decline and that their SA status is an artifact of superior baseline neurocognitive capacity. Although our analysis demonstrating that older age was associated with lower global scaled scores in CN and CI groups, but not the SA group, preliminarily supports the validity of our SA criteria, the magnitude of these age effects were small and did not significantly differ across groups. In addition to other factors importantly contributing to variance in global neurocognitive performance, these small effect sizes are likely influenced by the narrow age range of our sample.
Our results highlight clinically informative predictors and benefits of neurocognitive resilience; yet, the racial/ethnic composition of our sample was predominantly non-Hispanic white men and may limit the generalizability of our findings to more socio-demographically diverse populations. Furthermore, our cohort of older PLWH is relatively young compared to the healthy adult cohorts studied in the extant SA literature of persons not living with HIV, but the age range is indicative of some of the oldest PLWH with a sufficient sample size to be studied. Although the inclusion of an age-matched HIV-uninfected comparison group provided an informative anchor point for SA prevalence in healthy adults, this comparison group was not comparable to the PLWH group on other important demographic factors. Consequently, important questions remain regarding the extent to which our definition of SA in PLWH reflects resilience to the effects of HIV and aging into late adulthood, which may only be adequately addressed with data from ideal comparison groups. As the proportion of PLWH older than 65 years of age increases, longitudinal cohort studies of PLWH will be better equipped to address critical questions related to the prevalence, stability, and impact of SA in PLWH compared to healthy seniors.
Although we focused on evaluating the relationships between SA status and clinical correlates commonly assessed in PLWH, the absence of biomarker data indicative of central nervous system integrity (e.g., neuroimaging, cerebrospinal fluid assays) prevents us from determining the neurobiological correlates of SA status. Additionally, an assessment of modifiable behaviors (e.g., physical activity, neurocognitive activity, positive psychological outlook) that may mediate the relationships between SA status and psychosocial, medical, and everyday functioning correlates could help to prioritize research in clinical interventions to increase the fraction of SA in PLWH (Vance & Burrage, 2006).
Taken together, our results demonstrate that a substantial fraction of older, HIV-infected patients maintain their maximal neurocognitive abilities that confer real-world benefits even compared to patients with normal age-related cognitive decline. Although HIV disease negatively impacts the prevalence of SA, our findings highlight the clinical value in identifying neurocognitive resilience within PLWH and focus on the potential for positive outcomes despite aging with HIV. Examination of the stability of SA status through longitudinal analysis, exploration of biological and genetic markers of neuronal integrity, and assessment of modifiable lifestyle factors should enhance studies of future interventions to improve neurocognitive aging in older PLWH.
-----------------
Medical Comorbidities
Examination of medical comorbidities revealed significant group differences for rates of hepatitis C virus (HCV) seropositivity and diabetes. Post hoc comparisons indicated that SA (super aging) had significantly lower rates of HCV than the CN (cognitive impaired) group and lower rates of diabetes than both the CN and CI groups. No significant group differences were found for other markers of metabolic syndrome (i.e., hypertension, hyperlipidemia, body mass index). see table 2 below.
Psychiatric and Substance Use Characteristics
Significant group differences were observed for rates of lifetime cannabis use disorder and cocaine use disorder. SA had significantly higher rates of cannabis use disorder than CI individuals and CN individuals displayed higher rates of cocaine use disorder than the CI group (Table 3). Although lifetime and current diagnoses of major depressive disorder (MDD) did not differ across groups, SA endorsed significantly fewer depressive symptoms on the BDI-II than both the CN (d=-0.35) and CI (d=-0.46) groups. see Table 3 below.
Multinomial Regression Predicting Neurocognitive Status
A multinomial logistic regression was performed with the three neurocognitive groups in PLWH as the dependent variable. Predictors were all outcome variables from Tables 2 and 3 with a trend-level omnibus effect (excluding race/ethnicity, i.e., age, WRAT, current CD4<200, HCV, diabetes, cannabis use disorder, and BDI-II). Based on available data, the sample size for this model included 113 SA, 259 CN, and 287 CI participants. Overall, the model was significant ( (14,659)=83.73; p<.001; Nagelkerke pseudo- = 0.137). Likelihood ratio tests indicated that older age, lower WRAT scores, diagnosis of diabetes, and higher BDI-II scores all increased the likelihood of classification as either CN or CI compared to SA (Table 4). Furthermore, a lifetime diagnosis of cannabis use disorder decreased the likelihood of classification as CI compared to SA.
To focus on a clinically relevant subgroup, we reran the multinomial logistic regression among participants with undetectable levels of HIV plasma RNA. Of the 535 participants with an undetectable viral load, 97 (18%) were SA, 208 (39%) were CN, and 230 (43%) were CI. Age, WRAT, BDI-II, and diagnosis of lifetime cannabis use disorder remained significant predictors of neurocognitive status in this virally suppressed subgroup. Although diabetes increased likelihood of CN (odds ratio [OR]=1.74; p=.13) or CI (OR=1.63; p=.19) compared to SA, these associations were no longer statistically significant.
Everyday Functioning and HRQoL Correlates of Neurocognitive Status
A stair-step pattern was observed for most outcomes from the PAOFI, IADL, and MOS-SF-36 measures, with SA individuals endorsing the most favorable everyday functioning and HRQoL outcomes followed by CN then CI participants. SA individuals endorsed significantly fewer cognitive symptoms on the PAOFI than CN (d=-0.34; p<.001) and CI participants (d=-0.64; p<.0001) and fewer declines in IADLs than either CN (d=-0.42; p<.01) or CI participants (d=-0.70; p<.0001). The CN group also reported significantly fewer cognitive symptoms (d=-0.30; p<.05) and IADL declines (d=-0.33; p<.001) than the CI group. Figure 3 displays similar group differences on rates of unemployment and IADL dependence as well as the MOS-SF-36 physical and mental HRQoL composite scores.
Fig. 3 Everyday functioning and HRQoL by neurocognitive status. Risk ratio (RR) estimates represent the reduction in risk of IADL dependence or unemployment for each pair-wise comparison. Cohen's d effect size estimates reflect differences in HRQoL for each pair-wise comparison. All p-values are significant after Bonferroni-adjustment or Tukey's HSD. ***p<.001; **p<.01; *p<.05.
-----------------------------------
Neurocognitive SuperAging in Older Adults Living With HIV: Demographic, Neuromedical and Everyday Functioning Correlates
Abstract
Objectives: Studies of neurocognitively elite older adults, termed SuperAgers, have identified clinical predictors and neurobiological indicators of resilience against age-related neurocognitive decline. Despite rising rates of older persons living with HIV (PLWH), SuperAging (SA) in PLWH remains undefined. We aimed to establish neuropsychological criteria for SA in PLWH and examined clinically relevant correlates of SA.
Methods: 734 PLWH and 123 HIV-uninfected participants between 50 and 64 years of age underwent neuropsychological and neuromedical evaluations. SA was defined as demographically corrected (i.e., sex, race/ethnicity, education) global neurocognitive performance within normal range for 25-year-olds. Remaining participants were labeled cognitively normal (CN) or impaired (CI) based on actual age. Chi-square and analysis of variance tests examined HIV group differences on neurocognitive status and demographics. Within PLWH, neurocognitive status differences were tested on HIV disease characteristics, medical comorbidities, and everyday functioning. Multinomial logistic regression explored independent predictors of neurocognitive status.
Results: Neurocognitive status rates and demographic characteristics differed between PLWH (SA=17%; CN=38%; CI=45%) and HIV-uninfected participants (SA=35%; CN=55%; CI=11%). In PLWH, neurocognitive groups were comparable on demographic and HIV disease characteristics. Younger age, higher verbal IQ, absence of diabetes, fewer depressive symptoms, and lifetime cannabis use disorder increased likelihood of SA. SA reported increased independence in everyday functioning, employment, and health-related quality of life than non-SA.
Conclusions: Despite combined neurological risk of aging and HIV, youthful neurocognitive performance is possible for older PLWH. SA relates to improved real-world functioning and may be better explained by cognitive reserve and maintenance of cardiometabolic and mental health than HIV disease severity. Future research investigating biomarker and lifestyle (e.g., physical activity) correlates of SA may help identify modifiable neuroprotective factors against HIV-related neurobiological aging. (JINS, 2019, 00, 1–13)
INTRODUCTION
Antiretroviral therapy (ART) has facilitated increased life expectancy for people living with HIV (PLWH; Wing, 2016). In 2014, 45% of PLWH in the United States were over the age of 50 (Centers for Disease Control and Prevention, 2018) and this proportion is expected to increase (Smit et al., 2015). HIV-associated neurocognitive disorder (HAND) affects approximately half of PLWH (Heaton et al., 2010; Norman et al., 2011; Saloner & Cysique, 2017), and older PLWH are at three times higher risk for HAND compared to younger PLWH (Valcour et al., 2004). Furthermore, there is evidence to suggest that HIV accelerates and accentuates neurocognitive aging (Pathai, Bajillan, Landay, & High, 2014; Sheppard et al., 2017). Older PLWH are at increased risk for functional decline (Thames et al., 2011; Vance, Fazeli, & Gakumo, 2013), which is not only costly, but also negatively affects quality of life (Morgan et al., 2012). Identifying factors that promote successful cognitive aging with HIV and developing interventions to sustain or enhance them may avoid or reverse the adverse effects of aging.
While definitions of successful cognitive aging in PLWH differ slightly, all definitions require individuals to be neurocognitively unimpaired and functionally independent (Malaspina et al., 2011; Moore et al., 2017). Successful cognitive aging rates in older PLWH range from 19–32%, and translates into real-world benefits, including greater success in managing medication and medical appointments, less decline in activities of daily living, and better psychological health and health-related quality of life (HRQoL) (Malaspina et al., 2011; Moore et al., 2017, 2014). Given that the neuropsychological criteria for successful cognitive aging solely requires the absence of neurocognitive impairment, taking into consideration age, there likely remains considerable heterogeneity in neurocognitive performance (e.g., low average to superior) among the successful cognitive aging group. Thus, distinguishing older PLWH with superior neurocognitive abilities from those with average neurocognitive abilities may explain additional variance in everyday functioning outcomes.
Older adults with preserved cognition appear to resist "normal" age-related decline. The term SuperAger refers to older adults that perform equivalently to young or middle-aged adults on episodic memory tests (Harrison, Maass, Baker, & Jagust, 2018; Rogalski et al., 2013; Sun et al., 2016). Alternatively, others have researched "SuperNormals" or "Optimal Memory Performers" – older adults who demonstrate above-average episodic memory performance in comparison to average older adults (Dekhtyar et al., 2017; Lin et al., 2017; Mapstone et al., 2017; Wang et al., 2019). Both definitions provide evidence that older adults with superior memory perform better on other cognitive domains, particularly executive functioning (Dekhtyar et al., 2017; Gefen et al., 2015) and processing speed (Dekhtyar et al., 2017; Harrison et al., 2018).
Additionally, SuperAgers have larger volumes of the cerebral cortex, hippocampus, and cingulate cortex (Dekhtyar et al., 2017; Harrison et al., 2018; Lin et al., 2017; Rogalski et al., 2013; Sun et al., 2016; Wang et al., 2019) as well as slower rates of cortical atrophy (Cook et al., 2017). Furthermore, SuperAgers display lower levels of biomarkers of neurodegeneration such as oxidative stress (Mapstone et al., 2017), inflammation (Bott et al., 2017), and amyloid (Lin et al., 2017; Rogalski et al., 2013) and tau deposition (Gefen et al., 2015).
Despite not having a gold-standard definition of SuperAging (SA) or preserved cognition, commonalities exist among the definitions. Most studies have classified SuperAgers based on superior memory performance alone and only required either average age-adjusted performance for a few other neuropsychological measures (Harrison et al., 2018; Rogalski et al., 2013). Some have required that they be otherwise neurocognitively normal (Dekhtyar et al., 2017; Lin et al., 2017). Thus, SA studies predominantly focus on superior memory performance rather than superior global neurocognitive performance.
The majority of these studies, which consist of primarily septua- and octogenarians, require SuperAgers to perform equivalent to or better than those in their mid-50s; however, most neurocognitive abilities peak in the mid-20s and then begin to decline (Hartshorne & Germine, 2015; Heaton, Taylor, & Manly, 2003; Salthouse, 2003, 2009). Although SA is typically evaluated in healthy adults who are at least 60 years old, the aging population of PLWH is younger with 50 years old serving as a cutoff for defining a medically advanced age (Blanco et al., 2012). Nevertheless, neurocognitive aging studies have demonstrated substantial inter-individual variability in neurocognition for healthy adult cohorts below the age of 60 (Lachman, Teshale, & Agrigoroaei, 2015; Martin & Zimprich, 2005; Schaie & Willis, 2010). Importantly, this heterogeneity in neurocognition tracks with variation in biopsychosocial factors such that high neurocognitive performance correlates with high cognitive reserve and low comorbidity burden (Anstey, Sargent-Cox, Garde, Cherbuin, & Butterworth, 2014; Ferreira et al., 2017).
While current definitions of SA may be appropriate for studying healthy older adults resistant to the clinical expressions of biological aging and Alzheimer's disease, SA criteria should be tailored for study in older PLWH who are younger and at greater risk for multi-domain neurocognitive decline rather than focal memory deficits. Thus, we aimed to: (1) establish neuropsychological criteria for neurocognitive SA in PLWH; (2) identify clinical predictors of SA in PLWH; (3) assess the everyday functioning correlates of SA status.
METHODS
Participants
Participants included 734 PLWH and 123 HIV-uninfected controls aged 50–64 years. A total of 340 PLWH were enrolled in the NIH-funded CNS HIV Anti-Retroviral Therapy Effects Research (CHARTER) study, consisting of six participating university centers: Johns Hopkins University (Baltimore, MD; n=51); Mt. Sinai School of Medicine (New York, NY; n=92); University of California at San Diego (San Diego, CA; n=32); University of Texas Medical Branch (Galveston, TX; n=73); University of Washington (Seattle, WA; n=38); and Washington University (St. Louis, MO; n=54). The remaining 394 PLWH and 123 HIV-uninfected participants were enrolled in other NIH-funded research studies at the University of California, San Diego's HIV Neurobehavioral Research Program (HNRP). All participant visits for the present study took place between 2002 and 2017. All studies were approved by local Human Subjects Protection Committees, and all participants provided written informed consent. All PLWH were required to have ≥5 years of estimated duration of HIV disease to be considered for inclusion.
Exclusion criteria were: (1) diagnosis of psychotic or mood disorder with psychotic features, neurological, or medical condition that may impair neurocognitive functioning, such as traumatic brain injury, stroke, epilepsy, or advanced liver disease; (2) low verbal IQ of <70 as estimated by the reading subtest of the Wide Range Achievement Test (WRAT; Wilkinson & Robertson, 2006); or (3) evidence of intoxication by illicit drugs (except marijuana) or Breathalyzer test for alcohol on the day of testing by positive urine toxicology.
Procedures
Neurocognitive assessment
Participants were classified as SA based on their performance on a comprehensive and standardized battery of neurocognitive tests, which has been described in detail elsewhere (Carey et al., 2004; Heaton et al., 2010) (Table 1). Briefly, the battery covers seven neurocognitive domains commonly impacted in HIV-infected persons: verbal fluency, executive functioning, processing speed, learning, delayed recall, attention/working memory, and motor skills (Heaton et al., 2010). Since some participants had been exposed to the test battery at prior research visits, raw scores for each test were converted to practice effect-adjusted scaled scores (M=10; SD=3; Heaton et al., 2001). These demographically uncorrected scaled scores were converted to T scores (M=50; SD=10) that corrected for the effects of age, education, sex, and race/ethnicity on neurocognition (Heaton, Miller, Taylor, & Grant, 2004; Heaton et al., 2003; Norman et al., 2011).
To generate variables that reflect maximum neurocognitive performance at a younger age, a second set of adjusted T scores were computed in which the age of 25, instead of actual age, was entered into the demographic correction formulas along with actual education, sex, and race/ethnicity. These scores, referred to as "peak-age" T scores, consequently compare an individual's neurocognitive performance to normative standards for 25-year-olds of the same education, sex, and race/ethnicity (Heaton, Miller, et al., 2004; Heaton et al., 2003; Norman et al., 2011). Both the actual-age and peak-age T scores for each measure were averaged to compute global and domain-specific T scores within each cognitive ability area. T scores were converted to actual-age and peak-age domain-specific deficit scores (DDS) that give differential weight to impaired, as opposed to normal scores, on a scale ranging from 0 (T≥40; normal) to 5 (T<20; severe impairment). DDS were then averaged to generate an actual-age and peak-age global deficit score (GDS). Consistent with prior studies, the presence of global impairment was defined by GDS≥0.5 and domain-specific impairment by DDS>0.5 (Blackstone et al., 2012; Carey et al., 2004).
SuperAging criteria
To estimate intact and peak neurocognitive functioning, SA status was operationally defined as: (1) peak-age GDS<0.5; and (2) actual-age DDS ≤ 0.5 for all seven neurocognitive domains. Participants that did not meet SA criteria were classified as either cognitively normal (CN) or cognitively impaired (CI) using the standard actual-age GDS impairment cut-point of≥0.5 (Figure 1).
Fig. 1 Neurocognitive status criteria. SuperAging was operationalized as a peak-age global deficit score within normal limits (i.e., less than 0.5) and normal performance on all seven actual-age deficit scores (i.e., less or equal than 0.5).
Neuromedical and laboratory assessment
All participants underwent a comprehensive neuromedical assessment, including a medical history that included medications, Centers for Disease Control staging, and blood draw. HIV infection was diagnosed by enzyme-linked immunosorbent assay with Western blot confirmation. Routine clinical chemistry panels, complete blood counts, rapid plasma reagin, hepatitis C virus antibody, and CD4+ T cells (flow cytometry) were performed at each site's Clinical Laboratory Improvement Amendments (CLIA)–certified, or CLIA equivalent, medical center laboratory. Levels of HIV viral load in plasma were measured using reverse transcriptase-polymerase chain reaction (Amplicor, Roche Diagnostics, Indianapolis, IN, with a lower limit of quantitation 50 copies/mL).
Psychiatric assessment
678 PLWH had available data from the Composite International Diagnostic Interview (CIDI), a fully structured, computer-based interview, to determine DSM-IV diagnoses for current and lifetime mood and substance use disorders. (World Health Organization, 1998). Additionally, a subset of PLWH (n=712) completed the Beck Depression Inventory-II (BDI-II; Beck, Steer, & Brown, 1996) to assess current symptoms of depressed mood.
Everyday functioning and quality of life assessment
Instrumental activities of daily living (IADL) dependence was assessed using a revised version of the Lawton and Brody (1969) self-report measure of everyday functioning (Heaton, Marcotte, et al., 2004; Woods et al., 2008), in which participants rated current abilities compared to previous abilities across 13 everyday functioning domains. Two outcome variables were generated: (1) A continuous variable of the number of declines in IADL; and (2) a dichotomous variable for IADL dependence, defined as ≥2 declines at least partially attributable to cognitive problems.
The Patient's Assessment of Own Functioning Inventory (PAOFI) is a 33-item self-report measure used to measure perceived cognitive symptoms in everyday life (Chelune, Heaton, & Lehman, 1986). Items endorsed as fairly often or greater are considered clinically significant cognitive symptoms. A continuous variable for total number of clinically significant everyday cognitive symptoms and a dichotomous variable for employment status (i.e., employed/unemployed) were examined as outcome variables.
A subset of PLWH (n=490) completed the Medical Outcome Study 36 Item Short-Form version 1.0 (MOS-SF-36), which assesses HRQoL. The reliability and validity of the MOS-SF-36 has been extensively documented in PLWH (Henderson et al., 2010; Wu, Revicki, Jacobson, & Malitz, 1997). For this study, the physical and mental health composite scores were examined as primary outcome variables.
RESULTS
SuperAging Prevalence
Of the 734 PLWH, 124 (17%) met criteria for SA. Of the remaining 610 non-SA participants, 279 (38%) were CN and 331 (45%) were CI. Figure 2 displays differences in actual-age T and peak-age T scores within and across SA and CN PLWH with Cohen's d effect size estimates for actual-age T scores. The prevalence of SA and CN were significantly higher, and prevalence of CI was significantly lower, in the HIV-uninfected group ( = 63.7; p<.0001). Of the 123 HIV-uninfected participants, 43 (35%) were SA, 67 (55%) were CN, and 13 (11%) were CI.
Fig. 2. SuperAger (SA) versus cognitively normal (CN) differences in neurocognitive performance. Cohen's d effect size estimates reflect
differences in actual-age T scores.
Demographics
Table 2 displays PLWH neurocognitive status group differences in demographic, clinical, and neuromedical variables. Only percent non-Hispanic white differed significantly among demographic factors. Although the CN group exhibited the lowest proportion of non-Hispanic white, no significant pairwise differences were found. SA individuals were on average a year younger than their CN and CI counterparts and this difference approached significance, but this did not result in significant pairwise differences. Although groups did not differ with respect to education, SA displayed significantly higher WRAT scores than CN (d=0.43) and CI (d=0.61) participants.
Compared to PLWH, the HIV-uninfected comparison group had significantly higher rates of non-Hispanic white participants (81% vs. 58%; p<.0001), females (38% vs. 16%; p<.0001), mean years of education (14.4 vs. 13.6; p<.001), and higher mean WRAT scores (106 vs. 98; p<.0001). By design, the HIV-uninfected group did not significantly differ from PLWH in mean age (55.5 vs. 55.1; p=.87).
HIV Disease Characteristics
A stair-step pattern of indicators of HIV disease severity was commonly observed such that SA displayed the lowest amount of HIV disease burden followed by CN then CI individuals. Although this stair-step pattern occurred for history of AIDS diagnosis, detectable plasma HIV, current CD4 count, and nadir CD4<200; only omnibus group differences in current CD4<200 were significant. Post hoc comparisons indicated that the SA group had significantly lower rates of participants with current CD4<200 than the CI group. In the full sample, participants with current CD4<200 were more likely to be off ART (19.6%) compared to those with current CD4≥200 (9.8%; = 6.7; p=.01). No noteworthy group differences were found for estimated duration of HIV disease or receipt of ART.
Medical Comorbidities
Examination of medical comorbidities revealed significant group differences for rates of hepatitis C virus (HCV) seropositivity and diabetes. Post hoc comparisons indicated that SA had significantly lower rates of HCV than the CN group and lower rates of diabetes than both the CN and CI groups. No significant group differences were found for other markers of metabolic syndrome (i.e., hypertension, hyperlipidemia, body mass index).
Psychiatric and Substance Use Characteristics
Significant group differences were observed for rates of lifetime cannabis use disorder and cocaine use disorder. SA had significantly higher rates of cannabis use disorder than CI individuals and CN individuals displayed higher rates of cocaine use disorder than the CI group (Table 3). Although lifetime and current diagnoses of major depressive disorder (MDD) did not differ across groups, SA endorsed significantly fewer depressive symptoms on the BDI-II than both the CN (d=-0.35) and CI (d=-0.46) groups.
Psychiatric and Substance Use Characteristics
Significant group differences were observed for rates of lifetime cannabis use disorder and cocaine use disorder. SA had significantly higher rates of cannabis use disorder than CI individuals and CN individuals displayed higher rates of cocaine use disorder than the CI group (Table 3). Although lifetime and current diagnoses of major depressive disorder (MDD) did not differ across groups, SA endorsed significantly fewer depressive symptoms on the BDI-II than both the CN (d=-0.35) and CI (d=-0.46) groups.
Multinomial Regression Predicting Neurocognitive Status
A multinomial logistic regression was performed with the three neurocognitive groups in PLWH as the dependent variable. Predictors were all outcome variables from Tables 2 and 3 with a trend-level omnibus effect (excluding race/ethnicity, i.e., age, WRAT, current CD4<200, HCV, diabetes, cannabis use disorder, and BDI-II). Based on available data, the sample size for this model included 113 SA, 259 CN, and 287 CI participants. Overall, the model was significant ( (14,659)=83.73; p<.001; Nagelkerke pseudo- = 0.137). Likelihood ratio tests indicated that older age, lower WRAT scores, diagnosis of diabetes, and higher BDI-II scores all increased the likelihood of classification as either CN or CI compared to SA (Table 4). Furthermore, a lifetime diagnosis of cannabis use disorder decreased the likelihood of classification as CI compared to SA.
To focus on a clinically relevant subgroup, we reran the multinomial logistic regression among participants with undetectable levels of HIV plasma RNA. Of the 535 participants with an undetectable viral load, 97 (18%) were SA, 208 (39%) were CN, and 230 (43%) were CI. Age, WRAT, BDI-II, and diagnosis of lifetime cannabis use disorder remained significant predictors of neurocognitive status in this virally suppressed subgroup. Although diabetes increased likelihood of CN (odds ratio [OR]=1.74; p=.13) or CI (OR=1.63; p=.19) compared to SA, these associations were no longer statistically significant.
Age and Global Performance Relationship by Neurocognitive Status
To examine the relationship between age and global neurocognitive performance within each neurocognitive status group in PLWH, we performed Pearson's partial correlations between age and demographically-uncorrected global scaled scores, co-varying for education, sex, and race/ethnicity. Age negatively correlated with lower global scaled scores within the CN (partial r=-.24; p<.001) and CI (partial r=-.15; p<.001) groups. However, age did not significantly relate to global scaled scores among the SA group (partial r=-.11; p=.24). Despite this lack of significance, comparison of Fisher's r-to-z transformed correlations indicated that the effect size of age on global scaled scores in SA did not significantly differ from the effect sizes of age on global scaled scores in CN (z=1.23; p=.22) and CI (z=.38; p=.70). Similarly, the magnitude of the relationship between age and global scaled scores did not differ between CN and CI (z=-1.15; p=.25).
Everyday Functioning and HRQoL Correlates of Neurocognitive Status
A stair-step pattern was observed for most outcomes from the PAOFI, IADL, and MOS-SF-36 measures, with SA individuals endorsing the most favorable everyday functioning and HRQoL outcomes followed by CN then CI participants. SA individuals endorsed significantly fewer cognitive symptoms on the PAOFI than CN (d=-0.34; p<.001) and CI participants (d=-0.64; p<.0001) and fewer declines in IADLs than either CN (d=-0.42; p<.01) or CI participants (d=-0.70; p<.0001). The CN group also reported significantly fewer cognitive symptoms (d=-0.30; p<.05) and IADL declines (d=-0.33; p<.001) than the CI group. Figure 3 displays similar group differences on rates of unemployment and IADL dependence as well as the MOS-SF-36 physical and mental HRQoL composite scores.
Fig. 3 Everyday functioning and HRQoL by neurocognitive status. Risk ratio (RR) estimates represent the reduction in risk of IADL dependence or unemployment for each pair-wise comparison. Cohen's d effect size estimates reflect differences in HRQoL for each pair-wise comparison. All p-values are significant after Bonferroni-adjustment or Tukey's HSD. ***p<.001; **p<.01; *p<.05.
|
|
|
|
|
|
|