To the Editor:

Alzheimer’s disease is the most common form of dementia (50–70% of cases) [1] and one of the leading causes of death in England and Wales [2]. The importance of developing strategies aiming to reduce the incidence of Alzheimer’s disease through modifiable risk factors is underscored by the inability of current treatments to reverse or delay disease progression.

Previous work using genome-wide association summary (GWAS) test statistics and multivariable Mendelian randomisation identified that a higher level of educational attainment and a higher level of intelligence are likely causal factors in Alzheimer’s disease, with intelligence affording a protective effect independent from that of educational attainment [3]. Published in this journal, Hu et al. [4] examined the effects of another variable, viz. “cognitive performance”. Again using GWAS summary statistics, they reported that a 1 SD increase in cognitive performance caused a 0.907 SD (95% CI = 0.877–0.938) increase in intelligence; moreover, a 1 SD increase in intelligence resulted in a 0.957 SD (95% CI = 0.937–0.978) increase in their so-called cognitive performance. Importantly, this cognitive performance also had a protective effect against Alzheimer’s disease that was independent of both education and intelligence [4]. Furthermore, Hu et al. [4] asserted that, whereas intelligence is largely fixed in early life, cognitive performance can be improved by education, exercise, and maintaining an active lifestyle; therefore, designing appropriate prevention strategies to increase cognitive performance might have clinical and public health implications by contributing to the reduction of Alzheimer’s disease.

Next, by carefully describing the sources of the GWAS summary statistics used for ‘intelligence’ and ‘cognitive performance’ in the Hu et al. [4] report, we ask whether these constructs are separable.

The instrumental variable used by Hu et al. [4] for ‘cognitive performance’ (n = 257,841) was obtained using GWAS summary statistics from a report by Lee at al. [5]. The variable was derived using 222,543 participants who undertook the verbal numerical reasoning test (also known as the fluid intelligence test) in the UK Biobank sample (UK Biobank data field 20016 for in-person assessments and data field 20191 for online assessment) and 35,298 participants of the COGENT consortium where the phenotype measured was the first unrotated principal component of performance on at least three neuropsychological tests or at least two IQ-test subscales. These data were then meta-analysed by Lee et al. [5] using Multi-Trait Analysis of GWAS (MTAG) [6] to capture the genetically correlated variance from three other cognitive performance related traits (educational attainment, highest mathematics qualification, and self-rated math ability). When cognitive performance was used as an outcome by Hu et al. [4], only the meta-analysis of UK Biobank and COGENT were used due to data availability.

When ‘intelligence’ was used as an outcome by Hu et al. [4], the data reported by Sniekers et al. [7] were used. These data contain 78,308 individuals of which 54,119 (a subset of both the participants used in Savage et al. [8] and Lee et al. [5]) were participants of UK Biobank who took the same verbal numerical reasoning test used in the cognitive performance phenotype derived by Lee et al. [5]. However, Sniekers et al. [7] included a measure of socio-economic status (SES, The Townsend Score) as an additional covariate that was not used in the Lee et al. [5] analysis.

When intelligence was used as an exposure by Hu et al. [4], 242 independent SNPs from the GWAS by Savage et al. [8] were used (n = 269,867). The GWAS on intelligence that was conducted by Savage et al. [8] utilised 195,653 participants who undertook the verbal numerical reasoning test in UK Biobank and so form a highly overlapping subset of participants that were in the Lee et al. [5] cognitive performance data set. Again, in this analysis by Savage et al. [8], socio-economic status was added as a covariate. This intelligence GWAS also used the same 35,298 participants of COGENT who completed the same tests as those used in the ‘cognitive performance’ phenotype derived by Lee et al. [5]. Savage et al. [8] also included an additional 35,993 participants sourced from independent epidemiological cohorts of European ancestry in which the phenotype was measured using various neurocognitive tests, primarily measuring fluid domains of cognitive functioning (See Table 1).

Table 1 Showing the cohorts used to derive the genetic instruments and outcomes for ‘cognitive performance’ and ‘intelligence’ from Hu et al. [4].

In psychology—although it may be applied beyond that discipline—there is an error known as the jangle fallacy [9]. One commits this error if one assumes that two identical (or near-identical) constructs are different if they happen to have different names. Unfortunately, Hu et al. [4] appear to have committed the jangle fallacy error by using intelligence and cognitive performance as if they were different variables. In this specific case of the jangle fallacy error, two data sets were used to make differently-named variables even though they contained the same participants who undertook exactly the same cognitive tests. The jangle fallacy error would still have been committed if another data set measuring intelligence (such as Sniekers et al. [7]) had been used as an exposure. As described above, these data sets are not identical but they measure the same phenotype as may be seen by the cognitive tests that were used (i.e., cognitive performance in Hu et al. [4] was measured using a test of fluid intelligence, and COGENT [10] derived a g factor from tests such as fluid reasoning, the Wechsler Adult Intelligence Scale—Revised, the Wechsler Abbreviated Scale of Intelligence, the Wechsler Adult Intelligence Scale I, II, and III).

The slight differences identified by Hu et al. [4] in the relationship that Alzheimer’s disease has with ‘intelligence’ and ‘cognitive performance’ is attributable to: (1) the bulk of the ‘intelligence’ data sets used in their instruments being conditioned on SES (195,653 participants’ scores were corrected for SES out of a total of 269,867 participants), whereas their ‘cognitive performance’ data set was not; (2) differences in the number of unique participants (Savage et al. [8] included 35,993 participants drawn from non-UK Biobank cohorts whereas Lee et al. [5] included an additional 26,890 UK Biobank participants); (3) and the ‘cognitive performance’ data set being meta-analysed with additional traits that capture SES. In the Hu et al. [4] report, intelligence and cognitive performance are not two correlated, but separate traits; rather, they are two measures of the same trait. We make the small caveat that the genetic variance associated with both SES and intelligence is absent from one of Hu et al. variables (‘intelligence’), and the genetic variance shared with educational attainment (often used as a measure of SES), highest mathematics qualification, and self-rated mathematics ability is better captured in the second (‘cognitive performance’).