Article | Open | Published:

Fast and powerful genome wide association of dense genetic data with high dimensional imaging phenotypes

Nature Communicationsvolume 9, Article number: 3254 (2018) | Download Citation

Abstract

Genome wide association (GWA) analysis of brain imaging phenotypes can advance our understanding of the genetic basis of normal and disorder-related variation in the brain. GWA approaches typically use linear mixed effect models to account for non-independence amongst subjects due to factors, such as family relatedness and population structure. The use of these models with high-dimensional imaging phenotypes presents enormous challenges in terms of computational intensity and the need to account multiple testing in both the imaging and genetic domain. Here we present a method that makes mixed models practical with high-dimensional traits by a combination of a transformation applied to the data and model, and the use of a non-iterative variance component estimator. With such speed enhancements permutation tests are feasible, which allows inference on powerful spatial tests like the cluster size statistic.

Introduction

Genome-wide association studies (GWAS) of neuroimaging data can advance our understanding of human brain by discovering genetic variants associated with normal and disorder-related phenotypic variance in brain structure and function1,2,3,4,5,6. Quantitative brain phenotypes are based on structural images (i.e. brain volume, cortical thickness, white matter integrity) or functional images (brain response to particular cognitive task or resting state). Genetic association at a hundred thousand locations in the human brain present immense statistical challenges including: low statistical power, need for multiple comparisons correction and like, other association studies, correction for population structure, a term that encompasses cryptic/family relatedness and population stratification.

In the GWA studies of unrelated individuals, non-independence due to latent population stratification or unknown (cryptic) relatedness7,8 is generally thought to be a confounding factor that can lead to excessive false positives when ignored. This type of non-independence has been studied throughly in the recent GWA era9,10,11,12,13. While genomic data can be used to control for population stratification by including the top principal components as fixed effect covariates in a linear regression model14, usually individuals with close estimated relatedness from identity-by-state (IBS) matrix or different ethnicities are excluded from the study sample. This might not be a problem in genetic studies with six digit sample sizes, but may make substantial differences in GWA studies with neuroimaging phenotypes where sample size is much smaller. Also, even in a carefully designed GWA study, it is hard to avoid spurious associations because of population structure; in particular, it is likely that in studies with large sample sizes, such as the UK biobank possess some level of population structure. Although the emergence of large scale neuroimaging consortia like ENIGMA or CHARGE can help to conduct well-powered genetic association studies through meta analysis framework, still it is crucial to use a powerful statistical method at the site level. Hence, there is a compelling need for analytical techniques that addresses these challenges.

Linear mixed effect models (LMMs) using molecularly derived empirical relatedness measures have grown in use recently for both studies of related and unrelated individuals, since they do not require self-reported biological relatedness, providing a framework where such complexities are automatically accounted for. LMMs have been used as an alternative to ordinary least squares (OLS), providing a mechanism to model trait variance explained by a genetic relationship matrix (GRM); such variance captures the genome-wide similarity between “unrelated” individuals by modeling it as a random effect15,16,17,18,19,20,21,22,23,24,25,26. It has been shown that the correction for the problem of latent population structure in GWA with an LMM is both effective and power preserving25,26,27.

These merits motivate the application of LMM for genetic association with high-dimensional imaging phenotypes. However, fitting LMM at each voxel in the brain is computationally intensive or even intractable at the voxel level since variance component estimation relies on likelihood function optimization using numerical methods. Moreover, search for genetic association across the genome at different locations with imaging phenotypes requires intense multiple testing corrections both for number of elements in an image and number of markers. Whether the association analysis is conducted for a few regions of interest (ROIs) or every voxel, naive application of Bonferroni correction for number of hypothesis testing in the image with usual GWA P-value leads to conservative statistical inference procedure as it ignores complex spatial dependence in the imaging phenotypes. Despite the many analytical techniques that have been developed to accelerate the GWA with LMM, these advances do not eliminate problems related to numerical optimization nor multiple testing problem. Furthermore, commonly used spatial inference tools, like cluster size28 or threshold-free cluster enhancement (TFCE) statistics29, depend on resampling-based inference methods to ensure valid control of false positives30,31. Familywise error rate (FWE) correction, controlling the chance of one or more false positives across the whole set (family) of tests32 requires the distribution the maximum statistic, can be computed for either voxels, ROIs, clusters or TFCE with a permutation test33, a standard tool to conduct inference in neuroimaging.

This paper makes two major contributions to reduce the complexity of LMM for genetic association with imaging phenotypes. First, the computational cost of variance component estimation is reduced by using a non-iterative one-step random effect estimator34. Second, the complexity of association testing is dramatically decreased by projecting the model and phenotype to a lower dimension space, combined with use of a score statistic for association testing. This projection is based on the eigenvectors of the GRM adjusted for the fixed effect nuisance terms16,35. In this setting, the projected phenotype likelihood function is equivalent to that used with restricted maximum likelihood (REML) of the LMM (Method), going forward we call this approach simplified REML35. While both models have the same statistical properties, our particular projection provides several computational benefits that dramatically reduces LMM complexity: (i) The diagonalized covariance allows a non-iterative one-step variance component estimator34, taking the form of a weighted regression of squared projected data on eigenvalues of the GRM adjusted for nuisance fixed effect terms, an approach that we call weighted least squares REML (WLS-REML); (ii) The regression form of our estimator is easily vectorized, meaning that many image elements and SNPs can be tested in a single and fast computational test in several high-level programming languages (Method); (iii) Finally, the simplicity and fast computation of the score test statistic makes permutation testing feasible, allowing exact, non-parametric control over the FWE, accounting for the number of tests conducted over all image elements and genetic markers; we define two permutation schemes, free and constrained, where in the latter case the permutation is confined to exchangeability blocks defined based on the eigenvalues distribution.

Results

Computational complexity

The reduced computational complexity of our method represents a significant advance over existing methods. The complexity of LMM association has two components, one for the variance component estimation, the other is for fixed effect parameter estimation and test statistic computation. For a GWA over S markers and V imaging phenotype elements on N individuals, the variance component likelihood optimization complexity of FaST-LMM19,35 which, to the best of our knowledge, is the fastest implementation of LMM is O(N3 + INV), where I is the average number of iterations, while for WLS-REML the random effect estimator (see Methods) it is O(N3 + NV) (the common O(N3) term is the time complexity of the GRM eigendecomposition). More critically, the estimation and test statistic computation complexity of FaST-LMM is O(SPN2V), where P is the number of nuisance fixed effects, while for WLS-REML (Eq. (16)) this is O(SNV), a substantial reduction for imaging phenotypes when number of image elements V is much bigger than the sample size N. Even for a single trait GWA (V = 1), our proposed projection reduces the association (Eq. (13)) complexity to O(SN) which is significantly less than FaST-LMM for large sample GWA.

In our previous work34, we introduced the WLS-ML random effect estimator that exploits a one-step optimization approach combined with eigen-rotation of phenotype and model (see Method for more details). The non-iterative estimator has a simple form, with variance components and fixed parameters each estimated by weighted least-squares regressions. In this paper, we evaluate our non-iterative ML and REML estimators (WLS-ML and WLS-REML) with their fully converged counterparts (Full ML or Full REML), comparing score, likelihood ratio (LRT) and Wald tests on intensive simulation studies; the two permutation schemes are also compared. Direct comparisons are made between FaST-LMM and our score test with WLS-REML using simulation and real data.

Simulation results

Intensive simulation studies are conducted to evaluate the proposed methods for association estimation and testing. The aim of the first study is to compare fully converged and one-step random effect estimators based on the simplified ML and REML functions. In the second study, the performance of various test statistics for the association testing are compared using a fully converged or one-step random effect estimators for ML and REML functions (see Supplementary Note 1 for details). Finally, we compare FaST-LMM to our preferred test, the score test based on the simplified REML function, using both false positive error rates and empirical power using simulated genetic markers (see Supplementary Note 2 for details).

Simulation results on the accuracy of genetic random effect $$\left( {\sigma _{\mathrm{A}}^2} \right)$$ estimation shows that the non-iterative one-step approaches are similar to their fully converged counterparts (Supplementary Figure 1), using either likelihood or restricted likelihood functions. When the data are independent $$\left( {\sigma _{\mathrm{A}}^2 = 0} \right)$$, the methods are indistinguishable in terms of bias and mean squared error (MSE). When $$\sigma _{\mathrm{A}}^2 \,> \, 0$$, the fully converged methods have less bias, but the difference is modest in absolute value; in terms of MSE, the non-iterative one-step methods have just slightly worse performance. The first simulation also shows good performance of fixed effect (β1) estimation (Supplementary Figure 2). Both the non-iterative one-step and fully converged have similar bias and MSE, with WLS-REML again closely following fully converged REML.

Simulations show that the false positive rates for the fixed effect score test for H0:β1 = 0 (Supplementary Figure 3a) are nominal; for both simplified ML or REML functions, for all simulation settings considered, test statistic type and type of random effect estimator, the false positive rates lay within the Monte Carlo confidence interval (MCCI) (see Supplementary Figures 4a, b).

The simulation results on the power of score test reveal negligible differences between the random effect estimation methods (Supplementary Figure 3b). Similar findings are obtained for the power of LRT and Wald tests (Supplementary Figures 4c, d). Like the parametric approach, we found that both permutation schemes, free or permutation within exchangeability blocks, control the false positives at the nominal level 5% (Fig. 1a and Supplementary Figures 5a, b), and could provide nearly equivalent power Fig. 1b, Supplementary Figures 5c, d) for all statistics either based on the simplified ML or REML functions. However, for all test statistics and $$\sigma _{\mathrm{A}}^2$$, the free permutation scheme is slightly more powerful than the constrained permutation test when a kinship matrix is used.

Simulations show that the null distribution of the score test for H0:β1 = 0 based on the simplified models using the fully converged and non-iterative variance component estimators are valid and indistinguishable (Fig. 2 and Supplementary Figure 6). However, we stress that the latter is much faster to calculate. Based on all of these results, we selected the score test based on the simplified REML function as the computationally most efficient test to be considered for genome-wide simulations and real data analysis.

Genome wide simulations were conducted to compare the parametric P-values from FaST-LMM and the score test based on the simplified REML using non-iterative variance component estimator in terms of false positives and power. The simulation results reveal that both approaches provide overall valid error rates (FaST-LMM = 4.94% and the WLS-REML score test = 4.89%, Fig. 3a). Power simulation shows that FaST-LMM and the score test have largely similar power (FaST-LMM = 15.25%, WLS-REML score test = 15.22%), however, FaST-LMM is slower (Fig. 3b). Despite reasonable concordance of P-value and fixed effect parameter estimates (β1) between FaST-LMM and simplified REML (Supplementary Figure 7), FaST-LMM’s estimates of parameter estimate variance $$( {{\mathrm{var}}( {\hat \beta })})$$ exhibits some systematic bias (Supplementary Figure 8).

The final simulation study evaluates controlling for a heritable fixed effect nuisance covariate in the null simulation setting when there is neither a SNP effect nor a covariate effect on the phenotype. Although the LMM can accommodate fixed effect nuisance terms, we compare to an alternate approach where nuisance covariates are regressed out in advance and LMM is fitted to the residualized phenotypes for GWA. We note that imaging association studies routinely use intracranial volume (ICV) as a nuisance covariate1,2, and ICV is well known to have heritability as large as 0.836,37. Figure 4 compares performance of FaST-LMM, EMMAX and the score test based on the simplified REML function using non-iterative random effect estimator (NINGA). The simulation results show that the parametric P-value from all approaches when the nuisance covariate is included in the LMM is valid (Fig. 4 left panel). However, the null distribution of P-values from LMM fitted on residualised phenotypes can be conservative (Fig. 4 right panel).

Association analysis of fractional anisotropy (FA) data

We performed GWA of whole brain FA data, using a whole brain parcellation of 42 ROIs, as well as a voxel-wise analysis for 53,458 voxels (332 subjects, 1,376,877 SNPs; for full details see Methods), comparing the WLS-REML score test with the fully converged random effect estimators with FaST-LMM. We also evaluate the use of OLS with MDS as nuisance fixed effects regressors for control of population structure in GWA with unrelated individuals.

The random effect estimators, one-step and fully converged REML are compared directly in Fig. 5 with a scatter plot, showing an apparent trade-off between accuracy and running time as the non-iterative method has lower estimates of $$\sigma _{\mathrm{A}}^2$$ for some regions.

Even with the tendency for genetic variance to be underestimated with the non-iterative method, the association statistic show remarkable concordance, with both approaches having almost the same performance (Fig. 6). FaST-LMM comparisons with the score test using the simplified REML function shows slightly larger statistics consistently for all ROIs, regardless of random effect estimation method (Supplementary Figures 10 and 11). Furthermore, comparing different approaches genomic control shows that regardless of random effect estimation method, the score test based on the simplified REML has smaller genomic control values than OLS with MDS nuisance regressors for all ROIs consistently. The genomic control of OLS with MDS nuisance regressors is poor, while the score test using both fully converged, one-step estimators and FaST-LMM have similar values close to unity (Supplementary Figure 9).

Figure 7 compares QQ-plot of association statistics between our model, FaST-LMM and OLS with MDS. These plots show either an identical distribution or slightly larger values for the OLS approach; however, the OLS approach has poor genomic control (Supplementary Figure 9) and after adjustment we get essentially identical results (Supplementary Figures 12 and 13).

A permutation test was used to find FWE-corrected P-values for 42 ROIs and 1,376,877 SNPs to assess association significance. Among the 42 × 1,376, 877 ≈ 57 million statistics, eight passed the permutation based FWE threshold $$\left( {\chi _1^2 = 34.72} \right)$$. Application of a Bonferroni correction for 42 tests to the usual GWA alpha level (5 × 10−8) yields to a more stringent threshold $$\left( {\chi _1^2 = 36.98} \right)$$ where only one association survives, indicating the potential improved power from a permutation-based inference that accounts for dependency among the tests (Fig. 8). An alternate, approximate approach involves computing the effective number of independent tests among the 42 ROIs; while there is no unique definition for the number of independent tests, we used one approach based on eigenvalues of phenotype correlation matrix38; this gave an effective test count of 18.30. Application of a Bonferroni correction for 18 independent tests to the usual GWA alpha level (5 × 10−8). This Bonferroni threshold with effective number of tests yields a slightly more stringent threshold $$\left( {\chi _1^2 = 35.37} \right)$$, but finds the same eight association statistics as found with permutation.

Finally we performed voxel-wise genome-wide association analysis of 53,458 voxels with 1,376,877 SNPs, using our proposed WLS-REML score test for association. Cluster-wise inference was performed on each spatial association map; we used a threshold corresponding to a $$\chi _1^2$$ P-value of 0.01 to create clusters, and 1000 permutations were used to compute the maximum distribution of cluster size over space and SNPs, offering FWE control over the entire search space; voxel-wise FWE thresholds were also computed. The level 5% FWE-corrected voxel-wise statistic threshold was 66.42, producing six significant association out of 84 billion tests. The 5% FWE corrected cluster size threshold is 7370 but no SNP’s statistic map had a cluster exceeding this value; the largest observed cluster size is 6648, which had a image-wide, genome-wide FWE-corrected cluster size P-value of 0.09. We note that the effective number of independent tests is not applicable to cluster-wise inference30 and does not scale to voxel-wise inference. Also, it depends on an arbitrary GWAS threshold that depends on the chip used.

Benchmarking and running times

We compared running time of our WLS-REML score test to that of FaST-LMM, which to our knowledge is the fastest implementation of LMM. The comparison was done using simulated and read data with a Intel(R) core(TM) 3.4 GHz i7-2600 CPU and 16 GB RAM. Parametric association testing of 5000 phenotypes with 6000 simulated markers using a sample of 300 individuals took 1 h with FaST-LMM, however, our implementation of the score test (Eq. (14-15)) only took 3 s. On real data, parametric whole genome association on 42 ROIs, required 756 min using FaST-LMM while our approach took only 2 min.

Discussion

Neuroimaging genetics has moved from establishing a heritable phenotypes to finding genetic markers that are associated with imaging phenotypes. Despite emerging world-wide consortia to boost GWA studies power using the largest possible sample sizes, there is a compelling need for powerful and computationally efficient analytic techniques that control for population structure at the site level. Moreover, GWA studies with neuroimaging phenotypes require fitting a model at each point (voxel/element) in the brain, and the large number of measurements presents a challenge both in terms of computational intensity and the need to account for elevated false positive risk because of the multiple testing problems both in terms of number of elements in image and number of markers being tested.

There has been rapid advances in quantitative genetic statistical methods account for population structure in GWA studies of unrelated individuals. Linear mixed models allowing for the rigorous testing of genetic associations (and, more generally, any fixed effects) have long been employed in human genetics as the standard to correct for the non-independence among subjects due to known familial relatedness in pedigree-based studies39,40,41,42,43. The LMM has gained popularity as an alternate method for GWA of unrelated individuals to correct for population structure15,16,17,18,19,20,21,22,23,24,25,26. However, due to the required inversion of potentially large matrices, the general LMM is computationally intensive where the complexity includes the deriving of the GRM, variance component parameter estimation, fixed effect estimation, and the calculation of the required association statistic for each marker grows with sample size and number of candidate markers for association testing.

To tackle these problems, we used an orthogonal transformation that substantially reduced LMM complexity for GWA. The equivalence between projected model and REML function helped us to reduce complexity of association testing. Specifically, the projection reduces the information matrix to a scalar that enables efficient vectorized implementation of score test with time complexity O(SNV). Further improvements in speed can be achieved by using the WLS-REML random effect estimator with O(NV) that we found to be more accurate than the WLS-ML estimator.

We conducted intensive simulation studies, evaluating a broad set of test statistics for association testing using the simplified ML and REML functions accompanied by one-step and fully converged random effect estimators. The one-step random effect estimator using simplified REML function provides more accurate approximation of the fully converged one in comparison to the WLS-ML variance component estimator. The simulation and real data analysis shows that only minor differences in marker effect estimation and association test statistics between one-step and fully converged random effect estimator. However, the former requires less computational resources. Also, we could not observe any appreciable differences in performances in terms of the error rate and power using the GRM from unrelated individuals or kinship matrix from a family study.

The WLS-REML random effect estimator is fast enough to be used to estimate voxel-wise heritability. Although the proposed one-step random effect estimator is not as accurate as fully converged one, it can be used for filtering a small number of elements for further investigation with more computational intense tools. Furthermore, when restricted to individuals with European ancestry we found LMM had genomic controls values closer to 1 than OLS values, indicating the success of the LMM in dealing with population structure.

We selected the score test based on the simplified REML function for further investigations because it only requires a single variance component estimate, common to all markers under the null hypothesis. Furthermore, efficient vectorized implementation of score test for images accelerates association testing. The null distribution of WLS-REML score test P-values was nearly as accurate as for the fully converged REML score test, meaning that permutation is not required for element-wise inference.

Genome wide association analysis of high dimensional imaging phenotype requires a computationally efficient LMM that can accommodate large sample sizes, provide multiple testing correction and complex spatial dependence among image elements. Although we have not observed significant differences of parametric P-values between our proposed method and existing methods, current LMM approaches depend on numerical optimization and are not feasible for GWA of over 1000’s imaging phenotypes, such as found in the UKBiobank44 or Human connectome project which has related individuals. Moreover, a search for genetic association across the genome with imaging phenotypes requires intense multiple testing corrections both for number of elements in an image and number of markers. The reduced complexity of our proposed method due to a non-iterative random effect estimator and vectorized implementation of score statistic computation makes permutation-based FWE-corrected inferences feasible. Our permutation approach is more computationally efficient than others proposed in the genetics literature45, avoiding inversion of the phenotypic covariance on each permutation. Permutation provides flexible FWE inference for ROIs or the whole brain, voxel-wise or cluster-wise. Though methods based on effective number of independent tests are computationally efficient, they depend on an (1) arbitrary GWAS threshold that depends on the chip used, (2) these approaches are not applicable to cluster-wise inference and (3) do not scale to voxel-wise inference which are the standard spatial statistics in imaging. While our vectorized implementation of score statistics demands complete data, there are various imputation methods can be used prior to association analysis to address missingness.

Whether using the linear mixed model for controlling population structure or kinship, high dimensional imaging phenotypes presents challenges in terms of computational intensity and elevated false positive risk; growing sample sizes and whole genome sequence data add to the computational burden. Our contribution in the acceleration of the exact LMM can be seen at two steps. First, covariance matrix estimation using WLS-REML random effect estimator reduces time complexity from O(N3 + INV) to O(N3 + NV). Further improvement in speed is also obtained by using the vectorized implementation of the score test based on simplified REML function. Our proposed method allows efficient implementation that reduces running time complexity to O(SNV). In addition, the efficient score test computation is fast enough to allow the permutation test to control family-wise error rate for number of elements in image and number of markers, and allow the use of spatial statistics like cluster size or TFCE.

Methods

Reference methods

Several approximate or exact methods have been proposed to speed up LMM-based testing. Approximate methods assume the total polygenic random effect is same for all markers under the null hypothesis of no marker effect, hence the relevant residual genetic variance component is estimated only once using all markers. In contrast, exact methods, which is the recommended LMM practice19,20,25,35,46, estimate a residual variance component conditional on each marker’s effect. In studies of “unrelated” subjects, this residual variance component often involves re-estimation of the GRM which is constructed excluding the candidate marker and surrounding markers in linkage disequilibrium.

To the best of our knowledge, the fastest implementation of exact LMM is Fast-LMM, which transforms the phenotype and LMM model with the genetic similarity matrix (GRM) eigenvectors and uses a profile likelihood approach to simplify variance component estimation. The eigenvector matrix diagonalisation along with the profile likelihood with only one variance parameter reduces optimization time substantially. In Fast-LMM the covariance matrix is estimated only under the null hypothesis of no marker effect, and then a generalized least squares (GLS) is applied to estimate the marker effect and the LRT is used for hypothesis testing. Note that small sample size behavior of this approach has not been validated; for example, using it for association analysis of imaging phenotypes with only 300 subjects might not be valid. In addition to concerns about the finite sample validity, Fast-LMM requires numerical optimization for each element (voxel/ROI) of image that makes it computationally intensive or essentially impractical for large-scale imaging phenotypes.

Linear mixed effect models

At each voxel/element, a LMM for the genetic association for N individuals can be written as:

$$Y = X_1\beta _1 + X_2\beta _2 + g + \varepsilon ,$$
(1)

where Y is the N-vector of the measured phenotype; X1 is a N-vector of a given marker’s minor allele count, implementing an additive genetic model; X2 is the N × (P − 1) matrix containing an intercept and fixed effect nuisance variables like age and sex; β1 is the scalar genetic effect; β2 is the (P − 1)-vector of nuisance regression coefficients and g is the N-vector of latent (unobserved) additive genetic effects; and ε is the N-vector of residual errors. The trait covariance, var(Y) = var(g + ε) = Σ can be written

$$\Sigma = \sigma _{\mathrm {A}}^2(2{\mathrm{\Phi }}) + \sigma _{\mathrm {E}}^2I,$$
(2)

where $$\sigma _{\mathrm {A}}^2$$ and $$\sigma _{\mathrm {E}}^2$$ are the additive genetic and the environmental variance components, respectively; I is the identity matrix; and 2Φ is the GRM matrix where element (i, j) is calculated as:

$$\phi _{i,j} = \frac{1}{M}\mathop {\sum}\limits_{k = 1}^M \frac{{\left( {x_{ik} - 2p_k} \right)\left( {x_{jk} - 2p_k} \right)}}{{2p_k\left( {1 - p_k} \right)}},$$

where xik is the minor allele count of the ith subject’s kth marker, coded as coded as 0, 1 or 2; pk is frequency of the kth marker; and M is the total number of markers.

Under the assumption that the the data follows a multivariate normal distribution, the model specified by Eqs. (1) and (2) have a log-likelihood of

$$\begin{array}{l}\ell _{{\mathrm{ML}}}\left( {\beta _{{\mathrm{ML}}},\Sigma _{{\mathrm{ML}}};Y,X} \right) \\ = - \frac{1}{2}\left[ {{\mathrm{Const}} + {\mathrm{log}}\left( {\left| \Sigma \right|} \right) + \left( {Y - X\beta } \right)\prime \Sigma ^{ - 1}\left( {Y - X\beta } \right)} \right],\end{array}$$
(3)

and a REML function of

$$\begin{array}{l}\ell _{{\mathrm{REML}}}\left( {\Sigma _{{\mathrm{REML}}};Y,X} \right) \\ = - \frac{1}{2}\left( {{\mathrm{Const}} - {\mathrm {log}}\left| {X\prime X} \right| + {\mathrm{log}}\left| \Sigma \right| + {\mathrm{log}}\left| {X\prime \Sigma ^{ - 1}X} \right| + Y\prime PY} \right),\end{array}$$
(4)

where $$X = \left[ {\begin{array}{*{20}{c}} {X_1} & {X_2} \end{array}} \right]$$ and $$\beta = \left[ {\begin{array}{*{20}{c}} {\beta _1} & {\beta _2} \end{array}} \right]$$ are the full design matrix of fixed effects and their parameter estimate vector, respectively, and P = Σ−1(I − X(X′Σ−1X)−1X′Σ−1), the projection matrix. The fixed effect parameters are estimated using GLS

$$\hat \beta _{{\mathrm{REML}}} = \left( {X\prime \hat \Sigma _{{\mathrm{REML}}}^{ - 1}X} \right)^{ - 1}X\prime \hat \Sigma _{{\mathrm{REML}}}^{ - 1}Y,$$

where $$\hat \Sigma _{{\mathrm{REML}}}^{ - 1}$$ comes from optimized REML function (Eq. (4)).

Several algorithms have been proposed to accelerate ML or REML optimization by transforming the model with the eigenvectors of the GRM and/or using a different covariance matrix parametrisation16,19,21,34,35,46. Here we consider standard additive model covariance matrix parametrisation (Eq. (2)) as we can efficiently estimate it with our one-step, regression based approach34.

Simplified REML and ML functions

The simplified ML function for LMM is discussed in refs. 16,19,21. For completeness, we review shortly the simplified ML function, to be next followed by development of the simplified REML function. The simplified ML function is obtained by transforming the data and model with an orthogonal transformation S, the matrix of eigenvectors of 2Φ that crucially coincide with the eigenvectors of Σ:

$$S\prime Y = S\prime X\beta + S\prime g + S\prime \varepsilon$$

which we write as

$$Y^ \ast = X^ \ast \beta + g^ \ast + \varepsilon ^ \ast ,$$
(5)

where $$Y^{\ast}$$ is the transformed data, $$X^{\ast}$$ is the transformed covariate matrix, $$g^{\ast}$$ and $$\epsilon^{\ast}$$ are the transformed random components. The diagonalising property of the eigenvectors then gives a simplified form for the variance:

$${\mathrm{var}}\left( {\varepsilon ^ \ast } \right) = \Sigma ^ \ast = \sigma _{\mathrm{A}}^2D_{\mathrm{g}} + \sigma _{\mathrm{E}}^2I,$$

where, $$\Sigma^{\ast}$$ is the variance of the transformed data and Dg = diag{λgi} is a diagonal matrix of the eigenvalues of 2Φ.

The log likelihood takes on the exact same form as Eq. (3) for $$Y^{\ast}$$, $$X^{\ast}$$, β and $$\Sigma^{\ast}$$, except is easier to work with since $$\Sigma^{\ast}$$ is diagonal:

$$\begin{array}{l}\ell _{{\mathrm{ML}}}\left( {\beta ,\sigma _{\mathrm{A}},\sigma _{\mathrm{E}};Y^ \ast ,X^ \ast } \right) \\ = - \frac{1}{2}\left[ {N{\mathrm{log}}\left( {2\pi } \right) + \mathop {\sum}\limits_{i = 1}^N {\mathrm{log}}\left( {\sigma _{\mathrm{A}}^2\lambda _{{\mathrm{g}}i} + \sigma _{\mathrm{E}}^2} \right) + \mathop {\sum}\limits_{i = 1}^N \frac{{\left( {y_i^ \ast - x_i^ \ast \beta } \right)^2}}{{\sigma _{\mathrm{A}}^2\lambda _{{\mathrm{g}}i} + \sigma _{\mathrm{E}}^2}}} \right],\end{array}$$

where $$y_i^ \ast$$ is the ith element of $$Y^{\ast}$$, and $$x_i^ \ast$$ is the ith row of $$X^{\ast}$$.

The REML function (Eq. (4)) is simplified using projection matrix (P = Σ−1(I − X(X′Σ−1X)−1X′Σ−1)35). Also, an alternate simplified form of the REML function can be obtained as follows:

Let M = I − X(XX)−1X′ be the residual forming matrix based on the fixed effects regressors. Since M is idempotent, it can be decomposed as

$$\begin{array}{l}M = AA\prime ,\\ A\prime A = I,\end{array}$$

where A is the N × (N − P) matrix of the eigenvectors of M corresponding to the non-zero eigenvalues. Crucially, A also residualises the data, because it is orthogonal to the design matrix X:

$$\begin{array}{*{20}{l}} {A\prime X} \hfill & = \hfill & {A\prime AA\prime X,} \hfill \\ {} \hfill & = \hfill & {A\prime MX = 0.} \hfill \end{array}$$

Hence AY ~ N(0, A′ΣA) and the log likelihood of the transformed data is

$$\begin{array}{l}\ell \left( {A\prime Y,\Sigma } \right) \\ = - \frac{1}{2}\left[ {{\mathrm{Const}} + {\mathrm{log}}\left| {A\prime \Sigma A} \right| + Y\prime A\left( {A\prime \Sigma A} \right)^{ - 1}A\prime Y} \right].\end{array}$$
(6)

Now we show that this (Eq. (6)) is equivalent to Eq. (4), and thus we can use the eigenvectors of the residual-forming matrix to build the REML log likelihood.

It can be shown that

$${\mathrm{log}}\left| {A\prime \Sigma A} \right| = {\mathrm{log}}\left| \Sigma \right| + {\mathrm{log}}\left| {X\prime \Sigma ^{ - 1}X} \right| - {\mathrm{log}}\left| {X\prime X} \right|.$$

and using A(A′ΣA)−1A = P (47, M4.f p.451), we have that Eqs. (4) and (6) are equivalent and the transformed data likelihood function is exactly as same as the REML function.

As A is not unique, we seek to find one that diagonalises the covariance of the residualised data. The transformation matrix could be derived from eigendecomposition of GRM adjusted for the fixed effect covariates as follows:

$$M\left( {2{\mathrm{\Phi }}} \right)M = S_{\mathrm{r}}D_{{\mathrm{g}}_{\mathrm{r}}}S_{\mathrm{r}}\prime ,$$

where $$D_{{\mathrm{g}}_{\mathrm{r}}} = {\mathrm{diag}}\{ \lambda _{{\mathrm{g}}_{\mathrm{r}}i}\}$$ is the (N − P) × (N − P) diagonal matrix of non-zero eigenvalues; and Sr is the N × (N − P) matrix of eigenvectors that corresponds to non-zero eigenvalues. Firstly, Sr is a valid A, because its columns are orthogonal $$S_{\mathrm{r}}\prime S_{\mathrm{r}} = I$$ and $$M = S_{\mathrm{r}}S_{\mathrm{r}}\prime$$16. Thus we define the projected polygenic model by pre-multiplying $$S_{\mathrm{r}}\prime$$ both sides of polygenic model (Eq. (1)):

$$\begin{array}{*{20}{l}} {S_{\mathrm{r}}\prime Y} \hfill & = \hfill & {S_{\mathrm{r}}\prime X + S_{\mathrm{r}}\prime g + S_{\mathrm{r}}\prime \varepsilon ,} \hfill \\ {S_{\mathrm{r}}\prime MY} \hfill & = \hfill & {S_{\mathrm{r}}\prime MX + S_{\mathrm{r}}\prime Mg + S_{\mathrm{r}}\prime M\varepsilon .} \hfill \end{array}$$

which we write as

$$Y_{\mathrm{r}}^ \ast = g_{\mathrm{r}}^ \ast + \varepsilon _{\mathrm{r}}^ \ast ,$$
(7)

where $$Y_{\mathrm{r}}^ \ast$$, $$g_{\mathrm{r}}^ \ast$$ and $$\varepsilon _{\mathrm{r}}^ \ast$$ are N − P projected phenotype, genetic and residual vectors, respectively. In this fashion, the projected phenotype covariance matrix becomes diagonal:

$$\begin{array}{*{20}{l}} {{\mathrm{cov}}\left( {Y_{\mathrm{r}}^ \ast } \right)} \hfill & = \hfill & {\Sigma _{\mathrm{r}}^ \ast ,} \hfill \\ {} \hfill & = \hfill & {{\mathrm{cov}}\left( {S_{\mathrm{r}}\prime Y} \right),} \hfill \\ {} \hfill & = \hfill & {{\mathrm{cov}}\left( {S_{\mathrm{r}}\prime MY} \right),} \hfill \\ {} \hfill & = \hfill & {S_{\mathrm{r}}\prime \left( {M\Sigma M} \right)S_{\mathrm{r}},} \hfill \\ {} \hfill & = \hfill & {S_{\mathrm{r}}\prime \left( {\sigma _{\mathrm{A}}^2M(2{\mathrm{\Phi }})M + \sigma _{\mathrm{E}}^2M} \right)S_{\mathrm{r}},} \hfill \\ {} \hfill & = \hfill & {\sigma _{\mathrm{A}}^2S_{\mathrm{r}}\prime \left( {S_{\mathrm{r}}D_{{\mathrm{g}}_{\mathrm{r}}}S_{\mathrm{r}}\prime }\right)S_{\mathrm{r}} + \sigma _{\mathrm{E}}^2S_{\mathrm{r}}\prime \left( {S_{\mathrm{r}}S_{\mathrm{r}}\prime } \right)S_{\mathrm{r}},} \hfill \\ {} \hfill & = \hfill & {\sigma _{\mathrm{A}}^2D_{{\mathrm{g}}_{\mathrm{r}}} + \sigma _{\mathrm{E}}^2I,} \hfill \end{array}$$

where we have used the identity $$S_{\mathrm{r}}\prime M = S_{\mathrm{r}}\prime$$. That is, therefore the projected data, $$Y_{\mathrm{r}}^ \ast$$, loglikelihood takes on a simpler form:

$$\begin{array}{l}\ell _{{\mathrm{REML}}}\left( {\sigma _{\mathrm{A}}^2,\sigma _{\mathrm{E}}^2;Y_{\mathrm{r}}^ \ast } \right)\\ = - \frac{1}{2}\left[ {{\mathrm{Const}} + \mathop {\sum}\limits_{i = 1}^{N - P} {\mathrm{log}}\left( {\lambda _{{\mathrm{g}}_{\mathrm{r}}i}\sigma _{\mathrm{A}}^2 + \sigma _{\mathrm{E}}^2} \right) + \mathop {\sum}\limits_{i = 1}^{N - P} \frac{{y_{{\mathrm{r}}i}^{ \ast 2}}}{{\lambda _{{\mathrm{g}}_{\mathrm{r}}i}\sigma _{\mathrm{A}}^2 + \sigma _{\mathrm{E}}^2}}} \right].\end{array}$$
(8)

where $$y_{{\mathrm{r}}i}^{ \ast 2}$$ is the square of the ith element of $$Y_{\mathrm{r}}^ \ast$$. It is clear from the Eq. (8) that working with the simplified version of REML is computationally easier than the original one (Eq. (4)). Beside accelerating the REML optimisation, this approach facilitates performing LRT for fixed effects (βs) and leads to a computationally efficient estimator and test statistic, described below.

REML and ML parameter estimation

We choose Fisher’s scoring method to optimize the simplified ML and REML functions because it leads to computationally efficient variance component estimators. The score and the expected Fisher information matrices for the simplified models can be expressed as:

$$\begin{array}{l}S_{{\mathrm{ML}}}\left( {\beta ,\theta } \right) = \left[ {\begin{array}{*{20}{c}} {X^{ \ast} \prime \Sigma ^{ \ast - 1}\varepsilon ^ \ast } \\ { - \frac{1}{2}\left[ {U\prime \Sigma ^{ \ast - 1}1 - U\prime \Sigma ^{ \ast - 2}\varepsilon ^{ \ast 2}} \right]} \end{array}} \right],\\ I_{{\mathrm{ML}}}\left( {\beta ,\theta } \right) = \left[ {\begin{array}{*{20}{c}} {X^{ \ast} \prime \Sigma ^{ \ast - 1}X^ \ast } & 0 \\ 0 & {\frac{1}{2}U\prime \Sigma ^{ \ast - 2}U} \end{array}} \right],\end{array}$$

and

$$\begin{array}{*{20}{l}} {S_{{\mathrm{REML}}}\left( \theta \right)} \hfill & = \hfill & { - \frac{1}{2}\left[ {U_{\mathrm{r}}\prime \Sigma _{\mathrm{r}}^{ \ast - 1}1 - U_{\mathrm{r}}\prime \Sigma _{\mathrm{r}}^{ \ast - 2}Y_{\mathrm{r}}^{ \ast 2}} \right],} \hfill \\ {I_{{\mathrm{REML}}}\left( \theta \right)} \hfill & = \hfill & {\frac{1}{2}U_{\mathrm{r}}\prime \Sigma _{\mathrm{r}}^{ \ast - 2}U_{\mathrm{r}},} \hfill \end{array}$$

where $$\theta = \left( {\sigma _{\mathrm{E}}^2,\sigma _{\mathrm{A}}^2} \right)$$; U = [1, λg] and $$U_{\mathrm{r}} = [{\mathbf{1}}_{\mathrm{r}},\lambda _{{\mathrm{g}}_{\mathrm{r}}}]$$ are N × 2 and (N − P) × 2 matrices; and λg is the vector of eigenvalues of (2Φ); $$\lambda _{{\mathrm{g}}_{\mathrm{r}}}$$ the vector of eigenvalues of M(2Φ)M; 1 and 1r are N and (N − P)-vectors of one, respectively; $$Y_{\mathrm{r}}^{ \ast 2}$$ is the element wise square of $$Y_{\mathrm{r}}^ \ast$$; and $$\epsilon^{\ast2}$$ is the element wise square of $$\epsilon^{\ast}$$. Following Fisher’s scoring method it can be shown that at each iteration, maximum-likelihood estimation of β and θ are updated based on WLS regression of $$Y^{\ast}$$ on $$X^{\ast}$$ and $$\epsilon^{\ast2}$$ on U, respectively, as follows:

$$\begin{array}{l}\hat \beta _{{\mathrm{ML}},j + 1} = \left( {X^{ \ast} \prime \left( {\hat \Sigma _j^ \ast } \right)^{ - 1}X^ \ast } \right)^{ - 1}X^{^\ast} \prime \left( {\hat \Sigma _j^ \ast } \right)^{ - 1}Y^ \ast ,\\ \hat \theta _{{\mathrm{ML}},j + 1} = {\mathrm {max}}\left\{ {0,\,\left( {U\prime (\hat \Sigma _j^{ \ast 2})^{ - 1}U} \right)^{ - 1}U\prime \left( {\hat \Sigma _j^{ \ast 2}} \right)^{ - 1}\hat \varepsilon _j^{ \ast 2}} \right\},\end{array}$$
(9)

and REML estimation of θ is updated based on WLS regression of $$Y_{\mathrm{r}}^{ \ast 2}$$ on Ur as follows:

$$\hat \theta _{{\mathrm{REML}},j + 1} = {\mathrm {max}}\left\{ {0,\,\left( {{\mathrm {U}}_{\mathrm{r}}\prime \left( {\hat \Sigma _{{\mathrm{r}}j}^{ \ast 2}} \right)^{ - 1}U_{\mathrm{r}}} \right)^{ - 1}U_{\mathrm{r}}\prime \left( {\hat \Sigma _{{\mathrm{r}}j}^{ \ast 2}} \right)^{ - 1}Y_{\mathrm{r}}^{ \ast 2}} \right\},$$

where j indexes iteration; $$\Sigma _j^{ \ast 2}$$ and $$\Sigma _{{\mathrm{r}}j}^{ \ast 2}$$ are constructed with θML,j and θREML,j, respectively; $$\varepsilon _j^{ \ast 2}$$ is the element-wise square of $$\varepsilon _j^ \ast = Y^ \ast - X^ \ast \beta _{{\mathrm{ML}},j}$$; $$Y_{\mathrm {r}}^{ \ast 2}$$ is the element-wise square of $$Y_{\mathrm {r}}^ \ast$$; and the variance parameters θ must be positive, hence the maximum operator. As usual, these updates are iterated until convergence criteria holds.

It has been shown that when the initial value is a consistent estimator, the estimator based on the first iteration is asymptotically normal and consistent48. Such initial value for $$\hat \beta _{{\mathrm{ML}}}$$ and $$\hat \theta _{{\mathrm{ML}}}$$ could be derived from OLS regression coefficients of $$Y^{\ast}$$ on $$X^{\ast}$$ and squared residuals on U, respectively:

$$\begin{array}{l}\hat \beta _{{\mathrm{ML,OLS}}} = \left( {X^{ \ast} \prime X^ \ast } \right)^{ - 1}X^{ \ast} \prime Y^ \ast ,\\ \hat \theta _{{\mathrm{ML,OLS}}} = {\mathrm{max}}\left\{ {0,\,\left( {U\prime U} \right)^{ - 1}U\prime \varepsilon ^{ \ast 2}} \right\}.\end{array}$$

For REML, initial values for $$\hat \theta _{{\mathrm{REML,OLS}}}$$ can be found as OLS regression coefficient of $$Y_{\mathrm{r}}^{ \ast 2}$$ on Ur:

$$\hat \theta _{{\mathrm{REML,OLS}}} = {\mathrm{max}}\left\{ {0,\,\left( {U_{\mathrm{r}}\prime U_{\mathrm{r}}} \right)^{ - 1}U_{\mathrm{r}}\prime Y_{\mathrm{r}}^{ \ast 2}} \right\}.$$

Hence our one-step, non-iterative estimators are:

$$\hat \beta _{{\mathrm{ML,WLS}}} = \left( {X^{ \ast} \prime \left( {\hat \Sigma _{{\mathrm{OLS}}}^ \ast } \right)^{ - 1}X^ \ast } \right)^{ - 1}X^{ \ast} \prime \left( {\hat \Sigma _{{\mathrm{OLS}}}^ \ast } \right)^{ - 1}Y^ \ast ,$$
(10)
$$\hat \theta _{{\mathrm{ML,WLS}}} = {\mathrm{max}}\left\{ {0,\left( {U\prime \left( {\hat \Sigma _{{\mathrm{OLS}}}^{ \ast 2}} \right)^{ - 1}U} \right)^{ - 1}U\prime \left( {\hat \Sigma _{{\mathrm{OLS}}}^{ \ast 2}} \right)^{ - 1}\varepsilon _{{\mathrm{OLS}}}^{ \ast 2}} \right\},$$
(11)
$$\hat \theta _{{\mathrm{REML,WLS}}} = {\mathrm{max}}\left\{ {0,\left( {{\mathrm{U}}_{\mathrm{r}}\prime \left( {\hat \Sigma _{{\mathrm{OLS,r}}}^{ \ast 2}} \right)^{ - 1}U_{\mathrm{r}}} \right)^{ - 1}U_{\mathrm{r}}\prime \left( {\hat \Sigma _{{\mathrm{OLS,r}}}^{ \ast 2}} \right)^{ - 1}Y_{\mathrm{r}}^{ \ast 2}} \right\},$$
(12)

where $$\hat \Sigma _{{\mathrm{OLS}}}^ \ast$$ and $$\hat \Sigma _{{\mathrm{OLS,r}}}^ \ast$$ are formed by $$\hat \theta _{{\mathrm{ML,OLS}}}$$ and $$\hat \theta _{{\mathrm{REML,OLS}}}$$ respectively, and $$\hat \varepsilon _{{\mathrm{OLS}}}^{ \ast 2}$$ is the element-wise square of $$\hat \varepsilon _{{\mathrm{OLS}}}^ \ast = Y^ \ast - X^ \ast \hat \beta _{{\mathrm{OLS}}}$$. Going forward, we will use “ML” or “REML” to refer to the iterated estimators and “WLS” to refer to these one-step estimators.

Association testing

The score, LRT and the Wald tests can be used for the genetic association testing using either ML or REML functions of the model in Eq. (1).

The score statistic49 that requires the value of score and information matrices under the the null hypothesis constraint (H0:β1 = 0) for the simplified ML model (Eq. 5) can be written

$$T_{{\mathrm{S,ML}}} = \tilde \varepsilon\prime _{{\mathrm{ML}}} \tilde \Sigma _{{\mathrm{ML}}}^{ \ast - 1}X_1^ \ast \left[ {C\prime (X^{ \ast} \prime \tilde \Sigma _{{\mathrm{ML}}}^{ \ast - 1}X^ \ast )^{ - 1}C} \right]X_1^{ \ast} \prime \tilde \Sigma _{{\mathrm{ML}}}^{ \ast - 1}\tilde \varepsilon _{{\mathrm{ML}}},$$

where C is a P × 1 contrast vector; $$X^ \ast = [X_1^ \ast \quad X_2^ \ast ]$$ encompasses the full transformed covariate matrix; $$\tilde \varepsilon\prime _{{\mathrm{ML}}}$$ and $$\tilde \Sigma _{{\mathrm{ML}}}$$ are the ML residual and covariance matrix estimates under the null hypothesis constraint. The score statistic for the projected model (Eq. (7)) can be derived like TS,ML following the projection with respect to the H0 fixed effects, i.e. nuisance, terms X2,

$$T_{{\mathrm{S,REML}}} = Y_{\mathrm{r}}^{^\ast \prime }\tilde \Sigma _{\mathrm{r}}^{ \ast - 1}X_{1{\mathrm{r}}}^ \ast \left( {X_{1{\mathrm{r}}}^{\ast \prime }\tilde \Sigma _{\mathrm{r}}^{ - 1}X_{1{\mathrm{r}}}^ \ast } \right)^{ - 1}X_{1{\mathrm{r}}}^{\ast \prime }\tilde \Sigma _{\mathrm{r}}^{ \ast - 1}Y_{\mathrm{r}}^ \ast ,$$
(13)

where $$Y_{\mathrm{r}}^ \ast = S_{2{\mathrm{r}}}^\prime Y$$ and $$X_{1{\mathrm{r}}}^ \ast = S_{2{\mathrm{r}}}^\prime X_1$$ are (N − P + 1)-vectors of the projected phenotype and allele frequency count, respectively; and the projection matrix S2r is comprised of the eigenvectors of M2(2Φ)M2 with non-zero eigenvalues, $$M_2 = I - X_2\left( {X_2X_2\prime } \right)^{ - 1}X_2\prime$$; and $$\tilde \Sigma _{\mathrm{r}}^{ \ast - 1}$$ is the projected model covariance matrix estimation under the null model constraint.

The LRT50 statistic is twice the difference of the optimized log-likelihoods, unrestricted minus H0-restricted. For ML this requires optimizing the likelihood function twice, once under the null H0:β1 = 0, once under the alternative. We denote the test statistic for this test TL,ML. A well-known aspect of REML is that it cannot be used to tests of fixed effects, since the null hypothesis would represent a change of the projection that defines the REML model. However, we can consistently use the same projection S2r, under the unrestricted and restricted models, to diagonalise our covariance and carry out such a hypothesis test. To be precise, the unrestricted model is

$$S_{\mathrm {{2r}}}\prime Y = S_{\mathrm {{2r}}}\prime X_1\beta _1 + S_{\mathrm {{2r}}}\prime g + S_{\mathrm {{2r}}}\prime \varepsilon ,$$

where $$S_{\mathrm {{2r}}}\prime X_2\beta _2 = 0$$ by the construction of S2r, and the restricted model is

$$S_{\mathrm {{2r}}}\prime Y = S_{\mathrm {{2r}}}\prime g + S_{\mathrm {{2r}}}\prime \varepsilon .$$

Following the same procedure as ML, the test statistic is denoted by TL,REML.

For a scalar parameter, the Wald test49 is the parameter estimate divided by the standard deviation of the estimate under an unrestricted model. For an vector parameter β and contrast C, it takes the form

$$T_{\mathrm {W}} = C\hat \beta \left( {C(X\prime \hat \Sigma ^{ - 1}X)C\prime } \right)^{ - 1}\hat \beta \prime C\prime$$

where $$\hat \beta$$ and $$\hat \Sigma ^{ - 1}$$ are the parameter estimations under the alternative hypothesis; this form holds for both ML and REML. A test for genetic association testing can be calculated either using fully converged or one-step variance component estimators. In the parametric framework, all of the aforementioned tests null distribution follow chi square distribution with one degree of freedom asymptotically.

Inference using the permutation test

In neuroimaging the permutation test is a standard tool to conduct inference while controlling the family wise error rate (FWE)33. It only requires an assumption of exchangeability, that the joint distribution of the error is invariant to permutation, and provides exact inference in the absences of nuisance variables, or approximately exact inference with nuisance variables51. Control of the FWE of a voxel-wise or cluster-wise statistic is obtained from a maximum distribution of the corresponding statistic. However naive use of permutation test for genetic association testing, ignoring dependence structure between individuals, leads to invalid inferences45,52. To the best of our knowledge, the only work that uses a permutation test for association analysis in the context of LMM uses the estimated covariance matrix under the null hypothesis to whiten the phenotype vector, yielding exchangeable data45. This approach requires estimating the covariance matrix for each phenotype for each permutation, a significant computational burden with high dimensional imaging data. Here we propose two permutation schemes that account for dependence explained by our model, one free and one constrained permutation approach.

The genetic association testing in the context of LMM using a permutation test requires proper handling of fixed effect and random effect nuisance variables in order to respect the exchangeablity assumption. While there are a variety of methods for testing for a fixed effect when the errors are independent51. However, little work has been done for fixed effect inference using a permutation test in linear mixed models where the error term is correlated45.

Free permutation for the simplified ML model: For the simplified model (Eq. (5)) we create permuted data $$\tilde Y^ \ast$$ using the reduced, H0:β1 = 0 null model residuals to create surrogate null data,

$$\tilde Y^ \ast = X_2^ \ast \hat \beta _2 + \tilde P\hat \varepsilon ,$$

where $$\tilde P$$ is one of $$N!$$ possible $$N \times N$$ permutation matrices; $$\hat \beta _2$$ is the reduced model nuisance estimate found with either fully converged (Eq. (9)) or one-step (Eq (10)) methods; $$\hat \varepsilon$$ denotes the reduced model residuals likewise found with either fully converged or one-step estimators; and the tilde accent on the data ($$\tilde Y^ \ast$$) and permutation ($$\tilde P$$) denotes one of many null hypothesis realizations. The reduced null model is not exchangeable due to heteroscedasticity of $$\Sigma ^ \ast$$, but we account for this in each permutation step by fitting the simplified model (5) with the permuted covariance matrix

$${\mathrm{cov}}\left( {\tilde Y^ \ast } \right) = \tilde P\Sigma ^ \ast \tilde P\prime = \sigma _{\mathrm{A}}^2\tilde PD_{\mathrm{g}}\tilde P\prime + \sigma _{\mathrm{E}}^2I.$$

With this approach we obtain samples from the empirical null distribution of the maximum score, LRT and the Wald tests (or cluster-size, after thresholding one of these test statistics), where the maximum is taken over all voxels and SNPs to control FWE.

Free permutation for simplified REML model: While the previous approach creates null hypothesis realizations by permuting the null-model residuals and adding back on estimated nuisance, here we will permute data after reduced-model eigen-transformation. We do this because projection removes the nuisance fixed effect covariates. In both cases, though, we must account for the dependence existing under H0.

An alternate permutation scheme could be built based on projecting the LMM model (Eq. (1)) to the lower dimension space with respect to the null hypothesis reduced model, i.e. using only the nuisance fixed effect terms. Let M2 be the residual forming based on X2 alone and define S2r as the transformation based on the non-trivial eigenvectors of M2(2Φ)M2, creating a model with dimension N − (P + 1):

$$Y_{\mathrm{r}}^ \ast = X_{1{\mathrm{r}}}^ \ast \beta _1 + g_{\mathrm{r}}^ \ast + \varepsilon _{\mathrm{r}}^ \ast ,$$

where $$Y_{\mathrm{r}}^ \ast = S\prime_{2{\mathrm{r}}} Y$$ is the reduced transformed data; $$X_{1{\mathrm{r}}}^ \ast = S\prime_{2{\mathrm{r}}} X_1$$ is as above, the reduced transformed additive genetic effect; $$g_{\mathrm{r}}^ \ast = S\prime_{2{\mathrm{r}}} g$$ and $$\varepsilon _{\mathrm{r}}^ \ast = S\prime_{2{\mathrm{r}}} \varepsilon$$ are the latent genetic effect and random error terms, respectively, after the reduced transformation. Here we permute the data, producing $$\tilde Y_r^ \ast = \tilde PY_{\mathrm{r}}^ \ast$$, with permuted covariance matrix

$${\mathrm{cov}}\left( {\tilde Y_{\mathrm{r}}^ \ast } \right) = \sigma _{\mathrm{A}}^2\tilde PD_{{\mathrm{g}}_{\mathrm{r}}}\tilde P\prime + \sigma _{\mathrm{E}}^2I$$

fit in each permutation step. However, under the null hypothesis of no genetic effect, estimated random effects for permuted phenotype are exactly as same as the un-permuted phenotype and hence the variance components only need to be estimated once.

Constrained permutation for exchangeability blocks: In the free permutation approaches we permute despite the lack of exchangeability, but then permute the covariance structure to account for this. An alternate approach is to define exchangeability blocks where observations within each block can be regarded as exchangeable. Precisely, with exchangeability blocks, the assumption is that permutations altering the order of observations only within each block preserve the null hypothesis distribution of the full data.

While not feasible for the original correlated model (1), in the simplified ML (5) or simplified REML (7) model we can define approximate exchangeability blocks. In simplified models the sorted eigenvalues arrange the observations by variance (increasing or decreasing, depending on software conventions). Hence blocks of contiguous observations $$Y^{\ast}$$ or $$Y_{\mathrm{r}}^ \ast$$ will have variance that is as similar as possible and will be, under the null hypothesis, approximately exchangeable.

We propose to define exchangeability blocks such that the range of Dg or $$D_{{\mathrm{g}}_{\mathrm{r}}}$$ values within a block is no greater than 0.01. This cut off ensured the eigenvalues did not vary by more than a factor of 10% within a block. Permutation is constrained within these blocks and the test procedure is as described above for simplified ML and REML free permutation schemes above, except that the test statistic is computed using the unpermuted covariance matrix.

Efficient score statistic implementation for vectorized images

To fully exploit the computational advantage of our non-iterative, reduced-dimension projected model estimation method we require a vectorized algorithm. That is, even without iteration, the method will be relatively slow if the evaluation of the estimates is so complex that each phenotype must be looped over one-by-one. For fast evaluation with a high-level language like Matlab, the estimation process for a set of phenotypes must be cast as a series of matrix algebra manipulations. To exploit the computational advantage of this approach, the phenotype should be observed for all individuals. However, soft imputation methods can be used in case of missing phenotypes. An alternate approach has been proposed53 for multi-trait GWA. However, our proposed method is computationally more efficient due to NINGA.

In this section, we develop the vectorized algorithm for association one chromosome’s worth of SNPs and all image voxels/elements (subject to memory constraints). To avoid proximal contamination19 and efficient implementation of LMM, we follow leave one chromosome out approach where all markers on a chromosome being tested are excluded from the GRM25,27.

Let Yr and Xr be a (N − P) × V and (N − P) × G matrices of projected traits and allele frequencies, respectively, where V and G are number of elements in image and number of SNPs the tested chromosome, respectively. The score test requires parameter estimation under the null hypothesis constraints, and since X2 is the same for all SNPs, the estimated covariance matrix will be the same all markers the chromosome. Thus the covariance matrix only need to be estimated once as follows:

$$\begin{array}{l}{\mathbf{F}} = {\mathbf{Y}}_{\mathrm{r}} \odot {\mathbf{Y}}_{\mathrm{r}},\\ \theta = {\mathrm{max}}( {( {U\prime_{\mathrm{r}} U_{\mathrm{r}}})^{ - 1}U\prime_{\mathrm{r}} {\mathbf{F}},{\mathbf{0}}}),\\ W = {\mathbf{1}}_{NV} \oslash \left( {\left( {U_{\mathrm{r}}\theta } \right) \odot \left( {U_{\mathrm{r}}\theta } \right)} \right),\end{array}$$

where F and Yr are $$(N - P) \times V$$ matrices, where each column of Yr is $$Y_{\mathrm{r}}^ \ast$$ (Eq. (7)) for one image element and F is the element-wise squaring of Yr; $$\odot$$ denotes Hadamard product; $$\oslash$$ denotes element-wise division; θ is the 2 × V matrix of OLS solutions which is matrix counterpart of $$\hat \theta _{{\mathrm{REML,OLS}}}$$; 0 is the 2 × V matrix of zeros; and here $${\mathrm{max}}( \cdot , \cdot )$$ is an element-wise maximum between the two operands, evaluating to a 2 × V matrix; W and 1NV are the (NP) × V matrices, where each column of W is $${\mathrm{diag}}\left( {\hat \Sigma _{{\mathrm{OLS,r}}}^{ \ast - 2}} \right)$$ for the corresponding image element and 1NV is a matrix of ones. With the following notation,

$$\begin{array}{l}{\hskip-34pt}A = {\mathbf{1}}_V W,\\ {\hskip-34pt}B = D{\prime}_{\mathrm {g}} W,\\ C = \left( {D{\prime}_{\mathrm {g}} \odot \lambda {\prime}_{{\mathrm{g}}_{\mathrm{r}}} } \right)W,\\ {\hskip-12pt}D = {\mathbf{1}}_V \left( {W \odot F} \right),\\ {\hskip-13pt}E = D{\prime}_{\mathrm {g}} \left( {W \odot F} \right),\end{array}$$

where 1V is the length-V column vector of ones, we can compute the variance components of the vectorized image as

$${\mathbf{\sigma }}_{\mathrm{A}}^2 = {\mathrm{max}}\left( {\left( { - B \odot D + A \odot E} \right) \oslash \left( {A \odot C - B \odot B} \right),{\mathbf{0}}} \right),$$
(14)
$${\mathbf{\sigma }}_{\mathrm{E}}^2 = {\mathrm{max}}\left( {\left( {C \odot D - B \odot E} \right) \oslash \left( {A \odot C - B \odot B} \right),{\mathbf{0}}} \right),$$
(15)
$${\mathbf{S}} = U_{\mathrm{r}}\left[ {\begin{array}{*{20}{c}} {{\mathbf{1}}_V \oslash {\mathbf{\sigma }}_{\mathrm{E}}^2} \\ {{\mathbf{1}}_V \oslash {\mathbf{\sigma }}_{\mathrm{A}}^2} \end{array}} \right],$$

where $${\mathbf{\sigma }}_{\mathrm{A}}^2$$ and $${\mathbf{\sigma }}_{\mathrm{E}}^2$$ are the length-V column vectors of genetic and environmental variance components, respectively; and S is a (N − P) × V matrix which here each column of S is the element-wise reciprocal of the diagonal of the variance matrix of the corresponding image element’s data Yr for each element of image. In this fashion, the score statistic matrix for all markers being tested and the vectorized image can be expressed as:

$${\mathbf{T}}_{{\mathrm{S,R}}} = \left[ {\left( {{\mathbf{X}}\prime \left( {{\mathbf{S}} \odot {\mathrm{Y}}_{\mathrm{r}}} \right)} \right) \odot \left( {{\mathbf{X}}\prime \left( {{\mathbf{S}} \odot {\mathrm{Y}}_{\mathrm{r}}} \right)} \right)} \right] \oslash \left[ {\left( {{\mathbf{X}} \odot {\mathbf{X}}} \right){\mathbf{S}}} \right],$$
(16)

where TS,R is a G × V matrix of score statistics for all SNPs and traits.

Real data

To validate our proposed methods for association estimation and inference for imaging data, we applied them on a dataset from healthy and schizophrenic individuals to perform ROI and voxel-wise genome wide association analysis using cluster wise inference. The sample was 54% healthy individual (175 control/155 schizophrenic) and had a mean age of 39.12 (SD = 14.9) where 50% of the sample is male54.

Diffusion tensor imaging

Imaging data was collected using a Siemens 3T Allegra MRI (Erlangen, Germany) using a spin-echo, EPI sequence with a spatial resolution of 1.7 × 1.7 × 4.0 mm. The sequence parameters were: TE/TR = 87/5000 ms, FOV = 200 mm, axial slice orientation with 35 slices and no gaps, 12 isotropically distributed diffusion weighted directions, two diffusion weighting values (b = 0 and 1000 s/mm2), the entire protocol repeated three times.

ENIGMA-DTI protocols for extraction of tract-wise average FA values were used. These protocols are detailed elsewhere55 and are available online http://enigma.ini.usc.edu/protocols/dti-protocols/. Briefly, FA images from subjects were non-linearly registered to the ENIGMA–DTI target brain using FSL’s FNIRT55. This target was created as a “minimal de-formation target” based on images from the participating studies as previously described (Jahanshad et al., 2013b). The data were then processed using FSL’s tract-based spatial statistics (TBSS; http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/TBSS) analytic method56 modified to project individual FA values on the hand-segmented ENIGMA-DTI skeleton mask. The protocol, target brain, ENIGMA–DTI skeleton mask, source code and executables, are all publicly available (http://enigma.ini.usc.edu/ongoing/dti-working-group/). The FA values are normalized across individuals by inverse Gaussian transform57,58 to ensure normality assumption. Finally, we analyzed the voxel and cluster-wise FA values with applying along the ENIGMA skeleton mask.

Genetic quality control

In this study only genotyped single nucleotide polymorphisms (SNPs) from genome-wide information were included in the analysis. Visual inspection of multi-dimensional scaling analysis was used to extract individuals with European ancestry. SNPs from individuals with European ancestry that do not meet any of the following quality criteria were excluded: genotype call rate 95%, significant deviation from Hardy–Weinberg equilibrium P < 10−6 and minor allele frequency 0.1 was used to ensure that sufficient numbers of subjects would be found in our sample in each genotypic group (homozygous major allele, heterozygous, homozygous minor allele) using an additive genetic model. After all quality control steps, 962,885 out of 1,406,990 SNPs remain for genome-wide association analysis.

Data availability

The data that support the findings of this study are available from Dr. Peter Kochunov upon reasonable request.

Code availability

Software implementation of this method, Nonparametric Inference for Genetics Analysis (NINGA), is available at http://nisox.org/software/ninga/.

Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Change history

• 04 September 2018

This Article was originally published without the accompanying Peer Review File. This file is now available in the HTML version of the Article; the PDF was correct from the time of publication.

References

1. 1.

Hibar, D. P. et al. Common genetic variants influence human subcortical brain structures. Nature 520, 224–229 (2015). 04.

2. 2.

Stein, J. L. et al. Identification of common variants associated with human hippocampal and intracranial volumes. Nat. Genet. 44, 552–561 (2012). 05.

3. 3.

Stein, J. L. et al. Genome-wide analysis reveals novel genes influencing temporal lobe structure with relevance to neurodegeneration in alzheimer’s disease. Neuroimage 51, 542–554 (2010).

4. 4.

Stein, J. L. et al. Voxelwise genome-wide association study (vgwas). Neuroimage 53, 1160–1174 (2010). Imaging Genetics.

5. 5.

Potkin, S. G. et al. Hippocampal atrophy as a quantitative trait in a genome-wide association study identifying novel susceptibility genes for alzheimer’s disease. PLoS One 4, 1–15 (2009). 08.

6. 6.

Potkin, S. G. et al. A genome-wide association study of schizophrenia using brain activation as a quantitative phenotype. Schizophr. Bull. 35, 96–108 (2009).

7. 7.

Voight, B. F. & Pritchard, J. K. Confounding from cryptic relatedness in case-control association studies. PLoS Genet. 1, 09 (2005).

8. 8.

Weir, B. S., Anderson, A. D. & Hepler, A. B. Genetic relatedness analysis: modern data and new challenges. Nat. Rev. Genet. 7, 771–780 (2006).

9. 9.

Pritchard, J. K., Stephens, M., Rosenberg, N. A. & Donnelly, P. Association mapping in structured populations. Am. J. Hum. Genet. 67, 170–181 (2000).

10. 10.

Cardon, L. R. & Palmer, L. J. Population stratification and spurious allelic association. Lancet 361, 598–604 (2003).

11. 11.

Helgason, A., Yngvadóttir, B., Hrafnkelsson, B., Gulcher, J. & Stefánsson, K. An icelandic example of the impact of population structure on association studies. Nat. Genet. 37, 90–95 (2005).

12. 12.

Balding, D. A tutorial on statistical methods for population association studies. Nat. Rev. Genet. 7, 781–791 (2006).

13. 13.

Price, A. L., Zaitlen, N. A., Reich, D. & Patterson, N. New approaches to population stratification in genome-wide association studies. Nat. Rev. 11, 459–463 (2010).

14. 14.

Price, A. L. et al. Principal components analysis corrects for stratification in genome-wide association studies. Nat. Genet. 38, 904–909 (2006).

15. 15.

Yu, J. et al. A unified mixed-model method for association mapping that accounts for multiple levels of relatedness. Nat. Genet. 38, 203–208 (2006).

16. 16.

Kang, H. M. et al. Efficient control of population structure in model organism association mapping. Genetics 178, 1709–1723 (2008).

17. 17.

Zhang, Z. et al. Mixed linear model approach adapted for genome-wide association studies. Nat. Genet. 42, 355–360 (2010).

18. 18.

Kang, H. M. et al. Variance component model to account for sample structure in genome-wide association studies. Nat. Genet. 42, 348–354 (2010).

19. 19.

Lippert, C. et al. FaST linear mixed models for genome-wide association studies. Nat. Methods 8, 833–837 (2011).

20. 20.

Lippert, C. et al. Improved linear mixed models for genome-wide association studies. Nat. Methods 8, 833–835 (2011).

21. 21.

Zhou, X. & Stephens, M. Genome-wide efficient mixed-model analysis for association studies. Nat. Genet. 44, 821–824 (2012).

22. 22.

Svishcheva, G. R., Axenovich, T. I., Belonogova, N. M., Duijn, C. Mvan & Aulchenko, Y. S. Rapid variance components–based method for whole-genome association analysis. Nat. Genet. 44, 1166–1170 (2012).

23. 23.

Pirinen, M., Donnelly, P. & Spencer, C. C. A. Efficient computation with a linear mixed model on large-scale data sets with applications to genetic studies. Ann. Appl. Stat. 7, 369–390 (2013).

24. 24.

Listgarten, J., Lippert, C. & Heckerman, D. FaST-LMM-Select for addressing confounding from spatial structure and rare variants. Nat. Genet. 45, 470–471 (2013).

25. 25.

Widmer, C. et al. Further improvements to linear mixed models for genome-wide association studies. Sci. Rep. 4, 6874 (2014).

26. 26.

Kadri, N. K., Guldbrandtsen, B., Sørensen, P. & Sahana, G. Comparison of genome-wide association methods in analyses of admixed populations with complex familial relationships. PLoS One 9, e88926 (2014).

27. 27.

Yang, J., Zaitlen, N. A., Goddard, M. E., Visscher, P. M. & Price, A. L. Advantages and pitfalls in the application of mixed-model association methods. Nat. Genet. 46, 100–106 (2014).

28. 28.

Friston, K. J., Worsley, K. J., Frackowiak, R. S. J., Mazziotta, J. C. & Evans, A. C. Assessing the significance of focal activations using their spatial extent. Hum. Brain Mapp. 1, 210–220 (1994).

29. 29.

Smith, S. & Nichols, T. Threshold-free cluster enhancement: addressing problems of smoothing, threshold dependence and localisation in cluster inference. Neuroimage 44, 83–98 (2009).

30. 30.

Eklund, A., Nichols, T. E. & Knutsson, H. Cluster failure: Why fmri inferences for spatial extent have inflated false-positive rates. Proc. Natl Acad. Sci. 113, 7900–7905 (2016).

31. 31.

Ge, T., Feng, J., Hibar, D. P., Thompson, P. M. & Nichols, T. E. Increasing power for voxel-wise genome-wide association studies: The random field theory, least square kernel machines and fast permutation procedures. Neuroimage 63, 858–873 (2012).

32. 32.

Nichols, T. & Hayasaka, S. Controlling the familywise error rate in functional neuroimaging: a comparative review. Stat. Methods Med. Res. 12, 419–446 (2003).

33. 33.

Nichols, T. E. & Holmes, A. P. Nonparametric permutation tests for functional neuroimaging: a primer with examples. Hum. Brain Mapp. 15, 1–25 (2002).

34. 34.

Ganjgahi, H. et al. Fast and powerful heritability inference for family-based neuroimaging studies. Neuroimage 115, 256–268 (2015).

35. 35.

Lippert, C. et al. Greater power and computational efficiency for kernel-based association testing of sets of genetic variants. Bioinformatics 30, 3206–3214 (2014).

36. 36.

Glahn, D. C., Thompson, P. M., & Blangero, J. Neuroimaging endophenotypes: strategies for finding genes influencing brain structure and function. Hum. Brain Mapp. 28:488–501 (2007).

37. 37.

Peper, J. S., Brouwer, R. M., Boomsma, D. I., Kahn, R. S. & Pol, H. E. H. Genetic influences on human brain structure: a review of brain imaging studies in twins. Hum. Brain Mapp. 28, 464–473 (2007).

38. 38.

Li, M.-X., Yeung, J. M., Cherny, S. S. & Sham, P. C. Evaluating the effective numbers of independent tests and significant p-value thresholds in commercial genotyping arrays and public imputation reference datasets. Hum. Genet. 131, 747–756 (2012).

39. 39.

Lange, K., Westlake, J. & Spence, M. A. Extensions to pedigree analysis iii. variance components by the scoring method. Ann. Hum. Genet. 39, 485–491 (1976).

40. 40.

Hopper, J. L. & Mathews, J. D. Extensions to multivariate normal models for pedigree analysis. Ann. Hum. Genet. 46, 373–383 (1982).

41. 41.

Hasstedt, S. J. A mixed-model likelihood approximation on large pedigrees. Comput. Biomed. Res. 15, 295–307 (1982).

42. 42.

Boerwinkle, E., Chakraborty, R. & Sing, C. F. The use of measured genotype information in the analysis of quantitative phenotypes in man. Ann. Hum. Genet. 50, 181–194 (1986).

43. 43.

Almasy, L. & Blangero, J. Multipoint quantitative-trait linkage analysis in general pedigrees. Am. J. Hum. Genet. 62, 1198–1211 (1998).

44. 44.

Elliott, L., et al. The genetic basis of human brain structure and function: 1,262 genome-wide associations found from 3,144 gwas of multimodal brain imaging phenotypes from 9,707 uk biobank participants. Preprint at bioRxiv doi: https://doi.org/10.1101/178806 (2017).

45. 45.

Abney, M. Permutation testing in the presence of polygenic variation. Genet. Epidemiol. 39, 249–258 (2015).

46. 46.

Blangero, J., et al. A kernel of truth: statistical advances in polygenic variance component models for complex human pedigrees. Eds. Theodore, F., Jay, C. D., Stephen F. G. In Advances in Genetics, Vol. 81, 1–31 (Elsevier, New York, 2013).

47. 47.

Searle, S. R., Casella, G., & McCulloch, C. E. Variance Components, Vol. 391 (Wiley, Hoboken, New Jersey, 2009).

48. 48.

Amemiya, T. A note on a heteroscedastic model. J. Econom. 6, 365–370 (1977).

49. 49.

Rao, C. R. Linear Statistical Inference and its Applications (Wiley, Hoboken, New Jersey, 2008).

50. 50.

Neyman, J., & Pearson, E. S. On the problem of the most efficient tests of statistical hypothesesPhilos. Trans. R. Soc. Lond. 231, 289–337 (1933).

51. 51.

Winkler, A. M., Ridgway, G. R., Webster, M. A., Smith, S. M. & Nichols, T. E. Permutation inference for the general linear model. Neuroimage 92C, 381–397 (2014).

52. 52.

Cheng, R. & Palmer, A. A. A simulation study of permutation, bootstrap, and gene dropping for assessing statistical significance in the case of unequal relatedness. Genetics 193, 1015–1018 (2013).

53. 53.

Fabregat-Traver, D. & Bientinesi, P. Computing petaflops over terabytes of data: the case of genome-wide association studies. ACM Trans. Math. Softw. 40, 1–27 (2014). 22.

54. 54.

Kochunov, P. et al. Heterochronicity of white matter development and aging explains regional patient control differences in schizophrenia. Hum. Brain Mapp. 37, 4673–4688 (2016).

55. 55.

Jahanshad, N. et al. Multi-site genetic analysis of diffusion images and voxelwise heritability analysis: a pilot project of the enigma dti working group. Neuroimage 81, 455–469 (2013).

56. 56.

Smith, S. M. et al. Tract-based spatial statistics: voxelwise analysis of multi-subject diffusion data. Neuroimage 31, 1487–1505 (2006).

57. 57.

Servin, B. & Stephens, M. Imputation-based analysis of association studies: Candidate regions and quantitative traits. PLoS Genet. 3, 07 (2007).

58. 58.

Allison, D. B. et al. Testing the robustness of the likelihood-ratio test in a variance-component quantitative-trait loci-mapping procedure. Am. J. Hum. Genet. 65, 531–544 (1999).

59. 59.

Clopper, C. J. & Pearson, E. S. The use of confidence or fiducial limits illustrated in the case of the binomial. Biometrika 26, 404–413 (1934).

Acknowledgements

T.E.N. is supported by the Wellcome Trust. H.G, T.E.N. and P.K were supported by National Institutes of Health R01 EB015611-01.

Author information

Affiliations

1. Department of Statistics, University of Oxford, Oxford, UK

• Habib Ganjgahi
2. Medical Research Council Harwell Institute, Harwell, UK

• Habib Ganjgahi
3. Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK

• Anderson M. Winkler
•  & Thomas E. Nichols
4. Big Data Analytics Group, Hospital Israelita Albert Einstein, São Paulo, SP, Brazil

• Anderson M. Winkler
5. Department of Psychiatry, Yale University School of Medicine, New Haven, CT, USA

• David C. Glahn
6. South Texas Diabetes and Obesity Institute, University of Texas Rio Grande Valley School of Medicine, Brownsville, TX, USA

• John Blangero
7. Maryland Psychiatric Research Center, Department of Psychiatry, University of Maryland School of Medicine, Baltimore, MD, USA

• Brian Donohue
•  & Peter Kochunov
8. Oxford Big Data Institute, Li Ka Shing Centre for Health Information and Discovery, Nuffield Department of Population Health, University of Oxford, Oxford, UK

• Thomas E. Nichols
9. Department of Statistics, University of Warwick, Coventry, UK

• Thomas E. Nichols

Contributions

H.G. developed the statistical method, ran the simulations, analyzed the simulated and real data, and lead the writing of the manuscript. P.K. provided the human data and supervised its analysis. B.D. implemented the method both in CPU and GPU. A.M.W., D.C.G., J.B. and T.E.N. contributed to revisions of the paper. T.E.N. supervised the project. All authors contributed to critical review of the manuscript during its preparation.

Competing interests

The authors declare no competing interests.

Corresponding author

Correspondence to Thomas E. Nichols.