Our strategy builds from the above observations and estimates the ITV and ITC to detect latent genetic interactions. To derive estimates of these quantities, we first remove the additive genetic effect from the traits to ensure that any variance and covariance effects are not due to the additive effect. Let us denote the trait residuals as where we assume the effect size is known for simplicity. We can then express the ITV and ITC as a function of these residuals: the ITV of trait and the ITC between traits and is defined as and , respectively (Appendix 5.1.1). Thus, we can estimate the ITV by squaring the residuals, , and estimate the ITC between traits and by the pairwise product of the residuals (i.e., the cross products), . Aggregating the ITV and ITC estimates across all individuals, we denote the cross product (CP) terms in the matrix where the th row vector is , and the squared residual (SQ) terms in the matrix where .
Our inference goal is to assess whether the SNP, , is independent of the squared residuals and cross products,
where ‘.’ denotes all the rows (or individuals) and ‘’ denotes statistical independence. In the above regression model, this corresponds to testing the global null hypothesis versus the alternative hypothesis for at least one of the traits. While a regression model can be directly applied to the squared residuals and cross products to test the global null hypothesis (see Appendix for mathematical details), a univariate model approach does not adequately leverage pleiotropy and requires a multiple testing correction which reduces power.
To address these issues, we develop a new multivariate kernel-based framework, Latent Interaction Testing (LIT), that captures pleiotropy across the ITV and ITC terms to increase power for detecting latent interactions. There are three key steps in the LIT framework (Figure 1):
We expand on the above steps in detail below.
Step 1: In the first step, LIT standardizes the traits and then regresses out the additive genetic effects, population structure, and any other covariates. This ensures that any differential variance and/or covariance patterns are not due to additive genetic effects or population structure. Suppose there are measured covariates and principal components to control for structure. We denote these variables in the matrix . After regressing out these variables and the additive genetic effects, the matrix of residuals is , where is the standardized trait matrix, is a matrix of effect sizes and is a matrix of coefficients estimated using least squares. We also regress out population structure from the genotypes which we denote by .
The above approach only removes the mean effects and does not correct for variance effects from population structure which can impact type I error rate control [57]. A strategy to adjust for the variance effects is to standardize the genotypes with the estimated individual-specific allele frequencies (IAF), i.e., the allele frequencies given the genetic ancestry of an individual. However, it is computationally costly to standardize the genotypes for biobank-sized datasets as it requires estimating the IAFs of all SNPs using a generalized linear model [58, 59]. Therefore, in this work, we remove the mean effects from structure and then adjust the test statistics with the genomic inflation factor to be conservative. Our software includes an implementation to standardize the genotypes using the IAFs for smaller datasets.
Step 2: The second step uses the residuals, , to reveal any latent interactions by constructing estimates of the ITV and ITC. For the th individual’s set of trait residuals, the ITVs are estimated by squaring the trait residuals while the ITCs are estimated by calculating the cross products of the trait residuals. We express the squared residuals as , and the pairwise cross products as . Importantly, when the studentized residuals are used, then and represent an unbiased estimate of the ITVs and ITCs, respectively. We aggregate these terms across all individuals into the matrix .
Step 3: In the last step, we test for association between the adjusted SNP and the squared residuals and cross products (SQ/CP) using a kernel-based distance covariance framework [31–33]. Specifically, we apply a kernel-based independence test called the Hilbert-Schmidt independence criterion (HSIC), which has been previously used for GWAS data (see, e.g., [35–38]). The HSIC constructs two similarity matrices between individuals using the SQ/CP matrix and genotype matrix, then calculates a test statistic that measures any shared signal between these similarity matrices. To estimate the similarity matrix, a kernel function is specified that captures the similitude between the th and th individual.
Since our primary application is biobank-sized data, we use a linear kernel so that LIT is computationally efficient. The linear similarity matrix is defined as for the genotype matrix and for the SQ/CP matrix. The linear kernel is a scaled version of the covariance matrix and, for this special case, the HSIC is related to the RV coefficient. We note that one can choose other options for a kernel function, such as a polynomial kernel, projection kernel, and a Gaussian radial-basis function that can capture non-linear relationships [34, 35].
Once the similarity matrices and are constructed, we can express the HSIC test statistic as
which follows a weighted sum of Chi-squared random variables under the null hypothesis, i.e., where and are the ordered non-zero eigenvalues of the respective matrices and . Intuitively, the test statistic measures the ‘overlap’ between two random matrices where large values of imply the two matrices are similar (i.e., a latent genetic interactive effect) while small values of imply no evidence of similarity (i.e., no latent genetic interactive effects). We can approximate the null distribution of using Davies’ method, which is computationally fast and accurate for large [35, 38, 60].
For the linear kernel considered here, we implement a simple strategy to substantially improve the computational speed of LIT. We first calculate the eigenvectors and eigenvalues of the SQ/CP and genotype matrices to construct the test statistic. Since the number of traits, , is much smaller than the sample size, , we can perform a singular value decomposition to estimate the subset of eigenvectors and eigenvalues in a computationally efficient manner [61–63]. This allows us to circumvent direct calculation and storage of large similarity matrices. Let and be the singular value decomposition (SVD) of the similarity matrices where the matrix is a diagonal matrix of eigenvalues and is a matrix of eigenvectors of the respective kernel matrices. We can then express the test statistic in terms of the SVD components as , where is the outer product between the two eigenvectors. Thus, for a single SNP, the test statistic is , where is the rank of the SQ/CP matrix and is the rank of the genotype matrix such that .
We explore an important aspect of the test statistic in Equation 5, namely, the role of the eigenvalues in determining statistical significance. The above equations suggest that the eigenvalues of the kernel matrices are emphasizing the eigenvectors that explain the most variation in the test statistic. While this may be reasonable in some settings, the interaction signal can be captured by eigenvectors that explain the least variation and this can be very difficult to ascertain beforehand [40]. In this case, the testing procedure will be underpowered. Thus, we also consider weighting the eigenvectors equally in LIT, i.e., , where are the eigenvalues of the outer product matrix. In this work, we implement a linear kernel (scaled covariance matrix) and so, in this special case, weighting the eigenvectors equally is equivalent to the projection kernel.
In summary, there are two implementations of the LIT framework. The residuals are first transformed to calculate the SQ and CP to reveal any latent interactive effects. We then calculate the weighted and unweighted eigenvectors in the test statistic which we refer to as weighted LIT (wLIT) and unweighted LIT (uLIT), respectively. We also apply a Cauchy combination test (CCT) [41] to combine the -values from the LIT implementations to maximize the number of discoveries and hedge for various (unknown) settings where one implementation may outperform the other. More specifically, let denote the value for the implementations. In this case, the CCT statistic is , where is a mathematical constant. A corresponding -value is then calculated using the standard Cauchy distribution. Importantly, when applying genome-wide significance levels, the CCT -value provides control of the type I error rate under arbitrary dependence structures. In the Results section, we refer to the CCT -value as aggregate LIT (aLIT).
We can extend LIT to assess latent interactions within a genetic region (e.g., a gene) consisting of multiple SNPs. In the first step, we regress out the joint additive effects from the multiple SNPs along with any other covariates and population structure. In the second step, we calculate the squared residuals and cross products using the corresponding residual matrix. Finally, in the last step, we construct the similarity matrices and perform inference using the HSIC: the linear similarity matrix for the genotype matrix is and our test statistic is where is the rank of the genotype matrix.
Compared to the previous section, this extended version of LIT is a region-based test for interactive effects instead of a SNP-by-SNP test. A region-based test is advantageous to reduce the number of tests compared to a SNP-by-SNP approach. However, in this work, we demonstrate LIT on SNP-by-SNP genome-wide scan to demonstrate the scalability.
Do you have any questions about this protocol?
Post your question to gather feedback from the community. We will also invite the authors of this article to respond.