Resampling-based tests for Lasso in genome-wide association studies

Genome-wide association studies involve detecting association between millions of genetic variants and a trait, which typically use univariate regression to test association between each single variant and the phenotype. Alternatively, Lasso penalized regression allows one to jointly model the relationship between all genetic variants and the phenotype. However, it is unclear how to best conduct inference on the individual Lasso coefficients, especially in high-dimensional settings. We consider six methods for testing the Lasso coefficients: two permutation (Lasso-Ayers, Lasso-PL) and one analytic approach (Lasso-AL) to select the penalty parameter for type-1-error control, residual bootstrap (Lasso-RB), modified residual bootstrap (Lasso-MRB), and a permutation test (Lasso-PT). Methods are compared via simulations and application to the Minnesota Center for Twins and Family Study. We show that for finite sample sizes with increasing number of null predictors, Lasso-RB, Lasso-MRB, and Lasso-PT fail to be viable methods of inference. However, Lasso-PL and Lasso-AL remain fast and powerful tools for conducting inference with the Lasso, even in high-dimensions. Our results suggest that the proposed permutation selection procedure (Lasso-PL) and the analytic selection method (Lasso-AL) are fast and powerful alternatives to the standard univariate analysis in genome-wide association studies.


Background
Genome-wide association studies (GWASs) involve studying association between millions of genetic variants, called "single nucleotide polymorphisms (SNPs)" and different traits of interest. Essentially, a GWAS can be viewed as a high-dimensional variable selection problem with the goal of finding SNPs that are significantly associated with a phenotype of interest. GWASs have predominantly been analyzed using univariate regression, i.e. "single marker association" methods (SMA), where one analyzes the marginal effect between an individual SNP and the phenotype, while ignoring the influence of other SNPs. Assuming that the phenotype is affected by multiple SNPs, SMA neglects useful information *Correspondence: saonli@umn.edu 1 Division of Biostatistics, School of Public Health, University of Minnesota, 55455 Minneapolis, USA Full list of author information is available at the end of the article regarding the structure of genetic association and hence may lose power to detect relevant SNPs.
Alternatively, jointly modeling all SNPs may lead to more accurate inference due to decreased residual variance in the phenotype of interest. However, given that the number of SNPs greatly exceeds the sample size, standard multiple linear regression techniques are no longer viable. In contrast, penalized regression may be used to jointly estimate regression coefficients in such settings. However, developing valid methods for conducting inference on the penalized regression coefficients remains an open area of research.
Several papers have used penalized regression in GWASs to perform variable selection, without conducting any rigorous inference on the selected variables. For example, Waldmann et al. [1] applied penalized regression to a GWAS by using cross validation to select tuning parameters, then compute the true positive and false positive rates simply based on whether or not a coefficient is nonzero. Cross Validation attempts to optimize the predictive performance of the model, but gives no control on the type-I error rate. Furthermore, both the software PUMA [2] and LOO indices of Wu et al. [3] use penalized regression to perform variable selection, then fit a standard multiple linear regression model using only those variables selected by the initial penalized model. Their software return p-values for individual non-zero regression coefficients using standard likelihood-based tests for multiple linear regression. Wu et al. [3] admit that these "pseudo p-values" are invalid because they neglect the complex selection procedure of obtaining the reduced model. Although these methods are very fast, they cannot warrant control of the type-I error rate.
One could potentially use resampling techniques to generate correct tests for the individual coefficients in the reduced model. Meinshausen [4] proposed a datasplitting approach to p-values with FDR or FWER control. However, sample splitting procedures may be quite conservative compared to methods that directly use the full data [5], likely because they lose power to detect SNPs with smaller effect sizes and minor allele frequencies. Chatterjee and Lahiri [6], and Sartori [7] used cross validation to select penalty parameters, then residual bootstrap confidence intervals (and modifications thereof ) to conduct inference on the regression coefficients. Both [6] and [7] focus on bootstrapping the Lasso in lowdimensional settings. Through simulations, we investigate the performance of the bootstrap and modified residual bootstrap when n is fixed and the number of null predictors p 0 → ∞.
All previously mentioned methods treat penalty parameter selection and inference as two separate problems. In contrast, Ayers and Cordell [8] proposed a permutation method to select penalty parameters for direct control of the type-1-error rate. If the goal is to detect the genetic variants that meet a certain threshold for the level of significance, Ayers and Cordell [8]'s technique gives us a computationally efficient way of identifying these variants compared to previously mentioned resampling techniques. Similarly, Yi et al. [9] proposed both a permutation and analytic method to select penalty parameters for false-discovery rate control. We propose a modified version of the permutation method of Ayers and Cordell, and compare it with the original method, as well as the analytic method of Yi et al.
Other recently proposed methods of inference for penalized regression not considered in this paper are as follows: Zhang [10] and Javanmard [11] use a "debiased" Lasso, which attempts to remove the bias in the Lasso coefficients, then constructs normal-based confidence intervals using the transformed coefficients. "Post-selection inference" or "selective inference" [12,13] is another recent development in inference for penalized regression. Selective inference is described as "the assessment of significance and effect sizes from a dataset after mining the same data to find these associations". For example, suppose one has used Lasso to "select" relevant predictors among a pool of many potentially relevant predictors. In particular, only predictors with nonzero estimated coefficients are considered for inference, all other predictors are dismissed. Selective inference addresses the question of how to conduct valid inference on the subset of selected predictors while accounting for the complex data-dependent procedure that selected those predictors in the first place. In this paper, we decided to focus mainly on resamplingbased methods of inference for the Lasso, and thus will not consider the de-biased Lasso or post-selection inferential methods.
Furthermore, one major limitation of existing methods of inference for penalized regression, as they're currently implemented in the statistical software R [14], is that they require one to store the entire genotype matrix in RAM, thus leading to great computational costs. In contrast, the methods we consider take advantage of the bigmemory R package [15] which allows one to work with highdimensional file-backed datasets that are larger than the available RAM. An important feature of the bigmemory package is the ability for multiple cores to share access to the same big.matrix object, without having to create an additional copy of the matrix in RAM at each core. Thus the bigmemory package allows for memory efficient parallel computing with high-dimensional matrices. In addition, the biglasso R package [16] allows one to fit Lasso penalized regression models using the bigmemory R package.
Through simulations, we compare six methods for testing the individual Lasso coefficients: two permutation and one analytic selection procedures for type-1-error control, residual bootstrap, modified residual bootstrap, and a permutation test. As a benchmark, all methods are compared to a standard single marker analysis. First we consider the scenario where the sample size (n) is fixed, and the number of null SNPs p 0 → ∞. We show that the bootstrap methods and permutation test become unstable as p 0 → ∞; however, both the permutation and analytic selection procedures appear to be powerful tools of inference, even in high-dimensional settings. Lastly, we apply the Lasso with both the permutation and analytic selection methods to the Minnesota Center for Twins and Family Study [17,18], using 3853 subjects and 507,541 SNPs.

Methods
Consider a GWAS with n subjects, p SNPs (p > n), and a quantitative phenotype Y Y Y . Without loss of generality, assume there are no additional covariates. Let x ij denote the number of minor alleles the ith subject carries at the jth SNP (i.e. x ij = 0, 1 or 2). Then standardize all SNPs to have a mean of zero and variance of one. We focus on the scenario where the proportion of null SNPs is large, i.e p 0 p ≈ 1, and the individual causal SNPs have relatively small effect sizes. Typically GWASs are analyzed using a "single marker analysis" (SMA), which tests the marginal effect of the j th SNP as follows: Then a t-test may be used to test the null hypothesis H 0 : β j = 0. The SMA applies model (1) to all j = 1, . . . , p SNPs and obtains p-values from a t-distribution, then declares a SNP significant if its p-value is less than or equal to some desired significance level α. In practice, one may adjust for multiple testing by controlling the family-wise error rate or false-discovery rate (see [19] for a review), but we chose not to address the issue of multiple testing in this paper. Rather the aim of this paper is to compare a standard SMA with several multi-marker methods in their ability to detect individual genetic variants that meet a desired level of significance.
As an alternative to SMA, we consider jointly modeling the main effects of all SNPs: Here we are interested in testing the conditional effect of the j th SNP, i.e. the effect of SNP j conditional on the effect of all other SNPs. Thus the following null hypothesis is of interest: However, given that p > n, standard multiple linear regression techniques are no longer viable. In contrast, Lasso [20] penalized regression allows one to estimate β β β in high-dimensional settings as follows: The subscript λ indicates that the Lasso estimatorβ β β λ depends on the penalty parameter λ which controls the rate of penalization in the estimated coefficients. In general, as λ increases, the Lasso coefficients are shrunk closer to zero. Typically Lasso yields a sparse solution of nonzero coefficients, and thus may be viewed as a variable selection tool. When there is a group of highly correlated predictors, the Lasso tends to only select one predictor in the group [21]. It is unclear if this property is disadvantageous for GWAS. If there is a group of highly correlated predictors that represents a causal region and Lasso only detects one SNP in that region; we have detected the region nontheless.
A followup analysis could easily find SNPs that are correlated with the detected SNP, if such SNPs would be of interest.
Lastly, most of the methods discussed in this paper can easily be extended to other sparse penalized regression models such as SCAD [22], MCP [23], Elastic Net [24], or Adaptive Lasso [25]. However, it is not the aim of this paper to compare different penalty types, thus we decided to focus only on the Lasso.

Permutations to select λ for type-1-error control
First we describe how to fit a Lasso model using a modified permutation method to select λ for control of the type-1-error rate. Define the following decision rule: whereβ λ,j is the estimated Lasso coefficient of the j th SNP for a given value of λ. Suppose there exists a value of λ, called λ α , that controls the type-1-error rate at level α under decision rule (5). Ayers and Cordell [8] showed that permutations can be used to estimate λ α . We propose a modified version of Ayers' method: 1. Permute Y Y Y to obtain "permuted dataset": {Y Y Y P , X X X}. Fit a Lasso model to {Y Y Y P , X X X}, and then record the value of λ that results in exactly s nonzero coefficients, where s p = α and p is the total number of predictors. Define this value of λ asλ α .
(a) Note that if α < 1 p , the above method won't work. In this case, suppose we let α = 1 p·k for some positive integer k. Then we can permute Y , k additional times (resulting in k Lasso models) and record the λ value in each of the k replicates that results in exactly 1 nonzero coefficient, giving , which effectively allows for only 1 nonzero coefficient out of the k · p total null coefficients, thus controlling the type-1-error rate at level α.
In either case above,λ α will control the type-1-error rate at approximately level α, however, we can reduce the mean squared error ofλ α as follows: 2. Repeat step (1) B times to obtain {λ α,i } B i=1 . Then define the final estimator of λ α as follows: 3. Finally, fit the Lasso model to the original data with λ =λ α for type-1-error control at approximately level α.
The main modification we propose to Ayers' method is to estimate λ α B times, then use the sample mean (or median),λ α as the final estimate of λ α . For highdimensional datasets, estimating λ α a single time (as does Ayers and Cordell [8]) may result in an unstable model in terms of variable selection (see Fig. 4 for more details). In contrast, as B increases, our estimatorλ α will select a stable model where the number of selected variables converges to some constant. Secondly, we propose two practical modifications to Ayer's method: 1) use the bisection algorithm of Wu et al. [3] to efficiently find the target λ α , and 2) use the bigmemory and biglasso R packages for memory-efficient parallel computing with high-dimensional matrices.
Yi et al. [9] proposed a similar permutation and analytic method to select λ for false-discovery rate control, which may easily be modified for control of the overall type-1-error rate [5]. Their methods require one to fit a grid of penalty parameter values, with the hope that at least one value on the grid achieves the desired error rate. In contrast, our permutation method will efficiently find the value of λ that gives the desired error rate. Through simulations, we compare Lasso using our modified permutation method to select λ for type-1-error control ("Lasso-PL") with the original Ayers' method ("Lasso-Ayers"), as well as the analytic method of Yi et al. ("Lasso-AL").
Lastly, we present theoretical justification for using the permutation method to select λ.
Theorem 1 Consider the linear model of the form Y Y Y = X X Xβ β β + , where X X X is an n × p matrix of p independent SNPs from n independent subjects, each standardized to have a mean of 0 and variance of 1; and ∼ N 0 0 0, σ 2 I I I n .
Consider the Lasso penalized regression model which can be written in the form of (7), with parameter λ that controls the overall rate of penalization. Then under decision rule (5),λ α (6) will control the type-1-error rate at approximately level α.
Proof Without loss of generality, suppose we wish to control the type-1-error rate at level α where α ≥ 1 p (the proof will have to be modified when α < 1 p , see step (1a) above for more info). Yi et al. [9] showed that β Lasso λ,j can be defined as follows: where λ ∈ (0, ∞), andβ OLS j is defined as:β The key result of (7) and (8) is thatβ Lasso Then permuting Y Y Y approximates the global null scenario: H 0 : β β β = 0 0 0, which gives the following result: where the asymptotic variance ofβ OLS Pr β OLS j > λ α = α. Then we can estimate λ α as follows: Defineλ α as the value of λ such that exactly α% of the penalizedβ Lasso λ,j 's are nonzero. 2. Then by Eqs. (7) and (8), α% of theβ Lassô sample quantile estimate of λ α . The asymptotic distribution of the sample quantile estimatorλ α is well known: where φ(·) is the pdf of a N(0, 1) random variable. Thusλ α is a consistent estimator for λ α , and will control the type-1-error rate at approximately level α for large p: The variability in our estimator can be reduced as follows: 3. Repeat step (1), B times to obtain λ α,b B b=1 and Finally,λ α will control the type-1-error rate at approximately level α for large p : The above theory holds when all SNPs are independent, but may be conservative given dependency among SNPs (see Breheny [5] for more details).
Lastly, we summarize the computational cost of fitting a Lasso-PL model to a real dataset with 3853 subjects and 507,451 SNPs, using a combination of the glmnet [26], bigmemory [15], and biglasso [16] R packages.
1. Use glmnet to find a reasonable starting window for λ α (≈ 2.5 min): for a given significance level α, this step is a one-time computational cost. It is necessary to find a reasonable window for the target value λ α , in order to speed up computation in step 3. Technically this step could be done with the biglasso R package, however, the algorithm we use to find a reasonable starting window for λ α took around 2.5 min with glmnet and over 1 h with biglasso. Thus we recommend using glmnet for step 1. 2. Create the big.matrix object (≈ 10 min): creating a file-backed big.matrix object is a one-time computational cost. In future R sessions, one can instantaneously reload the big.matrix object without any overhead. 3. Fit a biglasso model while estimating λ α B times: parallel computing with a base-R matrix object would require one to create an additional copy of the matrix within each core, thus leading to high computational cost. In contrast, the bigmemory R package allows multiple cores to share access to a single copy of the dataset. Using 20 cores in parallel, this step took around 1 h for B = 100.

The residual bootstrap and modified residual bootstrap
For an introduction to the bootstrap, see Efron [27] and Hesterberg [28]. Chatterjee and Lahiri [29] proved that the standard residual bootstrap approximation to the Lasso distribution may be inconsistent whenever one or more components of the regression parameter vector are zero. Chatterjee and Lahiri [6] proposed a "modified residual bootstrap" for the Lasso as an attempt to overcome this problem. However, in low-dimensional settings, Sartori [7] found that the residual bootstrap worked "acceptably well, " and that the modified residual bootstrap appeared to offer no significant improvement. We consider the performance of both the residual bootstrap and modified residual bootstrap when n is fixed and p 0 → ∞. The basic setup of the residual bootstrap ("Lasso-RB") is as follows: 1. Use 10-fold cross-validation to select λ, then keep λ fixed throughout all remaining steps 2. Fit a Lasso model of the form Y Y Y = X X Xβ β β + and obtain the Lasso estimateβ β β λ 3. Calculate the residuals Then center the residuals: The modified residual bootstrap ("Lasso-MRB") makes the following changes: step (3) uses the modified bootstrap residualsẽ ẽ e = Y Y Y −Ỹ Ỹ Y , whereỸ Ỹ Y = X X Xββ β λ , and β λ,j =β λ,j I |β λ,j | > τ , for a given threshold τ .

Permutation test p-values
Anderson and Legendre [30] provide an overview of permutation tests for multiple linear regression. We extend the permutation test of Manly [31] to the Lasso ("Lasso-PT") as follows: where α is the specified significance level.

Simulations
All simulated genotypes were generated with Vanderbilt's "GWA Simulator" program [32] using HapMap Illumina300k CEU phased data (Utah Residents with Northern and Western European Ancestry). The GWA simulator can simulate genotypes for case-control designs. We used the simulator to simulate genotypes on a set of markers under the null hypothesis of no association with the disease, and then simulated a quantitative trait on these individuals using linear regression with a set of genetic variants associated with the quantitative trait.
Only common variants with minor allele frequencies greater than 0.05 were simulated.

Simulation 1
In Simulation 1, we are interested in the scenario where the sample size (n) is fixed, the true causal effects are relatively small, and the number of null SNPs p 0 → ∞. Sixhundred subjects were simulated with five independent causal SNPs, each coming from a different chromosome. Without loss of generality, we assume there are no additional covariates. The quantitative trait Y Y Y was simulated as follows: Y Y Y 600×1 = X X X 600×5 β β β 5×1 + 600×1 , ∼ N(0 0 0, I I I) (13) After standardizing each SNP to have a mean of zero and variance of one, β β β = 0.15J J J 5 (where J J J 5 is a vector of ones) was chosen so that each causal SNP explains ≈ 2% of the variation in Y Y Y for a total R 2 ≈ 0.10. Three-hundred datasets were simulated according to (13). Lastly, we consider five different settings where the number of null SNPs varies from 0, 45, 300, 900, and 20,000. The null SNPs were simulated from a chromosome independent of the causal SNPs, and contain varying levels of correlation. We compare a standard single marker analysis to the Lasso using six different methods of inference on the Lasso coefficients: two permutation and one analytic methods to select λ for type-1-error control, residual bootstrap, modified residual bootstrap, and a permutation test. In particular, we are interested in comparing the performance of these methods as the number of null SNPs p 0 → ∞. Lastly, all methods are compared in terms of their true positive rate (TPR): average proportion of causal SNPs detected across the 300 simulated datasets; and false positve rate (FPR): average proportion of null SNPs detected across 300 simulated datasets.
Notice in Table 1 that only Lasso-PL and Lasso-AL consistently have power greater than or equal to the standard SMA. In general, in scenarios where there are not many SNPs, Lasso-AL has slightly greater TPR than Lasso-PL, presumably because Lasso-PL requires a large number of SNPs in order to get a good estimate of λ α (see proof of Theroem 1). However, in the 20,000 null SNP scenario, Lasso-AL is slightly conservative compared to Lasso-PL.
In our simulation study, Lasso-Ayers is consistently less powerful than Lasso-PL (especially in the 0 and 45 null SNP scenario), and has slightly inflated type-1-error in the 45 null SNP scenario. A Wilcoxon signed rank test to assess if the average false-positive rate (across 300 datasets) for Lasso-Ayers differs from the expected type-1-error rate α = 0.01 produced a p-value of 0.016. Because Lasso-Ayers only estimates λ α a single time, there may be high variability inλ α compared to Lasso-PL which estimates λ α B times then uses the sample mean (or median) as its final estimate (this can clearly be seen in Table 2). Thus on average, Lasso-Ayers is more prone to missing potential causal SNPs by over-estimating λ α , or underestimating λ α and having excess false positives. Lasso-PL appears to correct this by obtaining a more stable estimate of λ α , thus having increased power and better control of the type-1-error relative to Lasso-Ayers. Table 2 presents a comparison ofλ α from Lasso-Ayers, Lasso-PL, and Lasso-AL. Notice that Lasso-Ayers and Lasso-PL, on average, select the same value of λ; however, Lasso-PL significantly reduces the variability inλ α , and thus may provide more accurate inference. Although the variability inλ α may appear small for all methods, the scale is relative, for small changes in λ can result in drastically different number of nonzero coefficients, especially in high-dimensional settings. In the low dimensional settings, the three different methods produce similar values ofλ α on average, with Lasso-Ayers being the most variable. In the high-dimensional 20,000 null SNP scenario, Lasso-AL appears to be slightly over-penalizing relative to Lasso-PL and Lasso-Ayers.
When there is no null SNP, Lasso-RB and Lasso-MRB are the most powerful models; but when there are 45 null SNPs, all of the bootstrap methods have significantly inflated FPR. The reason the FPR is inflated may be due to the theoretical work of [6,29] which shows that the bootstrap may become unstable given one or more null predictors. In addition, notice as the number of null SNPs increases, the bootstrap methods become increasingly conservative, such that the power in the 20,000 null SNP scenario is significantly less than the other competing methods.
Notice Figs. 1 and 2 show that the bootstrap methods perfectly approximate the true Lasso distribution when there are no null SNPs. However, as p 0 increases, the bootstrap methods area no longer able to approximate the true Lasso distribution; thus hypothesis testing using the bootstrap may fail to control the type-1-error at the correct level (as seen in the 45 and 20,000 null SNP scenarios). In addition, notice that the modified residual bootstrap is unable to provide significantly better approximations to the true Lasso distribution, compared to the standard residual bootstrap in Fig. 1. We tried a range of threshold values for the Lasso-MRB: t = 0.001, 0.005, 0.01, 0.03, and 0.05. For values of t ≤ 0.001, the MRB performed nearly identical to the RB, and for t > 0.05 the MRB degenerates to a large point mass at zero. However, for all t ∈ (0.001, 0.05), the MRB does not appear to significantly improve upon the standard residual bootstrap. Similar to the bootstrap methods, the permutation test (Lasso-PT) becomes significantly conservative in the setting with 20,000 null SNPs. One reason why the bootstrap and permutation test perform so poorly as p 0 increases may be because using 10-fold CV to select λ in highdimensional settings is too stringent, resulting in overpenalization of the coefficients.
According to Table 3, as the number of null SNPs increases, the average λ selected by 10-fold CV increases and gets very close to the true effect size of the causal SNPs (0.15). Thus using 10-fold CV to select λ in high-dimensional settings may be too stringent, resulting in over-penalization of the coefficients, and thus lead to conservative results. In the 20,000 null SNP setting, even in the best case scenario where if every non-zero coefficient was declared significant, Lasso-PT and the bootstrap methods would still all have TPR less than 0.6 (which is much less than the competing methods). Thus clearly the 10-fold CV is over-penalizing the coefficients and making it harder to distinguish the causal SNPs from the null SNPs in this setting.
Given that the performance of Lasso-RB, Lasso-MRB, and Lasso-PT significantly degenerates as p 0 increases (especially in the 20,000 null SNP scenario), we decided to omit these methods from the remainder of our paper.

Simulation 2
Simulation 2 is similar to Simulation 1, except now we allow each causal SNP to be correlated with several neighboring SNPs. Specifically, each causal SNP was allowed ten neighboring SNPs in varying levels of linkagedisequilibrium (LD) with the causal SNP. Thus we have five causal regions or "LD blocks" of SNPs that are associated with the disease trait. Kruglyak [33] and Pritchard [34] defined "useful LD" as having r > 0.316 or r 2 > 0.1, where r is defined in [35]. Thus when picking the   Thus the goal is to "detect" a causal region by detecting at least one SNP that is in significant LD with the latent causal SNP of that region. In reality, the true causal SNPs are often not sequenced, thus our goal is to "tag" the true causal SNP by detecting SNPs in significant LD with the latent causal SNP. In each simulated dataset, there are 1000 total SNPs. A "null" SNP is defined as any SNP that is not in significant LD with any of the five latent causal SNPs. We defined a SNP as being in "significant LD" with a latent causal SNP if r ≥ τ , and considered two different values for τ : 0.3 and 0.5.
Models are compared by their true positive rate (TPR), linked true positive rate (LTPR), and false positive rate (FPR). The TPR represents the average proportion of the five causal regions that are detected across the threehundred simulated datasets. To detect a causal region, one must detect at least one SNP that is in significant LD with that region's latent causal SNP. The LTPR is defined as the average proportion of SNPs detected that are in significant LD with at least one latent causal SNP. For example, in a given dataset, suppose there are 20 SNPs that are in significant LD with at least one latent causal SNP; and suppose a model detects 10 of these SNPs as having significant association with the disease trait. This would result in an LTPR of 10 20 = 0.5. In contrast, suppose I detect at least 1 SNP in 3 of the 5 causal regions that is in significant LD with that region's latent causal SNP. This would result in a TPR of 3 5 = 0.6.
Lastly, FPR(τ ) represents the average proportion of null hypotheses that are falsely rejected, where a null SNP is defined as not being in significant LD (using cutoff τ ) with any latent causal SNP.
Notice in Table 4 that Lasso-PL, Lasso-Ayers, and Lasso-AL have higher TPR than SMA for both cutoffs τ = 0.3 and 0.5. Recall here that TPR measures the ability of a model to detect the true latent causal SNPs. Secondly, notice that SMA has significantly higher LTPR than the Lasso models. We should expect this because given a group of highly correlated SNPs that are also correlated with a latent causal SNP, Lasso tends to only select one SNP within the group, whereas SMA is likely to select multiple SNPs. However, given that the Lasso models have higher TPR than SMA implies that many of the SNPs SMA is picking up are redundant and not offering much additional information about the latent causal SNP.
For a given correlation threshold τ , the penalized regression models appear to do a better job of controlling the FPR at level α. This is probably because SMA is more likely to detect spurious SNPs that are weakly correlated with the latent causal SNPs.
Interestingly, there appears to be negligible difference between Lasso-PL and Lasso-Ayers in Simulation 2; whereas in Simulation 1, Lasso-PL was consistently more powerful and better controlled the type-1-error rate. However, given the results from Simulation 1 and our real data analysis, Lasso-PL appears to be the better method despite the increased computational cost.

Minnesota Center for Twins and Family Study (MCTFS)
The Minnesota Center for Twins and Family Study [17,18] contains genotype information on over 520,000 SNPs using Illumina's Human 660W Quad Array, with 8405 subjects clustered into 4-member families (each with 2 parents and 2 children). The families are categorized by sibling relationship type: MZ twins, DZ twins, full siblings, adopted siblings, and mixed siblings (one adopted, one biological). The overall goal of the study is to explore the genetic and environmental factors of substance abuse.
After quality control procedures, we focused on 3853 caucasian parents and 507,541 SNPs with MAF > 1%, HWE p-values > 10 −6 , and genotype call rates > 99% (see [18] for more details). Remaining missing genotypes were imputed using a combination of Beagle [36] and minimac [37], since existing penalized regression software cannot handle missing data.
We decided to focus on two quantitative clinical phenotypes created by [18], which were derived using Each dataset contains n = 600 subjects, 1000 SNPs and five causal regions. All testing was done using significance level α = 0.01. See Section "Simulation 2" for a definition of "causal region", TPR, LTPR, and FPR the hierarchical factor analytic approach of [38]. These 2 phenotypes of interest are: (1) Alcohol Consumption (composite of measures of alcohol use frequency and quantity); and (2) Non-Substance Behavioral Disinhibition (composite of measures non-substance use behavioral disinhibition including symptoms of conduct disorder and aggression). For each phenotype, we first fit a mixed linear model with covariates: Sex, Age, top 10 principle components, and a random intercept for Family ID. In this case, two spouses are given the same Family ID. The random intercept is included because spouses may exhibit correlated substance abuse behavior. The conditional residuals (that account for both fixed and random effects) from this fit were used as the new response for all subsequent genetic testing. Since Lasso-PL can only use significance levels of the form α = s p or 1 p * k for positive integers s and k, we used the significance level α = 6 507,541 ≈ 1.18 * 10 −5 for all tests. In practice, one would choose α apriori (e.g. 10 −5 ), then pick s or k to get as close to the desired α level as possible.
Lastly, for each phenotype, we compared Lasso-PL (with B = 100), Lasso-AL, and Lasso-Ayers to the standard single marker analysis (SMA). Results for the GWAS of Alcohol Consumption and Non-Substance Behavioral Disinhibition can be found in Fig. 3 and Tables 5-6.
Notice for the GWAS of Alcohol Consumption (Table 5) that SMA, Lasso-AL, and Lasso-Ayers detected only 4 SNPs, whereas Lasso-PL detected 7 SNPs. For the GWAS of Non-Substance Behavioral Disinhibition (Table 6), SMA detected 7 SNPs, while Lasso-PL detected 6 of the 7 SNPs found by SMA, plus 3 additional SNPs. Note the only SNP found by SMA that was not found by Lasso-PL (rs831750), is significantly correlated with SNP rs1007227 (r 2 = 0.788), which was detected by all models. However, Lasso-AL failed to tag SNP rs7314533 which was detected by all other models. Thus there appears to be slight evidence that Lasso-PL and Lasso-Ayers are more powerful than Lasso-AL here. Secondly, the few SNPs that Lasso-PL detected that SMA missed, were still borderline significant for SMA. Thus overall, the three Lasso models performed very similar to the standard SMA, with evidence that Lasso-PL may be slightly more powerful, and Lasso-AL may be conservative. As expected, if there are two highly correlated SNPs that are associated with the trait of interest, SMA is more likely to detect both SNPs compared to penalized regression; but Lasso-PL and Lasso-Ayers are still detecting the associated region nontheless. Fig. 3 Comparison of selected number of SNPs in a GWAS with two different quantitative traits: alcohol consumption (top) and non-substance behavioral disinhibition (bottom). A total of 3853 subjects were used with 507,541 SNPs. All testing used significance level α = 1.18 * 10 −5 . Venn Diagrams were created using [39] Tables 5 and 6: "S" means "significant" and "N" means "not significant" using significance level α = 1.18 * 10 −5 . Note that Lasso-PL, Lasso-AL, and Lasso-Ayers cannot provide exact p-values, but selects significant SNPs while attempting to control the type-1-error rate at level α. One could fit multiple penalized regression models and estimate λ that controls the type-1-error rate at various orders of magnitude (e.g. 10 −5 , 10 −6 , etc) to get a better idea of the significance of each selected SNP (not done here) A natural question is whether B = 100 permutations is sufficient for obtaining an accurate estimate of λ α in the Lasso-PL models. We created a diagnostic tool to assess whether or not B is large enough (see Fig. 4).
In order to indicate that we have obtained a stable model in terms of variable selection (i.e. thatλ α is an accurate estimate of λ α ), we need the number of selected SNPs to converge to some constant and the number of discrepant SNPs to converge to 0, as the number of permutations → ∞. In Fig. 4, notice as the number of permutations used in the estimatorλ α increases, the number of SNPs our model selects converges to 7; and the number of discrepancies in selected SNPs between models using B and B − 1 permutations converges to 0. Thus it appears that using at least 15 permutations is sufficient for obtaining a stable estimate of λ α in this application. Note if we had estimated λ α only a single time, as does Ayers and Cordell [8], we would have identified only four nonzero SNPs, thus missing the three potentially associated SNPs thatλ α identifies when B > 15. Overall, B = 100 was more than enough to ensure an accurate estimate of λ α for the two Lasso-PL models.
Lastly, a comparison of the computation time required to fit each model can be found in Table 7. Notice that all of the methods considered in this paper have reasonable computational costs for a realistic large-scale GWAS, with SMA and Lasso-Ayers being the fastest methods. Although Lasso-AL is an analytical method, it still requires one to fit a Lasso model using a grid of λ values, then estimates the type-1-error rate for each value of λ analytically, with the hope that at least one value of λ within the grid obtains an estimated type-1-error rate near the desired level α. This calculation cannot be done apriori because it requires estimation of σ 2 in Eq. 2 for each value of λ in the grid. We attempted to find a reasonable window for the target value of λ beforehand, then fit a grid of λ values within this window in order to reduce the computational cost. However, it may be possible to further reduce the computational time needed to fit Lasso-AL models by picking a more optimal grid. Lastly, notice the computation time for Lasso-Ayers with the Behavioral Disinhibition trait is almost twice the time needed for the Alcohol Consumption trait. This is because the bisection algorithm [3] we used to find the target value of λ (in both Lasso-Ayers and Lasso-PL models) sometimes has high variability in computation time. Thus it may be possible to further reduce the computation time needed to fit Lasso-Ayers or Lasso-PL by developing more efficient  algorithms for finding the target value of λ. However, our current implementation of these methods seems reasonable enough using modest computational resources.

Discussion
Penalized regression is a useful tool for GWASs that allows one to simultaneously test the relationship between hundreds of thousands of SNPs and a phenotype of interest with a single model. Unlike the standard single marker analysis, penalized regression jointly models all SNPs, and thus provides a more realistic model of the structure of genotype-phenotype association. Through simulations, we compared six methods for conducting inference on the individual Lasso coefficients: two permutation and one analytic approachs to select λ for type-1-error control, residual bootstrap, modified residual bootstrap, and permutation test p-values. Simulation 1 showed that for fixed sample size and increasing number of null SNPs, the bootstrap methods fail to approximate the true Lasso distribution. In addition, the modified residual bootstrap gave no significant advantage over the standard residual bootstrap. In the simulation scenario with 20,000 null SNPs, the bootstrap methods (Lasso-RB, Lasso-MRB) and permutation test (Lasso-PT) become significantly conservative relative to competing methods of inference. Therefore, we do not recommend using the residual bootstrap, modified residual bootstrap, or permutation test with the Lasso in high-dimensional settings. Throughout our simulations, we found using our modified permutation approach or an analytic method to select λ for type-1-error control (Lasso-PL and Lasso-AL) were often the most powerful models, with power consistently greater than or equal to the standard SMA. Unlike the bootstrap or permutation test, Lasso-PL and Lasso-AL performed consistently well, even in high-dimensional settings. As to which method, Lasso-PL or Lasso-AL is more powerful, our study gave mixed results. In Simulation 1, Lasso-AL was more powerful in low-dimensional settings, but Lasso-PL was more powerful in the 20,000 null SNP setting. In Simulation 2 and our real data analysis, Lasso-PL was more powerful. For the real data GWAS of Alcohol Consumption, Lasso-AL failed to tag a SNP detected by the standard SMA, while both Lasso-PL and Lasso-Ayers successfully tagged this SNP. Overall, both Lasso-PL and Lasso-AL were consistently competitive with the standard SMA, thus it seems either method could be recommended in practice.
Simulation 1 and the real data application gave evidence that our modified permutation method to select λ (Lasso-PL) performs better than the original method (Lasso-Ayers) proposed by Ayers and Cordell [8]. Lasso-Ayers uses permutations to estimate the value of λ that controls the type-1-error rate at the desired level α. They use permutations to estimate the target λ α only a single time, whereas Lasso-PL estimates λ α B times then uses the sample mean (or median) as its final estimate, thus reducing the variability inλ α . Simulation 1 showed that Lasso-PL is consistently more powerful than Lasso-Ayers, and does a better job of controlling the type-1-error rate. Table 2 showed that Lasso-PL significantly reduces the variability inλ α compared to Lasso-Ayers. In addition, Lasso-PL always detected more SNPs in the real data analysis. By using permutations to estimate λ α only a single time, Lasso-Ayers is more prone to missing potential causal SNPs by over-estimating λ α , or under-estimating λ α and having excess false-positives. Lasso-PL corrects this by obtaining a more stable estimate of λ α .
Another key difference between Lasso-Ayers and Lasso-PL, is that Lasso-PL should lead to more consistent results. For example, if multiple Lasso-Ayers models are fit to the same high-dimensional dataset, there may be high variability in the number of selected relevant SNPs between each model. Whereas if multiple Lasso-PL models are fit to the same dataset, they should all obtain the same results given a sufficient number of permutations.
The main downsides of Lasso-PL, Lasso-AL, and Lasso-Ayers is that they provide no confidence intervals or exact p-values for individual SNPs. We are still guaranteed that the subset of selected SNPs maintains approximate type-1-error control at level α, but do not know exactly "how significant" each selected SNP is. One could fit multiple Lasso-PL or Lasso-AL models and estimate λ that controls the type-1-error rate at various orders of magnitude (e.g. 10 −5 , 10 −6 , etc) to get a better idea of the significance of each selected SNP, however, this would greatly increase the computational cost. Nonetheless, if our goal is to identify the genetic variants that meet a pre-specified level of significance, then Lasso-PL and Lasso-AL are fast and powerful alternatives to the standard single marker analysis.
Lastly, throughout this paper the methods we use control the overall type-1-error rate. However, these methods can easily be modified for control of the family-wise-error rate at level α by using the significance level α * = α p where p is the total number of SNPs or estimated number of effective tests. For controlling the false-discovery rate with penalized regression, see [9].

Conclusion
Developing valid methods to test Lasso coefficients in high-dimensional settings remains a challenging area of research. Through simulations, we've shown that the residual bootstrap (Lasso-RB), modified residual bootstrap (Lasso-MRB), and permutation test (Lasso-PT) become practically intractable in high-dimensional settings (p >> n). However, our modified permutation method to select λ for type-1-error control (Lasso-PL) and the analytic method of Yi et al. [9] (Lasso-AL) nearly always outperformed the standard univariate analysis in both simulations and real data application. The bigmemory and biglasso R packages may be used to fit high-dimensional Lasso-PL or Lasso-AL models with memory-efficient parallel computing. For a real dataset with 3853 subjects and 507,451 SNPs, Lasso-PL with B = 100 permutations took around one hour using 20 cores in parallel, while Lasso-AL took less than 30 min with a single core. Therefore, we recommend Lasso-PL or Lasso-AL as fast and powerful alternatives to the standard single marker analysis in genome-wide association studies.