<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Genet.</journal-id>
<journal-title>Frontiers in Genetics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Genet.</abbrev-journal-title>
<issn pub-type="epub">1664-8021</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fgene.2013.00270</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Genetics</subject>
<subj-group>
<subject>Original Research Article</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Evaluation of the lasso and the elastic net in genome-wide association studies</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Waldmann</surname> <given-names>Patrik</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>M&#x000E9;sz&#x000E1;ros</surname> <given-names>G&#x000E1;bor</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x0002A;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Gredler</surname> <given-names>Birgit</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Fuerst</surname> <given-names>Christian</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>S&#x000F6;lkner</surname> <given-names>Johann</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Division of Livestock Sciences, Department of Sustainable Agricultural Systems, University of Natural Resources and Life Sciences</institution> <country>Vienna, Austria</country></aff>
<aff id="aff2"><sup>2</sup><institution>Division of Statistics, Department of Computer and Information Science, Link&#x000F6;ping University</institution> <country>Link&#x000F6;ping, Sweden</country></aff>
<aff id="aff3"><sup>3</sup><institution>Qualitas AG</institution> <country>Zug, Switzerland</country></aff>
<aff id="aff4"><sup>4</sup><institution>ZuchtData EDV-Dienstleistungen GmbH</institution> <country>Vienna, Austria</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Henner Simianer, Georg-August University Goettingen, Germany</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Henner Simianer, Georg-August University Goettingen, Germany; Valentin Wimmer, Technische Universit&#x000E4;t M&#x000FC;nchen, Germany</p></fn>
<fn fn-type="corresp" id="fn001"><p>&#x0002A;Correspondence: G&#x000E1;bor M&#x000E9;sz&#x000E1;ros, Division of Livestock Sciences, University of Natural Resources and Life Sciences, Gregor-Mendel-Stra&#x000DF;e 33, A-1180 Vienna, Austria e-mail: <email>gabor.meszaros&#x00040;boku.ac.at</email></p></fn>
<fn fn-type="other" id="fn002"><p>This article was submitted to Livestock Genomics, a section of the journal Frontiers in Genetics.</p></fn>
</author-notes>
<pub-date pub-type="epreprint">
<day>08</day>
<month>07</month>
<year>2013</year>
</pub-date>
<pub-date pub-type="epub">
<day>04</day>
<month>12</month>
<year>2013</year>
</pub-date>
<pub-date pub-type="collection">
<year>2013</year>
</pub-date>
<volume>4</volume>
<elocation-id>270</elocation-id>
<history>
<date date-type="received">
<day>19</day>
<month>06</month>
<year>2013</year>
</date>
<date date-type="accepted">
<day>18</day>
<month>11</month>
<year>2013</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2013 Waldmann, M&#x000E9;sz&#x000E1;ros, Gredler, Fuerst and S&#x000F6;lkner.</copyright-statement>
<copyright-year>2013</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/3.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract><p>The number of publications performing genome-wide association studies (GWAS) has increased dramatically. Penalized regression approaches have been developed to overcome the challenges caused by the high dimensional data, but these methods are relatively new in the GWAS field. In this study we have compared the statistical performance of two methods (the least absolute shrinkage and selection operator&#x02014;lasso and the elastic net) on two simulated data sets and one real data set from a 50 K genome-wide single nucleotide polymorphism (SNP) panel of 5570 Fleckvieh bulls. The first simulated data set displays moderate to high linkage disequilibrium between SNPs, whereas the second simulated data set from the QTLMAS 2010 workshop is biologically more complex. We used cross-validation to find the optimal value of regularization parameter &#x003BB; with both minimum MSE and minimum MSE &#x0002B; 1SE of minimum MSE. The optimal &#x003BB; values were used for variable selection. Based on the first simulated data, we found that the minMSE in general picked up too many SNPs. At minMSE &#x0002B; 1SE, the lasso didn&#x00027;t acquire any false positives, but selected too few correct SNPs. The elastic net provided the best compromise between few false positives and many correct selections when the penalty weight &#x003B1; was around 0.1. However, in our simulation setting, this &#x003B1; value didn&#x00027;t result in the lowest minMSE &#x0002B; 1SE. The number of selected SNPs from the QTLMAS 2010 data was after correction for population structure 82 and 161 for the lasso and the elastic net, respectively. In the Fleckvieh data set after population structure correction lasso and the elastic net identified from 1291 to 1966 important SNPs for milk fat content, with major peaks on chromosomes 5, 14, 15, and 20. Hence, we can conclude that it is important to analyze GWAS data with both the lasso and the elastic net and an alternative tuning criterion to minimum MSE is needed for variable selection.</p></abstract>
<kwd-group>
<kwd>lasso</kwd>
<kwd>elastic net</kwd>
<kwd>simulation</kwd>
<kwd>GWAS</kwd>
<kwd>population structure</kwd>
<kwd>cattle</kwd>
</kwd-group>
<counts>
<fig-count count="4"/>
<table-count count="3"/>
<equation-count count="13"/>
<ref-count count="49"/>
<page-count count="11"/>
<word-count count="9536"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="introduction" id="s1">
<title>Introduction</title>
<p>The genome-wide association study (GWAS) is a well-established technique for identifying genetic variants of interest, not only for many common complex human diseases but also for traits of interest in animal and plant genetics (McCarthy et al., <xref ref-type="bibr" rid="B31">2008</xref>; Goddard and Hayes, <xref ref-type="bibr" rid="B16">2009</xref>; Heffner et al., <xref ref-type="bibr" rid="B21">2009</xref>). While the GWAS approach is extensively used and generally celebrated, there are still several unresolved statistical challenges in studying the joint effects of that huge number of genetic and environmental variables (Cantor et al., <xref ref-type="bibr" rid="B4">2010</xref>; Moore et al., <xref ref-type="bibr" rid="B33">2010</xref>). First, in a typical GWAS, genotypes from some thousands to several millions single nucleotide polymorphism (SNP) markers are determined in subjects in the order of a few thousands, leading to the small <italic>n</italic>, large <italic>p</italic> problem (i.e., many more predictor variables than response variables). It is common in GWAS to perform single SNP regression which leads to very high rates of Type I error (false positives). The standard procedure is then to adjust the significance threshold with some kind of multiple comparison criteria, for example Bonferroni or False Discovery Rate (FDR) corrections. Unfortunately, these corrections will unavoidably introduce Type II errors (false negatives) and true SNP-associations of moderate and small effects will be erroneously discarded. Therefore, one of the most important issues in contemporary statistical genetics is to find methods that provide a satisfactory balance between false positives and false negatives in large-scale GWAS (Rice et al., <xref ref-type="bibr" rid="B39">2008</xref>). Secondly, when a large number of SNPs are genotyped on a genome-wide scale, correlated variables (linkage disequilibrium between SNPs) occur both because of biological factors (e.g., population history) and because of high dimensionality (Fan and Lv, <xref ref-type="bibr" rid="B11">2010</xref>).</p>
<p>There are several applications of regularized regression methods to GWAS (reviewed in Szymczak et al., <xref ref-type="bibr" rid="B43">2009</xref>; Dasgupta et al., <xref ref-type="bibr" rid="B6">2011</xref>) and genomic selection, which is closely related to GWAS but concerned with prediction of genomic breeding values (reviewed in de Los Campos et al., <xref ref-type="bibr" rid="B7">2013</xref>). While in genomic selection the focus is on predicting future performance, the association studies are designed to find genetic makers connected to the trait of interest. Penalized regression methods have been previously used for variable selection in high dimensional studies focused on human genetic data (e.g., Sung et al., <xref ref-type="bibr" rid="B41">2009</xref>; Wu et al., <xref ref-type="bibr" rid="B48">2009</xref>; Cho et al., <xref ref-type="bibr" rid="B5">2010</xref>). Kooperberg et al. (<xref ref-type="bibr" rid="B26">2010</xref>) compared the performance of elastic net and lasso using uncorrelated predictor variables. Ayers and Cordell (<xref ref-type="bibr" rid="B1">2010</xref>) examined the influence of penalties in several penalized regression models.</p>
<p>To overcome the challenges mentioned above, penalized regression approaches, also called shrinkage or regularization methods, have been developed. Although shrinking some of the regression coefficients toward zero may result in biased estimates, these regression coefficient estimates will have smaller variance. This can result in enhanced prediction accuracy because of a smaller mean squared error (Hastie et al., <xref ref-type="bibr" rid="B19">2009</xref>). Regression coefficients are shrunk by imposing a penalty on their size, which is done by adding a penalty function to the least-squares model. Moreover, some of these procedures (e.g., the lasso) enable variable selection such that only the important predictor variables stay in the model (Szymczak et al., <xref ref-type="bibr" rid="B43">2009</xref>).</p>
<p>Tibshirani (<xref ref-type="bibr" rid="B44">1996</xref>) proposed the lasso estimator which estimates the regression coefficients through an &#x02113;<sub>1</sub>-norm penalized least-squares criterion. This is equivalent to minimizing the sums of squares of residuals plus an &#x02113;<sub>1</sub> penalty on the regression coefficients. While demonstrating promising performance for many problems, the lasso estimator does have some shortcomings (Zou and Hastie, <xref ref-type="bibr" rid="B49">2005</xref>). Firstly, the lasso tends to have problems when predictor variables are highly correlated. In the extreme case of <italic>k</italic> identical predictor variables, the lasso breaks down. Secondly, when there is some group or cluster structure among the predictor variables, the lasso estimator usually selects only one predictor from a group while ignoring others. Thirdly, the lasso method cannot select more predictor variables than the sample size. This could potentially be a problem in various genomic studies that involve many more, often highly correlated, predictor variables than response variables. From a Bayesian point of view, the lasso penalty corresponds to a Laplace (double exponential) prior over the regression coefficients, which expects many coefficients to be close to zero, and a small subset to be larger and non-zero (Kyung et al., <xref ref-type="bibr" rid="B27">2010</xref>). However, the Bayesian lasso doesn&#x00027;t set any variables to exactly zero and therefore needs to be combined with some other form of variable selection (Hans, <xref ref-type="bibr" rid="B18">2010</xref>).</p>
<p>On the other hand, ridge regression (Hoerl and Kennard, <xref ref-type="bibr" rid="B23">1970</xref>) estimates the regression coefficients through an &#x02113;<sub>2</sub>-norm penalized least-squares criterion. It is well-known that ridge regression shrinks the coefficients of correlated predictor variables toward each other, allowing them to borrow strength from each other (Friedman et al., <xref ref-type="bibr" rid="B13">2010</xref>). However, this behavior is not without its problems. For example, in the case of the <italic>k</italic> identical predictor variables mentioned above, they each get identical coefficients with 1/<italic>k</italic> the size that any single one would get if fit alone. The ridge penalty is ideal if there are many predictor variables, and all have non-zero coefficients (from a Bayesian perspective these are drawn from a Gaussian prior distribution). One important difference between the lasso and ridge regression occurs for the predictor variables with the highest regression coefficients. Whereas the &#x02113;<sub>2</sub> penalty pushes the regression coefficients toward zero with a force proportional to the value of the coefficient, the &#x02113;<sub>1</sub> penalty exerts the same force on all non-zero coefficients. Hence for the variables that are most valuable (i.e., that clearly should be in the model and where shrinkage toward zero is less desirable) an &#x02113;<sub>1</sub> penalty shrinks less (Hesterberg et al., <xref ref-type="bibr" rid="B22">2008</xref>). The extent of shrinkage in ridge regression is also dependent on allele frequency and sample size (Gianola, <xref ref-type="bibr" rid="B15">2013</xref>).</p>
<p>Due to the drawbacks of using the lasso and ridge regression on their own, Zou and Hastie (<xref ref-type="bibr" rid="B49">2005</xref>) proposed the elastic net penalty which is based on a combined penalty of lasso and ridge regression penalties. The penalty parameter &#x003B1; determines how much weight should be given to either the lasso or ridge regression. The elastic net with &#x003B1; set to 0 is equivalent to ridge regression. The elastic net with &#x003B1; close to 1 performs much like the lasso, but removes any degeneracies and odd behavior caused by high correlations. Studies have shown that analysis with the elastic net can result in lower mean squared errors than the lasso and ridge regression when predictor variables are correlated (B&#x000FC;hlmann and van de Geer, <xref ref-type="bibr" rid="B3">2011</xref>). Moreover, the elastic net produces a higher number of correctly identified influential variables than the lasso, and has much lower false positive rate than ridge regression (Tutz and Ulbricht, <xref ref-type="bibr" rid="B45">2009</xref>).</p>
<p>The purpose of this study is to compare the statistical properties of the lasso and the elastic net in a typical GWA framework. Ridge regression was not tested because it selects all variables. The analyses are performed with the R package glmnet which uses the fast cyclical coordinate descent (CCD) algorithm (Friedman et al., <xref ref-type="bibr" rid="B13">2010</xref>). Tests are performed on simulated SNP data that exhibit LD, on simulated pedigree data from the QTLMAS 2010 workshop (Szyd&#x00142;owski and Paczy&#x000F1;ska, <xref ref-type="bibr" rid="B42">2011</xref>) as well as on real data from a 50 K genome-wide SNP panel of 5570 individuals in a dairy cattle study on fat content in milk.</p>
</sec>
<sec sec-type="materials and methods" id="s2">
<title>Materials and methods</title>
<sec>
<title>Methods</title>
<p>Consider a standard multiple linear regression model</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mtext>X</mml:mtext><mml:mi>&#x003B2;</mml:mi><mml:mo>+</mml:mo><mml:mi>e</mml:mi></mml:mrow></mml:math></disp-formula>
<p>where <italic>y</italic> is a vector of length <italic>n</italic> including the response variable, X &#x0003D; (<italic>x</italic><sub><italic>i</italic>1</sub>,&#x02026;,<italic>x</italic><sub><italic>ip</italic></sub>) is a <italic>n</italic> &#x000D7; <italic>p</italic> matrix holding the predictor variables, &#x003B2;<sub>0</sub> is the intercept, &#x003B2; &#x0003D; (&#x003B2;<sub>1</sub>, &#x02026;,&#x003B2;<sub><italic>p</italic></sub>) is a column vector that contains the regression coefficients and <italic>e</italic> is a vector of error terms assuming normal distribution <italic>e</italic> &#x0007E; <italic>N</italic> (0, &#x003C3;<sup>2</sup><sub><italic>e</italic></sub>). For models where <italic>n</italic> &#x0003E; <italic>p</italic>, the values of the unknown parameters &#x003B2;<sub>0</sub> and &#x003B2; can be uniquely estimated by minimizing the residual sum of squares</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mtext>argmin</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:munder><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:munderover><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mo>.</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:math></disp-formula>
<p>The number of SNPs (coded either as 0, 1, 2 or &#x02212;1, 0, 1) is generally much larger than the number of observations in a typical GWAS. A penalized regression function is formulated as</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M3"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mtext>argmin</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:msubsup><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:msubsup><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mstyle><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>P</mml:mtext><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>&#x003BB;</mml:mi><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>where <italic>P</italic> (&#x003BB;, &#x003B2;) a general penalty function with regularization parameter &#x003BB;. The lasso penalty (Tibshirani, <xref ref-type="bibr" rid="B44">1996</xref>) regularizes the linear regression coefficients through an &#x02113;<sub>1</sub>-norm penalized least-squares criterion [i.e., <italic>P</italic> (&#x003BB;, &#x003B2;) &#x0003D; &#x003BB; &#x02016;&#x003B2;&#x02016;<sub>&#x02113;<sub>1</sub></sub>]</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M4"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mtext>argmin</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:munderover><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>|</mml:mo></mml:mrow></mml:mrow></mml:mstyle></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>The resulting regression problem is non-linear in <italic>y</italic> and results in a convex optimization problem. The regularization parameter &#x003BB; controls the amount of shrinkage and needs to be tuned or chosen based on some prior results.</p>
<p>Ridge regression (Hoerl and Kennard, <xref ref-type="bibr" rid="B23">1970</xref>) estimates the linear regression coefficients through an &#x02113;<sub>2</sub>-norm penalized least-squares criterion [i.e., <italic>P</italic> (&#x003BB;, &#x003B2;) &#x0003D; &#x003BB; &#x02016;&#x003B2;&#x02016;<sub>&#x02113;<sub>2</sub></sub>]</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M5"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mtext>argmin</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mtext>p</mml:mtext></mml:munderover><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:msub><mml:mtext>X</mml:mtext><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:munderover><mml:mstyle mathsize='140%' displaystyle='true'><mml:mo>&#x02211;</mml:mo></mml:mstyle><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mtext>p</mml:mtext></mml:munderover ></mml:mrow></mml:mstyle><mml:msubsup><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>This minimization problem can be solved analytically. In ridge regression, the regularization parameter &#x003BB; controls the amount of shrinkage, but no predictor variables are set to zero. The shrinkage makes the &#x003B2; estimates biased but with a smaller variance. The regularization also facilitates the conditioning of the underdetermined system of linear equations that occurs in the <italic>p</italic> &#x0003E;&#x0003E; <italic>n</italic> situation, thus enabling a numerical solution. For correlated predictor variables, ridge regression shrinks the coefficients toward each other.</p>
<p>The elastic net (EN) method (Zou and Hastie, <xref ref-type="bibr" rid="B49">2005</xref>) is based on a compromise between the lasso and ridge regression penalties</p>
<disp-formula id="E6"><label>(6)</label><mml:math id="M6"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mtext>argmin</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B2;</mml:mi></mml:mrow></mml:munder><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:munderover><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mstyle></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mrow><mml:mo>+</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>p</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:msubsup><mml:mi>&#x003B2;</mml:mi><mml:mi>j</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:mi>&#x003B1;</mml:mi><mml:mrow><mml:mo>|</mml:mo><mml:mrow><mml:msub><mml:mi>&#x003B2;</mml:mi><mml:mtext>j</mml:mtext></mml:msub></mml:mrow><mml:mo>|</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:mstyle></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where 0 &#x02264; &#x003B1; &#x02264; 1 is a penalty weight. The EN with &#x003B1; &#x0003D; 1 is identical to the lasso, whereas it turns out to be ridge regression with &#x003B1; &#x0003D; 0 (Friedman et al., <xref ref-type="bibr" rid="B13">2010</xref>). Setting &#x003B1; close to 1 makes the EN to behave similar to the lasso, but eliminates problematic behavior caused by high correlations. When &#x003B1; increases from 0 to 1, for a given &#x003BB; the sparsity of the minimization (i.e., the number of coefficients equal to zero) increases monotonically from 0 to the sparsity of the lasso estimation. The elastic net can select more variables than observations.</p>
<p>Recently, Friedman et al. (<xref ref-type="bibr" rid="B12">2007</xref>, <xref ref-type="bibr" rid="B13">2010</xref>) developed a computationally efficient cyclic coordinate descent (CCD) method for estimation and prediction in regression models with lasso, ridge regression and the elastic net penalties. The CCD method is similar to the forward, stepwise multiple regression approach. First, it finds the &#x003BB;<sub>max</sub> value along the regularization path for which the entire vector <inline-formula><mml:math id="M7"><mml:mrow><mml:mover accent='true'><mml:mi>&#x003B2;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow></mml:math></inline-formula>. The strategy is then to select a minimum value &#x003BB;<sub>min</sub> &#x0003D; &#x003B5;&#x003BB;<sub>max</sub>, and construct a sequence of <italic>K</italic>-values of &#x003BB; decreasing from &#x003BB;<sub>max</sub> to &#x003BB;<sub>min</sub> on the log scale. Typical values are &#x003B5; &#x0003D; 0.001 and <italic>K</italic> &#x0003D; 100. Variables are then added iteratively according to their importance along the &#x003BB; path. One iteration in the CCD algorithm is based on first computing the simple least-squares coefficient on the partial residual, then applying soft thresholding to take care of the lasso contribution to the penalty, and finally using a proportional shrinkage for the ridge penalty. We refer to Friedman et al. (<xref ref-type="bibr" rid="B12">2007</xref>, <xref ref-type="bibr" rid="B13">2010</xref>) for details regarding the algorithm, but we note here that there are options for <italic>na&#x000EF;ve</italic> updates, where a complete cycle through all <italic>p</italic> variables has a computational cost of <italic>O</italic>(<italic>pn</italic>) operations, as well as for <italic>covariance</italic> updates, with <italic>m</italic> non-zero terms in the model, where a complete cycle costs <italic>O</italic>(<italic>pm</italic>) operations if no new variables become non-zero and <italic>O</italic>(<italic>pn</italic>) for each new variable entered. The covariance updating algorithm is more efficient for small data sets (<italic>p</italic> &#x0003C; 500), whereas the na&#x000EF;ve updating is more efficient when <italic>n</italic> &#x0226A; <italic>p</italic>.</p>
<p>The optimal value of &#x003BB;, where the predictor selection should be done, can then be found by <italic>k</italic>-fold cross-validation to find the minimum mean squared error (minMSE) or minMSE &#x0002B; 1 standard error of minMSE backwards along the &#x003BB; path (minMSE &#x0002B; 1SE), i.e., the largest &#x003BB;-value such that the error is within 1SE of the minimum. In <italic>k</italic>-fold cross-validation, the original sample is randomly partitioned into <italic>k</italic> subsamples. One subsample is then taken as the validation data for testing the model, and the remaining <italic>k</italic> &#x02212; 1 subsamples are used as training data. The cross-validation procedure is repeated <italic>k</italic> times (the number of folds), with each of the <italic>k</italic> subsamples used only once as the validation data. The <italic>k</italic> results from the folds are then averaged to create a single estimation with standard errors. The methods above have been implemented in an R package called glmnet (Friedman et al., <xref ref-type="bibr" rid="B13">2010</xref>). minMSE and minMSE &#x0002B; 1SE are automatically calculated using arguments s &#x0003D; &#x0201C;lambda.min&#x0201D; and s &#x0003D; &#x0201C;lambda.1se&#x0201D; in the cv.glmnet function.</p>
<p>Single SNP regression is common in GWAS, but the significance tests need to be adjusted for multiple comparison. We performed one single regression analysis for each SNP on all data sets and collected the corresponding <italic>p</italic>-values in vectors. We then calculated the local false discovery rate following Efron (<xref ref-type="bibr" rid="B8">2010</xref>)</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M8"><mml:mrow><mml:mtext>fdr</mml:mtext><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mo>&#x0201C;</mml:mo><mml:mtext>null</mml:mtext><mml:mo>&#x0201D;</mml:mo><mml:mo>&#x0007C;</mml:mo><mml:mi>z</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>&#x003C0;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:msub><mml:mi>f</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where <italic>f</italic> (<italic>z</italic>) is a mixture density</p>
<disp-formula id="E8"><label>(8)</label><mml:math id="M9"><mml:mrow><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x003C0;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:msub><mml:mi>f</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x003C0;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:msub><mml:mi>f</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p><italic>z</italic><sub><italic>i</italic></sub> &#x0003D; &#x003A6;<sup>&#x02212;1</sup> (<italic>p</italic><sub><italic>i</italic></sub>) is the inverse cumulative normal distribution function of the <italic>p</italic>-values, &#x003C0;<sub>0</sub> &#x0003D; Pr {null} and &#x003C0;<sub>1</sub> &#x0003D; Pr {non &#x02212; null} (the prior probabilities of the null and non-null hypotheses), <italic>f</italic><sub>0</sub> (<italic>z</italic>) the null density and <italic>f</italic><sub>1</sub> (<italic>z</italic>) the non-null density. Estimates of fdr(<italic>z</italic>) can be obtained by</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M10"><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mi>f</mml:mi><mml:mi>d</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy='true'>&#x0005E;</mml:mo></mml:mover><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003C0;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:msub><mml:mover accent='true'><mml:mi>f</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mover accent='true'><mml:mi>f</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M11"><mml:mrow><mml:mover accent='true'><mml:mi>f</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math></inline-formula> is estimated by Poisson mixture regression of the <italic>z</italic>-values (section 5.2 in Efron, <xref ref-type="bibr" rid="B8">2010</xref>). The mean and variance of the empirical null <inline-formula><mml:math id="M12"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>f</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mtext>&#x02009;</mml:mtext><mml:mo>~</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>N</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>&#x003B4;</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x003C3;</mml:mi><mml:mn>0</mml:mn><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math></inline-formula> are together with <inline-formula><mml:math id="M13"><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>&#x003C0;</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> also obtained from the <italic>z</italic>-values, but with Maximum Likelihood Estimation according to section 6.3 in Efron (<xref ref-type="bibr" rid="B8">2010</xref>). We used the locfdr (Efron et al., <xref ref-type="bibr" rid="B9">2011</xref>) R package for estimating fdr(<italic>z</italic>) and set a threshold value of 0.2 according to recommendations (Efron, <xref ref-type="bibr" rid="B8">2010</xref>).</p>
</sec>
<sec>
<title>Simulated data</title>
<p>The purpose of the first simulation setting was to evaluate how different levels of correlation (LD) between the predictor variables (SNP genotypes) influence the statistical properties of the lasso, ridge regression and the elastic net with varying values of &#x003B1;. Three correlation settings were considered using a data set of 1000 records and 25 important predictor variables (SNPs) selected out of a total of 50,000 predictor variables. These 25 predictor variables were divided into blocks of 5 and centered at position 1000, 10,000, 20,000, 30,000, and 40,000, respectively. In the first correlation setting, all predictor variables should display high correlation with each other, both within and between groups. The second setting was designed for a mixture between high and medium correlations, and the third contained only medium correlations between all predictor variables. The selected positions for important variables had no meaning in terms of chromosomal locations, any position selected on purpose or random would lead to the same result of correctly or incorrectly identified predictor variables. The R script to generate and analyze the data is provided as a supplement to this paper.</p>
<p>The data for the first setting was simulated as follows: First, we generated 1000 continuous values (phenotypes), where the first half was associated with high <bold>y</bold><sub>[1:500]</sub> &#x0007E; <italic>N</italic> (2, 1) and the second half with low <bold>y</bold><sub>[501:1000]</sub> &#x0007E; <italic>N</italic> (&#x02212;2, 1) values without additional sorting of phenotypes. The genotypes were constructed by randomly simulating values (0 or 1) from a binomial distribution with frequency 0.5 into two matrices of size 1000 &#x000D7; 50,000. The numbers of these matrices were then added to obtain SNP genotypes (predictor variables) coded as 0, 1, and 2 which is common practice in GWAS. In order to get high effects and correlations of the 25 selected predictor variables, these were simply sorted, independently from each other, so that all SNP genotypes 2 and half of genotype 1 were associated with the higher sorted phenotypic values. Accordingly, the other half of genotypes 1 and all genotypes 0 were associated with low values. This procedure resulted in an average correlation coefficient, <italic>r</italic>, between the 25 important predictor variables of 0.97 (<italic>r</italic><sup>2</sup> &#x0003D; 0.94).</p>
<p>In the second correlation setting (with a mixture of high and medium correlations), we used the data from the first setting and randomly permuted 25% of the sorted values within each of the 5 predictor variables centered around positions 1000 and 40,000 (i.e., resulting in a total of 10 predictor variables with medium correlations and 15 with high correlations). The average correlation between these 25 predictor variables was 0.78 (<italic>r</italic><sup>2</sup> &#x0003D; 0.62). In the third correlation setting (with only medium correlations), we randomly permuted 25% of the sorted values within each of all the 25 important predictor variables. This resulted in a mean correlation between these 25 predictor variables of 0.55 (<italic>r</italic><sup>2</sup> &#x0003D; 0.30). A hundred replicates were generated for each of the three correlation settings. One smaller (5000 predictor variables) and one larger (100,000 predictor variables) high correlation data set was also produced to test the influence of the ratio between <italic>n</italic> and <italic>p</italic> on &#x003B1;. Finally, in order to evaluate the influence of number of important predictor variables on &#x003B1;, we generated 100 replicates of the high correlation setting with 100 important predictor variables to be selected.</p>
</sec>
<sec>
<title>QTLMAS 2010 data</title>
<p>This data was initially constructed for the QTLMAS 2010 workshop (Szyd&#x00142;owski and Paczy&#x000F1;ska, <xref ref-type="bibr" rid="B42">2011</xref>). The number of individuals in the simulated pedigree was 3226 individuals of 5 generations deep. Of the 20 founders 5 were males and 15 were females. The pedigree structure was created assuming that each female mates once (mainly with males from their own generation) and gives birth to approximately 30 progeny. Five autosomal chromosomes were simulated, each about 100 Mbp long. The biallelic SNP data was simulated using a neutral coalescent model. The genomes for founders were compiled by drawing a pair of haplotypes from the haplotype pool. The founders&#x00027; alleles were then dropped down the pedigree with a recombination rate of 1 cM/Mb and a mutation rate of 10<sup>&#x02212;8</sup> per base per generation. The simulation algorithm produced 10,031 markers, including 263 monomorphic and 9768 biallelic SNPs. Out of the 9345 SNPs with <italic>MAF</italic> &#x0003E;0.05, 3933 loci showed significant deviation from Hardy&#x02013;Weinberg equilibrium (Pearson test under individual test error rate of 1%). Mean LD (<italic>r</italic><sup>2</sup> calculated from unphased genotypes) between adjacent SNPs with <italic>MAF</italic> &#x0003E;0.05 was 0.100 (<italic>SD</italic> &#x0003D; 0.152).</p>
<p>The continuous quantitative trait used in our study was determined by 37 QTLs, including 9 controlled genes and 28 random genes. The controlled genes were selected based on their high polymorphism and high LD with markers. The random genes were drawn from the simulated SNPs (excluding chromosome 5), whereas their additive effects were sampled from a truncated normal distribution, <italic>N</italic>(0, 10), and then accepted if abs(add. eff.) &#x0003C;2. Each simulated QTL was surrounded by 19&#x02013;47 polymorphic SNPs (<italic>MAF</italic> &#x0003E;0.05) located within 1 Mb distance from the QTL. Of these, 364 SNPs were in moderate to high LD with the QTLs (<italic>r</italic><sup>2</sup> &#x0003E; 0.1). Residuals were assumed to be uncorrelated and sampled from a normal distribution with variance of 51.76. The narrow-sense heritability (<italic>h</italic><sup>2</sup>) was 0.52 for males and 0.39 for females. For data analyses SNPs with <italic>MAF</italic> &#x0003C;0.01 were discarded, leaving a final sample of 9723 SNPs. We did not check for SNPs that deviate from HWE because we did not want to lose loci that could be under selection or in LD with selected QTLs.</p>
</sec>
<sec>
<title>Cattle data</title>
<p>Deregressed breeding values (Garrick et al., <xref ref-type="bibr" rid="B14">2009</xref>) for milk fat content were calculated for dual purpose Fleckvieh bulls from the joint German-Austrian genomic evaluation. Genotypes of the bulls were scored as 0, 1, and 2 based on data from the Illumina bovine 54 K SNP chip, but only unambiguously mapped SNPs based on Fadista and Bendixen (<xref ref-type="bibr" rid="B10">2012</xref>) were kept for further analysis. The quality of genotypic data was checked for minimum call rate &#x0003E;90% and <italic>MAF</italic> &#x0003E;1%. After correction for missing genotypes and phenotypes, the number of bulls was 5570. Missing alleles were replaced with the average allele frequency. SNPs significantly deviating from the Hardy&#x02013;Weinberg equilibrium (<italic>p</italic> &#x0003C; 0.00001) were deleted. Hence, a total of 34,373 SNPs were used in the final analyses.</p>
</sec>
<sec>
<title>Estimation of population structure using spectral graphs</title>
<p>Population structure can lead to spurious false positive associations in GWAS. There are several techniques available to adjust for population structure (Price et al., <xref ref-type="bibr" rid="B37">2010</xref>; Sillanp&#x000E4;&#x000E4;, <xref ref-type="bibr" rid="B40">2011</xref>). We have chosen a new technique based on spectral graph theory that is related to the popular PCA method (Lee et al., <xref ref-type="bibr" rid="B28">2010</xref>). The main goal of the spectral graph approach is to estimate a significant number of eigenvectors from the genomic markers (SNPs) that can be used as fixed covariates in the penalized regression analyses. The underlying idea of this technique is to represent the population as a weighted graph, where the vertex set is comprised by the individuals in the study, and the weights reflect the degree of genetic similarity between pairs of subjects. The graph is then embedded in a lower dimensional space using the top eigenvectors of a function of the weight matrix. The number of significant eigenvectors can then be estimated by a simulation approach that generates a reference population without population structure, for details see Lee et al. (<xref ref-type="bibr" rid="B28">2010</xref>). We used the methods implemented in the R package GemTools (Klei et al., <xref ref-type="bibr" rid="B25">2011</xref>) for the QTLMAS 2010 and cattle data. The eigenvectors enter the glmnet model without any penalty, using the penalty.factor function argument.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec>
<title>Simulated data</title>
<p>The three simulation settings were evaluated at the following penalty weights (&#x003B1; &#x0003D; 1, 0.75, 0.5, 0.3, 0.1, 0.05, 0.01), where &#x003B1; &#x0003D; 1 is equivalent to the lasso. Ridge regression is obtained with &#x003B1; &#x0003D; 0, but we did not perform any analyses with this value because ridge regression doesn&#x00027;t perform any variable selection. All other &#x003B1; are elastic net penalties with varying degree of &#x02113;<sub>1</sub>- and &#x02113;<sub>2</sub>-norm influence. The optimal value of &#x003BB; (i.e., the stopping criterion) was obtained by performing 10-fold cross-validations to find the minMSE value and the minMSE &#x0002B; 1SE value. Friedman et al. (<xref ref-type="bibr" rid="B13">2010</xref>) recommend using the latter because this avoids over fitting, although we report results for both here. Figure <xref ref-type="fig" rid="F1">1</xref> shows both the number of SNPs in the model and MSE as functions of &#x003BB;. The figure also shows where the minMSE and minMSE &#x0002B; 1SE are located.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>Plot of the Mean-Squared Error (MSE) and the number of SNPs in the model as functions of -log( &#x003BB;) for the 10-fold cross-validation analyses with the EN01 penalty of one the mixed LD simulated data sets</bold>. The red dots are the mean form the cross-validation and the bars indicate mean &#x0002B; 1SE and mean <bold>&#x02212;</bold> 1SE, respectively. The minMSE &#x0002B; 1SE of minMSE (left) and minMSE (right) and are indicated by the dashed vertical lines.</p></caption>
<graphic xlink:href="fgene-04-00270-g0001.tif"/>
</fig>
<p>We note that &#x003BB; at minMSE provides results with too many selected predictor variables because all evaluated models yielded relatively large levels of false positives under all three correlation settings (Table <xref ref-type="table" rid="T1">1</xref>). Another interesting result for this stopping criterion is that the lasso and some elastic net procedures with high &#x003B1; values start to overfit (based on number of false positives) without finding the correct number of predictor variables. For the minMSE &#x0002B; 1SE stopping criterion, the lasso never picked up any false positives, but acquired too few correct predictor variables for all three simulation settings, with a median of only 3 and 2 correct predictor variables out of 25 for the high and mixed correlations, respectively (Table <xref ref-type="table" rid="T1">1</xref>). The elastic net generally performed better, with &#x003B1; &#x0003D; 0.1 and 0.05 resulting in no missing predictor variables and no false positives for the high and low correlation setting. In the mixed correlation setting, no method yielded a perfect result, but &#x003B1; &#x0003D; 0.05 picked up 24 out of the 25 predictor variables and had no false positives. The 5000 and 100,000 predictor data sets both produced results that were very similar to the 50,000 predictor setting (results not shown). For the data with 100 important predictor variables, the lasso selected a median of 4 correct predictor variables and 0 false positives. The elastic net with &#x003B1; &#x0003D; 0.01 selected a median of 100 correct and 0 false positives, whereas &#x003B1; &#x0003D; 0.05 resulted in a median of 97 correct and 0 false positives. This is a similar finding as for the 25 predictor variables in the previous simulations. Hence, these results indicate that &#x003B1; around 0.1 seems to be the best choice for the parameters that we have focused on in our simulation set up.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p><bold>Results from the first simulation where three correlation settings (High, Mixture and Low) were considered, all with 25 significant predictor variables out of a total of 50,000 predictor variables</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th/>
<th/>
<th align="left" valign="top"><bold>Lasso</bold></th>
<th align="left" valign="top"><bold>EN075</bold></th>
<th align="left" valign="top"><bold>EN05</bold></th>
<th align="left" valign="top"><bold>EN04</bold></th>
<th align="left" valign="top"><bold>EN03</bold></th>
<th align="left" valign="top"><bold>EN02</bold></th>
<th align="left" valign="top"><bold>EN015</bold></th>
<th align="left" valign="top"><bold>EN01</bold></th>
<th align="left" valign="top"><bold>EN005</bold></th>
<th align="left" valign="top"><bold>EN001</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" colspan="12"><bold>minMSE</bold></td>
</tr>
<tr>
<td align="left" valign="top">High LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">3 (0.99)</td>
<td align="right" valign="top">6 (1.69)</td>
<td align="right" valign="top">10 (2.36)</td>
<td align="right" valign="top">13 (2.42)</td>
<td align="right" valign="top">16 (2.43)</td>
<td align="right" valign="top">21 (1.99)</td>
<td align="right" valign="top">23 (1.89)</td>
<td align="right" valign="top">25 (0.66)</td>
<td align="right" valign="top">25 (0.00)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">5 (13.51)</td>
<td align="right" valign="top">5 (15.48)</td>
<td align="right" valign="top">8 (24.31)</td>
<td align="right" valign="top">8 (18.60)</td>
<td align="right" valign="top">7 (19.27)</td>
<td align="right" valign="top">10 (25.19)</td>
<td align="right" valign="top">10 (27.33)</td>
<td align="right" valign="top">15 (34.78)</td>
<td align="right" valign="top">35 (72.89)</td>
<td align="right" valign="top">689 (335)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">2.89 (0.11)</td>
<td align="right" valign="top">2.90 (0.11)</td>
<td align="right" valign="top">2.91 (0.11)</td>
<td align="right" valign="top">2.88 (0.10)</td>
<td align="right" valign="top">2.89 (0.11)</td>
<td align="right" valign="top">2.91 (0.11)</td>
<td align="right" valign="top">2.97 (0.08)</td>
<td align="right" valign="top">2.98 (0.08)</td>
<td align="right" valign="top">3.01 (0.08)</td>
<td align="right" valign="top">3.12 (0.09)</td>
</tr>
<tr>
<td align="left" valign="top">Mixed LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">3 (1.04)</td>
<td align="right" valign="top">5 (1.38)</td>
<td align="right" valign="top">9 (1.71)</td>
<td align="right" valign="top">11 (1.94)</td>
<td align="right" valign="top">14 (1.87)</td>
<td align="right" valign="top">17 (1.66)</td>
<td align="right" valign="top">18 (1.41)</td>
<td align="right" valign="top">20 (1.52)</td>
<td align="right" valign="top">24 (1.15)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">3 (23.47)</td>
<td align="right" valign="top">4 (16.78)</td>
<td align="right" valign="top">6 (19.96)</td>
<td align="right" valign="top">9 (21.46)</td>
<td align="right" valign="top">7 (20.20)</td>
<td align="right" valign="top">11 (32.30)</td>
<td align="right" valign="top">17 (30.23)</td>
<td align="right" valign="top">25 (43.00)</td>
<td align="right" valign="top">59 (87.53)</td>
<td align="right" valign="top">898 (369)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">2.92 (0.10)</td>
<td align="right" valign="top">2.93 (0.10)</td>
<td align="right" valign="top">2.94 (0.10)</td>
<td align="right" valign="top">2.94 (0.10)</td>
<td align="right" valign="top">2.96 (0.10)</td>
<td align="right" valign="top">2.97 (0.10)</td>
<td align="right" valign="top">2.98 (0.11)</td>
<td align="right" valign="top">2.99 (0.11)</td>
<td align="right" valign="top">3.02 (0.11)</td>
<td align="right" valign="top">3.16 (0.11)</td>
</tr>
<tr>
<td align="left" valign="top">Low LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">18 (1.87)</td>
<td align="right" valign="top">19 (1.88)</td>
<td align="right" valign="top">20 (1.78)</td>
<td align="right" valign="top">20 (1.44)</td>
<td align="right" valign="top">21 (1.33)</td>
<td align="right" valign="top">23 (1.23)</td>
<td align="right" valign="top">23 (0.91)</td>
<td align="right" valign="top">24 (0.66)</td>
<td align="right" valign="top">25 (0.24)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">7 (16.73)</td>
<td align="right" valign="top">8 (20.85)</td>
<td align="right" valign="top">10 (28.81)</td>
<td align="right" valign="top">9 (30.58)</td>
<td align="right" valign="top">10 (29.86)</td>
<td align="right" valign="top">13 (30.76)</td>
<td align="right" valign="top">27 (32.62)</td>
<td align="right" valign="top">30 (54.56)</td>
<td align="right" valign="top">73 (81.20)</td>
<td align="right" valign="top">1227 (406)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">3.10 (0.12)</td>
<td align="right" valign="top">3.10 (0.12)</td>
<td align="right" valign="top">3.09 (0.12)</td>
<td align="right" valign="top">3.08 (0.10)</td>
<td align="right" valign="top">3.09 (0.10)</td>
<td align="right" valign="top">3.09 (0.10)</td>
<td align="right" valign="top">3.10 (0.10)</td>
<td align="right" valign="top">3.11 (0.10)</td>
<td align="right" valign="top">3.14 (0.10)</td>
<td align="right" valign="top">3.31 (0.10)</td>
</tr>
<tr>
<td align="left" valign="top" colspan="12"><bold>minMSE &#x0002B; 1SE</bold></td>
</tr>
<tr>
<td align="left" valign="top">High LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">3 (0.91)</td>
<td align="right" valign="top">7 (1.80)</td>
<td align="right" valign="top">12 (2.38)</td>
<td align="right" valign="top">16 (2.32)</td>
<td align="right" valign="top">20 (2.05)</td>
<td align="right" valign="top">24 (1.22)</td>
<td align="right" valign="top">25 (0.81)</td>
<td align="right" valign="top">25 (0.10)</td>
<td align="right" valign="top">25 (0.00)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.71)</td>
<td align="right" valign="top">0 (0.35)</td>
<td align="right" valign="top">0 (1.61)</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.10)</td>
<td align="right" valign="top">0 (0.74)</td>
<td align="right" valign="top">35 (68)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">2.99 (0.12)</td>
<td align="right" valign="top">3.00 (0.11)</td>
<td align="right" valign="top">3.01 (0.12)</td>
<td align="right" valign="top">3.00 (0.12)</td>
<td align="right" valign="top">3.00 (0.12)</td>
<td align="right" valign="top">3.01 (0.10)</td>
<td align="right" valign="top">2.97 (0.08)</td>
<td align="right" valign="top">2.98 (0.09)</td>
<td align="right" valign="top">3.01 (0.08)</td>
<td align="right" valign="top">3.12 (0.10)</td>
</tr>
<tr>
<td align="left" valign="top">Mixed LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">3 (0.75)</td>
<td align="right" valign="top">6 (1.38)</td>
<td align="right" valign="top">10 (1.56)</td>
<td align="right" valign="top">12 (1.59)</td>
<td align="right" valign="top">14 (1.23)</td>
<td align="right" valign="top">16 (1.06)</td>
<td align="right" valign="top">17 (1.25)</td>
<td align="right" valign="top">18 (1.47)</td>
<td align="right" valign="top">24 (1.06)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.00)</td>
<td align="right" valign="top">0 (0.31)</td>
<td align="right" valign="top">0 (0.14)</td>
<td align="right" valign="top">0 (0.74)</td>
<td align="right" valign="top">0 (0.72)</td>
<td align="right" valign="top">0 (1.19)</td>
<td align="right" valign="top">0 (2.24)</td>
<td align="right" valign="top">38 (64.72)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">3.03 (0.11)</td>
<td align="right" valign="top">3.04 (0.11)</td>
<td align="right" valign="top">3.06 (0.11)</td>
<td align="right" valign="top">3.06 (0.11)</td>
<td align="right" valign="top">3.06 (0.11)</td>
<td align="right" valign="top">3.08 (0.11)</td>
<td align="right" valign="top">3.11 (0.11)</td>
<td align="right" valign="top">3.10 (0.12)</td>
<td align="right" valign="top">3.13 (0.12)</td>
<td align="right" valign="top">3.26 (0.11)</td>
</tr>
<tr>
<td align="left" valign="top">Low LD</td>
<td align="left" valign="top">Correct</td>
<td align="right" valign="top">17 (2.12)</td>
<td align="right" valign="top">18 (1.89)</td>
<td align="right" valign="top">19 (1.88)</td>
<td align="right" valign="top">20 (1.49)</td>
<td align="right" valign="top">21 (1.36)</td>
<td align="right" valign="top">23 (1.22)</td>
<td align="right" valign="top">24 (0.78)</td>
<td align="right" valign="top">25 (0.51)</td>
<td align="right" valign="top">25 (0.00)</td>
<td align="right" valign="top">25 (0.00)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">False positive</td>
<td align="right" valign="top">0 (0.39)</td>
<td align="right" valign="top">0 (0.65)</td>
<td align="right" valign="top">0 (0.48)</td>
<td align="right" valign="top">0 (0.56)</td>
<td align="right" valign="top">0 (0.80)</td>
<td align="right" valign="top">0 (2.45)</td>
<td align="right" valign="top">0 (1.21)</td>
<td align="right" valign="top">0 (2.06)</td>
<td align="right" valign="top">0 (1.91)</td>
<td align="right" valign="top">106 (109)</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">MSE</td>
<td align="right" valign="top">3.20 (0.13)</td>
<td align="right" valign="top">3.21 (0.13)</td>
<td align="right" valign="top">3.20 (0.13)</td>
<td align="right" valign="top">3.20 (0.12)</td>
<td align="right" valign="top">3.19 (0.11)</td>
<td align="right" valign="top">3.19 (0.12)</td>
<td align="right" valign="top">3.20 (0.11)</td>
<td align="right" valign="top">3.22 (0.10)</td>
<td align="right" valign="top">3.25 (0.10)</td>
<td align="right" valign="top">3.42 (0.11)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Reported values are medians and SD (within parentheses) over 100 replicates</italic>.</p>
<p><italic>The values of the elastic net (EN) refers to the penalty weight &#x003B1;</italic>.</p>
<p><italic>The stopping criteria for &#x003BB; were obtained with 10-fold cross-validation both at minimum MSE and minimum MSE plus 1 SE</italic>.</p>
<p><italic>The median and standard deviation of MSE at the stopping criteria is also reported</italic>.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec>
<title>QTLMAS 2010 data</title>
<p>We used penalty weights of &#x003B1; &#x0003D; 1, 0.9, 0.75, 0.5, 0.3, 0.1, and 0.05. The average &#x003BB; at minMSE &#x0002B; 1SE of ten 10-fold cross-validation runs at each &#x003B1; was considered to be optimal, to account for the minor differences in numbers of selected predictors between the repeats. Since the simulation procedure of this data makes it difficult to know the exact number of SNPs in LD with the 37 QTLs, we could not calculate the amount of correct or false positive predictor variables. However, we know that 364 SNPs were in moderate to high LD with the QTLs (<italic>r</italic><sup>2</sup> &#x0003E; 0.1), and that chromosome 5 has no QTLs.</p>
<p>We first report the results without correction for population structure. The lasso picked out the smallest number of predictor variables (161) at minMSE &#x0002B; 1SE (Table <xref ref-type="table" rid="T2">2</xref>). The elastic net with penalty weight &#x003B1; &#x0003D; 0.1 (EN01) selected 326 SNPs. These selected SNPs were located rather evenly over all 5 chromosomes, with the highest positive regression coefficient of SNP number 4609 on chromosome 3 (Figure <xref ref-type="fig" rid="F2">2A</xref>).</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p><bold>Results from the analysis of the simulated QTLMAS 2010 workshop data with and without correction for population structure (using eigenvectors from spectral graph analyses)</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th/>
<th/>
<th align="left" valign="top"><bold>Lasso</bold></th>
<th align="left" valign="top"><bold>EN09</bold></th>
<th align="left" valign="top"><bold>EN075</bold></th>
<th align="left" valign="top"><bold>EN05</bold></th>
<th align="left" valign="top"><bold>EN03</bold></th>
<th align="left" valign="top"><bold>EN01</bold></th>
<th align="left" valign="top"><bold>EN005</bold></th>
<th align="left" valign="top"><bold>fdr</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">No pop. struct. corr.</td>
<td align="left" valign="top">Selected SNPs</td>
<td align="right" valign="top">161</td>
<td align="right" valign="top">176</td>
<td align="right" valign="top">168</td>
<td align="right" valign="top">219</td>
<td align="right" valign="top">232</td>
<td align="right" valign="top">326</td>
<td align="right" valign="top">454</td>
<td align="right" valign="top">78</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">minMSE &#x0002B; 1SE</td>
<td align="right" valign="top">0.2825</td>
<td align="right" valign="top">0.3082</td>
<td align="right" valign="top">0.3822</td>
<td align="right" valign="top">0.5331</td>
<td align="right" valign="top">0.9087</td>
<td align="right" valign="top">2.6208</td>
<td align="right" valign="top">4.8283</td>
<td align="right" valign="top">&#x02013;</td>
</tr>
<tr>
<td align="left" valign="top">Pop. struct. corr.</td>
<td align="left" valign="top">Selected SNPs</td>
<td align="right" valign="top">82</td>
<td align="right" valign="top">87</td>
<td align="right" valign="top">87</td>
<td align="right" valign="top">92</td>
<td align="right" valign="top">98</td>
<td align="right" valign="top">161</td>
<td align="right" valign="top">240</td>
<td align="right" valign="top">134</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">minMSE &#x0002B; 1SE</td>
<td align="right" valign="top">0.2421</td>
<td align="right" valign="top">0.2594</td>
<td align="right" valign="top">0.3114</td>
<td align="right" valign="top">0.4673</td>
<td align="right" valign="top">0.7751</td>
<td align="right" valign="top">2.1707</td>
<td align="right" valign="top">4.0467</td>
<td align="right" valign="top">&#x02013;</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>The simulated pedigree consists of 3226 individuals from 5 generations</italic>.</p>
<p><italic>The continuous trait was controlled by 37 QTLs that had 364 SNPs with <italic>r</italic><sup>2</sup> &#x0003E; 0.1. The stopping criteria for &#x003BB; were obtained as the average of ten 10-fold cross-validation runs at minimum MSE plus 1 standard error</italic>.</p>
<p><italic>The values of the elastic net (EN) refers to the penalty weight &#x003B1; (e.g., EN005 is elastic net with &#x003B1; &#x0003D; 0.05)</italic>.</p>
<p><italic> fdr refers to the SNPs selected by the single marker regression local false discovery rate method (Efron, <xref ref-type="bibr" rid="B8">2010</xref>)</italic>.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p><bold>Plots of positions and regression coefficients of the selected SNPs from the elastic net (EN01) analysis of the QTLMAS 2010 data in relation to the positions of the 37 simulated QTLs (X)</bold>. The red, blue and green colors of the QTLs indicate additive positive, additive negative and epistatic (including imprinted) effects, respectively. The manhattan plot for the single marker regression shows the significance [-log10(<italic>p</italic>)-value] for all SNPs. <bold>(A)</bold> Without population structure correction. <bold>(B)</bold> With population structure correction using eigenvectors from spectral graph analysis. <bold>(C)</bold> Single marker regression with population structure correction. Highlighted markers (in magenta) are the important SNPs picked by the local false discovery rate method (Efron, <xref ref-type="bibr" rid="B8">2010</xref>).</p></caption>
<graphic xlink:href="fgene-04-00270-g0002.tif"/>
</fig>
<p>The spectral graph method yielded 76 significant eigenvectors that we used to correct for population structure. The eigenvectors were set up as covariates that were forced to be in the model permanently without being penalized. With this approach, the lasso selected only 82 SNPs and EN01 161 SNPs (Table <xref ref-type="table" rid="T2">2</xref>). The number of selected SNPs on chromosome 5 was reduced from 24 to 6 with the population structure correction for EN01 (Figure <xref ref-type="fig" rid="F2">2B</xref>). The single SNP regression results are shown in Figure <xref ref-type="fig" rid="F2">2C</xref>. For comparison purposes, the local false discovery rate was computed on the single marker regression <italic>p</italic>-values using the locfdr R package&#x02014;134 influential SNPs were identified.</p>
</sec>
<sec>
<title>Cattle data</title>
<p>In the Fleckvieh data set we used the same penalty weights as for the QTLMAS 2010 data, i.e., &#x003B1; &#x0003D; 1, 0.9, 0.75, 0.5, 0.3, 0.1, 0.05, and 0. Similarly to the QTLMAS2010 data set, ten 10-fold cross-validation runs at each &#x003B1; were used to find the average &#x003BB; value pointing to minMSE &#x0002B; 1SE.</p>
<p>The SNPs picked up by the various methods were between 1439 and 2689 for runs without population structure correction and between 1291 and 2504 for runs with correction for population structure (Table <xref ref-type="table" rid="T3">3</xref>). The number of SNPs increased by a decreasing penalty factor in fairly regular manner when no population structure was applied, whereas the number of selected SNPs increased sharply from &#x003B1; &#x0003D; 0.3 to &#x003B1; &#x0003D; 0.1. We obtained 117 significant eigenvectors as a description of population structure in this data using the spectral graph method. With adjustment for population structure, the number of selected SNPs was lower in all cases. The selected SNPs were situated on all 29 autosomes, mostly with very low regression coefficients close to 0. All selected effects with corresponding regression coefficients from lasso and the elastic net with penalty weight &#x003B1; &#x0003D; 0.1 are shown in Figures <xref ref-type="fig" rid="F3">3A,B</xref>. The single SNP regression results for the cattle data set are shown in Figure <xref ref-type="fig" rid="F3">3C</xref>, with 160 influential SNPs identified by the local false discovery rate method.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p><bold>Results from the analysis of the deregressed breeding value evaluation for fat content in Fleckvieh bulls</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th/>
<th/>
<th align="left" valign="top"><bold>Lasso</bold></th>
<th align="left" valign="top"><bold>EN09</bold></th>
<th align="left" valign="top"><bold>EN075</bold></th>
<th align="left" valign="top"><bold>EN05</bold></th>
<th align="left" valign="top"><bold>EN03</bold></th>
<th align="left" valign="top"><bold>EN01</bold></th>
<th align="left" valign="top"><bold>EN005</bold></th>
<th align="left" valign="top"><bold>fdr</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">No pop. struct. corr.</td>
<td align="left" valign="top">Selected SNPs</td>
<td align="right" valign="top">1439</td>
<td align="right" valign="top">1451</td>
<td align="right" valign="top">1452</td>
<td align="right" valign="top">1556</td>
<td align="right" valign="top">1603</td>
<td align="right" valign="top">2142</td>
<td align="right" valign="top">2689</td>
<td align="right" valign="top">251</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">minMSE &#x0002B; 1SE</td>
<td align="right" valign="top">0.0029</td>
<td align="right" valign="top">0.0033</td>
<td align="right" valign="top">0.0039</td>
<td align="right" valign="top">0.0057</td>
<td align="right" valign="top">0.0092</td>
<td align="right" valign="top">0.0240</td>
<td align="right" valign="top">0.0438</td>
<td align="right" valign="top">&#x02013;</td>
</tr>
<tr>
<td align="left" valign="top">Pop. struct. corr.</td>
<td align="left" valign="top">Selected SNPs</td>
<td align="right" valign="top">1291</td>
<td align="right" valign="top">1291</td>
<td align="right" valign="top">1297</td>
<td align="right" valign="top">1400</td>
<td align="right" valign="top">1460</td>
<td align="right" valign="top">1966</td>
<td align="right" valign="top">2504</td>
<td align="right" valign="top">160</td>
</tr>
<tr>
<td/>
<td align="left" valign="top">minMSE &#x0002B; 1SE</td>
<td align="right" valign="top">0.0028</td>
<td align="right" valign="top">0.0031</td>
<td align="right" valign="top">0.0038</td>
<td align="right" valign="top">0.0055</td>
<td align="right" valign="top">0.0090</td>
<td align="right" valign="top">0.0236</td>
<td align="right" valign="top">0.0433</td>
<td align="right" valign="top">&#x02013;</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>34,373SNPs were analyzed on 5570 individuals with and without correction for population structure (using eigenvectors from spectral graph analyses)</italic>.</p>
<p><italic>The stopping criteria for &#x003BB; were obtained as the average of ten 10-fold cross-validation runs at minimum MSE plus 1 standard error</italic>.</p>
<p><italic>The values of the elastic net refer to the penalty weight &#x003B1; (e.g., EN005 is elastic net with &#x003B1; &#x0003D; 0.05)</italic>.</p>
<p><italic> fdr refers to the SNPs selected by the single marker regression local false discovery rate method (Efron, <xref ref-type="bibr" rid="B8">2010</xref>)</italic>.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p><bold>Plots of positions and regression coefficients of the selected SNPs from (A) Lasso, (B) elastic net with &#x003B1; &#x0003D; 0.1, and (C) single marker regression with population structure correction for fat content in Fleckvieh bulls</bold>. The manhattan plot shows the significance [-log10(<italic>p</italic>)-value] for all SNPs. Highlighted markers (in magenta) are the important SNPs picked by the local false discovery rate method (Efron, <xref ref-type="bibr" rid="B8">2010</xref>).</p></caption>
<graphic xlink:href="fgene-04-00270-g0003.tif"/>
</fig>
<p>The standard errors were similar between methods, but we noticed an increasing trend in standard error with decreasing penalty factor. When comparing the same methods with and without population structure correction, the standard error was similar for lasso and the elastic net with high penalty factors.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>The goal of the first simulated data was to evaluate how different levels of correlation (linkage disequilibrium) between the predictor variables (SNP genotypes) influence the statistical properties of the lasso and the elastic net. Three correlation settings (high, mixed, and low LD) were considered, all with 25 influential predictor variables with known positions out of a total of 50,000 predictor variables. The response variable (phenotype) was chosen to eliminate complications in the variable selection that arise with high levels of error variance. When single SNP regression with a Bonferroni threshold was applied to the three LD scenarios it always resulted in 25 correctly selected variables and no false positives. However, it would be necessary to perform further simulation studies with other LD patterns before we can conclude that single SNP regression yields similar results as the elastic net.</p>
<p>The optimal value of the regularization parameter &#x003BB; was obtained by performing 10-fold cross-validation to find the minMSE value and the minMSE &#x0002B; 1SE value. When minMSE was used as stopping criteria, the results show that an excessive amount of false positive predictor variables were selected under all three correlation settings. Moreover, the lasso and some elastic net procedures with high &#x003B1; values started to overfit without finding the correct number of true predictor variables. For the minMSE &#x0002B; 1SE stopping criterion, the lasso produced no false positives, but picked up too few correct SNP predictor variables for all three simulation settings. The elastic net generally performed better, with &#x003B1; &#x0003D; 0.1 resulting in no missing predictor variables and no false positives for the high and low correlation setting and only a few missing correct predictor variables in the mixed correlation setting. However, this &#x003B1; value never produced the lowest minMSE or minMSE &#x0002B; 1SE. Hence, the minimum of minMSE &#x0002B; 1SE should not be used to select variables over different &#x003B1; values. It should be noted that many factors can influence the tuning of &#x003BB; and &#x003B1;. Waldron et al. (<xref ref-type="bibr" rid="B47">2011</xref>) used a 2D tuning approach for the elastic net penalties and found that a simultaneous tuning was required to differentiate it from the lasso and ridge regression. However, their approach is based on a combination of cross validation and quasi-Newton optimization to maximize the partial log-likelihood. It is unclear how our finding that the lowest minMSE doesn&#x00027;t correspond to the best penalty factors would influence their approach.</p>
<p>The purpose of the analyses of the second simulated data, from the QTLMAS 2010 workshop, was to evaluate the lasso and the elastic net methods on SNP and phenotype data that has been simulated under biologically more complex scenarios. Unfortunately, because of the complexity of the procedure generating this data, we have no exact number and positions of significantly associated SNPs. The continuous quantitative trait used in our study was determined by 37 QTLs, each QTL being surrounded by 19&#x02013;47 polymorphic SNPs located within 1Mb distance from the QTL. 364 SNPs were found to be in moderate to high LD with the QTLs (<italic>r</italic><sup>2</sup> &#x0003E; 0.1). Hence, the true number of associated SNPs is likely be much larger than 37. Based on these figures, it seems as if the lasso once again selects too few predictor variables. The elastic net with &#x003B1; values around 0.1 appears to have superior performance. However, this conclusion should be interpreted with care since we don&#x00027;t know the exact number of associated SNPs. Recently, Mucha et al. (<xref ref-type="bibr" rid="B34">2011</xref>) published a compilation of results from the analyses of the QTLMAS 2010 data set. Seven different methods, including partial least-squares and Bayesian variable selection, were used by different research groups. None of the methods detected all of the 37 putative QTLs. The highest number of associated SNPs was detected by partial least-squares and BayesC. It is clear that these methods underestimate the number of SNPs that are in LD with the QTLs. Ogutu et al. (<xref ref-type="bibr" rid="B35">2012</xref>) applied the lasso, ridge regression and the elastic net to the QTLMAS 2011 data, but from a genomic selection perspective. Figure <xref ref-type="fig" rid="F2">2C</xref> shows the manhattan plot resulting from the single marker regression with population structure correction. The results are similar compared to the elastic net, with major peaks above the QTL positions. The fdr method selected 134 SNPs, similarly to the elastic net with alpha value 0.1 which identified 161 markers.</p>
<p>For the real cattle data the largest effects were on BTA5 (at 100.6 Mb), multiple SNPs on BTA14 (0.6&#x02013;2.6 Mb), BTA15 (51.9 Mb) and BTA20 (27.2 Mb). Note that these conclusions are based on the elastic net with alpha value 0.1, but other alpha values tend to give similar results. The results were corresponding with those of Hayes et al. (<xref ref-type="bibr" rid="B20">2010</xref>) who found significant associations on BTA5, BTA14, and BTA20 in similar regions as in our study. Meredith et al. (<xref ref-type="bibr" rid="B32">2012</xref>) found moderate association on BTA15 in the 50 Mb region for protein content. The highest signal located on BTA14 points to the DGAT1, a gene with major effect on fat content (Grisart et al., <xref ref-type="bibr" rid="B17">2002</xref>), located at 1.8 Mb. The single marker regression for the cattle data set resulted into 160 selected markers using the fdr methodology. This is a much lower number compared to the result from any of the penalized regression methodologies. The main peaks (Figure <xref ref-type="fig" rid="F3">3C</xref>) corresponded with the regions with the largest effects, identified by lasso and elastic net, with the majority of SNPs on the beginning of BTA14. Figure <xref ref-type="fig" rid="F4">4</xref> shows the Venn diagram comparing the numbers of selected markers by single marker regression with the local false discovery rate method (Efron, <xref ref-type="bibr" rid="B8">2010</xref>), lasso and elastic net with alpha value 0.1. The area circles relate to the number of selected markers, the area of overlap between the circles is proportional to the number of commonly selected markers using different methods. About one third of these were on BTA14, the rest in smaller groups on 16 other chromosomes. The list of the 72 SNPs selected by all three methods can be found in the supplement file 2. Almost all markers selected by lasso were also selected by the elastic net, which can be explained by their methodological similarities.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>Venn diagram (BioVenn, Hulsen et al., <xref ref-type="bibr" rid="B24">2008</xref>) for number of overlapping markers selected by single marker regression (red), lasso (lime), and the elastic net (blue)</bold>. The area of overlap is proportional to the number of commonly selected markers.</p></caption>
<graphic xlink:href="fgene-04-00270-g0004.tif"/>
</fig>
<p>Although it is not expected that the single marker and penalized regression methods pick exactly the same SNPs, we note a relatively large difference in numbers of selected SNPs in the simulated and real data. We note that the deregressed breeding values that were used as phenotypes contain almost no error variance. This could be the reason for the large difference between the number of SNPs selected by single marker regression and the penalized regression methods. The &#x02212;log10(<italic>p</italic>) values for the most significant region on BTA14 were unconventionally high as well, possibly for the same reason. When reducing the numbers of genotypes, the numbers of selected SNPs also decrease. One thousand nine hundred animals instead of the full set yielded a much lower number of SNPs (in the range of 500 based on the elastic net with alpha value 0.1). Moreover, when looking at the effect size of the selected SNPs, we see relatively high effect of each marker in the QTLMAS data set, but close to zero effects in the cattle data for the majority of the SNPs.</p>
<p>There are several applications of the lasso, ridge regression and elastic net to GWAS data (Malo et al., <xref ref-type="bibr" rid="B30">2008</xref>; Wu et al., <xref ref-type="bibr" rid="B48">2009</xref>; Cho et al., <xref ref-type="bibr" rid="B5">2010</xref>). However, few studies have made statistical evaluations of these methods. Usai et al. (<xref ref-type="bibr" rid="B46">2009</xref>) tested the least angle regression version of the lasso on the QTLMAS 2008 data and found that 169 SNPs were needed to explain the variation of the 48 simulated QTLs. Yet, they used a rather ad hoc cross-validation approach where the highest correlation between genomic breeding values (sum of the regression coefficients of the SNPs) and true simulated breeding values was used as stopping criterion. This approach is difficult to generalize to real data because it relies on the fact that the breeding values are known or estimated without error. Ayers and Cordell (<xref ref-type="bibr" rid="B1">2010</xref>) compared the statistical properties of the lasso, ridge regression and elastic net on simulated data. However, the objectives of their study differ from our study. Firstly, Ayers and Cordell (<xref ref-type="bibr" rid="B1">2010</xref>) used a logistic model for binary phenotypes, whereas our study is focused on quantitative traits. Secondly, we have used cross-validation to obtain both minimum MSE and minimum MSE &#x0002B; 1SE, while Ayers and Cordell (<xref ref-type="bibr" rid="B1">2010</xref>) used a permutation approach aimed at controlling the Type I error rate. Thirdly, we have tried to argue that it is important to detect (map) all SNPs that have high effects on the phenotype, regardless if these are highly correlated with each other. Ayers and Cordell (<xref ref-type="bibr" rid="B1">2010</xref>) considered effects from groups of highly correlated variables as a single signal to prevent inflated false-positive rates, which we would regard more appropriate for prediction of future phenotypes.</p>
<p>It is well-known that population structure can lead to spurious false positive associations in GWA studies (Price et al., <xref ref-type="bibr" rid="B37">2010</xref>; Sillanp&#x000E4;&#x000E4;, <xref ref-type="bibr" rid="B40">2011</xref>). There are several ways to adjust for population structure. If the pedigree is known, it is possible to use the additive relationship matrix as a structured random effect in a mixed model. Similarly, one can use the genomic relationship matrix based on identity-by-state coefficients. However, both these approaches are computationally demanding because of the need to solve large mixed model equations. Another approach is based on principal component analysis (PCA) where a certain number of important eigenvectors are calculated from a similarity matrix (estimated either based on the pedigree or the SNPs) between the individuals, and are used as covariates in the SNP regression model (Patterson et al., <xref ref-type="bibr" rid="B36">2006</xref>). The computationally challenging part of this approach is the orthogonal decomposition of the similarity matrix. Lee et al. (<xref ref-type="bibr" rid="B28">2010</xref>) proposed spectral graph approach that in combination with simulation produces a certain number of significant eigenvectors that can be used for correction of population structure in GWAS. The correction was performed using the number of significant eigenvectors in our study, as we suspected some influence based on pedigree structure. It is important to fix the eigenvectors and not perform penalized variable selection on these because none of them were selected in our analyses if not treated as fixed. Further attempts to model population structure in a regularized regression framework include Puniyani et al. (<xref ref-type="bibr" rid="B38">2010</xref>).</p>
<p>Some recent studies have developed Bayesian versions of the elastic net. These methods are based on specifying shrinkage priors over the regression coefficients &#x003B2;. Bornn et al. (<xref ref-type="bibr" rid="B2">2010</xref>) suggested a hierarchical model with a prior distribution over &#x003B2; that combine and compromise between Laplace and Gaussian priors. This model requires computationally challenging tuning of two parameters that determines the amount of shrinkage. Li and Lin (<xref ref-type="bibr" rid="B29">2010</xref>) proposed a related Bayesian elastic net method with a slightly different specification of the prior where the two penalty parameters were chosen by the empirical Bayes method, whereas Kyung et al. (<xref ref-type="bibr" rid="B27">2010</xref>) showed that the two tuning parameters could be estimated within the Gibbs sampler by assigning hyperpriors to them. These methods are computationally demanding and it remains to be seen how large data they can handle can become. Simulation analyses in these two papers showed that the frequentist elastic net performed well in comparison with its Bayesian versions.</p>
<p>Finally, we would like to emphasize that the focus of this study has been on association analysis. GWAS can be interpreted as a mapping method where it is of importance to find all SNPs along the chromosomes that are associated with the phenotype. In statistical terms, GWAS should be based on the correct <italic>estimation</italic> of regression coefficients. On the other hand, in genome-wide selection (GS), the goal is to <italic>predict</italic> future observations based on a sub-set of SNPs that result in a high predictive accuracy. The number of SNPs and their chromosomal positions are of less concern in GS. It is difficult to draw any general conclusions about the predictive difference between the lasso and the elastic net based on our study. Further studies on the relation between estimation and prediction in highly multicollinear data and new criteria for variable selection and predictive performance are needed.</p>
</sec>
<sec sec-type="supplementary material" id="s5">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="http://www.frontiersin.org/journal/10.3389/fgene.2013.00270/abstract">http://www.frontiersin.org/journal/10.3389/fgene.2013.00270/abstract</ext-link></p>
<supplementary-material xlink:href="DataSheet1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="DataSheet2.XLSX" id="SM2" mimetype="application/vnd.openxmlformats-officedocument.spreadsheetml.sheet" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<sec>
<title>Conflict of interest statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
</sec>
</body>
<back>
<ack>
<p>We are grateful to F&#x000F6;rderverein Biotechnologieforschung, Rinderbesammungsgenossenschaft Memmingen, Gesellschaft zur F&#x000F6;rderung der Fleckviezucht in Niederbayern, Nutzvieh GmbH Miesbach, Rinderunion Baden-W&#x000FC;rttemberg eG, Zentrale Arbeitsgemeinschaft &#x000D6;sterreichischer Rinderz&#x000FC;chter, Arbeitsgemeinschaft S&#x000FC;ddeutscher Rinderzucht- und Besamungsorganisationen and ZuchtData EDV-Dienstleistungen GmbH for providing the cattle genotypes and phenotypes used in this analysis., The Vienna Scientific Cluster is acknowledged for providing computing resources. The financial support of the Austrian Science Fund (FWF) via the project TRP 46-B19 is greatly acknowledged. Part of the study was conducted using a travel grant provided by the European Science Foundation (ESF).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ayers</surname> <given-names>K. L.</given-names></name> <name><surname>Cordell</surname> <given-names>H. J.</given-names></name></person-group> (<year>2010</year>). <article-title>SNP selection in genome-wide and candidate gene studies via penalized logistic regression</article-title>. <source>Genet. Epidemiol</source>. <volume>34</volume>, <fpage>879</fpage>&#x02013;<lpage>891</lpage>. <pub-id pub-id-type="doi">10.1002/gepi.20543</pub-id><pub-id pub-id-type="pmid">21104890</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="other"><person-group person-group-type="author"><name><surname>Bornn</surname> <given-names>L.</given-names></name> <name><surname>Gottardo</surname> <given-names>R.</given-names></name> <name><surname>Doucet</surname> <given-names>A.</given-names></name></person-group> (<year>2010</year>). <source>Grouping priors and the Bayesian elastic net. Technical report 254, Department of Statistics</source>. <publisher-name>University of British Columbia</publisher-name>. arXiv:1001.4083.</citation>
</ref>
<ref id="B3">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>B&#x000FC;hlmann</surname> <given-names>P.</given-names></name> <name><surname>van de Geer</surname> <given-names>S.</given-names></name></person-group> (<year>2011</year>). <source>Statistics for High-Dimensional Data: Methods, Theory and Applications</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Springer-Verlag</publisher-name>. <pub-id pub-id-type="doi">10.1007/978-3-642-20192-9</pub-id></citation>
</ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cantor</surname> <given-names>R. M.</given-names></name> <name><surname>Lange</surname> <given-names>K.</given-names></name> <name><surname>Sinsheimer</surname> <given-names>J. S.</given-names></name></person-group> (<year>2010</year>). <article-title>Prioritizing GWAS results: a review of statistical methods and recommendations for their application</article-title>. <source>Am. J. Hum. Genet</source>. <volume>86</volume>, <fpage>6</fpage>&#x02013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1016/j.ajhg.2009.11.017</pub-id><pub-id pub-id-type="pmid">20074509</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cho</surname> <given-names>S.</given-names></name> <name><surname>Kim</surname> <given-names>K.</given-names></name> <name><surname>Kim</surname> <given-names>Y. J.</given-names></name> <name><surname>Lee</surname> <given-names>J. K.</given-names></name> <name><surname>Cho</surname> <given-names>Y. S.</given-names></name> <name><surname>Lee</surname> <given-names>J. Y.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>Joint identification of multiple genetic variants via elastic-net variable selection in a genome-wide association analysis</article-title>. <source>Ann. Hum. Genet</source>. <volume>74</volume>, <fpage>416</fpage>&#x02013;<lpage>428</lpage>. <pub-id pub-id-type="doi">10.1111/j.1469-1809.2010.00597.x</pub-id><pub-id pub-id-type="pmid">20642809</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dasgupta</surname> <given-names>A.</given-names></name> <name><surname>Sun</surname> <given-names>Y. V.</given-names></name> <name><surname>K&#x000F6;nig</surname> <given-names>I. R.</given-names></name> <name><surname>Bailey-Wilson</surname> <given-names>J. E.</given-names></name> <name><surname>Malley</surname> <given-names>J. D.</given-names></name></person-group> (<year>2011</year>). <article-title>Brief review of regression-based and machine learning methods in genetic epidemiology: the Genetic Analysis Workshop 17 experience</article-title>. <source>Genet. Epidemiol</source>. <volume>35</volume>(<supplement>Suppl. 1</supplement>), <fpage>S5</fpage>&#x02013;<lpage>S11</lpage>. <pub-id pub-id-type="doi">10.1002/gepi.20642</pub-id><pub-id pub-id-type="pmid">22128059</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Los Campos</surname> <given-names>G.</given-names></name> <name><surname>Hickey</surname> <given-names>J. M.</given-names></name> <name><surname>Pong-Wong</surname> <given-names>R.</given-names></name> <name><surname>Daetwyler</surname> <given-names>H. D.</given-names></name> <name><surname>Calus</surname> <given-names>M. P.</given-names></name></person-group> (<year>2013</year>). <article-title>Whole-genome regression and prediction methods applied to plant and animal breeding</article-title>. <source>Genetics</source> <volume>193</volume>, <fpage>327</fpage>&#x02013;<lpage>345</lpage>. <pub-id pub-id-type="doi">10.1534/genetics.112.143313</pub-id><pub-id pub-id-type="pmid">22745228</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Efron</surname> <given-names>B.</given-names></name></person-group> (<year>2010</year>). <source>Large-Scale Inference: Empirical Bayes Methods for Estimation, Testing, and Prediction</source>. <publisher-loc>Cambrige, UK</publisher-loc>: <publisher-name>Cambrige University Press</publisher-name>. <pub-id pub-id-type="doi">10.1017/CBO9780511761362</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Efron</surname> <given-names>B.</given-names></name> <name><surname>Turnbull</surname> <given-names>B. B.</given-names></name> <name><surname>Narasimhan</surname> <given-names>B.</given-names></name></person-group> (<year>2011</year>). <source>Locfdr: Computes Local False Discovery Rates. R package Version 1.1-7</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://CRAN.R-project.org/package&#x0003D;locfdr">http://CRAN.R-project.org/package&#x0003D;locfdr</ext-link></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fadista</surname> <given-names>J.</given-names></name> <name><surname>Bendixen</surname> <given-names>C.</given-names></name></person-group> (<year>2012</year>). <article-title>Genomic position mapping discrepancies of commercial SNP chips</article-title>. <source>PLoS ONE</source> <volume>7</volume>:<fpage>e31025</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0031025</pub-id><pub-id pub-id-type="pmid">22363540</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname> <given-names>J.</given-names></name> <name><surname>Lv</surname> <given-names>J.</given-names></name></person-group> (<year>2010</year>). <article-title>A selective overview of variable selection in high dimensional feature space</article-title>. <source>Stat. Sin</source>. <volume>20</volume>, <fpage>101</fpage>&#x02013;<lpage>148</lpage>. <pub-id pub-id-type="pmid">21572976</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friedman</surname> <given-names>J.</given-names></name> <name><surname>Hastie</surname> <given-names>T.</given-names></name> <name><surname>Hoefling</surname> <given-names>H.</given-names></name> <name><surname>Tibshirani</surname> <given-names>R.</given-names></name></person-group> (<year>2007</year>). <article-title>Pathwise coordinate optimization</article-title>. <source>Ann. Appl. Stat</source>. <volume>2</volume>, <fpage>302</fpage>&#x02013;<lpage>332</lpage>. <pub-id pub-id-type="doi">10.1214/07-AOAS131</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friedman</surname> <given-names>J.</given-names></name> <name><surname>Hastie</surname> <given-names>T.</given-names></name> <name><surname>Tibshirani</surname> <given-names>R.</given-names></name></person-group> (<year>2010</year>). <article-title>Regularization paths for generalized linear models via coordinate descent</article-title>. <source>J. Stat. Softw</source>. <volume>33</volume>, <fpage>1</fpage>&#x02013;<lpage>22</lpage>. <pub-id pub-id-type="pmid">20808728</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garrick</surname> <given-names>D. J.</given-names></name> <name><surname>Taylor</surname> <given-names>J. F.</given-names></name> <name><surname>Fernando</surname> <given-names>R. L.</given-names></name></person-group> (<year>2009</year>). <article-title>Deregressing estimated breeding values and weighting information for genomic regression analyses</article-title>. <source>Genet. Sel. Evol</source>. <volume>41</volume>, <fpage>55</fpage>. <pub-id pub-id-type="doi">10.1186/1297-9686-41-55</pub-id><pub-id pub-id-type="pmid">20043827</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gianola</surname> <given-names>G.</given-names></name></person-group> (<year>2013</year>). <article-title>Priors in whole-genome regression: the Bayesian alphabet returns</article-title>. <source>Genetics</source> <volume>194</volume>, <fpage>573</fpage>&#x02013;<lpage>596</lpage>. <pub-id pub-id-type="doi">10.1534/genetics.113.151753</pub-id><pub-id pub-id-type="pmid">23636739</pub-id></citation>
</ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goddard</surname> <given-names>M. E.</given-names></name> <name><surname>Hayes</surname> <given-names>B. J.</given-names></name></person-group> (<year>2009</year>). <article-title>Mapping genes for complex traits in domestic animals and their use in breeding programmes</article-title>. <source>Nat. Rev. Genet</source>. <volume>10</volume>, <fpage>381</fpage>&#x02013;<lpage>391</lpage>. <pub-id pub-id-type="doi">10.1038/nrg2575</pub-id><pub-id pub-id-type="pmid">19448663</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grisart</surname> <given-names>B.</given-names></name> <name><surname>Coppieters</surname> <given-names>W.</given-names></name> <name><surname>Farnir</surname> <given-names>F.</given-names></name> <name><surname>Karim</surname> <given-names>L.</given-names></name> <name><surname>Ford</surname> <given-names>C.</given-names></name> <name><surname>Berzi</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2002</year>). <article-title>Positional candidate cloning of a QTL in dairy cattle: identification of a missense mutation in the bovine DGAT1 gene with major effect on milk yield and composition</article-title>. <source>Genome Res</source>. <volume>12</volume>, <fpage>222</fpage>&#x02013;<lpage>231</lpage>. <pub-id pub-id-type="doi">10.1101/gr.224202</pub-id><pub-id pub-id-type="pmid">11827942</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hans</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>Model uncertainty and variable selection in Bayesian lasso regression</article-title>. <source>Stat. Comput</source>. <volume>20</volume>, <fpage>221</fpage>&#x02013;<lpage>229</lpage>. <pub-id pub-id-type="doi">10.1007/s11222-009-9160-9</pub-id><pub-id pub-id-type="pmid">22549815</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hastie</surname> <given-names>T. R.</given-names></name> <name><surname>Tibshirani</surname> <given-names>R.</given-names></name> <name><surname>Friedman</surname> <given-names>J.</given-names></name></person-group> (<year>2009</year>). <source>Elements of Statistical Learning: Data Mining, Inference and Prediction, 2nd Edn</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Springer-Verlag</publisher-name>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hayes</surname> <given-names>B. J.</given-names></name> <name><surname>Pryce</surname> <given-names>J.</given-names></name> <name><surname>Chamberlain</surname> <given-names>A. J.</given-names></name> <name><surname>Bowman</surname> <given-names>P. J.</given-names></name> <name><surname>Goddard</surname> <given-names>M. E.</given-names></name></person-group> (<year>2010</year>). <article-title>Genetic architecture of complex traits and accuracy of genomic prediction: coat colour, milk-fat percentage, and type in Holstein cattle as contrasting model traits</article-title>. <source>PLoS Genet</source>. <volume>6</volume>:<fpage>e1001139</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pgen.1001139</pub-id><pub-id pub-id-type="pmid">20927186</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heffner</surname> <given-names>E. L.</given-names></name> <name><surname>Sorrels</surname> <given-names>M. E.</given-names></name> <name><surname>Jannink</surname> <given-names>J. L.</given-names></name></person-group> (<year>2009</year>). <article-title>Genomic selection for crop improvement</article-title>. <source>Crop. Sci</source>. <volume>49</volume>, <fpage>1</fpage>&#x02013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.2135/cropsci2008.08.0512</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hesterberg</surname> <given-names>T.</given-names></name> <name><surname>Choi</surname> <given-names>N. H.</given-names></name> <name><surname>Meier</surname> <given-names>L.</given-names></name> <name><surname>Fraley</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>Least angle and &#x02113;<sub>1</sub> penalized regression: a review</article-title>. <source>Stat. Surv</source>. <volume>2</volume>, <fpage>61</fpage>&#x02013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1214/08-SS035</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hoerl</surname> <given-names>A. E.</given-names></name> <name><surname>Kennard</surname> <given-names>R.</given-names></name></person-group> (<year>1970</year>). <article-title>Ridge regression: biased estimation for non orthogonal problems</article-title>. <source>Technometrics</source> <volume>12</volume>, <fpage>55</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1080/00401706.1970.10488634</pub-id></citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hulsen</surname> <given-names>T.</given-names></name> <name><surname>de Vlieg</surname> <given-names>J.</given-names></name> <name><surname>Alkema</surname> <given-names>W.</given-names></name></person-group> (<year>2008</year>). <article-title>BioVenn - a web application for the comparison and visualization of biological lists using area-proportional Venn diagrams</article-title>. <source>BMC Genomics</source> <volume>9</volume>:<fpage>488</fpage>. <pub-id pub-id-type="doi">10.1186/1471-2164-9-488</pub-id><pub-id pub-id-type="pmid">18925949</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Klei</surname> <given-names>L.</given-names></name> <name><surname>Kent</surname> <given-names>B. P.</given-names></name> <name><surname>Melhem</surname> <given-names>N.</given-names></name> <name><surname>Devlin</surname> <given-names>B.</given-names></name> <name><surname>Roeder</surname> <given-names>K.</given-names></name></person-group> (<year>2011</year>). <source>GemTools: a Fast and Efficient Approach to Estimating Genetic Ancestry</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/pdf/1104.1162">http://arxiv.org/pdf/1104.1162</ext-link></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kooperberg</surname> <given-names>C.</given-names></name> <name><surname>LeBlanc</surname> <given-names>M.</given-names></name> <name><surname>Obenchain</surname> <given-names>V.</given-names></name></person-group> (<year>2010</year>). <article-title>Risk prediction using genome-wide association studies</article-title>. <source>Genet. Epidemiol</source>. <volume>34</volume>, <fpage>643</fpage>&#x02013;<lpage>652</lpage>. <pub-id pub-id-type="doi">10.1002/gepi.20509</pub-id><pub-id pub-id-type="pmid">20842684</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kyung</surname> <given-names>M.</given-names></name> <name><surname>Gill</surname> <given-names>J.</given-names></name> <name><surname>Ghosh</surname> <given-names>M.</given-names></name> <name><surname>Casella</surname> <given-names>G.</given-names></name></person-group> (<year>2010</year>). <article-title>Penalized regression, standard errors, and Bayesian lassos</article-title>. <source>Bay. Anal</source>. <volume>5</volume>, <fpage>369</fpage>&#x02013;<lpage>412</lpage>. <pub-id pub-id-type="doi">10.1214/10-BA607</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>A. B.</given-names></name> <name><surname>Luca</surname> <given-names>D.</given-names></name> <name><surname>Roeder</surname> <given-names>K.</given-names></name></person-group> (<year>2010</year>). <article-title>A spectral graph approach to discovering genetic ancestry</article-title>. <source>Ann. Appl. Stat</source>. <volume>4</volume>, <fpage>179</fpage>&#x02013;<lpage>202</lpage>. <pub-id pub-id-type="doi">10.1214/09-AOAS281</pub-id><pub-id pub-id-type="pmid">20689656</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Q.</given-names></name> <name><surname>Lin</surname> <given-names>N.</given-names></name></person-group> (<year>2010</year>). <article-title>The Bayesian elastic net</article-title>. <source>Bay. Anal</source>. <volume>5</volume>, <fpage>151</fpage>&#x02013;<lpage>170</lpage>. <pub-id pub-id-type="doi">10.1214/10-BA506</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Malo</surname> <given-names>N.</given-names></name> <name><surname>Libiger</surname> <given-names>O.</given-names></name> <name><surname>Schork</surname> <given-names>N. J.</given-names></name></person-group> (<year>2008</year>). <article-title>Accommodating linkage disequilibrium in genetic-association analyses via ridge regression</article-title>. <source>Am. J. Hum. Genet</source>. <volume>82</volume>, <fpage>375</fpage>&#x02013;<lpage>385</lpage>. <pub-id pub-id-type="doi">10.1016/j.ajhg.2007.10.012</pub-id><pub-id pub-id-type="pmid">18252218</pub-id></citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>McCarthy</surname> <given-names>M. I.</given-names></name> <name><surname>Abecasis</surname> <given-names>G. R.</given-names></name> <name><surname>Cardon</surname> <given-names>L. R.</given-names></name> <name><surname>Goldstein</surname> <given-names>D. B.</given-names></name> <name><surname>Little</surname> <given-names>J.</given-names></name> <name><surname>Ioannidis</surname> <given-names>J. P. A.</given-names></name> <etal/></person-group>. (<year>2008</year>). <article-title>Genome-wide association studies for complex traits: consensus,uncertainty and challenges</article-title>. <source>Nat. Rev. Genet</source>. <volume>9</volume>, <fpage>356</fpage>&#x02013;<lpage>369</lpage>. <pub-id pub-id-type="doi">10.1038/nrg2344</pub-id><pub-id pub-id-type="pmid">18398418</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Meredith</surname> <given-names>B. K.</given-names></name> <name><surname>Kearney</surname> <given-names>F. J.</given-names></name> <name><surname>Finlay</surname> <given-names>E. K.</given-names></name> <name><surname>Bradley</surname> <given-names>D. G.</given-names></name> <name><surname>Fahey</surname> <given-names>A. G.</given-names></name> <name><surname>Berry</surname> <given-names>D. P.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Genome-wide associations for milk production and somatic cell score in Holstein-Friesian cattle in Ireland</article-title>. <source>BMC Genet</source>. <volume>13</volume>:<fpage>21</fpage>. <pub-id pub-id-type="doi">10.1186/1471-2156-13-21</pub-id><pub-id pub-id-type="pmid">22449276</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Moore</surname> <given-names>J. H.</given-names></name> <name><surname>Asselbergs</surname> <given-names>F. W.</given-names></name> <name><surname>Williams</surname> <given-names>S. M.</given-names></name></person-group> (<year>2010</year>). <article-title>Bioinformatics challenges for genome-wide association studies</article-title>. <source>Bioinformatics</source>, <volume>26</volume>, <fpage>445</fpage>&#x02013;<lpage>455</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btp713</pub-id><pub-id pub-id-type="pmid">20053841</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mucha</surname> <given-names>S.</given-names></name> <name><surname>Pszczo&#x00142;a</surname> <given-names>M.</given-names></name> <name><surname>Strabel</surname> <given-names>T.</given-names></name> <name><surname>Wolc</surname> <given-names>A.</given-names></name> <name><surname>Paczy&#x000F1;ska</surname> <given-names>P.</given-names></name> <name><surname>Szydlowski</surname> <given-names>M.</given-names></name></person-group> (<year>2011</year>). <article-title>Comparison of analyses of the QTLMAS XIV common dataset. II: QTL analysis</article-title>. <source>BMC Proc</source>. <volume>5</volume>(<supplement>Suppl 3</supplement>):<fpage>S2</fpage>. <pub-id pub-id-type="doi">10.1186/1753-6561-5-S3-S2</pub-id><pub-id pub-id-type="pmid">21624172</pub-id></citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ogutu</surname> <given-names>J. O.</given-names></name> <name><surname>Schulz-Streeck</surname> <given-names>T.</given-names></name> <name><surname>Piepho</surname> <given-names>H. P.</given-names></name></person-group> (<year>2012</year>). <article-title>Genomic selection using regularized linear regression models: ridge regression, lasso, elastic net and their extensions</article-title>. <source>BMC Proc</source>. <volume>6</volume>(<supplement>Suppl. 2</supplement>):<fpage>S10</fpage>. <pub-id pub-id-type="doi">10.1186/1753-6561-6-S2-S10</pub-id><pub-id pub-id-type="pmid">22640436</pub-id></citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patterson</surname> <given-names>N.</given-names></name> <name><surname>Price</surname> <given-names>A. L.</given-names></name> <name><surname>Reich</surname> <given-names>D.</given-names></name></person-group> (<year>2006</year>). <article-title>Population structure and eigenanalysis</article-title>. <source>PLoS Genet</source>. <volume>2</volume>:<fpage>e190</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pgen.0020190</pub-id><pub-id pub-id-type="pmid">17194218</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Price</surname> <given-names>A. L.</given-names></name> <name><surname>Zaitlen</surname> <given-names>N. A.</given-names></name> <name><surname>Reich</surname> <given-names>D.</given-names></name> <name><surname>Patterson</surname> <given-names>N.</given-names></name></person-group> (<year>2010</year>). <article-title>New approaches to population stratification in genome-wide association studies</article-title>. <source>Nat. Rev. Genet</source>. <volume>11</volume>, <fpage>459</fpage>&#x02013;<lpage>463</lpage>. <pub-id pub-id-type="doi">10.1038/nrg2813</pub-id><pub-id pub-id-type="pmid">20548291</pub-id></citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Puniyani</surname> <given-names>K.</given-names></name> <name><surname>Kim</surname> <given-names>S.</given-names></name> <name><surname>Xing</surname> <given-names>E. P.</given-names></name></person-group> (<year>2010</year>). <article-title>Multi-population GWA mapping via multi-task regularized regression</article-title>. <source>Bioinformatics</source> <volume>26</volume>, <fpage>208</fpage>&#x02013;<lpage>216</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btq191</pub-id><pub-id pub-id-type="pmid">20529908</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rice</surname> <given-names>T. K.</given-names></name> <name><surname>Schork</surname> <given-names>N. J.</given-names></name> <name><surname>Rao</surname> <given-names>D. C.</given-names></name></person-group> (<year>2008</year>). <article-title>Methods for handling multiple testing</article-title>. <source>Adv. Genet</source>. <volume>60</volume>, <fpage>293</fpage>&#x02013;<lpage>308</lpage>. <pub-id pub-id-type="doi">10.1016/S0065-2660(07)00412-9</pub-id><pub-id pub-id-type="pmid">18358325</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sillanp&#x000E4;&#x000E4;</surname> <given-names>M. J.</given-names></name></person-group> (<year>2011</year>). <article-title>Overview of techniques to account for confounding due to population stratification and cryptic relatedness in genomic data association analyses</article-title>. <source>Heredity</source> <volume>106</volume>, <fpage>511</fpage>&#x02013;<lpage>519</lpage>. <pub-id pub-id-type="doi">10.1038/hdy.2010.91</pub-id><pub-id pub-id-type="pmid">20628415</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sung</surname> <given-names>Y. J.</given-names></name> <name><surname>Rice</surname> <given-names>T. K.</given-names></name> <name><surname>Shi</surname> <given-names>G.</given-names></name> <name><surname>Gu</surname> <given-names>C. C.</given-names></name> <name><surname>Rao</surname> <given-names>D. C.</given-names></name></person-group> (<year>2009</year>). <article-title>Comparison between single-marker analysis using Merlin and multi-marker analysis using LASSO for Framingham simulated data</article-title>. <source>BMC Proc</source>. <volume>3</volume>(<supplement>Suppl. 7</supplement>):<fpage>S27</fpage>. <pub-id pub-id-type="doi">10.1186/1753-6561-3-s7-s27</pub-id><pub-id pub-id-type="pmid">20018017</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szyd&#x00142;owski</surname> <given-names>M.</given-names></name> <name><surname>Paczy&#x000F1;ska</surname> <given-names>P.</given-names></name></person-group> (<year>2011</year>). <article-title>QTLMAS 2010: simulated dataset</article-title>. <source>BMC Proc</source>. <volume>5</volume>(<supplement>Suppl. 3</supplement>):<fpage>S3</fpage>. <pub-id pub-id-type="doi">10.1186/1753-6561-5-S3-S3</pub-id><pub-id pub-id-type="pmid">21624173</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szymczak</surname> <given-names>S.</given-names></name> <name><surname>Biernacka</surname> <given-names>J. M.</given-names></name> <name><surname>Cordell</surname> <given-names>H. J.</given-names></name> <name><surname>Gonz&#x000E1;lez-Recio</surname> <given-names>O.</given-names></name> <name><surname>K&#x000F6;nig</surname> <given-names>I. R.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2009</year>). <article-title>Machine learning in genome-wide association studies</article-title>. <source>Genet. Epidemiol</source>. <volume>33</volume>, <fpage>S51</fpage>&#x02013;<lpage>S57</lpage>. <pub-id pub-id-type="doi">10.1002/gepi.20473</pub-id><pub-id pub-id-type="pmid">19924717</pub-id></citation>
</ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tibshirani</surname> <given-names>R.</given-names></name></person-group> (<year>1996</year>). <article-title>Regression shrinkage and selection via the lasso</article-title>. <source>J. Roy. Stat. Soc. B</source> <volume>58</volume>, <fpage>267</fpage>&#x02013;<lpage>288</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-9868.2011.00771.x</pub-id></citation>
</ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tutz</surname> <given-names>G.</given-names></name> <name><surname>Ulbricht</surname> <given-names>J.</given-names></name></person-group> (<year>2009</year>). <article-title>Penalized regression with correlation-based penalty</article-title>. <source>Stat. Comp</source>. <volume>19</volume>, <fpage>239</fpage>&#x02013;<lpage>253</lpage>. <pub-id pub-id-type="doi">10.1007/s11222-008-9088-5</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Usai</surname> <given-names>M. G.</given-names></name> <name><surname>Goddard</surname> <given-names>M. E.</given-names></name> <name><surname>Hayes</surname> <given-names>B. J.</given-names></name></person-group> (<year>2009</year>). <article-title>LASSO with cross-validation for genomic selection</article-title>. <source>Genet. Res</source>. <volume>91</volume>, <fpage>427</fpage>&#x02013;<lpage>436</lpage>. <pub-id pub-id-type="doi">10.1017/S0016672309990334</pub-id><pub-id pub-id-type="pmid">20122298</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Waldron</surname> <given-names>L.</given-names></name> <name><surname>Pintilie</surname> <given-names>M.</given-names></name> <name><surname>Tsao</surname> <given-names>M. S.</given-names></name> <name><surname>Shepherd</surname> <given-names>F. A.</given-names></name> <name><surname>Huttenhower</surname> <given-names>C.</given-names></name> <name><surname>Jurisica</surname> <given-names>I.</given-names></name></person-group> (<year>2011</year>). <article-title>Optimized application of penalized regression methods to diverse genomic data</article-title>. <source>Bioinformatics</source> <volume>27</volume>, <fpage>3399</fpage>&#x02013;<lpage>3406</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btr591</pub-id><pub-id pub-id-type="pmid">22156367</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>T. T.</given-names></name> <name><surname>Chen</surname> <given-names>Y. F.</given-names></name> <name><surname>Hastie</surname> <given-names>T.</given-names></name> <name><surname>Sobel</surname> <given-names>E.</given-names></name> <name><surname>Lange</surname> <given-names>K.</given-names></name></person-group> (<year>2009</year>). <article-title>Genome-wide association analysis by lasso penalized logistic regression</article-title>. <source>Bioinformatics</source> <volume>25</volume>, <fpage>714</fpage>&#x02013;<lpage>721</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btp041</pub-id><pub-id pub-id-type="pmid">19176549</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zou</surname> <given-names>H.</given-names></name> <name><surname>Hastie</surname> <given-names>T.</given-names></name></person-group> (<year>2005</year>). <article-title>Regularization and variable selection via the elastic net</article-title>. <source>J. Roy. Stat. Soc. B</source> <volume>67</volume>, <fpage>301</fpage>&#x02013;<lpage>320</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-9868.2005.00503.x</pub-id></citation>
</ref>
</ref-list>
</back>
</article>