- About this Journal ·
- Abstracting and Indexing ·
- Advance Access ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Scientifica

Volume 2012 (2012), Article ID 519394, 9 pages

http://dx.doi.org/10.6064/2012/519394

## A Comparison of Two Classes of Methods for Estimating False Discovery Rates in Microarray Studies

^{1}Cancer Research and Biostatistics, Seattle, WA 98101, USA^{2}Department of Biostatistics, University of Washington, Seattle, WA 98195, USA

Received 9 May 2012; Accepted 27 May 2012

Academic Editors: M. Araki, P. Carbonell, J. A. Castro, and K. Jung

Copyright © 2012 Emily Hansen and Kathleen F. Kerr. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The goal of many microarray studies is to identify genes that are differentially expressed between two classes or populations. Many data analysts choose to estimate the false discovery rate (FDR) associated with the list of genes declared differentially expressed. Estimating an FDR largely reduces to estimating , the proportion of differentially expressed genes among all analyzed genes. Estimating is usually done through *P*-values, but computing *P*-values can be viewed as a nuisance and potentially problematic step. We evaluated methods for estimating directly from test statistics, circumventing the need to compute *P*-values. We adapted existing methodology for estimating from *t*- and *z*-statistics so that could be estimated from other statistics. We compared the quality of these estimates to estimates generated by two established methods for estimating from *P*-values. Overall, methods varied widely in bias and variability. The least biased and least variable estimates of , the proportion of differentially expressed genes, were produced by applying the “convest” mixture model method to *P*-values computed from a pooled permutation null distribution. Estimates computed directly from test statistics rather than *P*-values did not reliably perform well.

#### 1. Introduction

Gene expression microarrays are a standard tool for large-scale measurement of gene expression. Microarrays are widely used to detect genes that are differentially expressed (DE) across different groups. Methodology for detecting DE genes has matured over the past decade. Methods have evolved from simple fold-change rules, to the use of classical statistical methods, to the use of test statistics developed specifically for the microarray context (here termed Specialized Test Statistics).

The search for DE genes is usually done in the framework of statistical hypothesis testing. A hypothesis test is performed for each gene. Since microarray studies usually involve tens of thousands of genes, detecting DE genes automatically involves multiple testing issues. Rather than controlling false positives through the traditional family-wise error rate (FWER), most researchers prefer to consider the false discovery rate (FDR). The false discovery rate is an alternative to FWER that was originally proposed by Benjamini and Hochberg [1]. The FDR is the expected proportion of false positives among all the genes declared DE. For example, a FDR of 5% means that among all genes declared DE, 5% of these are truly non-DE on average.

Controlling the FWER is too conservative in the microarray setting, because usually investigators are willing to get a small proportion of false positives in exchange for a sizeable list of potentially DE genes for further study. There is a near-consensus that FDR-estimation procedures are the preferred method for addressing multiple testing in the microarray context [2, 3]. A typical procedure is as follows. (1) Compute a test statistic for every gene. (2) Obtain a -value for every gene. (3) For some threshold *α* close to 0 (e.g., ), call all genes with -values less than *α* significant. (4) Estimate the FDR associated with the list of significant genes. In reality steps (3) and (4) might be done iteratively, with the -value threshold adjusted depending on the estimated FDR resulting from step (4). Next, we review methodology for these 4 steps, which will introduce the question this paper will address.

Step (1) involves the choice of test statistic. The earliest approaches to identifying DE genes were simple fold-change rules. An example of a fold-change rule is to declare a gene DE if its average expression level varies by more than a factor of two between the comparison groups. However, such rules are generally considered unsatisfactory because they do not incorporate the variability of the data, and there is no associated level of confidence in the conclusion of declaring a gene DE [2]. It was natural for statisticians to propose classical test procedures instead of fold-change rules. Let be the (possibly log-transformed) level of expression of a given gene in sample of population . Let be the number of samples drawn from population . Define to be the sample average level of expression of the gene in population . Also define to be the sample variance of the expression level of the gene in population . Typically expression levels are compared across two populations or groups (). To identify DE genes, one could use the classical two-sample *t-*statistic:
However, because of the large number of genes and the nature of microarray data, this statistic is not well suited for discriminating DE genes. The denominator of the *t-*statistic relies on estimates of the expression variances. With the sample sizes typical of most microarray studies (only a few samples per group), these estimates are very unstable. Given the large number of genes in microarray studies, some genes will exhibit a low variance by chance. In real data one often finds that the genes with the largest *t-*statistics are those with the smallest denominators, not necessarily those that are DE between groups.

Tusher et al. [4] proposed a specialized statistic for microarrays, known as the “SAM” or *s-*statistic,
Note that the *s-*statistic is identical to the *t-*statistic with the addition of a constant, , to the denominator. The constant *δ* has the effect of stabilizing the denominator of the *t-*statistic. We will refer to as the stabilizing constant.

There is no consensus on the best way to choose *δ*. Tusher et al. [4] state that they choose to ensure that the variance of the *s-*statistic “is independent of gene expression.” Broberg [5] used the 5th percentile of all *t-*statistic denominators as *δ*, and Xie et al. [6] used the median. Efron et al. [7] compared 5 choices of : 0; 5th, 50th, and 90th percentiles, and the limit as it approached infinity and found the 90th percentile worked best.

In addition to the SAM statistic, other Specialized Test Statistics for microarrays have been developed. Specialized Test Statistics have overwhelmingly been found to outperform the *t-*statistic for detecting DE genes. In each of the papers mentioned above ([8–13]), simulations were included comparing the performance of the proposed Specialized Test Statistic to the classical *t-*statistic. Using real microarray data from a set of “spike-in” assays, Qin et al. [14] assessed the performance of six different statistics. Results unambiguously demonstrated superior performance of Specialized Test Statistics over the mean or *t-*statistic for identifying DE genes although there was no “clear winner” among the Specialized Test Statistics. Similarly, Zhang and Cao [15] employed both simulation and real “spike-in” data and demonstrated that Specialized Test Statistics performed comparably and clearly outperformed classical statistics like the *t-*statistic.

Step (4) involves estimating the FDR associated with a list of genes declared significant. Many different “Mixture Model Methods” (MMMs) [16] have been developed to estimate the FDR for a list of genes declared DE. MMMs assume that there is a valid -value computed for each gene to test the null hypothesis that the gene is not DE. MMMs consider these -values as a mixture of -values for genes for which the null hypothesis is true and -values for which the null hypothesis is false. Estimators of the FDR have the form
where *α* is the -value threshold, is the proportion of non-DE genes, is the proportion of DE genes, and is the cumulative distribution function of all -values. Most MMMs use the observed number of -values less than to estimate . Therefore, most MMMs differ only in their estimates of .

In order to compute a -value—step (2)—one needs to know the distribution of the test statistic under the null hypothesis. Obtaining an empirical null distribution by permutation is a very popular choice in the microarray context. However, Kerr [17] showed that -values resulting from permutation tests and MMMs may be incompatible since permutation-test -values may not satisfy all of the assumptions implicit in MMM methodology. Therefore, estimating the FDR directly from test statistics—skipping step (2) altogether—could be advantageous.

There are few tools available to estimate the FDR directly from test statistics. One such tool, “locfdr” [18], operates directly on test statistics, but has been found to be highly sensitive to minor changes or transformations of the test statistics [17]. Therefore, in this paper, we investigate whether the methodology of “fdrtool” [19, 20] can be easily adapted to estimate FDRs directly from Specialized Test Statistics.

There are many different Specialized Test Statistics to choose from; we use the SAM-statistic in our investigation due to its simplicity and popularity. We compare results to a procedure in current common practice, which is to use a variant of a permutation-test -value together with a MMM for -values. In this approach, a single null distribution of the test statistic is estimated by pooling all the permutation null test statistics across genes. Kerr [17] showed that such “pooled null -values” are different from permutation test -values. However, “pooled null -values” have the attractive feature that they have a monotone relationship with the test statistic. For our second class of methods for estimating , we computed pooled null permutation -values and then estimated the FDR with an MMM. We used two MMMs that we have seen to work well: “*q*value” [21] and “convest” [22].

#### 2. Results and Discussion

##### 2.1. Methodology

The methodology of “fdrtool” is not designed to take *s-*statistics or any other Specialized Test Statistics as input. Since “fdrtool” accepts *t-*statistics, and *s-*statistics are similar to *t-*statistics, we investigated whether we could apply “fdrtool” to *s-*statistics and get accurate results. Our idea was to compute both *s-*statistics and *t-*statistics on the same data, and then rescale the *s-*statistics to have the same spread as the corresponding *t-*statistics. We considered two measures of variability: standard deviation and interquartile range. In addition, we considered four different *s-*statistics, each using a different stabilizing constant term in the denominator. We designed simulations to investigate the accuracy of using “fdrtool” to estimate FDR’s from *s-*statistics in this way. Our simulated data were based on real microarray data.

As a simple example, suppose that we have 100 genes with measured expression levels on two samples from two groups we wish to compare. We compute a *t-*statistic and an *s-*statistic for each gene. Due to the addition of the stabilizing constant in the denominator of the *s-*statistic, the 100 *s-*statistic values will be closer to zero than the 100 *t-*statistic values. In other words, the addition of to the denominator pulls the *s-*statistic values toward zero, so that the variance of the *s-*statistics across genes is smaller than the variance of the *t-*statistics.

Accordingly, we used a measure of the difference in variability of the statistics across genes as the rescaling factor to rescale the *s-*statistics. We examined two measures of variability: the standard deviation (SD) and the interquartile range (IQR). We took the ratio of the variability of the *t-*statistic to the variability of the *s-*statistic as our rescaling factor. For the SD, we transformed the *s-*statistics by multiplying them by the factor , where SD_{T} and SD_{S} are the standard deviations across genes of *t-*statistics and *s-*statistics, respectively. The rescaling factor for the IQR was similarly defined, with IQR in the place of SD. Note that rescaling the *s-*statistics maintains their rank order, which retains the *s-*statistic’s advantages for accurately detecting DE genes.

In computing *s-*statistics one must choose the value of the stabilizing constant *δ* in the *s-*statistic denominator. We considered four choices for *δ*, and so defined four different *s-*statistics: the *s*30, *s*50, *s*70, and *s*90. The *s*30-statistic uses the 30th percentile of all *t-*statistic denominators as , the *s*50-statistic uses the 50th percentile of all *t-*statistic denominators as *δ*, and so forth. In summary, for each simulated data set, we computed the *t-*statistic and the four *s-*statistics. We transformed the *s-*statistics by two different rescaling factors (SD and IQR). When loading the rescaled *s-*statistics into “fdrtool,” we evaluated two different specifications to the software. We could specify that the input was *t-*statistics or *z-*statistics. We evaluated the performance of these five statistics using the four different combinations of rescaling factors and two different input options.

##### 2.2. Design of Simulation Study

The design of the simulation study is that same as Kerr (2009) [17], and we describe it here briefly. We based simulations on real microarray data of EBV-transformed lymphoblastoid cell line tissue from 60 individuals with European ancestry (CEU) and 45 ethnic Chinese (CHB). There are data on 47,293 transcripts.

For each gene, we calculated the sample mean and sample standard deviation of that gene in each population group (CEU and CHB). We rounded the sample mean values to the nearest tenth digit, so that means could be unambiguously declared equal or unequal between groups. We simulated CEU and CHB sample data (where we knew the “truth” regarding the degree of differential expression) from independent normal distributions with parameters based on the values from the actual data. We simulated datasets of 10,000 transcripts.

We ran three types of simulations: EV, UV1, and UV2. In each simulation, a proportion of genes () were differentially expressed in the mean. For the simulated DE genes, the difference in means for the simulated CEU and CHB samples was taken from the observed sample means in the real data. The variances of the simulated CEU and CHB samples differed between the three simulation types.

In the EV (“equal variance”) simulations, the standard deviation of both the simulated CEU sample and the simulated CHB sample came from the observed standard deviation from the CEU data. This was not the case in the UV (“unequal variance”) simulations. In the UV1 simulations, the standard deviation of the simulated CEU sample came from the observed standard deviation from the CEU data, and the standard deviation of the simulated CHB sample came from the observed standard deviation from the CHB data. In the UV2 simulations, this was reversed: the standard deviation of the simulated CEU sample came from the observed standard deviation from the CHB data, and the standard deviation of the simulated CHB sample came from the observed standard deviation from the CEU data.

We initially simulated data for four different values of : 0.01, 0.05, 0.10, and 0.25, and three different sample sizes: large, intermediate, and small. Letting nCEU and nCHB denote the sample sizes for CEU and CHB, respectively, our sample sizes were as follows: large (nCEU, nCHB) = (60, 45), intermediate (nCEU, nCHB) = (16, 12), and small (nCEU, nCHB) = (8, 6). Note that all of the sample sizes maintain the 4 : 3 ratio of the original data. With three sample sizes, four values of , three simulation types, two rescaling factors (SD and IQR), and two input specification options (*t-*score, *z-*score), there were a total of 144 different simulation scenarios. After examining the results, we performed additional EV simulations for , 0.02, 0.03, and 0.04 for the three sample sizes, two rescaling factors, and two input specification options, adding an additional 48 simulation scenarios, for a total of 192 simulation scenarios. We repeated each scenario 20 times.

##### 2.3. Evaluation of Adapted fdrtool Method

First, we verified that the *s-*statistics outperformed the *t-*statistic in identifying DE genes. Figure 1 shows that the *s-*statistics outperform the *t-*statistic across all sample sizes and values of . This difference in performance is more marked for the smaller sample sizes than the larger sample sizes, across all values of . The four *s-*statistics are generally close in performance. The *s*30-*, s*50-*, *and *s*70-statistics perform comparably. The *s*90-statistic (purple curve) stands slightly apart from the other three, giving less sensitivity at low false positive rates and better sensitivity at higher false positive rates. The difference in the ROC curves for the *s*90-statistic and the other *s-*statistics is more pronounced in the smaller sample sizes. However, the difference between the *s*90-statistic and the other *s-*statistics is not as large as the difference between the *t-*statistic and the four *s-*statistics.

We investigated the performance of “fdrtool” for *s-*statistics when “fdrtool” treated them as either *t-*statistics or *z-*statistics. Specifying in “fdrtool” that the inputted *s-*statistics were *t-*statistics (Figure 2) worked reasonably well for a low percentage of DE genes (, 0.05). In the first row of Figure 2, we see that the estimates of have a modest conservative bias for the 30- and 50-statistics in the two smaller sample sizes. However, for higher proportions of DE genes (, 0.25), the “fdrtool” estimates of are poor, with excessive conservative bias. This result held regardless of whether statistics were rescaled with IQR (Figure 1) or SD (see supplementary File 1 in supplementary material available online with doi:10.6064/2012/519394). However, SD-scaling tended to give less predictable results.

Other aspects of the results presented in Figure 2 are notable. First, when *π* =0.25, there is a curvilinear trend in estimates of as we move from the *t-*statistic at one extreme to the s90-statistic at the other. Second, the bias in the estimates of when *t-*statistics are computed on the data tends to be at least as large as the bias for the *s-*statistics. This is surprising since “fdrtool” was developed for *t-*statistics. Third, there is a substantial difference between simulations with and simulations with . When , estimates for the *s*70- and *s*90-statistics show anticonservative bias, whereas, when , these statistics show conservative bias. Figure 3 expands upon Figure 2, with simulation results for , 0.01, 0.02, 0.03, 0.04, and 0.05. There is a decrease from anticonservative bias to conservative bias as increases from 0.005 to 0.05.

In contrast to the results for the *t-*statistic input specification, telling “fdrtool” the inputted statistics were *z-*statistics (Figure 4) worked better for high percentages of DE genes (, 0.25), but showed anticonservative bias for low percentages of DE genes (). Bias and variability is mostly improved over Figure 2 except for . All results presented here are for the EV simulations; results for the UV1 and UV2 simulations were similar (see supplementary File 1).

#### 3. Evaluation of Mixture Model Methods on Pooled Permutation Null -Values

A popular approach in practice is to compute a variant of a permutation-test -value. In this approach, a single null distribution of the test statistic is estimated by pooling all the permutation null test statistics across genes. “Pooled null -values” can be computed from a single empirical null distribution. We also obtained estimates of using pooled null -values and then estimating the FDR with an MMM. We obtained estimates of using two MMMs that we have seen to work well: “*q*value” [21] and “convest” [22].

Figure 5 shows the “*q*value” results for the EV simulations. For a low percentage of DE genes (, 0.05), the “*q*value” results (top two rows of Figure 5) largely show less bias than the “fdrtool” results (top two rows of Figures 2 and 4). However, there are exceptions as, for example, the *s*50-and *s*70-statistics show less bias in the smaller sample sizes when in the “fdrtool” results than in the “*q*value” results. The “*q*value” estimates show considerably greater variability than the “fdrtool” estimates, so it is not entirely clear that one method is better. For higher percentages of DE genes (, 0.25), the story is similar, with “*q*value” estimates generally showing less bias than “fdrtool” estimates, but greater variability. However, the higher variability in the “*q*value” results seems more clearly acceptable when , 0.25, given the dramatic reduction in bias. Interestingly, “*q*value” does not appear to perform worse for the *t-*statistic than it does for the *s-*statistics, whereas “fdrtool” almost always performs better for *s-*statistics. The comparison of the “fdrtool” results to the “*q*value” results was similar for the UV1 and UV2 simulations (see Additional File 1).

Figure 6 shows the “convest” estimates of for the EV simulations. Overall, the “convest” estimates show less bias and comparable or less variability than the “fdrtool” or “*q*value” results. Thus, although it relies on -values that could be considered invalid as permutation test -values [17], “convest” might yield superior estimates of (and hence FDR) in terms of bias and variability. Similar to the “*q*value” results, “convest” performs comparably for *t-*statistics as for *s-*statistics.

#### 4. Discussion

The “convest” method took longer to compute than the “fdrtool” approach (approximately 20 times as long). However, computation time was on the order of 6.5 seconds for 1 set of 10,000 -values and was not prohibitive. The more important difference is that “convest” requires -values, which we computed via permutation, whereas the “fdrtool” is applied directly to test statistics and does not require any data permutations.

An important limitation of this study is that all simulations were based on one dataset. We also did not explore different correlation structures for the simulated gene expression.

#### 5. Conclusions

We compared approaches for estimating , the proportion of differentially expressed genes, from microarray data. The approaches were in two classes: (1) adapting the methodology of “fdrtool” to Specialized Test Statistics and (2) applying mixture model methods (MMMs) to -values computed from a pooled permutation null distribution. The best-performing method was in the second class, using the MMM “convest” to -values computed from a pooled permutation null distribution. Overall, estimates of exhibited the least bias and variability, and bias tended to be conservative rather than anti-conservative.

The first class of approaches for estimating adapted existing methodology of “fdrtool” to *s-*statistics. Interestingly, “fdrtool” generally performed better for *s-*statistics than it did for *t-*statistics, even though the empirical modeling is designed for *t-*statistics. The performance of the “fdrtool”-based approach varied substantially on the proportion of differentially expressed gene. However, the “convest”-based approach outperformed the fdrtool approach in almost all scenarios.

#### Authors’ Contributions

K. F. Kerr designed the study and simulated the microarray data. E. Hansen applied all the methods for estimating the proportion of DE genes and made all figures. Both authors wrote, edited, and approved the final manuscript.

#### Acknowledgment

This publication was made possible by Grant Number ES007033 from NIEHS. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIEHS.

#### References

- Y. Benjamini and Y. Hochberg, “Controlling the false discovery rate: a practical and powerful approach to multiple testing,”
*Journal of the Royal Statistical Society Series B*, vol. 57, pp. 289–300, 1995. View at Google Scholar - D. B. Allison, X. Cui, G. P. Page, and M. Sabripour, “Microarray data analysis: from disarray to consolidation and consensus,”
*Nature Reviews Genetics*, vol. 7, no. 1, pp. 55–65, 2006. View at Publisher · View at Google Scholar · View at Scopus - A. Dupuy and R. M. Simon, “Critical review of published microarray studies for cancer outcome and guidelines on statistical analysis and reporting,”
*Journal of the National Cancer Institute*, vol. 99, no. 2, pp. 147–157, 2007. View at Publisher · View at Google Scholar · View at Scopus - V. G. Tusher, R. Tibshirani, and G. Chu, “Significance analysis of microarrays applied to the ionizing radiation response,”
*Proceedings of the National Academy of Sciences of the United States of America*, vol. 98, no. 9, pp. 5116–5121, 2001. View at Publisher · View at Google Scholar · View at Scopus - P. Broberg, “Statistical methods for ranking differentially expressed genes,”
*Genome Biology*, vol. 4, no. 6, article R41, 2003. View at Google Scholar · View at Scopus - Y. Xie, W. Pan, and A. B. Khodursky, “A note on using permutation-based false discovery rate estimates to compare different analysis methods for microarray data,”
*Bioinformatics*, vol. 21, no. 23, pp. 4280–4288, 2005. View at Publisher · View at Google Scholar · View at Scopus - B. Efron, R. Tibshirani, J. D. Storey, and V. Tusher, “Empirical Bayes Analysis of a Microarray Experiment,”
*Journal of the American Statistical Association*, vol. 96, no. 456, pp. 1151–1160, 2001. View at Google Scholar · View at Scopus - P. Baldi and A. D. Long, “A Bayesian framework for the analysis of microarray expression data: regularized t-test and statistical inferences of gene changes,”
*Bioinformatics*, vol. 17, no. 6, pp. 509–519, 2001. View at Google Scholar · View at Scopus - I. Lönnstedt and T. Speed, “Replicated microarray data,”
*Statistica Sinica*, vol. 12, no. 1, pp. 31–46, 2002. View at Google Scholar · View at Scopus - G. K. Smyth, “Linear models and empirical bayes methods for assessing differential expression in microarray experiments,”
*Statistical Applications in Genetics and Molecular Biology*, vol. 3, no. 1, article 3, 2004. View at Google Scholar · View at Scopus - X. Cui, J. T. Hwang, J. Qiu, N. J. Blades, and G. A. Churchill, “Improved statistical tests for differential gene expression by shrinking variance components estimates,”
*Biostatistics (Oxford, England)*, vol. 6, no. 1, pp. 59–75, 2005. View at Google Scholar · View at Scopus - C. M. Kendziorski, M. A. Newton, H. Lan, and M. N. Gould, “On parametric empirical Bayes methods for comparing multiple groups using replicated gene expression profiles,”
*Statistics in Medicine*, vol. 22, no. 24, pp. 3899–3914, 2003. View at Publisher · View at Google Scholar · View at Scopus - G. W. Wright and R. M. Simon, “A random variance model for detection of differential gene expression in small microarray experiments,”
*Bioinformatics*, vol. 19, no. 18, pp. 2448–2455, 2003. View at Publisher · View at Google Scholar · View at Scopus - L.-X. Qin, K. F. Kerr, A. Boyles et al., “Empirical evaluation of data transformations and ranking statistics for microarray analysis,”
*Nucleic Acids Research*, vol. 32, no. 18, pp. 5471–5479, 2004. View at Publisher · View at Google Scholar · View at Scopus - S. Zhang and J. Cao, “A close examination of double filtering with fold change and t test in microarray analysis,”
*BMC Bioinformatics*, vol. 10, no. 1, article 402, 2009. View at Publisher · View at Google Scholar · View at Scopus - D. B. Allison, G. L. Gadbury, M. Heo et al., “A mixture model approach for the analysis of microarray gene expression data,”
*Computational Statistics and Data Analysis*, vol. 39, no. 1, pp. 1–20, 2002. View at Google Scholar · View at Scopus - K. F. Kerr, “Comments on the analysis of unbalanced microarray data,”
*Bioinformatics*, vol. 25, no. 16, pp. 2035–2041, 2009. View at Publisher · View at Google Scholar · View at Scopus - B. Efron, “Size, power and false discovery rates,”
*Annals of Statistics*, vol. 35, no. 4, pp. 1351–1377, 2007. View at Publisher · View at Google Scholar · View at Scopus - K. Strimmer, “A unified approach to false discovery rate estimation,”
*BMC Bioinformatics*, vol. 9, article 303, 2008. View at Publisher · View at Google Scholar · View at Scopus - K. Strimmer, “fdrtool: a versatile R package for estimating local and tail area-based false discovery rates,”
*Bioinformatics*, vol. 24, no. 12, pp. 1461–1462, 2008. View at Publisher · View at Google Scholar · View at Scopus - J. D. Storey and R. Tibshirani, “Statistical significance for genomewide studies,”
*Proceedings of the National Academy of Sciences of the United States of America*, vol. 100, no. 16, pp. 9440–9445, 2003. View at Publisher · View at Google Scholar · View at Scopus - M. Langaas, B. H. Lindqvist, and E. Ferkingstad, “Estimating the proportion of true null hypotheses, with application to DNA microarray data,”
*Journal of the Royal Statistical Society Series B*, vol. 67, no. 4, pp. 555–572, 2005. View at Publisher · View at Google Scholar · View at Scopus