- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Computational and Mathematical Methods in Medicine
Volume 2013 (2013), Article ID 860673, 11 pages
A Robust Rerank Approach for Feature Selection and Its Application to Pooling-Based GWA Studies
1Institute of Statistical Science, Academia Sinica, Taipei 11529, Taiwan
2Institute of Epidemiology and Preventive Medicine, National Taiwan University, Taipei 10055, Taiwan
3Research Center of Genes, Environment and Human Health, National Taiwan University, Taipei 10055, Taiwan
Received 12 January 2013; Accepted 8 March 2013
Academic Editor: Shinto Eguchi
Copyright © 2013 Jia-Rou Liu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Large--small- datasets are commonly encountered in modern biomedical studies. To detect the difference between two groups, conventional methods would fail to apply due to the instability in estimating variances in -test and a high proportion of tied values in AUC (area under the receiver operating characteristic curve) estimates. The significance analysis of microarrays (SAM) may also not be satisfactory, since its performance is sensitive to the tuning parameter, and its selection is not straightforward. In this work, we propose a robust rerank approach to overcome the above-mentioned diffculties. In particular, we obtain a rank-based statistic for each feature based on the concept of “rank-over-variable.” Techniques of “random subset” and “rerank” are then iteratively applied to rank features, and the leading features will be selected for further studies. The proposed re-rank approach is especially applicable for large--small- datasets. Moreover, it is insensitive to the selection of tuning parameters, which is an appealing property for practical implementation. Simulation studies and real data analysis of pooling-based genome wide association (GWA) studies demonstrate the usefulness of our method.
Recently, many researches encounter the problem where the data objects have an extremely large number of features while the available sample size is relatively small. In view of the number of features, unless there is a large sample size, conventional statistical methods that based on asymptotic theory are not applicable. This is so-called the curse of dimensionality. To improve the accuracy and efficiency of data analysis, it is helpful to reduce the number of features before fitting statistical models. Various dimension reduction methods have been proposed, which can be categorized into two categories: feature extraction and feature selection . The former is to transform a high dimensional data into a lower dimensional space, while the latter is to select a subset of the original features. For example, principle component analysis (PCA)  is one of the most commonly used feature extraction methods. However, the meanings of the extracted principle components are often difficult to interpret. Alternatively, feature selection methods aim to choose a subset of features which do not alter the original explanation [3, 4]. In the field of bioinformatics, statistical techniques are usually used to be a preprocessing step for the purpose of identifying relevant features with disease status. After identifying a set of susceptible genetic markers, their biological meanings or functions need to be verified by further studies. In this view, feature selection seems to be more appropriate than feature extraction due to the advantage of interpretability.
To detect the difference between two groups, one approach is fold change . It is calculated simply as the ratio of the sample means of two groups. Its drawback, however, is that it ignores the variance of each group so that the statistical power is usually poor. In contrast, -test takes into account the sample variances to detect mean change between two groups, and it is very powerful when the normality assumption holds. However, the data is not always normally distributed especially for the case of small sample size. The instability in estimating variances in this situation would also make type-I error inflated and fail to apply. The Significance Analysis of Microarrays (SAM)  is proposed to improve this drawback through adding a small positive number to the denominator of the -test statistic. However, it is still difficult to apply since its detection power is sensitive to value and, to the best of our knowledge, there is no optimal data-driven approach to select .
Nonparametric methods are distribution-free and robust to the presence of outliers. To achieve robustness, these methods consider the ranks of data instead of the original measurements. Unfortunately, traditional “rank-over-sample” methods such as Wilcoxon rank sum test or area under the receiver operating characteristic curve (AUC) would fail to deal with the large--small- datasets. This can be seen by considering a case-control study with cases and controls, where there are at most distinct AUC values which can be much smaller than the number of features. In this case, the high proportion of tied values would make it hard to construct a ranking list. To improve the detection power while keeping the robustness, methods based on the idea of “rank-over-variable” were proposed including Rank Product  and Rank Test . In these methods, the original data points are still replaced by their ranks, but the rank here is defined for each feature by its position in the list of sorted variables of a single subject. As a consequence, the possible ranking values would range between one and the number of variables, which decreases the occurrence of tied values while keeping the robustness. Compared with unstably estimating variances in -test, rank-over-variable methods do not involve the estimation of variance but it has been reported to be able to detect both changes in mean and correlation between two groups . Motivated by the advantages of rank-over-variable, the aim of this work is to propose a robust rerank approach for feature selection. As will become clear later, the proposed rerank approach is especially applicable for large--small- datasets and is not sensitive to the selection of tuning parameters.
The rest of this paper is organized as follows. In Section 2, based on the idea of rank-over-variable, we propose a robust rerank approach to create a ranking list for feature selection. Numerical studies are conducted in Section 3 to verify that our rerank approach does outperform AUC, -test, and SAM. The paper is ended with conclusions in Section 4.
2. Inference Procedure
2.1. Rerank Approach
Consider a case-control study that examines markers with cases and controls. Let and be the continuous measurement for marker , , of subject in the case and control groups, respectively. The goal is to identify those markers which are truly associated with disease. To construct a ranking list of markers, we consider a rank-based statistic by modifying the method of Alvo et al. . Define the centered markers by where is the overall sample mean of the th marker. Let be the rank of among for subject in decreasing order, and is similarly defined. The rank-based statistic of marker is then calculated to be the absolute value of the mean difference between and ; that is, A relevant marker then should possess a large value of . Note that the concept of “rank-over-variable” we adopt here to construct and is more appropriate than the traditional rank-over-sample methods to analyze large--small- datasets as described in Section 1.
In the construction of (3), all the data points are subtracted by the overall mean before ranking over variables within a subject. We note that this centering is critical. Figure 1 shows a simple example to illustrate its necessity. Consider four markers where one (with the symbol ) is relevant and the remaining (with the symbols , , ) are irrelevant to disease status. In control group, we assume that the mean of relevant marker () is larger than that of other three markers (Figure 1(a)). In case group, the means of irrelevant markers is identical to that in control group, while there is a large mean shift of marker . If we are only concerned about the order of the four markers in each group, we cannot observe any difference (Figure 1(a)). Obviously, rank-over-variable method without centering by means will fail to identify the relevant marker in this situation. Instead, supposing that all data points are subtracted by the overall means, the irrelevant markers will be close to zero while the relevant markers will be in the opposite directions and far away from zero as shown in Figure 1(b). In fact, the null hypothesis of the rank-over-variable method is that the interrelationships among the “centered markers” in case and control groups are the same; that is,
: The orders of centered markers within two groups are identical.
Violation of the null hypothesis then indicates the existence of some relevant features, and those features can be reasonably identified by the ranking score (3).
If there are fewer irrelevant markers, the relevant markers might be more likely to be ranked in the top list. In most cases, however, the proportion of relevant markers is much lower than that of irrelevant markers, and the performance of the ranking list directly based on may not be satisfactory. To enhance the detection power of the rank-based statistic , we further apply the techniques of “random subset” and “rerank” [9, 10] as described later. Here we use to denote an indicator function.
Algorithm 1 (random subset).
(1) Randomly select a marker index subset of without replacement to form the partial dataset with the selected markers.
(2) Calculate in (3) based on the dataset obtained in Step 1.
(3) Repeat Steps 1-2 for , and output the adjusted rank-based statistic
The idea of random subset is intuitive: calculation of with fewer irrelevant variables should be more efficient. Following the strategy of Chang and Chen , the size of subsets is chosen to be half of the number of markers in the original dataset. To take into account all combinations of markers and to ensure each marker is included with a sufficiently large number of times, the procedure is repeated times.
The adjusted rank-based statistic , however, is still calculated with the all markers being involved. Considering a ranking list of markers constructed by (4), we can reasonably regard the low-ranked markers as irrelevant markers. If we drop those irrelevant markers, the relevant markers might be more likely to be ranked in the top of the list. This fact motivates us to further consider the technique of “rerank,” and the algorithm is described later.
Algorithm 2 (rerank). Initialize. and
(1) Calculate defined in (4) and where is the number of markers in .
(2) Based on , collect the top markers in to form .
(3) Repeat Steps 1-2 until , and output
In the rerank procedure, we can reasonably expect that a relevant marker would be recalculated many times. It is straightforward to sum up all adjusted rank-based statistics from each iteration, and then a relevant marker would possess a large score. However, there are two parts that should be modified. Firstly, note that the magnitude of rank-over-variable statistics will be influenced by the number of markers () under consideration. To make statistics from different iteration comparable, we use instead of in the rerank algorithm. Secondly, if the averaged score in an iteration is large, it implies that this iteration includes more markers with good separability of disease status. To implement this idea, we use the weight defined in (5) to quantify the importance of each iteration. The final score from rerank technique is therefore defined as a weighted sum in (6). Based on the ranking list constructed by , researchers can select top-ranked markers as candidates for further evaluation. The flowchart of the proposed rerank approach is placed in Figure 2. In practice, the choice of depends on research funding, prior knowledge, and so forth. A data-driven approach to determine is developed in Section 2.2.
Remark 3. Both “random subset” and “rerank” are computationally demanded. To increase the computation speed, we suggest to select the top markers by -test, and the rerank approach is only implemented on these markers to identify candidate features. It is verified in our simulation studies that this preprocessing does not affect the performance heavily.
2.2. Selection of
Given a ranking list, researchers can select top-ranked markers as candidates for further evaluation. For example, in our bipolar study in Section 3.2, it is allowed to select due to the limited budget. In the case of having no prior knowledge about , value and false discovery rate (-value) are commonly used indices for feature selection. However, when the sample size is extremely small (e.g., 8 case and 8 control pools in our bipolar dataset), these methods may not be ready to be applied. In this study, we alternatively propose a method to directly estimate the number of truly relevant markers based on the constructed ranking list.
Consider a ranking list constructed by the rerank approach. If a marker is relevant, it implies that all of the higher-ranked markers in this ranking list are also relevant. Based on the idea of Cook and Yin , instead of computing a standard permutation-based value for each marker, we compute a modified -value to determine whether all of the higher-ranked markers are relevant. The algorithm is described later.
(1) Calculate defined in (6) and .
(2) Construct a ranking list based on . Let be the index such that is the th largest.
(3) For , (i)randomly permute the class labels on those markers with indices to form a partially permuted dataset, (ii)based on the dataset obtained from Step 3(i), calculate defined in (6) and , (iii)repeat Step 3(i)-(ii) for , and output
We now describe the rationale of this algorithm and how to use the values to determine . Assume is the number of truly relevant markers and we are given a correct ranking list. Firstly, in the population level, it is obvious that is an increasing function of provided the ranking list is correct. When , the algorithm actually permutes all irrelevant markers to form the permuted data, while the relevant markers are not permuted. In this case, the permuted data should behave very similarly to the original one and, hence, the distribution of should be identical to that of . We thus expect that the value of is close to 0.5. Moreover, when , is expected to be increasing uniformly in , since we are including markers without separation abilities, and finally to reach unity when . On the other hand, when , the relevant markers are permuted and the value of can be hardly as large as . In this case, the value should be lower than 0.5, and the pattern of should be far away from that of . Based on the previous properties, we thus suggest to choose The proposed selection criterion will be evaluated by a simulation study as described later.
We conduct a simulation study with 2000 markers for equal numbers of case and control groups. All of the markers in both groups follow standard normal distribution except markers in the case group that are distributed as . With 100 cases and 100 controls, Figure 3(a) shows that the curve rises rapidly before reaching 0.5. After passing , the curve increases uniformly with increasing . An obvious change point at with suggests that is a suitable choice. Figure 3(a) also shows that the recovery proportion is 100% when . When the sample size is merely 10 (Figure 3(b)), although there is no obvious change point, it still shows that the corresponding of is close to . The simulation results suggest that we can utilize the formula (8) to select a possible value of , after which the markers are treated as irrelevant.
3. Numerical Studies
The proposed rerank approach is applicable to analyze large--small- datasets with continuous measurements. One application is the pooling-based genomewide association (GWA) study dataset. Instead of individual genotyping which is more expensive, using pooled DNA samples is an effective strategy to reduce the costs of GWA studies. In a pooling-based GWA study, the sample is genotyped in pools of individuals instead of individually genotyping. In particular, the data points () from a pooling-based GWA study are the estimated allele frequencies for SNP of pool in the case (control) group. However, it will generate an ultrahigh dimensional dataset with extremely small sample size. For example, there are 249,473 markers but only 8 case and 8 control pools available in our bipolar dataset. Moreover, the additional measurement error from the pooling process and the existence of outliers have the potential to decrease detection power. In this situation, the proposed rerank approach is more suitable to deal with the large--small- datasets and is more robust to the pooling error and outliers. These facts will be confirmed by the following numerical studies.
3.1. Simulation Studies Using GAIN-MDD Dataset
We simulate DNA pooling datasets from a real individual genotype dataset called GAIN-MDD dataset, which was accessed through the Genetic Association Information Network (GAIN) studies database of Genotypes and Phenotypes (dbGaP) for major depressive disorder (MDD) [12, 13]. There are 416,170 SNPs with 1673 cases and 1721 controls after quality control. We first implement the basic case/control association test by PLINK  to the original GAIN-MDD dataset and then define the top 100 SNPs as the truly relevant SNPs. To simulate a pooling-based GWA dataset, case pools and control pools are constructed by randomly selecting cases and controls from GAIN-MDD dataset, where is the pooling size. Let and be the minor allele frequency (MAF) for SNP of pool in the case and control groups, respectively. To mimic the existence of pooling error and outliers, the observed MAF is generated by and is similarly defined. Collect and to form a simulated DNA pooling dataset. The anticipated aim of this simulation study is to recover the 100 truly relevant SNPs by analyzing the simulated DNA pooling dataset. We repeat simulation studies 100 times and report the averaged number of truly relevant SNPs identified in the top ranking list of each method. We use , , and for rerank approach. The SAM is implemented by the samr R package (from http://www-stat.stanford.edu/~tibs/SAM/) . As to the setting of , there are three situations considered: (A) without pooling error and outliers (); (B) with pooling error (); and (C) 1% data points are outliers with , and otherwise.
Simulation results for equal numbers of case and control pools are shown in Figure 4. When , under the case of (each with size ), Figure 4(a) shows that the performance of rerank approach is better than AUC for any and is better than -test except for large . Note that the 100 truly relevant SNPs are selected based on chi-square test, which is similar to -test, and it is reasonable for -test to obtain similar ranking list of susceptible SNPs, even when data was aggregated. For the case of (each with size ), Figure 4(b) shows that -test and AUC have worse performances due to the instability in estimating variances and high proportion of tied values, respectively. In contrast, the proposed rerank approach does outperform these two methods for any given . In the presence of pooling errors (Figures 4(c) and 4(d)) and outliers (Figures 4(e) and 4(f)), the performances of all methods become worse, but a similar pattern can be observed. The similar patterns can be also observed when the numbers of case and control pools are unequal, except that the performances of all methods become worse simultaneously (Figure 5).
In Figures 4 and 5, we also plot the results of SAM with various choices of , and with the estimated by samr R package . Observing the shaded area from SAM with various , SAM has a chance to identify more truly relevant SNPs than other methods, provided that we can accurately choose the optimal (which corresponds to the highest line of the shaded area). Unfortunately, is unknown in advance and there is no guarantee that the suggested algorithm by  can choose . See the dashed line in Figures 4 and 5, which is far from the optimal result of SAM. Moreover, the wide range of shaded areas indicates that the choice of is critical to the performance of SAM, especially for small sample size and in the presence of outliers. On the other hand, the performance of the rerank approach is similar to that of SAM with optimal for small and is better than SAM with the estimated for a wide range of .
Another advantage of our rerank approach is its insensitivity to the selection of tuning parameters. To see this, we further report the simulation results for various (the number of prescreened SNPs) and (the percentage of rerank) values under equal numbers of case and control pools and . Figure 6(a) suggests that we should choose a conservative value of (e.g., larger than 5000) especially for small sample size (), although there is no obvious difference for larger sample size (). Figure 6(b) shows that the performance is not sensitive to except for the case of . These simulation results then suggest to use a conservative value of and , and the performance of rerank approach is guaranteed. In summary, the rerank approach is more robust to small sample size, pooling error, and outliers and is insensitive to the selection of tuning parameters.
3.2. Bipolar Dataset
In this subsection, we demonstrate a real data analysis using the proposed rerank approach. The dataset is from a two-stage GWA study to identify common variants for the association with bipolar disorder . The bipolar disorder patients were recruited from three hospitals in southern Taiwan from 2008 to 2010. Healthy controls were recruited from the community through advertisements. At Stage 1, a genomewide screen using Illumina HumanOmini1-Quad chip with 970,342 SNPs was performed by DNA pooling with 8 case and 8 control pools constructed from 200 patients and 200 controls. Among the initial 970,342 SNPs, we exclude SNPs if they are (1) on sex chromosome, (2) failed genotyping, (3) monomorphic, (4) with call rate <0.8, or (5) with MAF <0.05. After quality control filtering, there remain 249,473 SNPs. The rerank approach is then applied to evaluate the association for each SNP and to construct a ranking list. One hundred top-ranked SNPs are selected to design and make a panel with 96 SNPs, which would be individually genotyped in Stage 2 with the original plus additional samples, with the aim of identifying relevant SNPs responsible for bipolar disorder. We also aim to see if the result based on pooling data can be reproduced by individual genotype data, to evaluate the performance of DNA pooling for SNP selection. The flow diagram of this analysis is shown in Figure 7.
Recall the aim of Stage 1 is to design a panel with 96 SNPs for that validation by individually genotyping in Stage 2. The selection process is shown in Figure 8. We first select top 100 SNPs by the rerank approach with , and . Among those SNPs, 52 SNPs do not map to any gene while the remaining SNPs can map to 43 genes totally. According to previous studies, etiology of bipolar disorder involves neurotransmitter, neuronal system, immune function, and brain development. Among the 43 genes, we only focus on 8 genes that are associated with brain or neuron. They can be categorized into different biological functions, such as brain-specific chemokines or neurokines and receptor or ligand that regulates neuronal positioning or axon guidance. We next select 81 tag SNPs for the 8 genes based on Tagger . In addition, we choose 15 SNPs that are top-ranked but cannot map to any gene. The total 96 SNPs are conducted in a panel for Stage 2. Using the individual genotype data from Stage 2, the association test for each SNP is implemented in PLINK by fitting simple logistic regression under allelic, dominant, recessive, and additive genetic models, respectively .
To evaluate the reproducibility of the findings from Stage 1, Table 1 shows the association analysis results for the 16 overlapping SNPs in both Stage 1 and Stage 2, where the odds ratio (OR) and -value in Stage 2 are based on the genetic model with the most significant result. Among the 16 SNPs, 13 of them attain significant level in Stage 2 wherein 6 markers are positively relevant to bipolar disorder (OR = 1.4~1.5) and 7 markers are negatively relevant (OR = 0.5~0.7). One can see that the analysis results from Stage 1 and Stage 2 are consistent. It implies that the susceptible markers identified by our rerank approach have high reproducibility even using the pooled DNA data. Those relevant markers (), however, are listed in the much lower rank of -test. The poor performance of -test can be improved by SAM, but it still could not perform as efficient as the rerank approach did. For example, the SNP 3, SNP 6, and SNP 8 have very small -values (0.00085, 0.00651, 0.00785), but not in the top list of SAM.
To validate the findings from Stage 1, we further conduct a set-based analysis for the 8 selected genes under different genetic models of association test by PLINK  using individual genotype data from the original plus additional samples. Table 2 shows that the Gene 5 attains 5% significance level under all models of set-based analysis from Stage 2. Under some genetic models, the Gene 4, Gene 6, Gene 7, and Gene 8 are also significant.
Based on the previously mentioned replication and validation steps by analyzing individual genotype data, our rerank approach indeed has the ability to efficiently detect associations using pooled DNA data. Researchers can focus on these candidate SNPs or genes for further biological studies.
In this study, we propose a robust rerank approach to create a ranking list for feature selection, which comprises three components: (1) rank-based statistic (rank-over-variable), (2) random subset, and (3) rerank. The rank-based statistic is the main scoring function for quantifying association strength, which is motivated by the Rank Test of Alvo et al. . We also apply the techniques of random subset and rerank [9, 10] iteratively to enhance the detection power of rank-based statistic. The combination of these three components demonstrates good performance and robustness in both simulation and real pooling-based GWA study datasets. In addition to the pooling-based GWA study datasets, our rerank approach can be applied to any large--small- datasets with continuous measurements to select differential features between two groups, such as gene expression datasets and biomarker datasets. It provides researchers a sizeable number of differential features for further studies.
In the rerank approach, it involves an important concept: rank-over-variable. The advantage is not only to avoid tied values for ranking in the large--small- situation, but the information of correlations among features can also be taken into account during the ranking process. In other words, although the rank-based statistic is defined as the mean difference of rank values, it is likely to have the ability to detect both mean and correlation changes between two groups . It is interesting to investigate this mechanism in a future study.
The GWA dataset used in Section 3.1 was accessed through the Genetic Association Information Network (GAIN) database of Genotypes and Phenotypes (dbGaP) for major depressive disorder (MDD), accession no. phs000020.v2.p1 (http://www.ncbi.nlm.nih.gov/projects/gap/cgi-bin/study.cgi?study_id=phs000020.v2.p1). The authors thank Po-Chang Hsiao for his IT assistance and Li-Chung Chuang and Chung-Feng Kao for helping in downloading and preprocessing the GWA dataset. This research was supported by the National Science Council (NSC 101-2118-M-002-003-) to Dr. H. Hung (PI) and National Health Research Institute (NHRI-EX102-9918NC) grants to Dr. P. H. Kuo (PI).
- P. Cunningham, Dimension Reduction, pp. 91–112, chapter 4, Springer, Berlin, Germany, 2008.
- H. Hotelling, “Analysis of a complex of statistical variables into principal components,” Journal of Educational Psychology, vol. 24, no. 6, pp. 417–441, 1933.
- I. Guyon and A. Elisseeff, “An introduction to variable and feature selection,” The Journal of Machine Learning Research, vol. 3, pp. 1157–1182, 2003.
- Y. Saeys, I. Inza, and P. Larrañaga, “A review of feature selection techniques in bioinformatics,” Bioinformatics, vol. 23, no. 19, pp. 2507–2517, 2007.
- J. L. DeRisi, V. R. Iyer, and P. O. Brown, “Exploring the metabolic and genetic control of gene expression on a genomic scale,” Science, vol. 278, no. 5338, pp. 680–686, 1997.
- V. G. Tusher, R. Tibshirani, and G. Chu, “Significance analysis of microarrays applied to the ionizing radiation response,” Proceedings of the National Academy of Sciences of the United States of America, vol. 98, no. 9, pp. 5116–5121, 2001.
- R. Breitling, P. Armengaud, A. Amtmann, and P. Herzyk, “Rank products: a simple, yet powerful, new method to detect differentially regulated genes in replicated microarray experiments,” FEBS Letters, vol. 573, no. 1–3, pp. 83–92, 2004.
- M. Alvo, Z. Liu, A. Williams, and C. Yauk, “Testing for mean and correlation changes in microarray experiments: an application for pathway analysis,” BMC Bioinformatics, vol. 11, no. 1, article 60, 2010.
- F. Chang and J.-C. Chen, “An adaptive multiple feature subset method for feature ranking and selection,” in Proceedings of the 15th Conference on Technologies and Applications of Artificial Intelligence (TAAI '10), pp. 255–262, IEEE Computer Society, November 2010.
- G. C. Tseng and W. H. Wong, “Tight clustering: a resampling-based approach for identifying stable and tight patterns in data,” Biometrics, vol. 61, no. 1, pp. 10–16, 2005.
- R. D. Cook and X. Yin, “Dimension reduction and visualization in discriminant analysis (with discussion),” Australian and New Zealand Journal of Statistics, vol. 43, no. 2, pp. 147–199, 2001.
- T. A. Manolio, L. L. Rodriguez, L. Brooks et al., “New models of collaboration in genome-wide association studies: the Genetic Association Information Network,” Nature Genetics, vol. 39, no. 9, pp. 1045–1051, 2007.
- M. D. Mailman, M. Feolo, Y. Jin et al., “The NCBI dbGaP database of genotypes and phenotypes,” Nature Genetics, vol. 39, no. 10, pp. 1181–1186, 2007.
- S. Purcell, B. Neale, K. Todd-Brown et al., “PLINK: a tool set for whole-genome association and population-based linkage analyses,” American Journal of Human Genetics, vol. 81, no. 3, pp. 559–575, 2007.
- P. H. Kuo, J. R. Liu, M. K. Lu, R. B. Lu, and H. Hung, “A genome-wide association study of bipolar disorder using DNA pooling,” Asian Journal of Psychiatry, vol. 4, Supplement 1, p. S38, 2011.
- J. C. Barrett, B. Fry, J. Maller, and M. J. Daly, “Haploview: analysis and visualization of LD and haplotype maps,” Bioinformatics, vol. 21, no. 2, pp. 263–265, 2005.