Research Article  Open Access
Ensembling Variable Selectors by Stability Selection for the Cox Model
Abstract
As a pivotal tool to build interpretive models, variable selection plays an increasingly important role in highdimensional data analysis. In recent years, variable selection ensembles (VSEs) have gained much interest due to their many advantages. Stability selection (Meinshausen and Bühlmann, 2010), a VSE technique based on subsampling in combination with a base algorithm like lasso, is an effective method to control false discovery rate (FDR) and to improve selection accuracy in linear regression models. By adopting lasso as a base learner, we attempt to extend stability selection to handle variable selection problems in a Cox model. According to our experience, it is crucial to set the regularization region in lasso and the parameter properly so that stability selection can work well. To the best of our knowledge, however, there is no literature addressing this problem in an explicit way. Therefore, we first provide a detailed procedure to specify and . Then, some simulated and realworld data with various censoring rates are used to examine how well stability selection performs. It is also compared with several other variable selection approaches. Experimental results demonstrate that it achieves better or competitive performance in comparison with several other popular techniques.
1. Introduction
Variable selection is a classical problem in statistics and has enjoyed increased attention in recent years due to a massive growth of highdimensional data across many scientific disciplines. In modern statistical applications, the number of variables or covariates often exceeds the number of observations . In such settings, the true model is often assumed to be sparse, in the sense that only a small proportion of the variables actually relates to the response. Thus, variable selection is fundamentally important in statistical analysis of highdimensional data. With a proper selection method and under suitable conditions, we are able to build a good model to interpret the relationship between covariates and our interested outcome more easily, to avoid overfitting in prediction and estimation, and to identify important variables for applications or further study.
For variable selection, many researchers focus on multiple linear regression models. To emphasize that variable selection methods are useful for other statistical models as well, we use a different statistical model, that is, a Cox’s proportional hazards model (abbreviated as Cox model) [1], as the platform in this context. The Cox model was first proposed for exploring the relationship between the survival of a patient and some explanatory variables. As a matter of fact, the Cox model [2, 3] nowadays is one of the most commonly used forms in semiparametric models and it can not only solve the issues of censored data, but also analyze the influence of various factors on survival time simultaneously. A brief mathematical description of the Cox model is given as follows.
Suppose that there are observations of survival data. For an individual , denotes its survival time and stands for the observed data for the covariates. At the same time, is a censoring indicator variable, where means that is rightcensored. Let be the hazard rate at time ; the generic form of a Cox’s proportional hazards model can be expressed aswhere is a dimensional unknown coefficient vector and is the baseline hazard function, that is, the hazard function at time when all covariates take value zero. In general, can be estimated by maximizing partial likelihood function. For convenience, we assume below.
Like linear regression models, traditional methods such as subset selection [4, 5], forward selection, backward elimination, and a combination of both are among the most common methods for selecting variables in a Cox model. However, these methods will have difficulty in computation when faced with highdimensional data. Therefore, some other methods have been proposed to overcome this problem. After lasso (least absolute shrinkage and selection operator) [6] was first proposed for linear regression models, Tibshirani [7] extended it to the Cox model. Later on, many scholars [2, 3, 8–12] developed some penalized shrinkage techniques like SCAD [13] and adaptive lasso [14] specially for Cox models.
Although the abovementioned variable selection methods have been shown to be successful in theoretical properties and numerous experiments, their performance strongly depends on the proper setup of the tuning parameter. On the other hand, these approaches may be unstable (especially in the situation of highdimensional data). Breiman [15] proved that uncertainty can lead to more prediction loss. What is more important, small changes in data can result in that the same method selects different models. This makes the subsequent interpretation difficult and unreliable. In order to obtain more stable, accurate, and reliable variable selection results, ensemble learning [16, 17] is one kind of extremely potential technologies.
As a hot research topic in machine learning, ensemble learning is used more and more widely in many fields of natural science and social science in last two decades. The powerful advantages of ensemble learning lie in improving the generalization capacity and enhancing robustness in the process of learning. Its main idea is to obtain a number of different base learning machines by running some simple learning algorithm and then combine these base machines into an ensemble learning machine in some way. Generally, the base learning machines should have strong generalization capability on one side, and they should also complement each other on the other hand.
The ensemble approach for statistical modeling was first proposed for solving prediction problems, aiming to maximize prediction accuracy. Inspired by this idea, Zhu and Chipman [18] applied bagging ensemble approach to handle variable selection problems, aiming at maximizing selection accuracy. Meanwhile, they pointed out that there is much difference between “prediction ensembles” (PEs) and “variable selection ensembles” (VSEs). More recently, ensemble learning methods have attracted more attention on coping with variable selection problems since they can greatly improve the selection accuracy and lessen the risk to falsely select unimportant variables and simultaneously overcome the instability of traditional methods in the highdimensional data analysis. Because of these benefits, there are more and more researches applying ensemble learning to variable selection and putting forward some novel approaches. As far as we know, existing VSE techniques mainly include PGA (parallel genetic algorithm) [18], stability selection [19], BSS (bagged stepwise search) [20], random lasso [21], ST2E (stochastic stepwise ensemble) [22], TCSL (tilted correlation screening learning) [23], RMSA (random splitting model averaging) [24], SCCE (stochastic correlation coefficient ensemble) [25], and PST2E (pruned stochastic stepwise ensemble) [26]. It is noteworthy that these algorithms are mainly designed for handling variable selection problems in linear regression models. Only Zhu and Fan [20] investigated the performance of BSS and PGA in the Cox model.
Through analyzing these VSE techniques, it can be found that their success primarily lies in producing multiple importance measures for each predictor. By simply averaging these measures across multiple trials, the noise variables can be more reliably distinguished from the informative ones. In this process, the strength to select important variables and the diversity between the importance measures need to be preserved simultaneously [20, 22]. Stability selection applies subsampling (or bootstrap) to a selection method like lasso to improve its performance. In fact, it is an extremely general ensemble learning technique for identifying important variables. Due to the characteristics of lasso, it is very efficient in highdimensional situations. Another good property of stability selection is that it provides an effective way to control false discovery rate (FDR) in finite sample cases provided that its tuning parameters are set properly. Due to its versatility and flexibility, stability selection has been successfully applied in many domains such as gene expression analysis [24, 27–29]. Nevertheless, we have not found any literature about applying stability selection to a Cox model. Therefore, in this paper we would like to extend it to the situation of Cox models. At the same time, we also discuss how to set appropriate values for the involved parameters so that stability selection achieves its best performance.
The remainder of the paper is described as follows. In Section 2, the details for applying stability selection to the Cox model are described. We also provide an explicit way to set its involved parameters. In Section 3, some numerical experiments were conducted to study the impact of on the behavior of stability selection and to compare its performance with other variable selection approaches for the Cox model. In Section 4, some real examples are analyzed to further study the effectiveness of stability selection. Finally, some conclusions are offered in Section 5.
2. Stability Selection Algorithm for the Cox Model
In this paper, we consider stability selection with lasso as its base learner. Lasso [6] is one of the most effective techniques to deal with highdimensional linear regression problems with . With respect to its application in Cox models, the core idea is to maximize the partial likelihood minus the penalty function. For convenience, suppose that there are unique failure times, say, , among the observations . Let denote the index of the observation failing at time . The lasso algorithm needs to maximizeunder the constraint . In (2), is the set of indices, , with (i.e., the observations are at risk at time ). Equivalently, the estimate of can be obtained aswhere is the regularization parameter which controls the tradeoff between the model fitting and the coefficient shrinkage degree. At present, there are several efficient algorithms [7, 30] (such as cyclical coordinate descent) to get in (3). We refer readers to the related literature for more details about the optimization strategy.
In applications, we need to first set a sensible region, say, , for the regularization parameter in lasso. Notice that lasso will choose all variables (i.e., full model) for while choosing none of the variables (i.e., null model) for . By taking candidate values in , that is, , lasso generally employs 5fold or 10fold crossvalidation to select an optimal value of , say . Then, the variables which have nonzero coefficient estimation under are determined as important variables. Although lasso with being specified in this way has good prediction performance, much evidence [14, 19, 21] has shown that it tends to choose more variables than necessary (i.e., higher FDR).
To eliminate this drawback of lasso, Meinshausen and Bühlmann [19] developed stability selection which works by choosing variables whose selection probabilities are large as important ones. In reality, the selection probability can be estimated by running lasso on multiple different sets. These sets can be obtained via subsampling from the given set. Specifically, stability selection first estimates the probability that variable is important for each regularization parameter , and then takes the maximum probability over as the important measure for . Eventually, it selects important variables by a preset threshold . The detailed steps of stability selection algorithm for the Cox model are listed in Algorithm 1.

As argued by Meishausen and Bühlmann [19], the prominent advantage of stability selection is to control FDR under finite sample size and simultaneously to weaken the theoretical assumptions that are required to achieve variable selection consistency (i.e., the probability that the fitted model includes only truly important variables is tending to one when ). Let be the number of falsely selected variables with stability selection; Meinshausen and Bühlmann [19] have proved that, under some mild assumptions, for arbitrary , the expectation of satisfieswhere represents the average number of variables selected by base learner. Roughly speaking, we can set any two parameters of , , and and determine the remaining one according to the above inequality. For example, let and ; then can be specified as in which denotes taking the smallest integer larger than or equal to . As stated in [19], is recommended to take value in the range of and the results tend to be similar. As far as is concerned, it can be set by users according to the level of FDR that they would like to control. In general, small means to control FDR strictly so that less noise variables are falsely included. Nevertheless, too small may cause some truly important variables omitting in the final model. On the other hand, can be larger if one can accept a little higher FDR to make sure that all important variables can be included. Regarding , it should be no less than the number of truly important variables. Because we have no means to know the number of truly important variables in advance, however, one can first specify and and let be determined automatically.
As mentioned earlier, the crucial role of stability selection is to reduce the FDR of lasso (i.e., to exclude noise variables more reliably). Intuitively, it is still difficult to identify the true sparse model if too much noise variables are falsely included every time. Thus, a minimum value of (or ) needs to be specified for stability selection so that every time at most variables are chosen when . Subsequently, only the ’s lying in the interval are taken as candidate values of to implement lasso in each trial.
According to our experience, the setting of as well as is crucial to the success of stability selection. However, we cannot find any detailed instruction in related literature [19, 27, 28] about how to set them. Moreover, all the existing literature related to stability selection has not discussed how to apply it in Cox models. Here, we would like to provide an explicit way to cope with this problem in the framework of Cox models. According to the proposal in [30], we can first set for lasso in a Cox model asin whichHere, is the number of subjects (observations) at risk at time and is the set of indices, , with (i.e., the times for which observation is still at risk). Subsequently, we can set with for and for . In order to create candidate values for , we can set for .
Next, the parameter in stability selection can be determined by
Equation (7) implies that must be chosen to ensure that lasso selects at most variables for each . Specifically, one can begin with and decrease gradually until lasso detects variables as important (i.e., variables having nonzero coefficients). The value of obtained at this point is exactly defined in (7). Then, only the candidate values lying in are considered as the candidate values for in lasso to execute variable selection.
3. Experimental Studies
With simulated data, some experiments are conducted in this section to investigate the impact of on the behavior of stability selection in a Cox model and to compare it with several other variable selection approaches. In order to maintain consistency and comparability, we set ensemble size as 200. Each simulation was run 100 times to estimate the evaluation of a method. To simplify notations, we abbreviated stability selection as StabSel. Regarding lasso, we made use of 10fold crossvalidation to determine its optimal regularization parameter.
3.1. Simulation 1: Influence of
Meinshausen and Bühlmann [19] stated that the threshold value is a tuning parameter whose influence is small as long as it is in the range of . According to our experience, has more significant effect in comparison with the parameter . When and are fixed, small will make lasso select more variables in each path. As a result, some noise variables may be falsely considered as important ones (i.e., high false positive rate). On the other hand, the noise variables can be safely filtered out by setting a large . However, this may lead us to miss some signal variables (i.e., high false negative rate). Thus, plays a role in controlling the tradeoff between false positive rate and false negative rate of StabSel. Due to this consideration, we fixed and report results for several values of in the first experiment.
Suppose that there are variables, , with each generated from the standard normal distribution . Furthermore, the variables are correlated with for all . The response was generated from an exponential distribution whose hazard function iswhere the true coefficient vector . Clearly, only three variables , , are truly important and the remaining ones are unimportant. We took and conducted three experiments with censoring rates 0%, 20%, and 40%, respectively. For the censoring mechanism, a censoring time is generated independently and uniformly from for each observation. If , we replaced with and then let . Here, the parameter was chosen to achieve some desired censoring rates. For example, corresponds to censoring rate and corresponds to censoring rate. Aiming at evaluating the performance of StabSel for a given , we computed the selection frequency of StabSel in each case. Specifically, the selection frequency was calculated as, among 100 simulations, the minimum, median, and maximum number of times that the important and unimportant variables (and) are selected by StabSel, respectively. Interested readers can refer to [26] for the detailed definition of selection frequency. Table 1 summarizes the results for the cases with different centering rates.

The results in Table 1 demonstrate that StabSel using a relatively large performs slightly better in excluding unimportant variables. However, the side effect is that it more likely misses some truly important variables. In other words, StabSel controls false discovery rate (or false positive rate) quite effectively with a relatively large , but this will cause it to behave poorly in terms of catching important variables. To improve its selection accuracy, we must reduce . Nevertheless, this inevitably allows more false discoveries. In practice, it is worthy of choosing an appropriate value for depending on whether our emphasis is more on false positive rate or false negative rate. Moreover, we need to pay more attention to the tuning of if the censoring rate is high.
3.2. Simulation 2: Performance Comparison on a Cox Model with HighDimensional Data
In this subsection, we concentrated on applying StabSel and lasso to a Cox model with highdimensional data. To generate the design matrix, the following two simulated datasets were generated by following the strategy in [19].
Case 1. , where and , .
Case 2. , for , where , and , .
Moreover, we created sparse regression vectors by setting for all , except for a small variable set . For all , we chose the coefficient independently and uniformly in and let the size varying between 4 and 10. Here, we employed the method used in Section 3.1 to achieve the censoring rates 0% and 20%. Then, a Cox model was constructed by (8).
To compare the power of StabSel and lasso to ranking variables, we adopted the strategy utilized by [19], that is, focusing on the probability that variables in can be recovered correctly, where . For lasso, this means that there is a regularization parameter such that at least variables in are selected while all variables in are not selected. For stability selection, it stands for the fact that variables with highest selection frequency are all in . In this example, we fixed the threshold value and to determine a proper value for .
The top two subplots in Figure 1 correspond to the situation of while the bottom two subplots illustrate the results for . Notice that the latter task is more challenging than the former one. When the covariates are independent in Case 1, lasso performs satisfactorily and the advantage of StabSel is not significant. In Case 2, the dominance of StabSel over lasso to identify important variables more correctly can be clearly seen, especially when faced with censored data. In the more challenging task in which more important variables are required to be ranked ahead (i.e., ), the superiority of StabSel is more significant. In conclusion, this experiment shows that StabSel is indeed helpful to enhance the ranking ability of lasso.
3.3. Simulation 3: Performance Comparison with Several Other Methods
Finally, we considered a simulated dataset used in [20]. There are observations and predictor variables. Each predictor was generated according to The response vector was generated from an exponential distribution with hazard function As for the variables other than , , , the coefficient is zero. Altogether, three simulation studies were conducted with censoring rates 0%, 20%, and 40%, respectively. For StabSel, we fixed and . As mentioned in Section 2, the number of variables that lasso selects in each trial should be at least larger than the number of truly important variables. Thus, we increased the factor multiplying in because is small in this simulation. We compared it with traditional stepwise search as well as some VSE techniques including BSS [20], PGA [18], RSMA [24], and ST2E [22]. The parameters involved in these methods were set according to the related literature.
Table 2 summarizes the selection frequencies of IV and UIV for each approach. The results demonstrate that although PGA performs better to exclude unimportant variables, it may miss some truly important variables. On the other hand, RSMA, ST2E, and StabSel can identify almost the same number of important variables; the difference only lies in the exclusion of unimportant ones. In this aspect, StabSel is observed to behave the best. As for BSS, its ability to guard against noise variables seems to be worse than the others although it works well to identify IVs.

In order to see more clearly the differences among the considered approaches, we computed the average selection rate of IV and UIV. For IV, it was computed as the selection probabilities averaged over all important variables. The metric was similarly estimated for UIV. The results are illustrated in Figure 2. The top three subplots are IVs while the bottom three ones are for UIVs. From Figure 2, we can come to some conclusions similar to those drawn from Table 2.
At the same time, we also utilized several other metrics to extensively evaluate each method. First, we computed the selection success rate [13]. Given an algorithm, it refers to the fraction of times among 100 runs that the algorithm correctly identifies the true model (i.e., the model only includes ). Second, the true positive rate (TPR) and true negative rate (TNR) of each method were considered. In particular, TPR and TNR are as follows:where is the estimated coefficient vector in the th simulation. In addition, and represent the size of IV and UIV, respectively. The method “Oracle” corresponds to fitting a Cox model with only variables , , and . Usually, a good variable selection method should produce results as close as possible to those of Oracle.
It can be seen from Table 3 that stepwise method is hopeless to select variables since it can hardly find the true model. Among the VSE algorithms, StabSel always reaches the largest selection success rate, especially when the censoring rate is high. On the other hand, StabSel tends to achieve a model size closest to that of Oracle. As far as the prediction performance is concerned, StabSel almost always outperforms the other approaches.

4. RealWorld Applications
In this section, we applied the compared VSE techniques to three realworld datasets, that is, PBC [31], Lung [32], and Rats [33]. These real datasets were taken from the R package survival. For the original PBC and lung sets, we simply ignored the observations containing missing data. In these situations, there are no means to know which variables are truly important or not. Aiming at evaluating the selection behavior of each method, we took the original variables as truly important ones (i.e., IVs). Then, some irrelevant variables were artificially added to these sets by following the strategy used in [25, 34]. These irrelevant variables were generated from a uniform distribution on the interval . Table 4 lists the main characteristics of the used three datasets.

Analogous to the situation of simulation studies, the ensemble size was set as . The parameters involved in each method were set similarly to those used in simulations. For each dataset, the experiment was repeated 100 times. In each replication, a training set was randomly drawn from the given set with size being specified in Table 4. The rest of observations was then used as a test set to evaluate the prediction performance measured with Cindex [35] (i.e., concordance index). In particular, we applied each algorithm to the training set to perform variable selection. Based on the selected variables, the parameters in the corresponding model were estimated and the Cindex was estimated on the test set. Table 5 shows the results obtained with each algorithm.

In terms of selection rate, it can be observed from Table 5 that BSS performs well to identify IVs. Nevertheless, it behaves worse to exclude UIVs. On the contrary, PGA shows the lowest selected rate of UIVs while it has the lowest selected rate of IVs. Therefore, BSS and PGA are not ideal selection methods. For the remaining methods, StabSel, RSMA, and ST2E behave similarly in identifying IVs. But when compared with StabSel, RSMA and ST2E include more irrelevant variables. In conclusion, StabSel achieves better performance on variable selection when being evaluated with selection rate.
Furthermore, the results of Cindex in Table 5 reveal that the prediction performance of StabSel is competitive although it is not the best one. Furthermore, almost all ensemble methods tend to have low TPR values in these three real datasets. This is largely due to the fact that we directly consider all the original covariates as IVs among which some are actually uninformative.
5. Conclusions
As an ensemble method, StabSel [19] is the marriage of subsampling with a variable selection algorithm such as lasso. Due to its property of controlling false discovery rate, StabSel has a flexible manner to choose a proper amount of regularization. Another superiority of StabSel over lasso is that it requires less assumptions to achieve variable selection consistency. In this article, we extended StabSel to the Cox model. The specification of significantly affects the performance of StabSel since it controls the balance between false positive rate and false negative rate. We provide an explicit way to set a proper value for in the situation of Cox models. In comparison with other VSE techniques including PGA, BSS, RSMA, and ST2E, StabSel exhibits better selection ability to correctly identify important variables in a highdimensional Cox model. At the same time, StabSel has satisfactory prediction performance. When the censoring rate is high, its advantage is even more significant. Therefore, StabSel can be considered as an alternative to explore the relationship between covariates and survival times in survival analysis.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This research was supported in part by the National Natural Science Foundation of China (Grant nos. 11671317, 11601412, and 11501438), the Natural Science Basic Research Plan in Shaanxi Province of China (Grant no. 2017JQ1034), and the Science Foundation of Xi’an University of Architecture and Technology (Grant nos. RC1438 and QN1508).
References
 D. R. Cox, “Regression models and lifetables,” Journal of the Royal Statistical Society, vol. 34, no. 2, pp. 187–220, 1972. View at: Google Scholar  MathSciNet
 J. Fan and R. Li, “Variable selection for Cox's proportional hazards model and frailty model,” The Annals of Statistics, vol. 30, no. 1, pp. 74–99, 2002. View at: Publisher Site  Google Scholar  MathSciNet
 J. Q. Fan, Y. Feng, and Y. C. Wu, “Highdimensional variable selection for Cox’s proportional hazards model,” IMS Collections, vol. 6, pp. 70–86, 2010. View at: Google Scholar
 D. R. Wang and Z. Z. Zhang, “Variable selection for linear regression models: a survey,” Journal of Applied Statistics and Management, vol. 29, no. 4, pp. 615–627, 2010. View at: Google Scholar
 A. Miller, Subset Selection in Regression, Chapman and Hall/CRC Press, New Work, NY, USA, 2nd edition, 2002.
 R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the Royal Statistical Society: Series B, vol. 58, no. 1, pp. 267–288, 1996. View at: Google Scholar
 R. Tibshirani, “The lasso method for variable selection in the cox model,” Statistics in Medicine, vol. 16, no. 4, pp. 385–395, 1997. View at: Publisher Site  Google Scholar
 H. H. Zhang and W. Lu, “Adaptive Lasso for Cox's proportional hazards model,” Biometrika, vol. 94, no. 3, pp. 691–703, 2007. View at: Publisher Site  Google Scholar  MathSciNet
 J. Huang, S. Ma, and C.H. Zhang, “Adaptive Lasso for sparse highdimensional regression models,” Statistica Sinica, vol. 18, no. 4, pp. 1603–1618, 2008. View at: Google Scholar  MathSciNet
 A. Antoniadis, P. Fryzlewicz, and F. Letué, “The Dantzig Selector in Cox's Proportional Hazards Model,” Scandinavian Journal of Statistics, vol. 37, no. 4, pp. 531–552, 2010. View at: Publisher Site  Google Scholar
 P. Du, S. Ma, and H. Liang, “Penalized variable selection procedure for Cox models with semiparametric relative risk,” The Annals of Statistics, vol. 38, no. 4, pp. 2092–2117, 2010. View at: Publisher Site  Google Scholar  MathSciNet
 C. Liu, Y. Liang, X.Z. Luan et al., “The L1/2 regularization method for variable selection in the Cox model,” Applied Soft Computing, vol. 14, pp. 498–503, 2014. View at: Publisher Site  Google Scholar
 J. Fan and R. Li, “Variable selection via nonconcave penalized likelihood and its oracle properties,” Journal of the American Statistical Association, vol. 96, no. 456, pp. 1348–1360, 2001. View at: Publisher Site  Google Scholar  MathSciNet
 H. Zou, “The adaptive lasso and its oracle properties,” Journal of the American Statistical Association, vol. 101, no. 476, pp. 1418–1429, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 L. Breiman, “Heuristics of instability and stabilization in model selection,” The Annals of Statistics, vol. 24, no. 6, pp. 2350–2383, 1996. View at: Publisher Site  Google Scholar  MathSciNet
 L. I. Kuncheva, Combining Pattern Classifiers, Methods and Algorithms, John Wiley and Sons, New Jersey, NJ, USA, 2014.
 Z. H. Zhou, Machine Learning, Qinghua University Press, Beijing, China, 2016.
 M. Zhu and H. A. Chipman, “Darwinian evolution in parallel universes: a parallel genetic algorithm for variable selection,” Technometrics, vol. 48, no. 4, pp. 491–502, 2006. View at: Publisher Site  Google Scholar
 N. Meinshausen and P. Bühlmann, “Stability selection,” Journal of the Royal Statistical Society: Series B, vol. 72, no. 4, pp. 417–473, 2010. View at: Publisher Site  Google Scholar  MathSciNet
 M. Zhu and G. Z. Fan, “Variable selection by ensembles for the Cox model,” Journal of Statistical Computation and Simulation, vol. 81, no. 12, pp. 1983–1992, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 S. J. Wang, B. Nan, S. Rosset et al., “Random Lasso,” The Annals of Applied Statistics, vol. 5, no. 1, pp. 468–485, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 L. Xin and M. Zhu, “Stochastic stepwise ensembles for variable selection,” Journal of Computational and Graphical Statistics, vol. 21, no. 2, pp. 275–294, 2012. View at: Publisher Site  Google Scholar  MathSciNet
 B. Lin and Z. Pang, “Tilted correlation screening learning in highdimensional data analysis,” Journal of Computational and Graphical Statistics, vol. 23, no. 2, pp. 478–496, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 B. Lin, Q. Wang, J. Zhang, and Z. Pang, “Stable prediction in highdimensional linear models,” Statistics and Computing, vol. 27, no. 5, pp. 1401–1412, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 J. Che and Y. Yang, “Stochastic correlation coefficient ensembles for variable selection,” Journal of Applied Statistics, vol. 44, no. 10, pp. 1721–1742, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 C. Zhang, J. Zhang, and Q. Yin, “A rankingbased strategy to prune variable selection ensembles,” KnowledgeBased Systems, vol. 125, pp. 13–25, 2017. View at: Publisher Site  Google Scholar
 B. Hofner, L. Boccuto, and M. Göker, “Controlling false discoveries in highdimensional situations: boosting with stability selection,” BMC Bioinformatics, vol. 16, no. 1, article 144, 2015. View at: Publisher Site  Google Scholar
 A. Beinrucker, Ü. Dogan, and G. Blanchard, “Extensions of stability selection using subsamples of observations and covariates,” Statistics and Computing, vol. 26, no. 5, pp. 1059–1077, 2016. View at: Publisher Site  Google Scholar  MathSciNet
 K. He, Y. Li, J. Zhu et al., “Componentwise gradient boosting and false discovery control in survival analysis with highdimensional covariates,” Bioinformatics, vol. 32, no. 1, pp. 50–57, 2015. View at: Publisher Site  Google Scholar
 N. Simon, J. Friedman, T. Hastie, and R. Tibshirani, “Regularization paths for Cox's proportional hazards model via coordinate descent,” Journal of Statistical Software , vol. 39, no. 5, pp. 1–13, 2011. View at: Google Scholar
 T. Therneau and P. Grambsch, Modeling Survival Data: Extending the Cox Model, SpringerVerlag, New York, NY, USA, 2000.
 C. L. Loprinzi, J. A. Laurie, H. S. Wieand et al., “Prospective evaluation of prognostic variables from patientcompleted questionnaires,” Journal of Clinical Oncology, vol. 12, no. 3, pp. 601–607, 1994. View at: Publisher Site  Google Scholar
 N. Mantel, N. R. Bohidar, and J. L. Ciminera, “MantelHaenszel analyses of littermatched time to response data, with modifications for recovery of interlitter information,” Cancer Research, vol. 37, no. 11, pp. 3863–3868, 1977. View at: Google Scholar
 A. Mkhadri and M. Ouhourane, “A group VISA algorithm for variable selection,” Statistical Methods and Applications, vol. 24, no. 1, pp. 41–60, 2015. View at: Publisher Site  Google Scholar
 H. Uno, T. Cai, M. J. Pencina, R. B. D'Agostino, and L. J. Wei, “On the Cstatistics for evaluating overall adequacy of risk prediction procedures with censored survival data,” Statistics in Medicine, vol. 30, no. 10, pp. 1105–1117, 2011. View at: Publisher Site  Google Scholar  MathSciNet
Copyright
Copyright © 2017 QingYan Yin et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.