Emerging Trends in Soft Computing Models in Bioinformatics and Biomedicine
View this Special IssueResearch Article  Open Access
Itziar Irigoien, Basilio Sierra, Concepción Arenas, "Towards Application of OneClass Classification Methods to Medical Data", The Scientific World Journal, vol. 2014, Article ID 730712, 7 pages, 2014. https://doi.org/10.1155/2014/730712
Towards Application of OneClass Classification Methods to Medical Data
Abstract
In the problem of oneclass classification (OCC) one of the classes, the target class, has to be distinguished from all other possible objects, considered as nontargets. In many biomedical problems this situation arises, for example, in diagnosis, image based tumor recognition or analysis of electrocardiogram data. In this paper an approach to OCC based on a typicality test is experimentally compared with reference stateoftheart OCC techniques—Gaussian, mixture of Gaussians, naive Parzen, Parzen, and support vector data description—using biomedical data sets. We evaluate the ability of the procedures using twelve experimental data sets with not necessarily continuous data. As there are few benchmark data sets for oneclass classification, all data sets considered in the evaluation have multiple classes. Each class in turn is considered as the target class and the units in the other classes are considered as new units to be classified. The results of the comparison show the good performance of the typicality approach, which is available for high dimensional data; it is worth mentioning that it can be used for any kind of data (continuous, discrete, or nominal), whereas stateoftheart approaches application is not straightforward when nominal variables are present.
1. Introduction
In oneclass classification (OCC), the problem is to classify data when information is available for only one group of observations. Specifically, given one set of data, called the target class, the aim of the OCC methods is to distinguish data belonging to the target class from other possible classes. OCC can be seen as a special type of twoclass classification problem, when data from only one class is considered. This is an interesting problem because there are many real situations where a representative set of labeled examples for the second class is too costly, difficult to obtain, or not available at all. This situation can occur, for instance, in medical diagnosis, where data from healthy or even from nonhealthy patients are extremely hard or impossible to obtain: for example, through mammograms for breast cancer detection [1, 2], the oneclass recognition of cognitive brain functions [3], in prediction of proteinprotein interactions [4], in the lung tissue categorization of patients affected with interstitial lung diseases [5], or in the identification of patients with one or more Nosocomial infections using clinical and other data collected during the survey [6]. Several approaches to OCC have been presented and good overviews can be found in [7–10]. Some of the OCC approaches estimate the density of the reference data and set a threshold on this density, using a Gaussian model, a mixture of Gaussians models, or the Parzen density estimators [11, 12]. Boundary methods, as the centers, NNd [13, 14], and support vector machine SVM [15–18], cover the data set with small balls with equal radii and they make assumptions about the clustering characteristics of the data or their distribution in subspaces. These methods only achieve good results when the target data have the same distribution tendency in all orientations [19]. The reconstruction methods (mean clustering, selforganizing maps, PCA, mixtures of PCAs, and diabolo networks density) make assumptions about the clustering characteristics of the data or their distribution in subspaces, and a set of prototypes is needed (see, e.g., [20]). Many of these methods have dataspecific parameters or assume that data follow a specific model; therefore data knowledge is necessary. Oneclass classification can also be considered as outlier detection, where the classification model can be used to detect the units deviating significantly from the target class. There are some distancebased outlier detection methods [21, 22], which need the computation of the distances between units in the target class and distances between a new unit and their neighbors in the target class, but in contrast with other OCC methods they are more flexible. Some other stateoftheart methods are neural networks [23], Bayesian neural networks [24], or Naive Bayesian classifiers [25]. Recently, in [26] the authors formulated a typicality test and this approach is here applied to the OCC problem. Thus, objects in the target class can be considered as typical, while objects in the negative class can be considered as atypical. In order to evaluate the viability of the typicality approach a comparative study is presented. Five reference stateoftheart techniques, two parametric density methods, the Gaussian and mixture of Gaussians procedures; two nonparametric density methods, the Parzen and naive Parzen procedures; a boundary method, the support vector data description method, are experimentally compared with the typicality approach using biomedical data sets.
The paper is organized as follows. Section 2 presents the six considered OCC procedures that are evaluated for twelve real biomedical data set. The experimental study is summarized in Section 3, while conclusions are drawn in Section 4.
2. OneClass Classification
In this section, the oneclass classification problem is formally stated, and the six considered procedures are reviewed.
2.1. The OneClass Classification Problem
Consider a class , the target class, containing objects and represented by a random vector with probability density function with respect to a suitable measure . Let an object of be represented by a vector containing the values of the measures in features, not necessarily continuous. The OCC problem can be defined as the problem of assigning or not a new object to the target class , when data only from the target class is available. Thus, from data in the target class a classification model should be constructed. The OCC procedures usually consider a training phase using the socalled training data set; that is, either the probability density function or the parameters of the classifier’s model should be determined. In OCC the training data set contains only the observations belonging to , while the testing data set includes the observations from class and other possible class . As in medical care correct diagnosis is very important, it is necessary to evaluate the OCC models, which can be considered as a case/noncase diagnosis where the target class is, for instance, the case class. This diagnosis will misclassify some cases as noncases and some noncases as cases. These two types of misclassifications lead to two important aspects of the performance of the diagnosis, sensitivity, and specificity. As it is known, the sensitivity or true positive rate is the probability that occurs if an object in class is classified as belonging to this class. The specificity or true negative rate is the probability that occurs if an object not belonging to is classified as not belonging to . A very common way of displaying the values of the sensitivity and specificity is by the ROC curve (Receiver Operating Characteristic), which represents the pairs (1specificity, sensitivity). Therefore, the area under the ROC curve, the AUC, lies between 0 and 1 and takes value 1 for a perfect diagnosis and the value 0.5 for random diagnosis, so that AUC values will be useful to evaluate the performance of OCC models [27].
It is important to note that in oneclass classifiers the ability to learn the true characteristics of the data in presence of noise or errors in the feature values is specially important. Furthermore, the number of parameters to be estimated by users should be minimized, and the computational and storage requirements must be in consideration, as there are limiting factors in the use of some of the methods. Finally, oneclass classifiers are determined in the training phase using the training data set; thus the standard OCC procedures may be affected by initial settings.
Next, six oneclass classification methods will be reviewed. We consider five well known and reference OCC methods: two parametric density methods, the Gaussian and mixture of Gaussians; two nonparametric density methods, the Parzen and naive Parzen; a boundary method, the support vector data description. For these methods, we summarized some of their characteristics and references for more details about the construction of the classification model and properties are given. Finally, a nonparametric typicality approach based on distances is considered. As this method has not yet been considered as a oneclass classification procedure, more details about the classification model and properties will be included.
2.2. Gaussian and Mixture of Gaussians
The Gaussian and mixture of Gaussians methods assume that the data is distributed according to the normal distribution or to a mixture of normal distributions [9]. The parameters of the Gaussian model can be found by maximizing the likelihood function over the training data set, being the learning process computationally inexpensive. For the mixture of Gaussians, the parameters can be found efficiently by the EM algorithm. Thus, the learning process using the EM algorithm is more computationally demanding as a number of interactions should be done before the algorithm converges. The methods based on Gaussian models are sensitive to the noise in the training data set, as the noise introduces a significant bias to the estimate covariance matrix. Furthermore, these procedures present a rather high sensitivity to errors in feature values and outliers. In the learning phase, the storage requirements are rather high but very low in the classification phase.
2.3. Parzen and Naive Parzen
Parzen and naive Parzen density estimation are nonparametric procedures and do not need any assumption about the data distribution [6, 28, 29]. The density is estimated directly from the training data and is a function of the number of objects situated in a region of a specific volume with a value as the length of an edge. The value of plays the role of a smoothing parameter. An advantage of the method is that it does not need any estimation of parameters. However, too long values of the smoothing parameter imply an oversmoothed estimated density. When is too small then the estimated density contains noise. Furthermore, the method needs to store all the observation vectors and it makes it slower, presenting very low computational requirements of learning but rather high in classification. The method is relatively robust to the outliers in the training data, choosing appropriate distance, and presents rather high sensitivity to errors in feature values. These procedures need to estimate one parameter by the users.
2.4. OneClass Support Vector Data Description
Support vector data description (SVDD) is a boundary method [9, 17]. It defines a hypersphere with a minimum volume covering the entire training data set. The minimization is solved as a quadratic programming problem and can be solved efficiently by introducing Lagrange multipliers [30, 31]. The method is relatively resistant to noise. The number of parameters that are to be estimated is equal to the size of the training data set; thus it is not useful for large training data sets. SVDD presents rather low sensitivity to errors in feature values and outliers. The method presents very high computational requirements of learning but very low in classification and needs to estimate one parameter by the user and learnt the other parameters.
2.5. Typicality Approach
Consider a target class containing units measured on features. Let be a distance [32] function on . It is said that is an Euclidean distance function if the metric space can be embedded in an Euclidean space , such that , and we may understand as the mean of . There are various ways of achieving this situation, the most common probably being classical metric scaling, also known as principal coordinate analysis [33, 34]. Given the realvalued coordinates , it is possible to apply any standard multivariate technique. Such an approach was used by different authors [35–42]. In this context a general measure of dispersion of , the geometric variability of , with respect to can be defined by which is a variant of Rao’s diversity coefficient [43]. The proximity function of a unit to is defined as
In applied problems, the distance function is a datum, but the probability distribution for the population is unknown. Natural estimators given a sample coming from are for the geometric variability of and the proximity function of unit to , respectively.
See [44] and references therein for a review of these concepts, their application, different properties, and proofs.
Let be a new observation and consider the OCC problem to decide whether belongs to the target class or, on the contrary, it is an outlier or an atypical observation, belonging to some different and unknown class. Therefore, the OCC problem can be formulated as a hypothesis test with : comes from the target class with mean , : comes from another unknown class.
This test can be considered as a test of typicality, as is formulated in [26]. In our context, with only one known class, the typicality test reduces to compute . If is significant it means that comes from a different and unknown class.
Sampling distribution of can be difficult to find for mixed data, but nevertheless it can be obtained by resampling methods, in particular drawing bootstrap samples: draw units with replacement from and calculate the corresponding values; repeat this process 10P times, with . In this way, the bootstrap distribution under is obtained.
It is worth to point out that this procedure can be used for any kind of data (continuous, discrete, or nominal), whereas other approaches application is not straightforward when nominal variables are present. As the procedure needs the computation of the distances between units in the target class and distances between a new unit and the units in the target class the storage requirements are rather high but very low in the classification phase. The method is relatively robust to the outliers in the training data.
3. Results of the Experimental Study
As there are few benchmark data sets for OCC, we use data sets containing two or multiple classes. Each class in turn is considered as the target class and the units in the other classes are considered as new units to be classified. On the one hand, we used 10 biomedical data sets, none of them containing nominal variables, from the UCI machine learning repository [45] to evaluate the performance of all the above procedures. In order to perform the comparison, the selected data sets are the biomedical data used in [46]; only the target classes considered in that referred work are taken into account in this paper as well. On the other hand, we also applied the typicality approach to two data sets with mixed variables.
In our experiments, we followed the procedure stated in [46]. Thus, all multiclass problems are transformed to oneclass classification problems by setting a chosen class as a target class and all remaining classes as nontargets. The target class was randomly split into equal parts between the training and test sets. All oneclass classifiers were only trained on the target data, that is, the half of the target data, and tested on the test data, the remaining half of the target data and the nontarget data. The experiments were repeated 10 times and the AUC average and standard deviation values are reported.
3.1. Results of Data Sets without Nominal Variables
In Table 1, a brief description of these well known data sets is presented. For the typicality method, a suitable distance is selected for each data set, according to the type of data (see Table 2 last column). The considered distances were the Euclidean distance, the Euclidean distance after standardized the data, the Mahalanobis distance, or the correlation distance.


With breast Wisconsin prognostic, E. coli, hepatitis, and liver disorders data sets, the typicality model obtained similar AUC average values than the other procedures, as we can see in Table 2. For the breast Wisconsin origin data set and taking benign class as target class, the typicality procedure obtained very good results () and similar to the obtained by the other procedures. It is worth noting that, for the malignant class as target class, it obtained similar results () than the naive Parzen procedure () and much better results than the obtained for the other procedures. With the colon data set, while mixture of Gaussians is not available and Parzen or SVDD methods gave poor results with high variability (, for classes 1 and 2, resp.), the typicality method obtained clearly better results (, for classes 1 and 2, resp.) than Gaussian (, for class 1 and 2, resp.) or naive Parzen (, for classes 1 and 2, resp.) methods. When the leukemia data set was analyzed, mixture of Gaussians and Parzen procedures were not available at all, and SVDD procedure presented a large variability ( and , resp.). However, similar results were found for Gaussian, naive Parzen, or typicality procedures. For the METAS data set, it must point out that when the second class was the target class, the best results were obtained with the typicality procedure (), showing its good performance with high dimensional data sets. With the SPECT heart data set and using the typicality method, a little worse results were found when class 0 was the target class. However, when the target class was class 1, clearly the typicality procedure obtained the best results (). Finally, for the thyroid data set, the typicality results were similar or slightly better than those obtained by the other procedures.
In summary, from the results presented in Table 2 it is clear that, in general, the typicality approach performs equal or better than the other well known procedures, for all the considered UCI data sets. The results show that, while other procedures are affected by small target classes, the typicality approach is more robust. Furthermore, it performs well with highdimensional data. On the other hand, as shown in Table 2, stateoftheart algorithms give “NaN”—Not a Number—in some cases; this fact does not appear when the typicality approach is used. Additional statistics on the AUC average values are provided in Figure 1 under the form of boxplots. Black lines correspond to the median values and black segments to the minimum and maximum values of each method. As we can see, the typicality procedure is the more robust for all data sets and it is in the top best methods.
3.2. Results on Mixed Variables Data Sets
Next we report the results obtained using two data sets with mixed variables. That means that there are some quantitative, binary, and nominal variables. Therefore, methods that implicitly are based on the Euclidean distance are not adequate. Thus, only the typicality approach was performed with these two data sets. In presence of mixed variables, it is known that Gower’s distance is an appropriate distance, presenting good properties in terms of missing values [47, 48].
Statlog (Heart) Data Set. This data set is available in the UCI dataset repository. It is composed by 270 units classified in two classes: absence or presence of heart disease, with 150 and 120 units, respectively. There are 13 variables, 6 quantitative, 1 ordered, 3 binary, and 3 nominal, and no missing values are present. Taking in turn, absence and presence class as the target class, the typicality approach reported AUC average and standard deviation values and , respectively. Furthermore, Table 3 reports the results obtained when we attempt to achieve a fixed False Alarm Rate (FAR) or false negative rate (1sensitivity), namely, 0.1. Note that for the two target class, we obtain good results.

Liver Cancer Data Set. We apply the typicality approach to a liver cancer data set [49]. It consists of 213 cases described by 4 nominal variables (type of hepatitis, categorized age, sex, and whether cirrhosis is present) plus 1993 genes. It is worth to mention that for each case at least one missing value is present (9.6% of the values are missing). The data set is divided in three groups. Group T formed by 107 samples from tumors on liver cancer patients, group NT formed by 76 samples from nontumor tissues of liver cancer patients and group N formed by 30 samples from normal livers. In [42] it was shown that there exists a high degree of confusion between groups, so bad oneclass classification results are expected. Taking groups N, NT, and T as target classes, the typicality approach obtained AUC average values and standard deviation values , , and , respectively. Results obtained for a fixed FAR equal to 0.1 are reported in Table 4. From Table 4, we can observe that when T is the target class, the method cannot distinguish the other groups. When NT is the target class, units from N group are not distinguished and only half the units from T are distinguished properly as nontarget. When N is the target class, units from T are very well distinguished as nontarget.

4. Conclusions
A noticeable attention has been devoted to the oneclass classification problem in the last years. This type of classification is characterized by the use of observations belonging to only one known class. These methods are particularly useful in biomedical studies, when observations belonging to other classes are difficult or impossible to obtain. In this paper, reference stateofthe art oneclass classification methods have been reviewed, and their suitability has been compared with a recent typicality procedure. To assess the efficiency of this new typicality application, experiments have been conducted on several public data sets from the UCI repository and has been compared to five of the most OCC used procedures, namely, Gaussian, mixture of Gaussians, naive Parzen, Parzen, and support vector DD models [46]. The results show that the typicality approach performs equally well or better than these stateofthe art procedures, thus it will be very valuable in many biomedical applications. The typicality approach does not need any knowledge about the data distribution, does not estimate any parameter, and is applicable to any kind of data, not necessarily continuous. This approach performs well with high dimensional data and it is robust in front of small target classes, whereas other OCC method accuracy rates are not so stable. For all these reasons, the typicality approach can be very useful in many biomedical applications where clinical, pathological, or biological noncontinuous data can be found and where data from healthy or even from nonhealthy patients are extremely hard or impossible to obtain.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work was supported by the Basque Government Research Team Grant (IT31310) SAIOTEK Project SA2013/00397 and the University of the Basque Country UPV/EHU (Grant UFI11/45 (BAILab)).
References
 L. Tarassenko, P. Hayton, N. Cerneaz, and M. Brady, “Novelty detection for the identification of masses in mammograms,” in Proceedings of the 4th International Conference on Artificial Neural Networks, pp. 442–447, June 1995. View at: Google Scholar
 M. Costa and L. Moura, “Automatic assessment of scintmammographic images using a novelty filter,” Proceedings of the Annual Symposium on Computer Applications in Medical Care, pp. 537–541, 1995. View at: Google Scholar
 O. Boehm, D. R. Hardoon, and L. M. Manevitz, “Classifying cognitive states of brain activity via oneclass neural networks with feature selection by genetic algorithms,” International Journal of Machine Learning and Cybernetics, vol. 2, no. 3, pp. 125–134, 2011. View at: Publisher Site  Google Scholar
 J. A. Reyes and D. Gilbert, “Prediction of proteinprotein interactions using oneclass classification methods and integrating diverse biological data,” Journal of Integrative Bioinformatics, vol. 4, no. 3, p. 77, 2007. View at: Google Scholar
 A. Depeursinge, J. Iavindrasana, A. Hidki et al., “Comparative performance analysis of stateoftheart classification algorithms applied to lung tissue categorization,” Journal of Digital Imaging, vol. 23, no. 1, pp. 18–30, 2010. View at: Publisher Site  Google Scholar
 G. Cohen, H. Sax, and A. Geissbuhler, “Novelty detection using oneclass Parzen density estimator. An application to surveillance of nosocomial infections,” Studies in Health Technology and Informatics, vol. 136, pp. 21–26, 2008. View at: Google Scholar
 D. M. J. Tax, Oneclass classification [Ph.D. thesis], Delft University of Technology, 2001.
 C. Dsir, S. Bernard, C. Petitjean, and L. Heutte, “One class random forests,” Pattern Recognition, vol. 46, pp. 3490–3506, 2013. View at: Google Scholar
 O. Mazhelis, “Oneclass classifiers: a review and analysis of suitability in the context of mobilemasquerader detection,” South African Computer Journal, vol. 36, pp. 29–48, 2006. View at: Google Scholar
 S. S. Khan and M. G. Madden, “A survey of recent trends in one class classification,” in Proceedings of the 20th Irish Conference on Artificial Intelligence and Cognitive Science, 2009. View at: Google Scholar
 C. Bishop, Neural Networks for Pattern Recognition, Oxford University Press, Oxford, UK, 1995.
 R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, John Wiley & Sons, New York, NY, USA, 2000.
 A. Ypma, D. M. J. Tax, and R. P. W. Duin, “Robust machine fault detection with independent component analysis and support vector data description,” in Proceedings of the 9th IEEE Workshop on Neural Networks for Signal Processing (NNSP '99), Y. H. Hu, J. Larsen, E. Wilson, and S. Douglas, Eds., pp. 67–76, August 1999. View at: Google Scholar
 R. O. Duda and P. E. Hart, Pattern Classification and Scene Analysis, John Wiley & Sons, New York, NY, USA, 1973.
 V. Vapnik, The Nature of Statistical Learning Theory, Springer, New York, NY, USA, 1995.
 D. M. J. Tax and R. P. W. Duin, “Data domain description using support vectors,” in Proceedings of the 7th European Symposium on Artificial Neural Networks, pp. 251–256, 1999. View at: Google Scholar
 D. M. J. Tax and R. P. W. Duin, “Support Vector data description,” Machine Learning, vol. 54, no. 1, pp. 45–66, 2004. View at: Publisher Site  Google Scholar
 I. W. Tsang, J. T. Kwok, and P.M. Cheung, “Core vector machines: fast SVM training on very large data sets,” Journal of Machine Learning Research, vol. 6, pp. 363–392, 2005. View at: Google Scholar
 D. Wang, D. S. Yeung, and E. C. C. Tsang, “Structured oneclass classification,” IEEE Transactions on Systems, Man, and Cybernetics, B: Cybernetics, vol. 36, no. 6, pp. 1283–1295, 2006. View at: Publisher Site  Google Scholar
 F. Angiulli, “Prototypebased domain description for oneclass classification,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 6, pp. 1131–1144, 2012. View at: Publisher Site  Google Scholar
 E. M. Knorr, R. T. Ng, and V. Tucakov, “Distancebased outliers: algorithms and applications,” VLDB Journal, vol. 8, no. 34, pp. 237–253, 2000. View at: Google Scholar
 F. Angiulli, S. Basta, and C. Pizzuti, “Distancebased detection and prediction of outliers,” IEEE Transactions on Knowledge and Data Engineering, vol. 18, no. 2, pp. 145–160, 2006. View at: Publisher Site  Google Scholar
 C. M. Bishop, Neural Networks for Pattern Recognition, Clarendon Press, 1995.
 D. Barber and C. M. Bishop, “Ensemble learning in Bayesian neural networks,” in Neural Networks and Machine Learning, C. M. Bishop, Ed., vol. 168 of Series F: Computer and Systems Sciences, pp. 215–237, Springer, 1998. View at: Google Scholar
 B. Lerner and N. D. Lawrence, “A comparison of stateoftheart classification techniques with application to cytogenetics,” Neural Computing and Applications, vol. 10, no. 1, pp. 39–47, 2001. View at: Publisher Site  Google Scholar
 I. Irigoien and C. Arenas, “INCA: new statistic for estimating the number of clusters and identifying atypical units,” Statistics in Medicine, vol. 27, no. 15, pp. 2948–2973, 2008. View at: Publisher Site  Google Scholar
 A. P. Bradley, “The use of the area under the ROC curve in the evaluation of machine learning algorithms,” Pattern Recognition, vol. 30, no. 7, pp. 1145–1159, 1997. View at: Google Scholar
 R. P. W. Duin, “On the choice of smoothing parameters for Parzen estimators of probabilitydensity functions,” IEEE Transactions on Computers, vol. 25, no. 11, pp. 1175–1179, 1976. View at: Google Scholar
 M. Kraaijveld and R. Duin, “A criterion for the smoothing parameter for parzenestimators of probability density functions,” Tech. Rep., Delft University of Technology, 1991. View at: Google Scholar
 V. Vapnik, Statistical Learning Theory, Wiley Interscience, 1998.
 K. R. Müller, S. Mika, G. Rätsch, K. Tsuda, and B. Schölkopf, “An introduction to kernelbased learning algorithms,” IEEE Neural Networks, vol. 12, pp. 181–201, 2001. View at: Google Scholar
 J. C. Gower, “Measures of similarity, dissimilarity and distance,” in Encyclopedia of Statistical Sciences, S. Kotz, N. L. Johson, and C. B. Read, Eds., pp. 307–316, John Wiley & Sons, New York, NY, USA, 1985. View at: Google Scholar
 J. C. Gower, “Some distance properties of latent root and vector methods used in multivariate analysis,” Biometrika, vol. 53, pp. 325–338, 1966. View at: Google Scholar
 W. J. Krzanowski and F. H. C. Marriott, Multivariate Analysis. Part 1: Distributions, Ordination and Inference, Kendall’s Library of Statistics, Edward Arnold, London, UK, 1994.
 C. M. Cuadras and C. Arenas, “A distance based regressionmodel for prediction with mixed data,” Communications in Statistics A. Theory and Methods, vol. 19, pp. 2261–2279, 1990. View at: Google Scholar
 P. Legendre and M. J. Anderson, “Distancebased redundancy analysis: testing multispecies responses in multifactorial ecological experiments,” Ecological Monographs, vol. 48, pp. 505–519, 1999. View at: Google Scholar
 M. J. Anderson and J. Robinson, “Generalized discriminant analysis based on distances,” Australian and New Zealand Journal of Statistics, vol. 45, no. 3, pp. 301–318, 2003. View at: Google Scholar
 M. J. Anderson and T. J. Willis, “Canonical analysis of principal coordinates: a useful method of constrained ordination for ecology,” Ecology, vol. 84, no. 2, pp. 511–525, 2003. View at: Google Scholar
 W. J. Krzanowski, “Biplots for multifactorial analysis of distance,” Biometrics, vol. 60, no. 2, pp. 517–524, 2004. View at: Publisher Site  Google Scholar
 I. Irigoien, C. Arenas, E. Fernández, and F. Mestres, “GEVA: geometric variabilitybased approaches for identifying patterns in data,” Computational Statistics, vol. 25, pp. 241–255, 2010. View at: Publisher Site  Google Scholar
 I. Irigoien, B. Sierra, and C. Arenas, “ICGE: an R package for detecting relevant clusters and atypical units in gene expression,” BMC Bioinformatics, vol. 13, article 30, 2012. View at: Publisher Site  Google Scholar
 I. Irigoien, F. Mestres, and C. Arenas, “The depth problem: identifying the most representative units in a data group,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 10, pp. 161–172, 2013. View at: Google Scholar
 C. R. Rao, “Diversity: its measurement, decomposition, apportionment and analysis,” Sankhyā A, vol. 44, pp. 1–22, 1982. View at: Google Scholar
 C. Arenas and C. M. Cuadras, “Some recent statistical methods based on distances,” Contributions to Science, vol. 2, pp. 183–191, 2002. View at: Google Scholar
 C. Blake, E. Keogh, and C. Merz, “UCI Repository of Machine Learning Database,” http://archive.ics.uci.edu/ml/. View at: Google Scholar
 P. Juszczak, D. M. J. Tax, E. Pȩkalska, and R. P. W. Duin, “Minimum spanning tree based oneclass classifier,” Neurocomputing, vol. 72, no. 7–9, pp. 1859–1869, 2009. View at: Publisher Site  Google Scholar
 J. C. Gower, “A general coefficient of similarity and some of its properties,” Biometrics, vol. 27, pp. 857–871, 1971. View at: Google Scholar
 A. Montanari and S. Mignari, “Notes on the bias of dissimilarity indices for incomplete data sets: the case of archaelogical classifications,” Qüestiió, vol. 18, pp. 39–49, 1994. View at: Google Scholar
 K. Kato, “Adaptortagged competitive PCR: a novel method for measuring relative gene expression,” Nucleic Acids Research, vol. 25, no. 22, pp. 4694–4696, 1997. View at: Google Scholar
Copyright
Copyright © 2014 Itziar Irigoien et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.