Computational Intelligence Techniques in MedicineView this Special Issue
Research Article | Open Access
Soo Beom Choi, Won Jae Kim, Tae Keun Yoo, Jee Soo Park, Jai Won Chung, Yong-ho Lee, Eun Seok Kang, Deok Won Kim, "Screening for Prediabetes Using Machine Learning Models", Computational and Mathematical Methods in Medicine, vol. 2014, Article ID 618976, 8 pages, 2014. https://doi.org/10.1155/2014/618976
Screening for Prediabetes Using Machine Learning Models
The global prevalence of diabetes is rapidly increasing. Studies support the necessity of screening and interventions for prediabetes, which could result in serious complications and diabetes. This study aimed at developing an intelligence-based screening model for prediabetes. Data from the Korean National Health and Nutrition Examination Survey (KNHANES) were used, excluding subjects with diabetes. The KNHANES 2010 data were used for training and internal validation, while data from KNHANES 2011 were used for external validation. We developed two models to screen for prediabetes using an artificial neural network (ANN) and support vector machine (SVM) and performed a systematic evaluation of the models using internal and external validation. We compared the performance of our models with that of a screening score model based on logistic regression analysis for prediabetes that had been developed previously. The SVM model showed the areas under the curve of 0.731 in the external datasets, which is higher than those of the ANN model (0.729) and the screening score model (0.712), respectively. The prescreening methods developed in this study performed better than the screening score model that had been developed previously and may be more effective method for prediabetes screening.
The prevalence of type 2 diabetes is dramatically increasing, resulting in a global public health issue . The prevalence of diabetes was estimated at 285 million or 6.4% of adults in the world in 2010 , and this prevalence is expected to rise to 552 million by 2030 . The increasing rates of obesity are expected to result in a faster increase in the prevalence of type 2 diabetes in the future . However, owing to the absence of symptoms and/or disease-related knowledge, diabetes often goes undetected, and approximately one-third of people with diabetes are not aware of their status [5–7]. Therefore, development of a simple accurate screening method is needed. Historically, the majority of the clinical screening methods consisted of surveys developed using logistic regression analyses to predict diabetes [8–13].
Prediabetes was first recognized as an intermediate diagnosis and indication of a relatively high risk for the future development of diabetes by the Expert Committee on Diagnosis and Classification of Diabetes Mellitus in 1997 , and it has been reported that approximately 5–10% of patients with untreated prediabetes subsequently develop diabetes [15, 16]. This is significant considering that prediabetes based on impaired fasting glucose (IFG) was estimated to affect 4.9 million people, accounting for 17.4% of Korean adults in 2005 , with a further 35% of adults in the US with prediabetes in 2008 . The definition of prediabetes includes a fasting plasma glucose (FPG) level in the range of 100–125 mg/dL (5.6–6.9 mmol/L), impaired glucose tolerance (IGT) (oral glucose tolerance test (OGTT) 2 h measurement in the range of 140–199 mg/dL (7.8–11.0 mmol/L)), or HbA1c level in the range of 5.7–6.4% (39–46 mmol/mol). Similar to diabetes, the risk of microvascular complications is increased with prediabetes , and the risk for cardiovascular disease and total mortality is almost twice as high in individuals with prediabetes [19, 20]. Early diagnosis and intervention for prediabetes could prevent these complications, prevent delay, or prevent the transition to diabetes  and be cost-effective .
Machine learning is an area of artificial intelligence research, which uses statistical methods for data classification. Several machine learning techniques have been applied in clinical settings to predict disease and have shown higher accuracy for diagnosis than classical methods . Support vector machines (SVM) and artificial neural networks (ANN) have been widely used approaches in machine learning. They are the most frequently used supervised learning methods for analyzing complex medical data .
In this study, we aimed to develop and validate models to predict prediabetes using artificial neural network (ANN) and support vector machine (SVM) methods, which could be effective as simple and accurate screening tools. The model performance was compared to that of the screening score model that we modified for prediabetes based on the screening score for diabetes by Lee et al. , with respect to accuracy and area under the curve (AUC) of the receiver operating characteristic (ROC).
2. Materials and Methods
2.1. Data Source and Subjects
Data from the Korean National Health and Nutrition Examination Survey (KNHANES) 2010  and 2011  were used to develop and validate, respectively, the ANN and SVM models for prediabetes. The KNHANES is a cross-sectional survey that includes approximately 800 questions; it is conducted by the Division of Chronic Disease Surveillance, Korea Centers for Disease Control and Prevention. The survey represents the entire nation using rolling sampling survey method. The following exclusion criteria applied to the subjects in both datasets: <20 years of age, missing data for waist circumference, smoking status, alcohol intake, body mass index (BMI), physical activity, family history of diabetes, undetermined diabetic status or hypertension status, and diagnosed diabetes or undiagnosed diabetes. Diagnosed and undiagnosed patients were excluded in order to focusing on predicting prediabetes. Undiagnosed diabetes was defined as a FPG ≥ 126 mg/dL without diagnosis by clinician. Of the 8,958 subjects in the KNHANES 2010, 4,685 were included as shown in Figure 1. Of the 8,518 subjects in the KNHANES 2011, 4,566 subjects were included using the same flow chart as Figure 1. The subjects of the KNHANES 2010 and 2011 data sets were not overlapped.
Figure 2 illustrates the study flow. The development dataset from KNHANES 2010 was randomly divided into training and internal validation sets using a 2 : 1 ratio. The training set () was used to construct the ANN and SVM models. The internal validation set () was used to assess the ability to predict prediabetes. Additionally, data from KNHANES 2011 were used as an external validation set (). All individuals in the surveys participated voluntarily and informed consent was obtained from all participants. The survey protocol was approved by the Institutional Review Board of the Korean Centers for Disease Control and Prevention.
2.2. Risk Factors
We adopted the most frequently used nine variables from previous studies regarding diabetes prediction models: age, gender, family history of diabetes, hypertension, alcohol intake, BMI, smoking status, waist circumference, and physical activity [8–13]. FPG was determined using glucose levels that were collected following ≥ 8 hours of fasting. We considered only FPG although there are three methods to diagnose prediabetes. In KNHANES 2010, FPG was obtained from every subject, but OGTT was not tested, and HbA1c was tested only for subjects with diabetes. A family history of diabetes was limited to parents and siblings. Hypertension was defined as systolic blood pressure (SBP) > 140 mmHg, diastolic blood pressure (DBP) > 90 mmHg, or use of medication for blood pressure control . Alcohol intake was calculated using 2 questions: (1) alcohol consumption frequency during the previous 12 months and (2) average number of drinks on those days. The amount of alcohol was calculated based on the number of glasses, regardless of the kind of beverage, assuming that the amount of alcohol was approximately the same in each glass (approximately 8 g alcohol per glass). Smoking status was divided into “currently smoking regularly” and “others,” with the latter group including subjects who had never smoked or had quit smoking. The subjects who answered more than “moderate” to the question “how intense is your everyday activity?” were considered as physically active.
2.3. Artificial Neural Network
ANN is an artificial intelligence technology, inspired by the architecture of biological neurons such as that in the human brain . The technology is specialized for classification, and it is mostly used to identify underlying patterns for risk factors in medicine. When trained properly, neural networks are known to have more accurate predictive abilities than conventional methods such as logistic regression. There have been a number of recent advances in ANN methodology that enable automatic detection of an optimal predictive model [28, 29]. Unlike logistic regression, ANNs are able to detect complex nonlinear relationships between multiple predictors and diseases, which make them useful in support systems for medical decisions [30, 31].
The ANN models were constructed using NeuroSolution version 6.0 (NeuroDimension, Gainesville, FL), which is professional software that simplifies the construction of ANN . This software allows simultaneous testing of various types of neural networks, including generalized regression neural network, multilayer perceptron, probabilistic neural network, radial basis neural network, feedforward neural network, and support vector machine. To avoid overfitting, the prediction models were internally validated using cross-validation. Performance of the prediction models was monitored during training and cross-validation to obtain optimal algorithm and parameters, such as learning rate, momentum, and number of hidden nodes. The ANN was trained with 7 predictors including age, gender, waist circumference, BMI, family history of diabetes, hypertension, and alcohol intake, which were selected using backward logistic regression. The model chosen for prediabetes prediction was a multilayer perceptron model with 1 hidden layer, batch training, and momentum learning (MLP-1-BM) of backpropagation feedforward algorithm, which demonstrated the best performance as a desired ANN.
2.4. Support Vector Machine
SVM maps data to a higher dimensional space through a kernel function that linearly separates data patterns. The data are divided into two groups by the training data referred to as a support vector. SVM models are determined by choosing the maximum-margin hyperplane with the nearest support vector of the two groups . SVM improves the accuracy of a model through the optimization of separating space using the kernel function, but one of the disadvantages of SVM is that it requires many trials to construct an optimal SVM model in comparison with other machine learning techniques .
The same seven risk factors as those in the ANN model were employed for the SVM. To obtain the optimal model, we adopted a grid search in which a range of parameter values (penalty parameter  of 0.01, 0.1, 1, 10, and 100 and scaling factor [σ] of 0.001, 0.01, 0.1, 1, 10, and 100) was tested using the 10-fold cross-validation strategy. The optimal parameter values with a of 10 and σ of 10 for SVM using the Gaussian kernel function were obtained. The SVM models were constructed using MATLAB Version 2012a (Mathworks Inc., Natick, MA).
2.5. Screening Score of Our Models for Prediabetes
The models constructed by ANN and SVM were compared with a previously developed screening survey to illustrate performance of our models and the possibility of their use in real situations. For this purpose, we used a screening score model for the Korean population constructed by Lee et al. ; we felt this was appropriate because both studies constructed models for the Korean population. Lee et al. used data from KNHANES 2001 and 2005 for training and data from KNHANES 2007 and 2008 for external validation. In addition, the screening score model by Lee et al. used very similar risk factors to ours, with the exception of current smoking status. Those 6 variables independently associated with undiagnosed diabetes were chosen for their model: age, family history of diabetes, hypertension, waist circumference, smoking, and alcohol intake.
The risk score was assigned according to the odds ratio for each risk factor in the logistic regression model defined by Lee et al. . Within the total score range of 0–11 points, a cut-off score of ≥5 points was selected to indicate an individual at high risk for undiagnosed diabetes; this cut-off resulted in the highest value for the Youden index. The 6 risk factors jointly yielded an AUC of 0.730 for both the internal and external validation sets . To compare with our models for prediabetes, we constructed a new screening score model for prediabetes by adjusting the cut-off point value based on our definition of prediabetes (100 mg/dL ≤ FPG < 126 mg/dL), given that the screening score for diabetes used by Lee et al. was based on FPG ≥ 126 mg/dL . The screening score for prediabetes was designed with the same risk score model of the 6 risk factors using our training set for prediabetes (KNHANES 2010) and the Youden index; as a result, a cut-off score of ≥5 points was identified to indicate an individual with prediabetes.
2.6. Statistical Analyses
The weighted characteristics of the data from the KNHANES 2010 to represent the entire normal and prediabetes people in Korea are summarized by descriptive statistics in Table 1. For comparison of the factors between normal and prediabetes, the continuous and categorical characteristics were tested using -test and chi-square test, respectively.
|BMI: body mass index; FPG: fasting plasma glucose.|
Table values are given as mean ± standard error or % (standard error) [95% confidence interval] unless otherwise indicated. were obtained by -test or chi-square test.
Impaired fasting glucose was considered with values ≥ 100 mg/dL and <126 mg/dL.
To obtain the optimal variables for the prediction model, backward logistic regression was performed with the training set. Each step of the backward regression excluded the variables without a statistically meaningful correlation with the outcome, prediabetes. Three steps of backward regression were executed, and the selected 7 variables were age, BMI, hypertension, gender, alcohol intake, waist circumference, and family history. The Hosmer-Lemeshow test resulted in a value of 0.132, indicating that the chosen variables were well fitted.
ROC curve analysis is the most commonly used method in clinical analysis to establish an optimal cut-off point . Therefore, we generated ROC curves and the selected cut-off points that maximized the Youden index  to compare the performance of our optimal machine learning models with that of the screening score model for prediabetes based on the screening score by Lee et al. , using our internal and external validation sets. Following the ROC analysis, the AUC, accuracy, sensitivity, and specificity of our models and screening score model for prediabetes were calculated. The classification accuracy measured the proportion of cases correctly classified. Sensitivity measured the fraction of positive cases that were classified as positive. Specificity measured the fraction of negative cases that were classified as negative. We used SPSS 20.0 (IBM Corp, Armonk, NY) for statistical analysis and MedCalc 12.4 (MedCalc Inc., Mariakerke, Belgium) for ROC analysis. Statistical significance was set at .
The weighted characteristics of the KNHANES 2010 data are summarized in Table 1. The factors that were significantly related to prediabetes were age, gender, family history of diabetes, alcohol intake, BMI, waist circumference, FPG, systolic and diastolic blood pressures, and hypertension.
When the prediction performance of the ANN model using 10-fold cross-validation was assessed for the training set, the final model showed an AUC of 0.706 and an accuracy of 65.6% for prediabetes. Cross-validation of the optimal SVM parameters with the training set resulted in an AUC of 0.742 and accuracy of 69.9%. These results are not included in Table 2. The similar performance observed between the training and validation sets in Table 2 indicates that the trained models were not overfitting.
|AUC: area under the curve; ANN: artificial neural network; SVM: support vector machine.|
The internal validation set was comprised of data from the Korean National Health and Nutrition Examination Survey (KNHANES) 2010, and the external validation set included data from KNHANES 2011. *The chosen model was a multilayer perceptron model with 1 hidden layer, batch training, and momentum learning (MLP-1-B-M) of backpropagation feedforward algorithm. †The optimal model was found using Gaussian kernel function with a penalty parameter (C) of 10 and scaling factor (σ) of 10. ‡The performance was calculated by applying the screening score model for prediabetes based on that of Lee et al.  to the data from KNHANES 2010 and 2011.
With both the internal and the external validation sets, our ANN and SVM models showed better performance than the existing screening score model using logistic regression, especially in terms of AUC, which is known as a better measure than accuracy in evaluating learning algorithms  (Table 2). In the external validation set, the accuracy of the SVM model was 5.4% and 6.2% higher than that of the ANN and screening score models, respectively. The ROC curves of the ANN, SVM, and screening score models are depicted for the internal and external validation sets in Figure 3.
Table 3 shows the performance obtained by applying the screening score model by Lee et al.  to the data from KNHANES 2010 and 2011 for predicting prediabetes and undiagnosed diabetes. For all performance parameters (AUC, accuracy, sensitivity, and specificity) in both datasets, the ability to predict prediabetes was inferior to that for diabetes. In particular, the AUC and accuracy for prediabetes in the external validation set were lower than those for undiagnosed diabetes by 0.039 and 4.7%, respectively. AUC and accuracy of the SVM model for external validation are higher than those of the screen score model for prediabetes by 0.019 and 6.2%, respectively.
|AUC: area under the curve; KNHANES: Korean National Health and Nutrition Examination Survey. |
Prediabetes was defined as fasting plasma glucose, with values ≥100 mg/dL and <126 mg/dL. *Internal and external validation sets to evaluate the screening score for prediabetes ( = 1,551 for KNHANES 2010 and = 4,566 for KNHANES 2011). †Internal and external validation sets to evaluate the screening score for undiagnosed diabetes ( = 1,585 for KNHANES 2010 and = 4,683 for KNHANES 2011).
The results of the present study indicate that the ANN and SVM models that we developed to predict prediabetes, defined as IFG, performed better than the existing clinical screening score model, as indicated by the AUC and accuracy measures (Table 2). Although logistic regression analysis and ANN share common roots in statistical pattern recognition, the latter is a generalization of the former , which might explain why our ANN model performed better than the screening score model, which was based on logistic regression analysis. The SVM model performed particularly well due to the ability of SVM to efficiently find a unique optimal solution, incorporate multiple types of data with a degree of flexibility, and model nonlinear patterns . We also investigated SVM models with different numbers of risk factors to find optimal parameters. The best performance of SVM model with six risk factors including age, body mass index, hypertension, gender, daily alcohol intake, and waist circumference was an AUC of 0.743 and accuracy of 70.2% in training set, which was almost the same as the performance of SVM model with seven risk factors in our paper, resulting in an AUC of 0.742 and accuracy of 69.9%.
Meng et al.  compared the performance of logistic regression, ANNs, and decision tree models for predicting diabetes or prediabetes using common risk factors in China population. In Meng et al. study, the ANNs model was the poorest of the three models, with 73.23% accuracy. This result is consistent with ours that the performance of ANN model was lower than SVM model. However, the performance of ANN can be case-by-case depending on characteristics of data or developers.
Although similar statistical analyses were conducted (i.e., backward regression models), there were slight differences in the variables included in the present study and those in the study by Lee et al. . Lee et al. included current smoking status as a risk factor in the training set based on the data from KNHANES 2001 and 2005; however, current smoking status was not included in our training set using data from KNHANES 2010. This may have resulted from lifestyle changes in the Korean population between those years, including a decline in the overall smoking rate and stronger antismoking laws . Although several screening score models have been developed and used clinically, our prediction model is unique in several ways.
First, owing to the similarity between our machine learning models and the existing screening score models, we were able to compare the performance of our machine learning models with the existing models. Second, to the best of our knowledge, there are few studies investigating prediabetes; instead, the majority of the other models have been developed to predict undiagnosed diabetes. However, prediabetes is increasingly becoming a significant public health issue. Using our model to screen patients for prediabetes would enable interventions at an earlier stage, which would be easier to implement and more successful than interventions implemented following diabetes screening.
Prediabetes was more difficult to predict than diabetes using any of the parameters across all of the models, which is not unexpected. AUC and accuracy of the SVM model for external validation are higher than those of the screen score model for prediabetes by 0.019 and 6.2%, respectively (Table 2). Therefore, we demonstrated that the machine learning methods could help to overcome the difficulty in predicting prediabetes.
This study has certain limitations. First, FPG level was the only measurement that we used to define prediabetes and diabetes; OGTT and HbA1c were not taken into consideration. Data were lacking for these measurements; however, the use of FPG level was consistent with the model developed by Lee et al. , with which we compared our models.
Second, the screening score model for diabetes developed by Lee et al.  did not correspond perfectly with our model for prediabetes. For a more precise comparison in future studies, a screening score model for prediabetes should be constructed using the new regression equation with different risk factors for prediabetes. In spite of this limitation, the suggested model with the new cut-off point is considered a good model for predicting prediabetes with AUCs of 0.734 and 0.712 in the internal and external validation sets, respectively.
Last, the new models that we developed are limited in terms of convenience and potential widespread use. Although the screening score model is not the most effective one for disease prediction, it is simple and accessible. However, machine learning models could also become more accessible through the use of calculator software, particularly with the widespread use of devices such as computers, smart phones, and tablet PCs. Future studies could develop a calculator in which the values are entered via a website or application and the results are immediately delivered to the end user. The decision tree method is also warranted for easy interpreting tree-like plot in the future.
Our study constructed a reasonably good model to predict prediabetes in the Korean population. By applying similar methods in other countries, researchers could develop country-specific machine learning models for nationwide use. The creation of a user-friendly calculator program would enable access to screening by the general population, in addition to medical professionals. This widespread use could result in early diagnosis and treatment for people with prediabetes and diabetes, helping to relieve the public health diabetes burden and reducing the number of people who remain undiagnosed.
Conflict of Interests
The authors have no potential conflict of interests relevant to this paper.
This work was supported by the National Research Foundation of Korea (NRF) Grant funded by the Korean Government (MEST) (NRF-2012R1A2A2A03045612).
- B. C. Zyriax, R. Salazar, W. Hoeppner, E. Vettorazzi, C. Herder, and E. Windler, “The association of genetic markers for type 2 diabetes with prediabetic status - cross-sectional data of a diabetes prevention trial,” PLoS ONE, vol. 8, Article ID e75807, 2013.
- J. E. Shaw, R. A. Sicree, and P. Z. Zimmet, “Global estimates of the prevalence of diabetes for 2010 and 2030,” Diabetes Research and Clinical Practice, vol. 87, no. 1, pp. 4–14, 2010.
- D. R. Whiting, L. Guariguata, C. Weil, and J. Shaw, “IDF Diabetes Atlas: Global estimates of the prevalence of diabetes for 2011 and 2030,” Diabetes Research and Clinical Practice, vol. 94, no. 3, pp. 311–321, 2011.
- P. Zimmet, K. G. M. M. Alberti, and J. Shaw, “Global and societal implications of the diabetes epidemic,” Nature, vol. 414, no. 6865, pp. 782–787, 2001.
- S. M. Kim, J. S. Lee, J. Lee et al., “Prevalence of diabetes and impaired fasting glucose in Korea: Korean National Health and Nutrition Survey 2001,” Diabetes Care, vol. 29, no. 2, pp. 226–231, 2006.
- J. C. Yong, C. K. Hyeon, M. K. Hee, W. P. Seok, K. Jongoh, and J. K. Dae, “Prevalence and management of diabetes in Korean adults: Korea National Health and Nutrition Examination Surveys 1998–2005,” Diabetes Care, vol. 32, no. 11, pp. 2016–2020, 2009.
- C. C. Cowie, K. F. Rust, D. D. Byrd-Holt et al., “Prevalence of diabetes and impaired fasting glucose in adults in the U.S. population: national Health and Nutrition Examination Survey 1999–2002,” Diabetes Care, vol. 29, no. 6, pp. 1263–1268, 2006.
- Y. Lee, H. Bang, H. C. Kim, H. M. Kim, S. W. Park, and D. J. Kim, “A simple screening score for diabetes for the Korean population: development, validation, and comparison with other scores,” Diabetes Care, vol. 35, no. 8, pp. 1723–1730, 2012.
- J. Lindström and J. Tuomilehto, “The diabetes risk score: a practical tool to predict type 2 diabetes risk,” Diabetes Care, vol. 26, no. 3, pp. 725–731, 2003.
- M. B. Schulze, K. Hoffmann, H. Boeing et al., “An accurate risk score based on anthropometric, dietary, and lifestyle factors to predict the development of type 2 diabetes,” Diabetes Care, vol. 30, no. 3, pp. 510–515, 2007.
- H. Bang, A. M. Edwards, A. S. Bomback et al., “Development and validation of a patient self-assessment score for diabetes risk,” Annals of Internal Medicine, vol. 151, no. 11, pp. 775–783, 2009.
- C. A. Baan, J. B. Ruige, R. P. Stolk et al., “Performance of a predictive model to identify undiagnosed diabetes in a health care setting,” Diabetes Care, vol. 22, no. 2, pp. 213–219, 1999.
- S. Griffin, P. S. Little, C. N. Hales, A. L. Kinmonth, and N. J. Wareham, “Diabetes risk score: towards earlier detection of type 2 diabetes in general practice,” Diabetes Metabolic Research Review, vol. 16, no. 3, pp. 164–171, 2000.
- American Diabetes Association, “Diagnosis and classification of diabetes mellitus (Position Statement),” Diabetes Care, vol. 35, supplement 1, pp. S64–S71, 2012.
- F. De Vegt, J. M. Dekker, A. Jager et al., “Relation of impaired fasting and postload glucose with incident type 2 diabetes in a Dutch population: the Hoorn study,” Journal of the American Medical Association, vol. 285, no. 16, pp. 2109–2113, 2001.
- The Diabetes Prevention Program (DPP) Research Group, “The diabetes prevention program (DPP): description of lifestyle intervention,” Diabetes Care, vol. 25, no. 12, pp. 2165–2171, 2002.
- T. L. Sentell, G. He, E. W. Gregg, and D. Schillinger, “Racial/ethnic variation in prevalence estimates for United States prediabetes under alternative 2010 American diabetes association criteria: 1988–2008,” Ethnicity & Disease, vol. 22, no. 4, pp. 451–458, 2012.
- A. G. Tabák, C. Herder, W. Rathmann, E. J. Brunner, and M. Kivimäki, “Prediabetes: a high-risk state for diabetes development,” The Lancet, vol. 379, no. 9833, pp. 2279–2290, 2012.
- M. Coutinho, H. C. Gerstein, Y. Wang, and S. Yusuf, “The relationship between glucose and incident cardiovascular events: a metaregression analysis of published data from 20 studies of 95,783 individuals followed for 12.4 years,” Diabetes Care, vol. 22, no. 2, pp. 233–240, 1999.
- S. C. Port, M. O. Goodarzi, N. G. Boyle, and R. I. Jennrich, “Blood glucose: a strong risk factor for mortality in nondiabetic patients with cardiovascular disease,” American Heart Journal, vol. 150, no. 2, pp. 209–214, 2005.
- L. Tian, J. Zhu, L. Liu, Y. Liang, J. Li, and Y. Yang, “Prediabetes and short-term outcomes in nondiabetic patients after acute ST-elevation myocardial infarction,” Cardiology, vol. 127, no. 1, pp. 55–61, 2014.
- W. H. Herman, T. J. Hoerger, M. Brandle et al., “The cost-effectiveness of lifestyle modification or metformin in preventing type 2 diabetes in adults with impaired glucose tolerance,” Annals of Internal Medicine, vol. 142, no. 5, pp. 323–332, 2005.
- T. K Yoo, S. K. Kim, D. W. Kim et al., “Osteoporosis risk prediction for bone mineral density assessment of postmenopausal women using machine learning,” Yonsei Medical Journal, vol. 54, no. 6, pp. 1321–1330, 2013.
- C. Hsieh, R. Lu, N. Lee, W. Chiu, M. Hsu, and Y. Li, “Novel solutions for an old disease: diagnosis of acute appendicitis with random forest, support vector machines, and artificial neural networks,” Surgery, vol. 149, no. 1, pp. 87–93, 2011.
- Korea Centers for Disease Control and Prevention, The Fifth Korea National Health and Nutrition Examination Survey (KNHANES V), Ministry of Health and Welfare, Seoul, South Korea, 2010.
- Korea Centers for Disease Control and Prevention, The Fifth Korea National Health and Nutrition Examination Survey (KNHANES V-2), Ministry of Health and Welfare, Seoul, Republic of Korea, 2011.
- J. Zou, Y. Han, and S. So, “Overview of artificial neural networks,” Methods in Molecular Biology, vol. 458, pp. 15–23, 2008.
- A. A. Motsinger-Reif and M. D. Ritchie, “Neural networks for genetic epidemiology: past, present, and future,” BioData Mining, vol. 1, p. 3, 2008.
- W. Yao, X. Chen, Y. Zhao, and M. van Tooren, “Concurrent subspace width optimization method for RBF neural network modeling,” IEEE Transactions on Neural Networks and Learning Systems, vol. 23, no. 2, pp. 247–259, 2012.
- W. G. Baxt, “Application of artificial neural networks to clinical medicine,” The Lancet, vol. 346, no. 8983, pp. 1135–1138, 1995.
- C. Eller-Vainicher, I. Chiodini, I. Santi et al., “Recognition of morphometric vertebral fractures by artificial neural networks: Analysis from gismo Lombardia database,” PLoS ONE, vol. 6, no. 11, Article ID e27277, 2011.
- A. Heiat, “Comparison of artificial neural network and regression models for estimating software development effort,” Information and Software Technology, vol. 44, no. 15, pp. 911–922, 2002.
- C. Cortes and V. Vapnik, “Support-vector networks,” Machine Learning, vol. 20, no. 3, pp. 273–297, 1995.
- K. Kim, J. Y. Choi, T. K. Yoo, S. K. Kim, K. Chung, and D. W. Kim, “Mortality prediction of rats in acute hemorrhagic shock using machine learning techniques,” Medical and Biological Engineering and Computing, vol. 51, no. 9, pp. 1059–1067, 2013.
- A. K. Akobeng, “Understanding diagnostic tests 3: receiver operating characteristic curves,” Acta Paediatrica, vol. 96, no. 5, pp. 644–647, 2007.
- R. Fluss, D. Faraggi, and B. Reiser, “Estimation of the Youden index and its associated cutoff point,” Biometrical Journal, vol. 47, no. 4, pp. 458–472, 2005.
- J. Huang and C. X. Ling, “Using AUC and accuracy in evaluating learning algorithms,” IEEE Transactions on Knowledge and Data Engineering, vol. 17, no. 3, pp. 299–310, 2005.
- S. Dreiseitl and L. Ohno-Machado, “Logistic regression and artificial neural network classification models: a methodology review,” Journal of Biomedical Informatics, vol. 35, no. 5-6, pp. 352–359, 2002.
- R. C. Thurston, K. A. Matthews, J. Hernandez, and F. de la Torre, “Improving the performance of physiologic hot flash measures with support vector machines,” Psychophysiology, vol. 46, no. 2, pp. 285–292, 2009.
- X. H. Meng, Y. X. Huang, D. P. Rao, Q. Zhang, and Q. Liu, “Comparison of three data mining models for predicting diabetes or prediabetes by risk factors,” Kaohsiung Journal of Medical Sciences, vol. 29, no. 2, pp. 93–99, 2013.
- Y. Khang, S. Yun, H. Cho, and K. Jung-Choi, “The impact of governmental antismoking policy on socioeconomic disparities in cigarette smoking in South Korea,” Nicotine and Tobacco Research, vol. 11, no. 3, pp. 262–269, 2009.
Copyright © 2014 Soo Beom Choi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.