Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2016 / Article

Research Article | Open Access

Volume 2016 |Article ID 7485250 | 11 pages | https://doi.org/10.1155/2016/7485250

Retreatment Predictions in Odontology by means of CBR Systems

Academic Editor: Michele Migliore
Received10 Aug 2015
Revised14 Sep 2015
Accepted15 Sep 2015
Published14 Jan 2016

Abstract

The field of odontology requires an appropriate adjustment of treatments according to the circumstances of each patient. A follow-up treatment for a patient experiencing problems from a previous procedure such as endodontic therapy, for example, may not necessarily preclude the possibility of extraction. It is therefore necessary to investigate new solutions aimed at analyzing data and, with regard to the given values, determine whether dental retreatment is required. In this work, we present a decision support system which applies the case-based reasoning (CBR) paradigm, specifically designed to predict the practicality of performing or not performing a retreatment. Thus, the system uses previous experiences to provide new predictions, which is completely innovative in the field of odontology. The proposed prediction technique includes an innovative combination of methods that minimizes false negatives to the greatest possible extent. False negatives refer to a prediction favoring a retreatment when in fact it would be ineffective. The combination of methods is performed by applying an optimization problem to reduce incorrect classifications and takes into account different parameters, such as precision, recall, and statistical probabilities. The proposed system was tested in a real environment and the results obtained are promising.

1. Introduction

Bioinformatics can be applied to various fields of medicine, although it is normally used in fields associated with genetic expressions [1, 2], protein analysis [3], sequencing [4], and so forth. Its use is not as commonly applied to more restricted levels such as private medical consultations. Nevertheless, while bioinformatics does fall within this scope, medical doctors often use what is referred to as knowledge extraction, which is based on experience gained over time from experts in the field. Expert knowledge is composed of the prediction or classification of pathologies in relation to a set of symptoms exhibited by the patient. The decisions made by odontologists have been traditionally based on past experiences of previous treatment cases. There are normally too many variables to consider, which has in fact resulted in the high failure rate of retreatments and the inability to easily create expert knowledge, particularly from doctors recently new to the profession. Consequently, it is necessary to provide new solutions that facilitate the decision-making process of odontologists and can lead to decisions that minimize the failure of endodontic treatments and retreatments. Decision support systems can notably help odontologists make decisions, and case-based reasoning is especially appropriate for this kind of problems.

In odontology, the success rate of endodontic therapy is 90%, which leaves a failure rate of 10%. Thus, an odontologist would greatly appreciate the ability to use artificial intelligence techniques to analyze the cases falling within this 10% and determine whether retreatment or extraction is preferable. The problem in 40% of these cases is the result of root crown fractures, which in turn represent 5% of all dental fractures. The bacterial recolonization of the root canal and the subsequent appearance of radiological symptoms represent 15% of endodontic failure [57]. It is not possible to find many published studies within the field of bioinformatics that address the problem previously presented. Existing works are limited to statistical analysis, which extracts the variables that are differentiated in different groups of patients and, according to the results obtained, make it possible to characterize relevant variables. However, this method does not permit the simultaneous analysis of the influence of the different variables. Statistical analysis is limited to the application of specific tests such as chi-square [8], Mann-Whitney [9], or Kruskal-Wallis test [10]. Nevertheless, it is necessary to create a new process that can combine all the information gathered in an intelligent way in order to perform a final classification and prediction that can help the odontologists to make more precise decisions.

This work proposes an innovative reasoning system to predict the success of retreatments. The proposed reasoning system uses past experiences to propose new solutions. CBR (case-based reasoning) systems execute a CBR cycle composed of 4 stages: retrieve (to recover past experiences), reuse (to obtain a new solution based on the retrieved past experiences), revise (to evaluate the obtained solution), and retain (to learn from the new experience). The CBR system proposed in this paper recovers a set of variables for a group of patients. This dataset is used as an input for the reuse phase of the CBR system. The reuse phase incorporates new classification techniques during the reuse phase, not previously used for this kind of problem, in order to generate a classification for the new patient. During the reuse phase, the CBR system incorporates classifiers based on Bayesian networks. The combination of both methods is achieved by applying an optimization problem in which the functional objective is defined in order to reduce false negatives (not advising retreatment when it is in fact advisable). Traditional statistical techniques are applied during the revise phase to facilitate the interpretation of the results by selecting the variables that present different characteristics from those in the groups of individuals. One of the advantages of the proposed system is that it can be adjusted to human behavior, given that they are based on the analysis of previous information in order to provide new predictions.

The remainder of the paper is structured as follows: Section 2 revises related works, focusing on prediction systems used in this kind of problem; Section 3 presents the proposed predictive mechanism, describing in detail the stages of the CBR system; Section 4 presents a case study and the results obtained; finally, Section 5 presents the conclusions obtained.

2. Prediction Systems

The use of predictive techniques in medicine and especially in the field of odontology has been studied since the late 1980s, at which time a statistical analysis of clinical data was the primary technique applied.

In 2001, Chugal et al. published data related to a study of teeth extracted after unsuccessful endodontic treatments at the University of Connecticut’s School of Dental Medicine. The patients included in this study were treated between 1988 and 1992 in the graduate program and had experienced unsuccessful endodontic treatment within the previous four years. Variables were taken from both the clinical trial and the X-rays taken at the time of the endodontic treatment. The data obtained in this case were studied with contingency tables and the chi-squared test. The risk factors were compared using -tests for independent groups, or with nonparametric tests (Mann-Whitney or Kruskal-Wallis) [11].

In 2010, using the same characteristics, Givol et al. published the results of his study performed in patients from private clinics in Israel. In this case, all the possible clinical variables prior and subsequent to the endodontic treatment were taken from 5,217 patients treated between 1992 and 2008. The data were also studied using chi-squared [12] statistical tests.

In July of 2011, Song et al. presented the data relative to a study performed on patients from the Department of Conservative Dentistry at the Dental College of Yonsei University, Seoul, Korea, between August 2004 and December 2008. Included in this study were patients who had undergone unsuccessful endodontic treatment and were in need of periapical surgery. Song took into account the clinical and X-ray data from prior to the treatment, demographic data, and data subsequent to the failed treatment. To analyze the factors that could predict the endodontic failure, he applied a chi-squared statistical study [13].

Of the previously cited works, none used artificial intelligence or case-based reasoning; nor did any use predictive tools other than the application of statistical studies to analyze risk factors. The use of this type of system offers, therefore, a wide area of study within the field of odontology and in particular with the prediction of unsuccessful endodontic treatments.

3. Proposed Reasoning System

The purpose of CBR is to solve new problems by adapting solutions that have been used to solve similar problems in the past [14]. The primary concept when working with CBRs is the concept of case. A case can be defined as a past experience and is composed of three elements: a problem description which describes the initial problem, a solution which provides the sequence of actions carried out in order to solve the problem, and the final state which describes the state achieved once the solution was applied. A CBR manages cases (past experiences) to solve new problems. The way cases are managed is known as the CBR cycle and consists of four sequential steps which are recalled every time a problem needs to be solved: retrieve, reuse, revise, and retain. Each of the steps of the CBR life cycle requires a model or method in order to perform its mission. The algorithms selected for the retrieval of cases should be able to search the case base and select the problem and corresponding solution most similar to the new situation. Once the most important cases have been retrieved, the reuse phase begins, in which the solutions for the retrieved cases are adapted and a new solution is generated. During this stage a mixture based on Bayesian networks is used to carry out the final classification. The revise phase consists of an expert revision for the proposed solution. Finally, the retain phase allows the system to learn from the experiences obtained in the three previous phases, consequently updating the cases memory.

In this work, we propose a predictive system based on the CBR paradigm, specifically designed to be applied in the field of odontology. Figure 1 depicts the CBR presented in this paper. As seen in Figure 1, the most innovative algorithms are included in the reuse phase, where a mixture of Bayesian networks is used. Another innovation can be observed in the revise phase, where the relevant variables are recovered by applying statistical tests to facilitate the process of reviewing the results provided during the reuse phase.

Figure 1 shows the four stages of the proposed CBR system: retrieve, reuse, revise, and retain, which are described in detail in the following subsections.

The cases are defined according to the variables and the final classification of the case; the cases are defined according to the following expression:where with represent the input variables and the predicted variable with the final classification.

3.1. Retrieve

During the retrieve phase, existing cases in which a retreatment was performed are selected from the case memory. This eliminates all cases that involve only an initial treatment. During the reuse phase, all of the cases are selected to generate the prediction model.

3.2. Reuse

During the reuse phase, previously retrieved cases are selected and an associated classifier is built. The classification algorithm proposes a mixture of experts, where different methods are taken into consideration, including decision trees, decision rules, probabilistic models, fuzzy models, function-based algorithms, and ensemble. The system selects these algorithms for each kind of method: decision rules RIPPER [15], OneR [16], M5 [17], decision trees J48 [18], CART [19] (Classification and Regression Trees) [20], probabilistic models naive Bayes [21], fuzzy models -NN (-Nearest Neighbors) [20], Bayesian networks [22], Support Vector Machine (SVM) [23], and finally ensemble, such as Bagging [24] and Ada-Boosting [25]. In this paper, the technique selected to carry out the classification phase corresponds to a mixture of classifiers based on Bayesian networks. The mixture of classifiers minimizes a specific functional objective, which prioritizes the option for retreatment. The new case is then introduced and classified according to the classifier built in this phase. The classifiers and the mixture used in this study are explained in the following subsections.

3.2.1. Bayesian Network

In order to build Bayesian networks, it is first necessary to establish search mechanisms that can generate the DAG (Directed Acyclic Graph) using a set of heuristics that can reduce the number of combinations and generate the final Bayesian network. There are various Bayesian network search mechanisms, including tabu search [22], conditional independence [26], K2 [22], HillClimber [22], and TAN (Tree Augmented Naive Bayes) [27].

(1) Tabu Search. Tabu search can perform heuristic searches to select the structure from the Bayesian network best suited to a specific problem. A Tabu search can reduce the search area but does not guarantee finding the optimal solution. Algorithm 1 shows the algorithm used to calculate the graph for the Bayesian network. is the set of variables, is the variable and corresponds to node . is the set of parents for node , tabuList is the queue that stores the most recent movements, and is the expression used to calculate the quality of node . The study presented by [22] shows some of the proposals for this function.

Input:
Output:
deque tabuList  ;
() , 0 eoc;
while  notstop  do
    where ;
   if   then  ;
   else  ;
   tabuList.removeFront();
   tabuList.addBack();
   //Update  the values of A based on the new edge (i, j)
   UpdateMatrix();
   if    then
  ;
   end
end

Once the algorithm is complete, it returns the set of parents for each of the children nodes.

(2) Conditional Independence Test. This algorithm is based on the calculation of the conditional independence test for the variables to generate a DAG that can obtain the probability estimates. If the variables being studied are independent, it will not be possible to generate a Bayesian network with good results. During the first phase, a graph containing the relationships between the variables is built; a DAG is then generated based on the previous graph. It is therefore necessary to take into consideration the number of categories of variables so that the analyses of independence are significant. Algorithm 2 shows the procedure for creating a Bayesian network established for the conditional independence algorithm [26]. The function test in Algorithm 2 applies the chi-square statistical test [8] when 80% of the expected counts from the contingency table are greater than 5. Otherwise, Fisher’s exact test [28] is applied. No statistical tests were applied for nominal variables since the data were discretized for the study, and the variables were converted to qualitative ordinals.

Input:
Output:
;
foreach   (  do
  foreach
  , with   do
     if test() < -value then
    
     end
  end
end
//Rules DAG
foreach    do
  
end
foreach  (  do
  
end
foreach  (  do
  
end
foreach  (  do
  
end

(3) Mixture of Bayesian Networks. This work carries out a mixture of experts to minimize both the classification process and the false positives. False positives are defined by the value . The objective function is established according to (2). In order to give greater weight to false negatives, the parameter, which has a default value of 1, must be modified:where defines the weight of the classifier , is the predicted value of the classifier , is the real output value, and is the weighted value of the false positives (default value 1).

In order to give the correct weight to the output of the classifiers, (3) must be considered:To optimize the problem, the method of Lagrange multipliers, as defined by (5), is applied, wherebyThe optimum value is calculated as in the following equations:

3.3. Revise and Retain

The revise phase is carried out by applying techniques that attempt to express the classifications performed by the Bayesian network. The explanation of the Bayesian network includes statistical techniques to extract relevant variables during the classification process: the chi-square [8], the Yates correction tests [29], the chi-square with the Monte Carlo simulation [30], and Fisher’s exact test [28] are applied to select the variables of interest that characterize the various pathologies. It is important to note that in order for the expected frequency to be less than 5, the result may be incorrect; consequently, Yates correction would be applied in an attempt to mitigate this issue. The statistical results from chi-squared test are also provided, applying the Monte Carlo simulation to verify the results. Finally, an exact Fisher test is applied, which is the recommended method when the sample size is small and it is not possible to ensure that 80% of the data from a contingency table have a value greater than 5. Medical studies such as [31] use a process similar to the one presented for selecting variables that affect malformations; other biomedical studies include [2830]. There are many alternatives for correcting data, such as that in [32].

4. Case Study and Results Obtained

A case study was designed using the data from the patient files at the Faculty of Odontology, Masters of Endodontics, at the Complutense University of Madrid. All patients received root canal treatments between September 2000 and May 2014. Among all the patients treated during this time, we selected 205 cases (205 failures) that satisfied the inclusion criteria and were interested in a follow-up appointment. Success of retreatment of root canal therapy is defined as no presence of radiographic and clinical symptoms in a period of five years after the treatment was performed; failure of retreatment of root canal therapy is defined as the presence of radiographic lesions around the tooth retreated and presence of signs such as pain, movility, fistula, and inflammation. The retreatments were reviewed every year.

None of the patients from the selected cases who came for a follow-up treatment refused to participate in the study. The selected 205 cases contained all the information needed to complete the 72 variables being considered with 105 failures in retreatment. Some of the 72 variables were recombined in categorical values because they were binaries while the others were removed, for example, address of the treatment and sender. Certain initial variables included a high number of categories, which resulted in their recodification to ensure that the final number of categories per variable had around 3 or 4 different values. The final list of variables is summarized and described in Table 1. The variable to predict is highlighted in bold. These variables take into account all information relevant to the patient: medical and dental history and habits. Data relative to the state of the tooth prior to treatment were also included: the evolution, the clinical technique used, and the posttreatment results.


VariableClass

Habits-parafunctionsCategorical 2 values
General pathologyBinary
Total current treatmentsBinary
AllergyBinary
SessionsDiscrete
Mechanical/manual instrumentationBinary
Lateral or verticalBinary
AnestheticCategorical 3 values
ClampsCategorical 4 values
Ranking difficulty levelCategorical 3 values
Student courseDiscrete 4 values
Tooth positionCategorical 3 values
Anatomical characteristics of the crownCategorical 3 values
Root anatomyCategorical 3 values
AnomaliesBinary
Type of restoration: PernoBinary
PernoBinary
TypeBinary
Diámetro diameterCategorical 5 values
LengthCategorical 3 values
Time endodontics-restorationCategorical 4 values
Type of painCategorical 4 values
InflammationBinary
FistulaBinary
Number of rootsDiscrete 3 values
Number of tubesDiscrete 4 values
Root morphologyBinary
Curvatures Binary
DegreeCategorical 3 values
Bone levelCategorical 3 values
Stable occlusionCategorical 3 values
Fracture typeCategorical 2 values
LocationCategorical 5 values
Signs of fissure/fractureBinary
ProbingBinary
MovilityBinary
visible crackBinary
LevelBinary
Time to failureCategorical 4 values
RetreatmentsBinary
Percha solventBinary
Use of rotaryBinary
Failures in retreatmentBinary

We then analyzed the system explained in Figure 1. The system was tested with the 205 cases to predict the failures in retreatment depending on the variables shown in Table 1. In summary, the reuse phase was analysed according to different configurations and using the accuracy rate and the area under the ROC curve (AUC); the relevant variables extracted in the revise phase are shown in this section.

In the retrieve phase, the system extracted the cases with retreatment and used them to generate the classifiers during the reuse phase, The system was compared with different classifiers applied in the reuse phase; those specifically applied include BayesNet, NaiveBayes, AdaBoostM1, Bagging, DecisionStump, J48, IBK, JRip, LMT, Logistic, LogitBoost, OneR, SMO, and Stacking. The results obtained by applying the leave-one-out technique are shown in Table 2. In summary, the test was carried out as follows: we extracted a case in the memory and then proceeded with the CBR system explained in Figure 1. We can observe that the accuracy rate of the system is greater than the other classifiers, although the procedure is insufficient to determine whether the differences are statistically significant.


ClassifierCorrect

CBR system with Bayesian networks173
NaiveBayes157
AdaBoostM1162
Bagging149
DecisionStump141
J48154
IBK154
JRip158
LMT161
Logistic152
LogitBoost168
OneR141
SMO163

To evaluate the significance of the different techniques presented in Table 2, a cross-validation was established following Dietterich’s -cross-validation paired -test algorithm [33]. Instead of using the accuracy rate, the AUC was used, since the classification problem is not symmetrical. Value 5 in the algorithm represents the number of replications of the training process, and value 2 is the number of sets into which the global set is divided (2-fold). Thus, for each technique, the global dataset was divided into two groups and as follows: and . The learning and estimation stages were then carried out. This process was repeated 5 times for each technique and included the following steps: the classifier was trained using and was then used to classify and . In a second step, the classifier was trained using and was then used to classify and . The results obtained are shown in Table 3, where the columns represent the success rate obtained for , (-A trained with ) and , (-B trained with ) for each repetition. The rows of Table 3 show the different classifiers previously shown in Table 2.


-A-B-A-B-A-B-A-B-A-B Average test samples

NaiveBayes0.900.890.910.840.920.830.900.770.920.790.920.890.920.850.890.860.910.840.900.880.84
AdaBoostM10.890.890.910.810.930.770.910.780.950.820.900.820.930.790.910.870.920.820.920.810.82
Bagging0.950.870.950.840.970.700.960.730.980.800.940.800.920.750.940.830.970.800.920.730.79
DecisionStump0.690.690.640.570.690.570.740.640.690.590.730.650.720.630.690.690.700.680.720.630.63
J480.920.790.960.790.910.760.930.730.970.700.910.770.970.690.880.700.940.610.920.720.73
IBk1.000.771.000.781.000.741.000.701.000.711.000.831.000.791.000.821.000.751.000.740.76
JRip0.790.670.800.690.760.530.900.620.860.670.910.710.900.700.690.690.890.720.850.740.68
LMT1.000.850.970.810.940.770.980.810.990.790.960.830.980.800.940.891.000.830.960.830.82
Logistic1.000.701.000.641.000.651.000.691.000.621.000.621.000.711.000.681.000.731.000.790.68
LogitBoost0.930.900.950.820.960.780.950.810.970.810.940.860.950.780.920.890.950.800.940.830.83
OneR0.690.690.690.660.690.570.740.640.690.590.730.650.720.630.690.690.700.680.720.630.64
SMO0.920.770.950.730.960.720.910.720.980.720.910.750.960.750.910.750.970.740.900.710.74
CBR system0.970.890.960.860.980.860.970.810.990.800.970.900.980.880.970.910.980.900.980.870.87

Figure 2 shows the box plot associated with the AUC for each of the methods. As shown, the interquartile range for the CBR system is less than that for the other methods.

Once the results presented in Table 3 were obtained, a study on the significance of the different classification techniques was performed by applying the Mann-Whitney test. It was a nonparametric test in which it is not necessary to make assumptions on the data distribution, as in the -test. The test determines two values: and . shows whether the AUC value of the column classifier is greater than that of the file method, whereas determines if the AUC for the ROC curve of the column is lower than the row. The values above the diagonal in Table 4 show the level of significance for the statistical test; therefore, if the column classifier has an area under curve greater than the file (i.e., level of significance > 0.05) it is shown in normal type; otherwise, it is shown in bold type. Clearly, the CBR system approach has a greater AUC for the ROC curve than the other methods.


NaiveBayesAdaBoostM1BaggingDecisionStumpJ48IBkJRipLMTLogisticLogitBoostOneRSMOCBR system

NaiveBayes0.050.010.000.000.000.000.050.000.180.000.000.96
AdaBoostM10.060.120.000.000.010.000.630.000.740.000.000.99
Bagging0.010.080.000.010.180.000.940.000.940.000.021.00
DecisionStump0.000.000.001.001.000.981.000.981.000.701.001.00
J480.000.000.011.000.940.031.000.041.000.000.601.00
IBk0.000.000.161.000.950.001.000.001.000.000.061.00
JRip0.000.000.000.950.030.001.000.461.000.031.001.00
LMT0.090.580.941.001.001.001.000.000.570.000.000.99
Logistic0.000.000.000.980.040.000.600.001.000.051.001.00
LogitBoost0.170.700.961.001.001.001.000.641.000.000.000.98
OneR0.000.000.000.690.000.000.080.000.040.001.001.00
SMO0.000.000.011.000.690.040.990.000.990.001.001.00
CBR system0.911.001.001.001.001.001.001.001.000.981.001.00

The analysis of the cross-validation is completed using the Dietterich’s -cross-validation paired -test [33]. The results obtained are shown in the lower diagonal of Table 4. It is possible to observe that the results are very similar to those previously shown in the upper diagonal (Mann-Whitney test). The values below the diagonal contain the results. In this case, the file classifier is compared to the column using the same hypotheses as the Mann-Whitney case. The values greater than 0.05 indicate that the area of the row classifier is greater than that of the column. The CBR system provides the best results in the test.

The mixture was compared to other Bayesian network search algorithms in order to analyze its results. As with the comparisons of other methods, a cross-validation was performed, which provided the values shown in Table 5 for the AUC of the ROC curve. In this case, the average value appears to be less than the value for the methods shown in Table 3. The mixture increases the AUC provided for the other methods. By applying statistical tests as with Table 2, we can conclude that the value of the AUC for the mixture is statistically different for all methods with a significance level of 0.1. This was to be expected since the mixture is composed of both methods.


-A-B-A-B-A-B-A-B-A-BAverage test samples

Global HillClimber1.000.731.000.841.000.841.000.791.000.721.000.761.000.831.000.781.000.851.000.840.80
Global TabuSearch0.990.790.980.850.990.820.990.850.980.860.990.790.980.860.990.850.990.820.980.870.84
Global K21.000.811.000.801.000.831.000.891.000.841.000.821.000.851.000.821.000.861.000.850.84
Global TAN0.990.780.990.831.000.870.990.880.990.820.990.810.990.851.000.820.990.851.000.860.84
CISearchAlgorithm0.930.830.900.790.920.840.920.830.900.870.920.820.900.800.930.850.900.820.920.840.83
Local TabuSearch 0.970.830.960.830.960.880.970.870.960.860.970.840.970.830.970.870.970.850.980.830.84
Local K20.990.831.000.790.990.871.000.861.000.801.000.821.000.891.000.821.000.811.000.870.84
Local TAN0.980.840.990.830.980.860.990.860.990.821.000.810.980.861.000.860.990.850.990.870.84
Mixture  
Local TabuSearch  
CISearchAlgorithm
0.990.851.000.830.990.881.000.900.990.860.990.840.990.900.990.881.000.841.000.870.87

In order to explain the relevant variables during the reuse phase in the CBR system, the difference between the values of the variables for the categories of successful retreatments and extractions were analysed in the revise phase. To perform this analysis, the chi-square, Yates correction, chi-square with Monte Carlo simulation, and Fisher’s exact tests were applied. Table 6 displays the set of variables that were considered relevant by any of the three methods. We can see how the selection of variables coincides to a great degree for the different methods. This method for extracting relevant variables makes it possible to determine, to a large extent, the relevance of the variables that the Bayesian networks will have, particularly since the analysis of the dependence between variables is also used.


Variable value
Chi-squared testExact Fisher test
YatesMonte Carlo

Allergy0.019087880.020989510.01739722
Mechanical/manual instrumentation0.00359340.0019990.0029985
Anesthetic0.003062690.00299850.00149925
Ranking difficulty level0.000499750.00049975
Tooth number0.004655140.006496750.00449775
Root anatomy0.000499750.00049975
Type of restoration: Perno0.000499750.00049975
Perno0.000172390.00049975
Type0.005374580.0039980.00385604
Time endodontics-restoration0.000499750.00049975
Length0.006019850.005497250.00549725
Number of roots0.023521220.025987010.02198901
Curvatures yes/no0.023172780.029985010.02240645
Adjacent remaining0.001093890.00299850.00149925

5. Conclusions

This paper has presented an innovative system specially designed to help odontologists make decisions about retreatment. The medical staff that participated in the experiments have remarked on the usefulness of the proposed approach and have noted that the system can be very helpful for their work.

The results obtained show that, with the CBR analysis, the data obtained were relevant because by ordering the established variables, particularly those with the highest risk factor, we could predict the final solution for treatment and retreatment in 84.4% of the cases, by applying the leave-one-out techniques.

The combination applied in the mixture increases the AUC for the ROC curve, thus increasing the rate of accuracy for the results, which is important when working with nonsymmetrical case studies. The mixture also makes it possible to reduce the number of false negatives by placing great importance on the possibility of false positives. Moreover, the objective function can be modified depending on the case study, thus allowing for an increase in the relevance of some metrics.

In other case studies, it could be necessary to analyze the classifiers in the mixture in order to optimize the objective functions. In this case, the mixture of the Bayesian networks provided good results in some cases, while in others we could use alternative techniques such as decision trees in order to provide some rules to explain the classification in the revise phase.

Furthermore, the system makes it possible to extract the relevant variables that can distinguish the different types of retreatments. Nevertheless, more cases are required to contrast the results with greater accuracy.

The system can reduce the number of unsuccessful retreatments because it predicts the rate of success or failure, thus avoiding unnecessary extractions.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

This work has been carried out by the Project Sociedades Humano-Agente en Entornos Cloud Computing (Soha+C) SA213U13, a project cofinanced with Junta Castilla y León funds.

References

  1. J. F. De Paz, J. Bajo, V. Vera, and J. M. Corchado, “MicroCBR: a case-based reasoning architecture for the classification of microarray data,” Applied Soft Computing Journal, vol. 11, no. 8, pp. 4496–4507, 2011. View at: Publisher Site | Google Scholar
  2. J. M. Corchado, J. F. De Paz, S. Rodríguez, and J. Bajo, “Model of experts for decision support in the diagnosis of leukemia patients,” Artificial Intelligence in Medicine, vol. 46, no. 3, pp. 179–200, 2009. View at: Publisher Site | Google Scholar
  3. S. S. Sahu and G. Panda, “Identification of protein-coding regions in DNA sequences using a time-frequency filtering approach,” Genomics, Proteomics & Bioinformatics, vol. 9, no. 1-2, pp. 45–55, 2011. View at: Publisher Site | Google Scholar
  4. J. Zhang, R. Chiodini, A. Badr, and G. Zhang, “The impact of next-generation sequencing on genomics,” Journal of Genetics and Genomics, vol. 38, no. 3, pp. 95–109, 2011. View at: Publisher Site | Google Scholar
  5. I. Jurisica and J. Glasgow, “Applications of case-based reasoning in molecular biology,” AI Magazine, vol. 25, no. 1, pp. 85–95, 2004. View at: Google Scholar
  6. C. Canalda and E. Brau, Endodoncia: técnicas clínicas y bases científicas, vol. 2, Masson, Barcelona, Spain, 2006.
  7. J. M. Casanellas, Restauración del Diente Endodonciado, Pues, Madrid, Spain, 1st edition, 2006.
  8. J. F. Kenney and E. S. Keeping, Mathematics of Statistics, part 2, Van Nostrand, Princeton, NJ, USA, 2nd edition, 1951.
  9. M. John and C. E. Priebe, “A data-adaptive methodology for finding an optimal weighted generalized Mann-Whitney-Wilcoxon statistic,” Computational Statistics and Data Analysis, vol. 51, no. 9, pp. 4337–4353, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  10. W. H. Kruskal and W. A. Wallis, “Use of ranks in one-criterion variance analysis,” Journal of the American Statistical Association, vol. 47, no. 260, pp. 583–621, 1952. View at: Publisher Site | Google Scholar
  11. N. M. Chugal, J. M. Clive, and L. S. W. Spångberg, “A prognostic model for assessment of the outcome of endodontic treatment: effect of biologic and diagnostic variables,” Oral Surgery, Oral Medicine, Oral Pathology, Oral Radiology, and Endodontics, vol. 91, no. 3, pp. 342–352, 2001. View at: Publisher Site | Google Scholar
  12. N. Givol, E. Rosen, S. Taicher, and I. Tsesis, “Risk management in endodontics,” Journal of Endodontics, vol. 36, no. 6, pp. 982–984, 2010. View at: Publisher Site | Google Scholar
  13. M. Song, I.-Y. Jung, S.-J. Lee, C.-Y. Lee, and E. Kim, “Prognostic factors for clinical outcomes in endodontic microsurgery: a retrospective study,” Journal of Endodontics, vol. 37, no. 7, pp. 927–933, 2011. View at: Publisher Site | Google Scholar
  14. J. Kolodner, Case-Based Reasoning, Morgan Kaufmann Publishers, 1993.
  15. W. W. Cohen, “Fast effective rule induction,” in Proceedings of the 12th International Conference on Machine Learning (ICML '95), pp. 115–123, Morgan Kaufmann Publishers, Tahoe City, Calif, USA, July 1995. View at: Google Scholar
  16. G. Holmes, M. Hall, and E. Prank, “Generating rule sets from model trees,” in Advanced Topics in Artificial Intelligence, vol. 1747 of Lecture Notes in Computer Science, pp. 1–12, Springer, Berlin, Germany, 1999. View at: Publisher Site | Google Scholar
  17. R. C. Holte, “Very simple classification rules perform well on most commonly used datasets,” Machine Learning, vol. 11, no. 1, pp. 63–91, 1993. View at: Publisher Site | Google Scholar
  18. J. R. Quinlan, C4.5: Programs for Machine Learning, Morgan Kaufmann Publishers, 1993.
  19. L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classification and Regression Trees, Wadsworth International Group, 1984. View at: MathSciNet
  20. D. W. Aha, D. Kibler, and M. K. Albert, “Instance-based learning algorithms,” Machine Learning, vol. 6, no. 1, pp. 37–66, 1991. View at: Publisher Site | Google Scholar
  21. R. O. Duda and P. Hart, Pattern Classification and Scene Analysis, John Wisley & Sons, New York, NY, USA, 1973.
  22. R. R. Bouckaert, Bayesian Belief Networks: from Construction to Inference, University of Utrecht, Utrecht, The Netherlands, 1995.
  23. V. Vapnik and A. Lerner, “Pattern recognition using generalized portrait method,” Automation and Remote Control, vol. 24, pp. 774–780, 1963. View at: Google Scholar
  24. L. Breiman, “Bagging predictors,” Machine Learning, vol. 24, no. 2, pp. 123–140, 1996. View at: Google Scholar
  25. Y. Freund and R. E. Schapire, “Experiments with a new boosting algorithm,” in Proceedings of the 13th International Conference on Machine Learning (ICML '96), pp. 148–156, Bari, Italy, July 1996. View at: Google Scholar
  26. T. Verma and J. Pearl, “An algorithm for deciding if a set of observed independencies has a causal explanation,” in Proceedings of the 8th Conference on Uncertainty in Artificial Intelligence, pp. 323–330, Stanford, Calif, USA, July 1992. View at: Google Scholar
  27. N. Friedman, D. Geiger, and M. Goldszmidt, “Bayesian network classifiers,” Machine Learning, vol. 29, no. 2-3, pp. 131–163, 1997. View at: Publisher Site | Google Scholar
  28. X. Yang, Y. Huang, M. Crowson, J. Li, M. L. Maitland, and Y. A. Lussier, “Kinase inhibition-related adverse events predicted from in vitro kinome and clinical trial data,” Journal of Biomedical Informatics, vol. 43, no. 3, pp. 376–384, 2010. View at: Publisher Site | Google Scholar
  29. D. B. Shaul, B. Scheer, S. Rokhsar et al., “Risk factors for early infection of central venous catheters in pediatric patients,” Journal of the American College of Surgeons, vol. 186, no. 6, pp. 654–658, 1998. View at: Publisher Site | Google Scholar
  30. B. Nilsson, “A compression algorithm for pre-simulated Monte Carlo p-value functions: application to the ontological analysis of microarray studies,” Pattern Recognition Letters, vol. 29, no. 6, pp. 768–772, 2008. View at: Publisher Site | Google Scholar
  31. O. Himmetoglu, M. B. Tiras, R. Gursoy, O. Karabacak, I. Sahin, and A. Onan, “The incidence of congenital malformations in a Turkish population,” International Journal of Gynecology and Obstetrics, vol. 55, no. 2, pp. 117–121, 1996. View at: Publisher Site | Google Scholar
  32. A. M. Andrés and A. S. Mato, “Optimal correction for continuity and conditions for validity in the unconditional chi-squared test,” Computational Statistics and Data Analysis, vol. 21, no. 6, pp. 609–626, 1996. View at: Publisher Site | Google Scholar
  33. T. G. Dietterich, “Approximate statistical tests for comparing supervised classification learning algorithms,” Neural Computation, vol. 10, no. 7, pp. 1895–1923, 1998. View at: Publisher Site | Google Scholar

Copyright © 2016 Livia Campo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

832 Views | 373 Downloads | 3 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19. Sign up here as a reviewer to help fast-track new submissions.