The Scientific World Journal

Volume 2014 (2014), Article ID 398269, 13 pages

http://dx.doi.org/10.1155/2014/398269

## Efficiently Hiding Sensitive Itemsets with Transaction Deletion Based on Genetic Algorithms

^{1}Innovative Information Industry Research Center (IIIRC), School of Computer Science and Technology, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen 518055, China^{2}Shenzhen Key Laboratory of Internet Information Collaboration, School of Computer Science and Technology, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen 518055, China^{3}Medical School, Shenzhen University, Shenzhen 518060, China^{4}Department of Computer Science and Information Engineering, National University of Kaohsiung, Kaohsiung 811, Taiwan^{5}Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung 804, Taiwan

Received 28 May 2014; Revised 7 August 2014; Accepted 8 August 2014; Published 1 September 2014

Academic Editor: Shifei Ding

Copyright © 2014 Chun-Wei Lin et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

Data mining is used to mine meaningful and useful information or knowledge from a very large database. Some secure or private information can be discovered by data mining techniques, thus resulting in an inherent risk of threats to privacy. Privacy-preserving data mining (PPDM) has thus arisen in recent years to sanitize the original database for hiding sensitive information, which can be concerned as an NP-hard problem in sanitization process. In this paper, a compact prelarge GA-based (cpGA2DT) algorithm to delete transactions for hiding sensitive itemsets is thus proposed. It solves the limitations of the evolutionary process by adopting both the compact GA-based (cGA) mechanism and the prelarge concept. A flexible fitness function with three adjustable weights is thus designed to find the appropriate transactions to be deleted in order to hide sensitive itemsets with minimal side effects of hiding failure, missing cost, and artificial cost. Experiments are conducted to show the performance of the proposed cpGA2DT algorithm compared to the simple GA-based (sGA2DT) algorithm and the greedy approach in terms of execution time and three side effects.

#### 1. Introduction

With the rapid growth of data mining technologies in recent years, useful and meaningful information can thus be easily discovered for the purpose of decision making in different domains. The discovered information can be mostly classified into association rules [1–5], sequential patterns [6–9], classification [10–12], clustering [13, 14], and utility mining [15–18], among others. Among them, mining association rules method is the most common way to find the potential relationships between the purchased items or goods in a very large database. Some applications require protection against the disclosure of private, confidential, or secure data. For example, social security numbers, address information, credit card numbers, and purchasing behaviors of customers can be considered as the confidential, private, or privacy information.

Instead of personal information, privacy issue can be extended to business. Based on business purposes, shared information among companies may be extracted and analyzed by other partners, thus causing the security threats. Privacy-preserving data mining (PPDM) [19–22] was proposed to reduce privacy threats by hiding sensitive information while allowing required information to be discovered from databases. Such data may implicitly contain confidential information that will lead to privacy threats if it is misused. Heuristic methods [20, 21, 23–26] have been proposed to choose the appropriate data for sanitization in order to hide the sensitive information. During the procedure to hide the sensitive information, side effects of missing cost and artificial cost are thus generated and should be concerned in PPDM. The optimal way to select the sensitive information to be hidden is, however, concerned as the NP-hard problem in sanitization process [22, 27]. Genetic algorithms (GAs) [28] are able to find optimal solutions using the principles of natural evolution. The amount of chromosomes is thus required to process the several operations in evaluation process of simple GAs.

To solve the limitations of traditional GA-based algorithms with high requirements of memory and computations at each evolutionary process, the compact GA (cGA) mechanism [29] and the prelarge concept [30] are adopted in the proposed cpGA2DT algorithm. Based on the cGA mechanism, only two chromosomes are competed to each other at each iteration. The probabilities of transactions to be selected are increased along with the winner chromosome. The probabilities of transactions to be selected are, however, decreased along with the loser chromosome. Since only two chromosomes are generated for the competition, the memory requirements of populations can be greatly reduced. In addition, a flexible fitness function is designed to evaluate three side effects at each evolutionary process. This procedure causes the computations of multiple database rescans. The prelarge concept is adopted in the proposed cpGA2DT algorithm to find the prelarge itemsets [30, 31] in advance, thus reducing the computations of multiple database rescans at each evolution. To the best of our knowledge, this is the first approach to solve the limitations by considering both the time and the space complexities with transaction deletion for hiding sensitive itemsets. A straightforward approach (greedy) and a simple GA-based algorithm are also designed as a benchmark to evaluate the performance of the proposed cpGA2DT in regard to the execution time and the number of three side effects in the experiments. Contributions of this paper can be illustrated as follows.(1)Most past approaches applied heuristic ways to sanitize the original database for the purpose of hiding sensitive itemsets by deleting partial items. In this paper, a GA-based approach is thus proposed to optimize the selected transactions to be deleted, thus minimizing the side effects in PPDM.(2)It requires the amount of memory in evaluation process based on traditional GA approach. In this proposed approach, cGA is applied to reduce the population size based on probability distribution to select the appropriate transactions to be deleted.(3)The prelarge concept is used in the proposed algorithm to reduce the execution time for database rescan in chromosome evaluation.(4)An evaluation function with three adjustable weights is designed in the evaluation process to minimize the side effects of PPDM.

The remainder parts of this paper are organized as follows. Related works are described in Section 2; preliminary of PPDM is mentioned in Section 3. The proposed approach is illustrated in Section 4. An example is given in Section 5. Experiments are conducted in Section 6. Conclusion is given in Section 7.

#### 2. Review of Related Works

Related works of genetic algorithms, data sanitization, and prelarge concept are briefly reviewed in this section.

##### 2.1. Genetic Algorithms

Holland applied the natural selection and the survival of the fittest of Darwin theory and proposed the evolutionary computation of genetic algorithms (GAs) [28]. GAs are the search techniques, which are designed and developed to find a set of feasible solutions in a limited amount of time [32, 33]. According to the principle of survival of the fittest, GAs generate the next population by various operations with each individual in the population representing a set of possible solutions. Three basic operations including crossover, mutation, and selection are performed on chromosomes for the next generations. Each chromosomea is then evaluated by the designed fitness function. This procedure is recursively processed until the predefined termination criteria are achieved. Flowchart of GAs is shown in Figure 1.

Traditional GAs have to generate the size of populations for the purpose of performing crossover, mutation, and selection operations for the next generations, thus causing memory lack problem. Compact genetic algorithm (cGA) was thus proposed to simulate traditional GAs with only the probability vector for selection operation and population size without the crossover and mutation operations in order to generate two individuals (or chromosomes) at competition [29]. The probability of the* i*th vector in the winner chromosome is increased, but the loser probability is decreased. A cGA algorithm can reduce the memory requirements without the crossover and mutation operations but still can approximately mimic the behaviors of traditional GAs.

##### 2.2. Data Sanitization

Data mining [1, 34–37] is progressively developed to extract useful and meaningful information or rules from a very large database. The misuse of data mining techniques may, however, lead to security threats and privacy concerns. Privacy-preserving data mining (PPDM) [19, 23, 24, 38] was thus proposed to hide the confidential, private, or secure information before it is published in public or shared among alliances. Most approaches were proposed to perturb the original database for the purpose of hiding sensitive information in PPDM. Agrawal and Srikant introduced a quantitative measure to evaluate the utility of PPDM methods [19]. Lindell and Pinkas stated hiding confidential information on the union of shared databases among two parties without revealing any unnecessary information [20]. Oliveira and Zaïane, respectively, designed the multiple-rule hiding MinFIA, MaxFIA, and IGA algorithms to efficiently hide sensitive itemsets and introduced the performance measures for three side effects [39]. Dasseni et al. then proposed a hiding approach based on the hamming-distance approach to decrease the confidence or support values of association rules for hiding sensitive information [40]. Three heuristic algorithms are designed, respectively, to increase the supports of antecedent parts, to decrease the supports of consequent parts, and to decrease the support of either the antecedent or the consequent parts until the supports or confidences of association rules below the threshold values. Amiri then proposed aggregate, disaggregate, and hybrid approaches to hide multiple sensitive rules [23]. The designed aggregate approach computes the union of the supporting transactions for all sensitive itemsets. The transactions with the most sensitive and the least sensitive itemsets are thus removed to hide the sensitive information. The disaggregate approach aims to remove individual items from transactions and then remove whole transactions, thus reducing side effects of PPDM. Hybrid one is to combine the previous designed algorithms to firstly identify sensitive transactions and secondly to delete items from those of transactions until the sensitive information has been hidden. Many heuristic approaches are still being developed in progress for the purpose of hiding different types of knowledge in PPDM [21, 26, 41].

The optimal sanitization of databases is regarded to be an NP-hard problem [22, 27]. Genetic algorithms (GAs) were usually used to find optimal solutions in the least amount of time [28]. Fewer studies have adopted GAs to find optimal solutions to hide sensitive information. Han and Ng proposed secure protocols for rule discovery based on private arbitrarily partitioned data among two parties without compromising their data privacy using GAs [42]. It uses the true positive rate multiplied by the true negative rate to define the fitness function for evaluating the goodness of each decision rule. Dehkordi et al. designed three multiobjective methods to partially remove the items from the original database [43]. Only the number of modified transactions is considered in the fitness function for evaluation. The other side effects of missing cost and artificial cost thus arose in the evaluation process. In this paper, three side effects are concerned in the designed fitness function for hiding sensitive itemsets with transaction deletion based on cGA algorithm.

##### 2.3. Prelarge Concept

Data mining techniques are used to discover useful and meaningful information or rules to aid managers in making efficient decisions in many different domains. Most data mining techniques handle, however, the static database to extract the required information. Cheung et al., respectively, designed FUP [44] and FUP2 [45] concepts to maintain and update the discovered information in dynamic databases. The original database is still, however, required to be rescanned based on the FUP and FUP2 concepts in the updating process. Hong et al. proposed prelarge concepts [30, 31] for the purpose of efficiently updating the discovered information without rescanning the original database each time. Prelarge itemset is not large itemset but has high potential to be large in the future through the data insertion or deletion process. Upper (the same as the minimum support threshold in conventional mining algorithms) and lower support thresholds are used to define the large and prelarge itemsets. Prelarge itemsets are used as a buffer to reduce the movement of an itemset directly from large to small and vice versa. For transaction deletion based on prelarge concept [30], nine cases thus arose and are shown in Figure 2.

From Figure 2, cases 2, 3, 4, 7, and 8 do not affect the final frequent itemsets of association rules. Case 1 may remove some discovered frequent itemsets of association rules. Cases 5, 6, and 9 may produce new frequent itemsets of association rules. If all frequent or prelarge itemsets are prestored from the original database, cases 1, 5, and 6 can be easily maintained and updated. An itemset in Case 9 cannot possibly be a large itemset in the updated database as long as the number of deleted transactions is a considerably small proportion of the original databases, which can be defined as [30] where is a lower support threshold, is an upper support threshold, and is the number of transactions in databases. If the number of deleted transactions satisfies the above condition, which is smaller than the safety bound , an itemset in Case 9 is absolutely not large in the updated databases. It is thus unnecessary to rescan the original databases. In the proposed cpGA2DT, the prelarge concepts are adopted to reduce the database rescan in the evaluation process, thus speeding up computations.

#### 3. Preliminaries

Before sanitization process to hide the sensitive itemsets, frequent itemsets can be discovered by data mining techniques. Let be the set of items in the database ; a database consists of several transactions as , in which each transaction is a set of items. A minimum support threshold is set at . Denote a support of an item (itemset) by . An item (itemset) is denoted by if it is considered as a large or frequent item (itemset) as .

In PPDM, it is required not only to hide sensitive itemsets but also to minimize the side effects. The relationship of itemsets before and after the PPDM process can be seen in Figure 3, where represents the large itemsets of , represents the sensitive itemsets defined by users that are large, ~ represents the nonsensitive itemsets that are large, and is the large itemsets after some transactions are deleted.

Let be the number of sensitive itemsets that fail to be hidden. Thus, the number of sensitive itemsets should ideally be zero after the database is sanitized. The set of sensitive itemsets is shown in Figure 4, in which part is the interaction of and .

*Definition 1. *The hiding failure of the sensitive itemsets in PPDM is defined as , in which .

Another evaluation criterion is the number of missing itemsets, which is denoted by . A missing itemset is a nonsensitive large itemset in the original database but is not extracted from the sanitized database. This side effect is shown in Figure 5, in which the part is the difference of ~ and .

*Definition 2. *The missing itemsets in PPDM are defined by , in which .

The last evaluation criterion is the number of artificial itemsets, which is denoted by . It represents the set of large itemsets appearing in the sanitized database but not belonging to the large itemset in the original database. This side effect is shown in Figure 6, in which the part is the difference of and .

*Definition 3. *The artificial itemsets in PPDM are defined as , in which .

Hiding sensitive itemsets or information is not only one purpose of PPDM but also minimizing the above side effects for data sanitization.

#### 4. Proposed Compact Prelarge Genetic Algorithm to Delete Transactions (cpGA2DT)

In this paper, a cpGA2DT approach is thus proposed to find the appropriate transactions to be deleted for hiding sensitive itemsets. The sensitive itemsets to be hidden can be defined below.

*Definition 4. *Suppose that a set of HS consist of the amounts of sensitive itemsets to be hidden; thus .

In the proposed cpGA2DT for hiding the sensitive itemsets through transaction deletion, the support count of a sensitive itemset must be below the minimum support threshold, in which each transaction to be deleted must contain any of the sensitive itemsets in HS.

*Definition 5. *Suppose an original database ; a database is thus projected from , in which each in must consist of any of the sensitive itemsets in HS.

In GAs, a chromosome corresponds to a possible solution. Suppose that is appropriate transactions from to be deleted for hiding the sensitive itemsets. A chromosome with genes is thus designed. Each gene represents a possible transaction to be deleted as a positive integer of transaction ID (TID) value or* null*.

*Definition 6. *Suppose a projected database , in which each represents a transaction ID. Suppose that is appropriate transactions to be deleted; a chromosome is a set of gens. Each in is represented as a transaction or* null*.

In GAs, a flexible fitness function with three adjustable weights to evaluate the goodness of chromosomes is thus designed.

*Definition 7. *A fitness function to evaluate the goodness of a chromosome is defined as
where , , and are the weighting parameters. The , , and are the hiding failure, missing cost, and artificial cost. Details of the notations and the proposed cpGA2DT algorithm are described in Algorithm 1.

##### 4.1. Proposed cpGA2DT Algorithm

The designed cpGA2DT algorithm is described in Algorithm 1.

For the proposed cpGA2DT, it adopts both the compact GA and prelarge concepts to reduce not only the computations of database rescan but also the population size at each evaluation. Prelarge itemsets (PL) act like buffers and are used to reduce the movement of itemsets directly from large to small and vice versa when transactions are deleted (in steps (1) and (2)). In competition process, only two individuals are used for competition (in step (8)). This approach can reduce the population size to speed up the evaluation process. When the termination condition is not satisfied, two chromosomes are then generated again, respectively, to increase the probability of selected transactions in the winner chromosome but decrease the probability of selected transactions in the loser chromosome.

#### 5. An Illustrated Example

In this section, an example is given to demonstrate the proposed cpGA2DT for privacy-preserving data mining. Assume that an original database contains 10 transactions shown in Table 1.

Also assume that the set of sensitive itemsets is defined as to be hidden. The minimum support threshold is set at 40%. The proposed algorithm is then processed as follows. The transactions with any of the sensitive itemsets in Table 1 are then projected. In this example, transactions 2, 3, 4, 5, 7, and 8 are then projected to form another projected database. The initial probabilities of those five transactions are initially set at 0.5. The lower support threshold for deriving the prelarge itemsets in this example is calculated as (). The database is scanned to find the large and prelarge itemsets. The results are, respectively, shown in Tables 2 and 3.

Two chromosomes (individuals) are then generated randomly according to the probability vector with 4 genes. The results are then shown in Table 4.

The chromosomes in Table 4 are then competed by the designed fitness function. In this example, the weights for three factors are, respectively, set as 0.5, 0.3, and 0.2. Take as an example to illustrate the evolutionary process. The number of hiding failures for is 0 since all sensitive itemsets (*be*,* bce*) are completely hidden; the number of missing itemsets of is 3 (itemsets* e*,* bc*, and* ce* are missing), and the number of artificial itemsets of is 1 (itemset* ac* arose). The fitness value of is calculated as (=1.1). The is processed in the same way, and (=0.9). In the competition process, the is better than ; the probabilities of transactions 2, 3, 4, and 7 are then, respectively, increased and updated in the probability vector by 0.5 + 1/6 (=0.667); the probabilities of transactions 2, 5, 7, and 8 are then, respectively, decreased and updated in the probability vector by 0.5 – 1/6 (=0.33). After that, the probability vector is updated and shown in Table 5.

Steps (5) to (8) are then, recursively, processed until the termination condition is satisfied. In this example, three criteria are used as the termination conditions. The criteria are as follows. The fitness function value of the best chromosome is 0; or a predefined number of generations is achieved; or the probability vector is converged. After the evolutionary process, the top-4 transactions with high probabilities in the probability vector are then selected as the transactions to be deleted in the sanitization process.

#### 6. Experimental Results

Experiments are conducted to show the performance of the proposed cpGA2DT, which was performed on a Pentium IV processor at 2 GHz and 512 M of RAM running on the Mandriva platform. A greedy approach and a simple GA-based algorithm [46] are also designed as a benchmark to be compared with the proposed algorithm. For the greedy approach, it scans the transactions from top to down to directly delete the transactions with sensitive itemsets. The termination of the greedy algorithm is the number of the deleted transactions, which is predefined by users. A simple GA-based approach uses simple GAs to hide the sensitive information. Three real databases mushroom [47], BMS-WebView1 [48], and BMS-WebView2 [48] are used to evaluate the performance of the proposed cpGA2DT in terms of the execution time and the number of three side effects. The weights for three side effects , , and are set at 0.5, 0.25, and 0.25, which can be adjusted by users. Details of the three databases used in the experiments are shown in Table 6.

##### 6.1. Execution Time

Execution times obtained the proposed cpGA2DT; greedy and simple GA-based algorithms are then compared at various sensitivity percentages of the sensitive itemsets for three databases. Results are shown in Figures 7, 8, and 9. The is initially set at 1.5%. According to predefined number of transactions to be deleted (the size of chromosome) in the original database, the is easily retrieved for deriving the prelarge itemsets, thus speeding up the execution time without computations of database rescan.

From Figures 7 to 9, it is obvious to see that the straightforward greedy approach has the best performance in execution time since it does not consider any side effects but directly delete the transactions for the purpose of hiding sensitive itemsets. The proposed cpGA2DT can greatly reduce the execution time compared to the simple GA-based algorithm since for cpGA2DT it is unnecessary to rescan the original database for evaluating fitness at each iteration. Experiments are then conducted to show the execution times for three algorithms at various minimum support thresholds. The results are then shown in Figures 10, 11, and 12.

Form Figures 10 and 12, it is obvious to see that the greedy approach has the best performance of execution time at various minimum support thresholds. The proposed cpGA2DT has the best performance in BMSWebview-1 database. The simple GA-based algorithm still has the worst performance in execution time since it requires to rescan the original database to evaluate the goodness of fitness at each iteration. The side effects of hiding failure, missing cost, and the artificial cost are also evaluated to show the performance of the proposed cpGA2DT. The descriptions are given as follows.

##### 6.2. Hiding Failure (HF)

The hiding failure is one of the side effects to evaluate whether the sensitive information has been successfully hidden before and after sanitization process, which can be calculated as where is the number of sensitive itemsets after sanitization process and the is the number of sensitive itemsets before sanitization process. The hiding failure obtained three algorithms at various sensitivity percentages of the sensitive itemsets for three databases with (= 1.5%). The results are then shown in Figures 13, 14, and 15.

From Figures 13 to 15, it is obvious to see that the greedy approach has the worst performance for hiding the sensitive itemsets in three databases. The proposed cpGA2DT generally has the best performance for hiding the sensitive itemsets in three databases except when the sensitive percentage is set at 10% of frequent itemsets in BMSWebview-2 database. Experiments are then conducted to show that the performance of hiding failure obtained three algorithms at various minimum support thresholds. The results are then shown in Figures 16, 17, and 18.

From Figures 16 to 18, it is easily found that the proposed cpGA2DT generally has the best performance of hiding failure at various minimum support thresholds for three databases and is better than the greedy and the simple GA-based algorithms in most cases at various minimum support thresholds for three databases.

##### 6.3. Missing Cost (MC)

The side effects of missing cost are also evaluated to show the performance of the proposed cpGA2DT, which is calculated as where is the number of frequent itemsets before data sanitization and is the number of frequent itemsets after data sanitization. Note that even sensitive itemsets are the frequent itemsets but not considered here to calculate the missing cost. The missing cost obtained three algorithms which are then compared at various sensitivity percentages of the sensitive itemsets for three databases with (= 1.5%). The missing cost that obtained three algorithms has, however, zero for the mushroom database since the mushroom database is too small for data sanitization. All sensitive itemsets can thus be successfully hidden without any missing cost in mushroom database. The results for the other two databases are then shown in Figures 19 to 20.

In the experiments of the proposed cpGA2DT, the weight of hiding failure is set at 0.5, which is higher than the missing cost and artificial cost. From Figure 19, the proposed cpGA2DT has generated some missing costs at 15% and 20% sensitive percentages of frequent itemsets. The proposed cpGA2DT has not any missing cost in BMSWebview-2 database. Experiments are then conducted to show that the performance of missing cost obtained three algorithms at various minimum support thresholds for three databases. Again, the missing cost is zero for the obtained three algorithms for mushroom database. The results for the other two databases are then shown in Figures 21 to 22.

From Figure 21, the proposed cpGA2DT algorithm has no missing cost for the BMSWebview-1 database. The greedy approach slightly outperforms better than the proposed cpGA2DT in the BMSWebview-2 but the proposed cpGA2DT still achieves good performance at the 1.5% and 1.6% minimum support thresholds with zero missing cost. In the experimental process, we have also found that the greedy approach is executed to delete transactions from top transactions to down ones, and the deleted transactions of the greedy approach in BMSWebview-2 have fewer numbers of items within it. Thus, the missing cost of the greedy approach is a little bit better than the proposed algorithm at 1.65% minimum support threshold.

##### 6.4. Artificial Cost (AC)

The side effects of artificial cost are also evaluated to show the performance of the proposed cpGA2DT, which is calculated as In three databases that obtained three algorithms in various sensitivity percentages of the frequent itemsets and various minimum support thresholds, there are not any side effects of artificial cost. For the greedy approach in the experiments, the deleted transactions have short length with lower support items; thus the artificial cost is not shown. For the proposed cpGA2DT, instead of the above reason of the greedy approach, the artificial cost is also considered as a factor in the evaluation process, thus avoiding the side effects of artificial cost.

#### 7. Conclusion

In this paper, a compact GA-based cpGA2DT algorithm is thus proposed to hide the sensitive itemsets through transaction deletion. A flexible fitness function with three adjustable weights is also designed to consider the general side effects of hiding failure, missing cost, and the artificial cost to determine the goodness of the chromosomes. The prelarge concept is adopted in the proposed algorithm to reduce the computations of database rescan. The size of the populations is also reduced by the compact GA approach, thus reducing the memory lack problems of traditional GAs. Experiments are conducted to show that the proposed cpGA2DT algorithm outperforms better than the greedy and simple GA-based algorithms considering all criteria of side effects but the execution time.

#### Notations

Original database to be sanitized | |

: | Number of transactions in |

: | Projected database from in which each transaction in contains any sensitive itemsets in HS |

: | Sanitized database after the designed algorithm |

HS: | A set of sensitive itemsets to be hidden, |

: | Number of transactions to be deleted for hiding sensitive itemsets |

: | Upper support threshold |

: | Lower support threshold, |

: | A set of large itemsets in which the count of each itemset is larger than or equal to |

PL: | A set of prelarge itemsets in which the count of each itemset lies between and |

: | Probability vector of transactions in |

: | Two competition chromosomes. |

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

This research was partially supported by the Shenzhen Peacock Project, China, under Grant KQC201109020055A, by the Natural Scientific Research Innovation Foundation, Harbin Institute of Technology, under Grant HIT.NSRIF.2014100, and by the Shenzhen Strategic Emerging Industries Program under Grant ZDSY20120613125016389.

#### References

- R. Agrawal and R. Srikant, “Fast algorithms for mining association rules in large databases,” in
*Proceedings of the International Conference on Very Large Data Bases*, pp. 487–499, 1994. - T. F. Gharib, H. Nassar, M. Taha, and A. Abraham, “An efficient algorithm for incremental mining of temporal association rules,”
*Data and Knowledge Engineering*, vol. 69, no. 8, pp. 800–815, 2010. View at Publisher · View at Google Scholar · View at Scopus - J. Han, J. Pei, Y. Yin, and R. Mao, “Mining frequent patterns without candidate generation: a frequent-pattern tree approach,”
*Data Mining and Knowledge Discovery*, vol. 8, no. 1, pp. 53–87, 2004. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - T. Hong, C. Lin, and Y. Wu, “Incrementally fast updated frequent pattern trees,”
*Expert Systems with Applications*, vol. 34, no. 4, pp. 2424–2435, 2008. View at Publisher · View at Google Scholar · View at Scopus - C.-W. Lin, T.-P. Hong, and W.-H. Lu, “The Pre-FUFP algorithm for incremental mining,”
*Expert Systems with Applications*, vol. 36, no. 5, pp. 9498–9505, 2009. View at Publisher · View at Google Scholar · View at Scopus - R. Agrawal and R. Srikant, “Mining sequential patterns,” in
*Proceedings of the IEEE 11th International Conference on Data Engineering*, pp. 3–14, San Jose, Calif, USA, March 1995. View at Scopus - C. Kim, J. Lim, R. T. Ng, and K. Shim, “SQUIRE: sequential pattern mining with quantities,”
*Journal of Systems and Software*, vol. 80, no. 10, pp. 1726–1745, 2007. View at Publisher · View at Google Scholar · View at Scopus - J. Pei, J. Han, B. Mortazavi-Asl et al., “Mining sequential patterns by pattern-growth: The prefixspan approach,”
*IEEE Transactions on Knowledge and Data Engineering*, vol. 16, no. 11, pp. 1424–1440, 2004. View at Publisher · View at Google Scholar · View at Scopus - R. Srikant and R. Agrawal, “Mining sequential patterns: generalizations and performance improvements,” in
*Proceedings of the International Conference on Extending Database Technology: Advances in Database Technology*, pp. 3–17, 1996. - S. B. Kotsiantis, “Supervised machine learning: a review of classification techniques,” in
*Proceedings of the Conference on Emerging Artificial Intelligence Applications in Computer Engineering: Real Word AI Systems with Applications in eHealth, HCI, Information Retrieval and Pervasive Technologies*, pp. 3–24, 2007. - J. R. Quinlan, “Induction of decision trees,”
*Machine Learning*, vol. 1, no. 1, pp. 81–106, 1986. View at Publisher · View at Google Scholar · View at Scopus - J. R. Quinlan,
*C4.5: Programs for Machine Learning*, Morgan Kaufmann, 1993. - P. Berkhin, “A survey of clustering data mining techniques,” in
*Grouping Multidimensional Data*, pp. 25–71, 2006. View at Google Scholar - R. A. Jarvis and E. A. Patrick, “Clustering using a similarity measure based on shared near neighbors,”
*IEEE Transactions on Computers*, vol. 22, no. 11, pp. 1025–1034, 1973. View at Publisher · View at Google Scholar · View at Scopus - R. Chan, Q. Yang, and Y.-D. Shen, “Mining high utility itemsets,” in
*Proceedings of the 3rd IEEE International Conference on Data Mining (ICDM '03)*, pp. 19–26, November 2003. View at Publisher · View at Google Scholar · View at Scopus - C. W. Lin, G. C. Lan, and T. P. Hong, “An incremental mining algorithm for high utility itemsets,”
*Expert Systems with Applications*, vol. 39, no. 8, pp. 7173–7180, 2012. View at Publisher · View at Google Scholar · View at Scopus - Y. Liu, W. K. Liao, and A. Choudhary, “A two-phase algorithm for fast discovery of high utility itemsets,” in
*Advances in Knowledge Discovery and Data Mining*, pp. 689–695, 2005. View at Google Scholar - U. Yuna, H. Ryanga, and K. H. Ryub, “High utility itemset mining with techniques for reducing overestimated utilities and pruning candidates,”
*Expert Systems with Applications*, vol. 41, pp. 3861–3878, 2014. View at Google Scholar - R. Agrawal and R. Srikant, “Privacy-preserving data mining,”
*SIGMOD Record*, vol. 29, no. 2, pp. 439–450, 2000. View at Google Scholar · View at Scopus - Y. Lindell and B. Pinkas, “Privacy preserving data mining,” in
*Advances in Cryptology—CRYPTO 2000, 20th Annual International Cryptology Conference, Santa Barbara, California, USA, August 20–24, 2000*, vol. 1880 of*Lecture Notes in Computer Science*, pp. 36–54, 2000. View at Publisher · View at Google Scholar - S. M. Oliveira, O. Zaïane, and Y. Saygin, “Secure association rule sharing,”
*Advances in Knowledge Discovery and Data Mining*, vol. 3056, pp. 74–85, 2004. View at Google Scholar - V. S. Verykios, E. Bertino, I. N. Fovino, L. P. Provenza, Y. Saygin, and Y. Theodoridis, “State-of-the-art in privacy preserving data mining,”
*SIGMOD Record*, vol. 33, no. 1, pp. 50–57, 2004. View at Publisher · View at Google Scholar · View at Scopus - A. Amiri, “Dare to share: protecting sensitive knowledge with data sanitization,”
*Decision Support Systems*, vol. 43, no. 1, pp. 181–191, 2007. View at Publisher · View at Google Scholar · View at Scopus - M. Atallah, A. Elmagarmid, M. Ibrahim, E. Bertino, and V. Verykios, “Disclosure limitation of sensitive rules,” in
*Proceedings of the Workshop on Knowledge and Data Engineering Exchange (KDEX '99)*, pp. 45–52, Chicago, Ill, USA, November 1999. View at Publisher · View at Google Scholar - A. Evfimievski, R. Srikant, R. Agrawal, and J. Gehrke, “Privacy preserving mining of association rules,” in
*Proceedings of the 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '02)*, pp. 217–228, July 2002. View at Scopus - Y. Wu, C. Chiang, and A. L. P. Chen, “Hiding sensitive association rules with limited side effects,”
*IEEE Transactions on Knowledge and Data Engineering*, vol. 19, no. 1, pp. 29–42, 2007. View at Publisher · View at Google Scholar · View at Scopus - C. C. Aggarwal, J. Pei, and B. Zhang, “On privacy preservation against adversarial data mining,” in
*Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining*, pp. 510–516, August 2006. View at Scopus - J. H. Holland,
*Adaptation in Natural and Artificial Systems*, MIT Press, 1992. View at MathSciNet - G. R. Harik, F. G. Lobo, and D. E. Goldberg, “The compact genetic algorithm,”
*IEEE Transactions on Evolutionary Computation*, vol. 3, no. 4, pp. 287–297, 1999. View at Publisher · View at Google Scholar · View at Scopus - T. P. Hong and C. Y. Wang, “Maintenance of association rules using pre-large itemsets,” in
*Intelligent Databases: Technologies and Applications*, pp. 44–60, 2007. View at Google Scholar - T. P. Hong, C. Y. Wang, and Y. H. Tao, “A new incremental data mining algorithm using pre-large itemsets,”
*Intelligent Data Analysis*, vol. 5, pp. 111–129, 2001. View at Google Scholar - D. E. Goldberg,
*Genetic Algorithms in Search, Optimization and Machine Learning*, Addison-Wesley Longman, 1989. - X. Wang, Q. He, D. Chen, and D. Yeung, “A genetic algorithm for solving the inverse problem of support vector machines,”
*Neurocomputing*, vol. 68, no. 1–4, pp. 225–238, 2005. View at Publisher · View at Google Scholar · View at Scopus - M. Chen, J. Han, and P. S. Yu, “Data mining: an overview from a database perspective,”
*IEEE Transactions on Knowledge and Data Engineering*, vol. 8, no. 6, pp. 866–883, 1996. View at Publisher · View at Google Scholar · View at Scopus - M. Mohamed and M. Darwieesh, “Efficient mining frequent itemsets algorithms,”
*International Journal of Machine Learning and Cybernetics*, 2013. View at Publisher · View at Google Scholar - B. Nath, D. K. Bhattacharyya, and A. Ghosh, “Incremental association rule mining: a survey,”
*Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery*, vol. 3, no. 3, pp. 157–169, 2013. View at Publisher · View at Google Scholar · View at Scopus - B. Vo, T. Le, F. Coenen, and T.-P. Hong, “Mining frequent itemsets using the n-list and subsume concepts,”
*International Journal of Machine Learning and Cybernetics*, 2014. View at Publisher · View at Google Scholar - E. Bertino, D. Lin, and W. Jiang, “A survey of quantification of privacy preserving data mining algorithms,” in
*Privacy-Preserving Data Mining*, C. Aggarwal and P. Yu, Eds., vol. 34, pp. 183–205, Springer, New York, NY, USA, 2008. View at Google Scholar - S. R. M. Oliveira and O. R. Zaïane, “Privacy preserving frequent itemset mining,” in
*Proceedings of the IEEE International Conference on Privacy, Security and Data Mining*, pp. 43–54, 2002. - E. Dasseni, V. S. Verykios, A. K. Elmagarmid, and E. Bertino, “Hiding association rules by using confidence and support,” in
*Proceedings of the International Workshop on Information Hiding*, pp. 369–383, 2001. - T. P. Hong, C. W. Lin, K. T. Yang, and S. L. Wang, “Using TF-IDF to hide sensitive itemsets,”
*Applied Intelligence*, vol. 38, no. 4, pp. 502–510, 2013. View at Publisher · View at Google Scholar - S. Han and W. Ng, “Privacy-preserving genetic algorithms for rule discovery,” in
*Data Warehousing and Knowledge Discovery*, I. Song, J. Eder, and T. Nguyen, Eds., vol. 4654, pp. 407–417, Springer, Berlin, Germany, 2007. View at Google Scholar - M. N. Dehkordi, K. Badie, and A. K. Zadeh, “A novel method for privacy preserving in association rule mining based on genetic algorithms,”
*Journal of Software*, vol. 4, no. 6, pp. 555–562, 2009. View at Google Scholar · View at Scopus - D. W. Cheung, J. Han, V. T. Ng, and C. Y. Wong, “Maintenance of discovered association rules in large databases: an incremental updating technique,” in
*Proceedings of the IEEE 12th International Conference on Data Engineering*, pp. 106–114, March 1996. View at Scopus - D. W. L. Cheung, S. D. Lee, and B. Kao, “A general incremental technique for maintaining discovered association rules,” in
*Proceedings of the International Conference on Database Systems for Advanced Applications*, pp. 185–194, 1997. - T. Hong, I. Yang, C. Lin, and S. Wang, “Evolutionary privacy-preserving data mining,” in
*Proceedings of the World Automation Congress (WAC "10)*, pp. 1–7, September 2010. View at Scopus - Frequent itemset mining dataset repository, 2012, http://fimi.ua.ac.be/data/.
- Z. Zheng, R. Kohavi, and L. Mason, “Real world performance of association rule algorithms,” in
*Proceedings of the 7th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '01)*, pp. 401–406, San Francisco, Calif, USA, August 2001. View at Scopus