Table of Contents
Advances in Artificial Intelligence
Volume 2009, Article ID 134807, 13 pages
Research Article

A New Information Measure Based on Example-Dependent Misclassification Costs and Its Application in Decision Tree Learning

Faculty of Electrical Engineering and Computer Science, University of Technology Berlin, Sekr. FR 5-8, Franklinstraße 28/29, D-10587 Berlin, Germany

Received 29 December 2008; Revised 2 June 2009; Accepted 21 July 2009

Academic Editor: Rattikorn Hewett

Copyright © 2009 Fritz Wysotzki and Peter Geibel. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Linked References

  1. D. Michie, D. H. Spiegelhalter, and C. C. Taylor, Machine Learning, Neural and Statistical Classification, Series in Artificial Intelligence, Ellis Horwood, 1994.
  2. L. Saitta, Ed., Machine Learning—A Technological Roadmap, University of Amsterdam, Amsterdam, The Netherlands, 2000.
  3. P. D. Turney, “Types of cost in inductive concept learning,” in Proceedings of the Workshop on Cost-Sensitive Learning at the 17th International Conference on Machine Learning, Stanford University, Stanford, Calif, USA, 2000.
  4. Ch. X. Ling and V. S. Sheng, “Cost–sensitive learning and the class imbalance problem,” in Encyclopedia of Machine Learning, C. Sammut, Ed., Springer, 2008. View at Google Scholar
  5. Ch. X. Ling, Q. Yang, J. Wang, and S. Zhang, “Decision trees with minimal costs,” in Proceedings of International Conference on Machine Learning (ICML '04), 2004.
  6. C. Drummond and R. Holte, “Exploiting the cost (in)sensitivity of decision tree splitting criteria,” in Proceedings of the 17th International Conference on Machine Learning (ICML '00), pp. 239–246, Stanford, Calif, USA, June 2000.
  7. D. D. Margineantu and T. G. Dietterich, “Bootstrap methods for the cost-sensitive evaluation of classifiers,” in Proceedings of the 17th International Conference on Machine Learning, pp. 583–590, Morgan Kaufmann, San Francisco, Calif, USA, 2000.
  8. C. Elkan, “The foundations of cost-sensitive learning,” in Proceedings of the 17th International Conference on Artificial Intelligence (IJCAI '01), B. Nebel, Ed., pp. 973–978, Morgan Kaufmann, San Francisco, Calif, USA, August 2001.
  9. M. Kukar and I. Kononenko, “Cost-sensitive learning with neural networks,” in Proceedings of the 13th European Conference on Artificial Intelligence (ECAI '98), H. Prade, Ed., pp. 445–449, John Wiley & Sons, Chichester, UK, 1998.
  10. Y. Lin, Y. Lee, and G. Wahba, “Support vector machines for classification in nonstandard situations,” Machine Learning, vol. 46, no. 1–3, pp. 191–202, 2002. View at Google Scholar
  11. A. Lenarcik and Z. Piasta, “Rough classifiers sensitive to costs varying from object to object,” in Proceedings of the 1st International Conference on Rough Sets and Current Trends in Computing (RSCTC '98), L. Polkowski and A. Skowron, Eds., vol. 1424 of Lecture Notes in Computer Science, pp. 222–230, Springer, Berlin, Germany, June 1998.
  12. P. Geibel and F. Wysotzki, “Perceptron based learning with example dependent and noisy costs,” in Proceedings of the 20th International Conference on Machine Learning (ICML '03), Washington, DC, USA, 2003.
  13. P. Geibel and F. Wysotzki, “Learning perceptrons and piecewise linear classifiers sensitive to example dependent costs,” Applied Intelligence, vol. 21, no. 1, pp. 45–56, 2004. View at Publisher · View at Google Scholar
  14. P. Geibel, U. Brefeld, and F. Wysotzki, “Perceptron and SVM learning with generalized cost models,” Intelligent Data Analysis, vol. 8, no. 5, pp. 439–455, 2004. View at Google Scholar
  15. R. Klinkenberg and S. Rüping, “Concept drift and the importance of examples,” in Text Mining—Theoretical Aspects and Applications, J. Franke, G. Nakhaeizadeh, and I. Renz, Eds., Springer, 2003. View at Google Scholar
  16. P. Geibel, Risk-Sensitive Approaches for Reinforcement Learning, Shaker, 2006.
  17. L. Li, V. Bulitko, and R. Greiner, “Batch reinforcement learning with state importance,” in Proceedings of the European Conference on Machine Learning (ECML '04), pp. 566–568, Springer, 2004.
  18. A. Bendadi, O. Benn, P. Geibel, M. Hudik, T. Knebel, and F. Wysotzki, “Lernen von Entscheidungsbäumen bei objektabhängigen Fehlklassifikationskosten,” Technischer Bericht 2004-18, Technischen Universität Berlin, Berlin, Germany, 2005. View at Google Scholar
  19. F. Wysotzki and P. Geibel, “Computer-supported decision making with object dependent costs for misclassifications,” in Human Interaction with Machines, Proceedings of the 6th International Workshop Held at the Shanghai Jiao Tong University, G. Hommel and S. Huanye, Eds., pp. 129–139, Springer, March 2005.
  20. S. Unger and F. Wysotzki, Lernfähige Klassifizierungssysteme (Classifier Systems Which Are Able to Learn), Akademie-Verlag, Berlin, Germany, 1981.
  21. W. Müller and F. Wysotzki, “Automatic construction of decision trees for classification,” Annals of Operations Research, vol. 52, no. 4, pp. 231–247, 1994. View at Publisher · View at Google Scholar
  22. W. Müller and F. Wysotzki, “The decision-tree algorithm CAL5 based on a statistical approach to its splitting algorithm,” in Machine Learning and Statistics, G. Nakhaeizadeh and C. C. Taylor, Eds., pp. 45–65, John Wiley & Sons, New York, NY, USA, 1997. View at Google Scholar
  23. C. E. Shannon and W. Weaver, The Mathematical Theory of Communication, University of Illinois Press, Urbana, Ill, USA, 1949.
  24. A. Ethem, Machine Learning, MIT Press, 2004.
  25. J. Körner and F. Wysotzki, “Die Rolle der Übergeneralisierung in der Neurosenbildung,” Forum der Psychoanalyse, vol. 22, no. 4, pp. 321–341, 2006. View at Publisher · View at Google Scholar
  26. F. Wysotzki, W. Müller, and B. Schulmeister, “Automatic construction of decision trees and neural nets for classification using statistical considerations,” in Learning, Networks and Statistics, G. Della Riccia, H.-J. Lenz, and R. Kruse, Eds., CISM Courses and Lectures no. 382, pp. 121–134, Springer, New York, NY, USA, 1997. View at Google Scholar
  27. N. Tóth and B. Pataki, “On classification confidence and ranking using decision trees,” in Proceedings of the 11th International Conference on Intelligent Engineering Systems (INES '07), pp. 133–138, Budapest, Hungary, June-July 2007. View at Publisher · View at Google Scholar
  28. N. Tóth and B. Pataki, “Classification confidence weighted majority voting using decision tree classifiers,” International Journal of Intelligent Computing and Cybernetics, vol. 1, no. 2, pp. 169–192, 2008. View at Google Scholar
  29. J. R. Quinlan, C4.5: Programs for Machine Learning, Morgan Kaufmann, San Mateo, Calif, USA, 1993.
  30. L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classification and Regression Trees, Wadsworth and Brooks, Monterey, Calif, USA, 1984.
  31. B. Schulmeister and F. Wysotzki, “DIPOL-a hybrid piecewise linear classifier,” in Machine Learning and Statistics: The Interface, R. Nakeiazadeh and C. C. Taylor, Eds., pp. 133–151, John Wiley & Sons.
  32. T. Joachims, “Making large-scale SVM learning practical,” in Advances in Kernel Methods—Support Vector Learning, B. Schölkopf, C. Burges, and A. Smola, Eds., MIT-Press, 1999. View at Google Scholar
  33. P. K. Chan and S. J. Stolfo, “Toward scalable learning with non-uniform class and cost distributions: a case study in credit card fraud detection,” in Proceedings of the Knowledge Discovery and Data Mining (KDD '98), pp. 164–168, New York, NY, USA, 1998.
  34. B. Zadrozny and C. Elkan, “Learning and making decisions when costs and probabilities are both unknown,” in Proceedings of the 7th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '01), F. Provost and R. Srikant, Eds., pp. 204–214, ACM Press, New York, NY, USA, August 2001.