ABSTRACT
Classification is one of the most useful techniques for extracting meaningful knowledge from databases. Classifiers, e.g. decision trees, are usually extracted from a table of records, each of which represents an example. However, quite often in real applications there is other knowledge, e.g. owned by experts of the field, that can be usefully used in conjunction with the one hidden inside the examples. As a concrete example of this kind of knowledge we consider causal dependencies among the attributes of the data records. In this paper we discuss how to use such a knowledge to improve the construction of classifiers. The causal dependencies are represented via Bayesian Causal Maps (BCMs), and our method is implemented as an adaptation of the well known C4.5 algorithm.
- L. Breiman, J. Friedman, R. Olshen, and C. Stone. Classification and Regression Trees. Kluwer Academic Publishers, 1984.Google Scholar
- C. Eden, F. Ackermann, and S. Copper. The analysis of cause maps. Journal of Management Studies, 29(3):309/323, 1992.Google ScholarCross Ref
- D. Heckerman. Bayesian networks for data mining. Data Mining and Knowlwdge Discovery, 1:79--119, 1997. Google ScholarDigital Library
- B. Kemmerer, S. Mishra, and P. P. Shenoy. Bayesian causal maps as decision aids in venture capital decision making: Methods and applications. In In proceedings of the Accademy of Management Conference, 2002.Google ScholarCross Ref
- T. Mitchell. Machine Learning. McGraw-Hill Companies, 1997. Google ScholarDigital Library
- J. R. Quinlan. Induction of decision trees. Machine Learning, 1:81--106, 1986. Google ScholarCross Ref
- J. R. Quinlan. Simplifying decision trees, 1986. AI Memo No 930. Google ScholarDigital Library
- J. R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers, 1993. Google ScholarDigital Library
- Z. Zheng. Construction New Attributes for Tree Learning. PhD thesis, Basser Departement of Computer Science, University of Sydney, NSW 2006, Australia, March 1996.Google Scholar
Recommendations
Chinese text classification by the Naïve Bayes Classifier and the associative classifier with multiple confidence threshold values
Each type of classifier has its own advantages as well as certain shortcomings. In this paper, we take the advantages of the associative classifier and the Naive Bayes Classifier to make up the shortcomings of each other, thus improving the accuracy of ...
Specializing for predicting obesity and its co-morbidities
We present specializing, a method for combining classifiers for multi-class classification. Specializing trains one specialist classifier per class and utilizes each specialist to distinguish that class from all others in a one-versus-all manner. It ...
AdaBoost classifiers for pecan defect classification
Highlights The performance of AdaBoost algorithms were compared with support vector machine and Bayesian classifiers for pecan defect classification. AdaBoost classifiers took least time and gave best classification accuracy. AdaBoost classifiers ...
Comments