skip to main content
10.1145/1090193.1090196acmotherconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Learning Bayesian networks of rules with SAYU

Published:21 August 2005Publication History

ABSTRACT

Inductive Logic Programming (ILP) is a popular approach for learning in a relational environment. Given a set of positive and negative examples, an ILP system finds a logical description of the underlying data model that differentiates between the positive and negative examples. The key question becomes how to combine a set of rules to obtain a useful classifier. Previous work has shown that an effective approach is to treat each learned rule as an attribute in a propositional learner, and to use the classifier to determine the final label of an example [3]. This methodology defines a two step process. In the first step, an ILP algorithm learns a set of rules. In the second step, a classifier combines the learned rules. One weakness of this approach is that the rules learned in the first step are being evaluated by a different metric than how they are ultimately scored in the second step. ILP traditionally scores clauses through a coverage score or compression metric. Thus we have no guarantee that the rule learning process will select the rules that best contribute to the final classifier.We propose an alternative approach, based on the idea of constructing the classifier as we learn the rules [2, 4]. In our approach, rules are scored by how much they improve the classifier, providing a tight coupling between rule generation and rule usage. We call this novel methodology Score As You Use (SAYU) [2].In order to implement SAYU we defined an interface that allows an ILP algorithm to control a propositional learner. Second, we developed a greedy algorithm that uses the interface to decide whether to retain a candidate clause. We implemented this interface using Aleph to learn ILP rules, and Bayesian networks as the combining mechanism. We used two different Bayes net structure learning algorithms, Naïve Bayes and Tree Augmented Naïve Bayes (TAN) as propositional learners. We score the network by computing area under the precision recall curve for levels of recall greater than 0.2. Aleph proposes a candidate clause, which is introduced as a new feature in the training set. A new network topology is learned using the new training set, and then the new network is evaluated on a tuning set. If the score of the new network exceeds the previous score we retain the new rule in the training set. Otherwise the rule is discarded. The figure compares performance on the Breast Cancer dataset [1]. These results show that, given the same amount of CPU time, SAYU can clearly outperform the original two step approach. Furthermore, SAYU learns smaller theories. These results were obtained even though SAYU considers far fewer rules than standard ILP.

References

  1. J. Davis, E. Burnside, I. Dutra, D. Page, R. Ramakrishnan, V. Santos Costa, and J. Shavlik. View learning for statistical relational learning: With an application to mammography. In Proceedings of the 19th IJCAI, Edinburgh, Scotland, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. J. Davis, E. Burnside, I. C. Dutra, D. Page, and V. Santos Costa. An integrated approach to learning bayesian networks of rules. In To appear in ECML05, Porto, Portugal, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. J. Davis, V. Santos Costa, I. M. Ong, D. Page, and I. C. Dutra. Using Bayesian Classifiers to Combine Rules. In 3rd Workshop on Multi-Relational Data Mining, Seattle, USA, August 2004.Google ScholarGoogle Scholar
  4. N. Landwehr, K. Kersting, and L. D. Raedt. nFOIL: Integrating Naive Bayes and FOIL. In National Conference on Artificial Intelligene (AAAI), 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    MRDM '05: Proceedings of the 4th international workshop on Multi-relational mining
    August 2005
    73 pages
    ISBN:1595932127
    DOI:10.1145/1090193

    Copyright © 2005 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 21 August 2005

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • Article

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader