Representing Conditional Independence Using Decision Trees

Jiang Su, Harry Zhang

While the representation of decision trees is fully expressive theoretically, it has been observed that traditional decision trees has the replication problem. This problem makes decision trees to be large and learnable only when sufficient training data are available. In this paper, we present a new representation model, conditional independence trees (CITrees), to tackle the replication problem from probability perspective. We propose a novel algorithm for learning CITrees. Our experiments show that CITrees outperform naive Bayes, C4.5, TAN, and AODE significantly in classification accuracy.

Subjects: 15.6 Decision Trees; 12. Machine Learning and Discovery

Submitted: May 3, 2005

This page is copyrighted by AAAI. All rights reserved. Your use of this site constitutes acceptance of all of AAAI's terms and conditions and privacy policy.