Generalized Entropy for Splitting on Numerical Attributes in Decision Trees

Mingyu Zhong, Michael Georgiopoulos, Georgios Anagnostopoulos, Mansooreh Mollaghasemi

Decision Trees are well known for their training efficiency and their interpretable knowledge representation. They apply a greedy search and a divide-and-conquer approach to learn patterns. The greedy search is based on the evaluation criterion on the candidate splits at each node. Although research has been performed on various such criteria, there is no significant improvement from the classical split approaches introduced in the early decision tree literature. This paper presents a new evaluation rule to determine candidate splits in decision tree classifiers. The experiments show that this new evaluation rule reduces the size of the resulting tree, while maintaining the tree's accuracy.

Subjects: 15.6 Decision Trees; 12. Machine Learning and Discovery

Submitted: Feb 7, 2006

This page is copyrighted by AAAI. All rights reserved. Your use of this site constitutes acceptance of all of AAAI's terms and conditions and privacy policy.