Abstract
The trees constructed by decision tree induction programs are often unnecessarily large and containing substantial degrees of duplication because such programs typically build a separate subtree for each value of a categorical attribute. Attribute value grouping procedures attempt to avoid this problem by partitioning attribute values into groups, each of which gives rise to only one subtree. In this paper we raise and attempt to answer a number of questions about the performance of such procedures. We review the limited amount of research that has been done in this area and propose a number of novel attribute value grouping procedures. We then present the results of a systematic comparative study in which eight alternative attribute grouping procedures were evaluated using both artificial and real data sets. We found that attribute value grouping can produce substantial reductions in tree size and that the best methods produce average reductions approaching 50% found that there was no effect on the classification accuracy of the trees produced but the time required to produce them was reduced. The most surprising finding was that global methods, which group attribute values once prior to tree construction were superior to local methods, which repartition values throughout tree construction: they produce substantially smaller trees in less time that are marginally more accurate classifiers.
Get full access to this article
View all access options for this article.
