IGTree: Using trees for compression and classification in lazy learning algorithms

W.M.P. Daelemans, A.P.J. Bosch, van den, A.J.M.M. Weijters

Research output: Contribution to journalArticleAcademicpeer-review

117 Citations (Scopus)
1 Downloads (Pure)


We describe the IGTree learning algorithm, which compresses an instance base into a tree structure. The concept of information gain is used as a heuristic function for performing this compression. IGTree produces trees that, compared to other lazy learning approaches, reduce storage requirements and the time required to compute classifications. Furthermore, we obtained similar or better generalization accuracy with IGTree when trained on two complex linguistic tasks, viz. letter–phoneme transliteration and part-of-speech-tagging, when compared to alternative lazy learning and decision tree approaches (viz., IB1, information-gain-weighted IB1, and C4.5). A third experiment, with the task of word hyphenation, demonstrates that when the mutual differences in information gain of features is too small, IGTree as well as information-gain-weighted IB1 perform worse than IB1. These results indicate that IGTree is a useful algorithm for problems characterized by the availability of a large number of training instances described by symbolic features with sufficiently differing information gain values.
Original languageEnglish
Pages (from-to)407-423
JournalArtificial Intelligence Review
Issue number1-5
Publication statusPublished - 1997


Dive into the research topics of 'IGTree: Using trees for compression and classification in lazy learning algorithms'. Together they form a unique fingerprint.

Cite this