License: Creative Commons Attribution 3.0 Unported license (CC BY 3.0)
When quoting this document, please refer to the following
DOI: 10.4230/LIPIcs.ITCS.2020.44
URN: urn:nbn:de:0030-drops-117295
URL: http://dagstuhl.sunsite.rwth-aachen.de/volltexte/2020/11729/
Blanc, Guy ;
Lange, Jane ;
Tan, Li-Yang
Top-Down Induction of Decision Trees: Rigorous Guarantees and Inherent Limitations
Abstract
Consider the following heuristic for building a decision tree for a function f : {0,1}^n → {± 1}. Place the most influential variable x_i of f at the root, and recurse on the subfunctions f_{x_i=0} and f_{x_i=1} on the left and right subtrees respectively; terminate once the tree is an ε-approximation of f. We analyze the quality of this heuristic, obtaining near-matching upper and lower bounds:
- Upper bound: For every f with decision tree size s and every ε ∈ (0,1/2), this heuristic builds a decision tree of size at most s^O(log(s/ε)log(1/ε)).
- Lower bound: For every ε ∈ (0,1/2) and s ≤ 2^Õ(√n), there is an f with decision tree size s such that this heuristic builds a decision tree of size s^Ω~(log s).
We also obtain upper and lower bounds for monotone functions: s^O(√{log s}/ε) and s^Ω(∜{log s}) respectively. The lower bound disproves conjectures of Fiat and Pechyony (2004) and Lee (2009).
Our upper bounds yield new algorithms for properly learning decision trees under the uniform distribution. We show that these algorithms - which are motivated by widely employed and empirically successful top-down decision tree learning heuristics such as ID3, C4.5, and CART - achieve provable guarantees that compare favorably with those of the current fastest algorithm (Ehrenfeucht and Haussler, 1989), and even have certain qualitative advantages. Our lower bounds shed new light on the limitations of these heuristics.
Finally, we revisit the classic work of Ehrenfeucht and Haussler. We extend it to give the first uniform-distribution proper learning algorithm that achieves polynomial sample and memory complexity, while matching its state-of-the-art quasipolynomial runtime.
BibTeX - Entry
@InProceedings{blanc_et_al:LIPIcs:2020:11729,
author = {Guy Blanc and Jane Lange and Li-Yang Tan},
title = {{Top-Down Induction of Decision Trees: Rigorous Guarantees and Inherent Limitations}},
booktitle = {11th Innovations in Theoretical Computer Science Conference (ITCS 2020)},
pages = {44:1--44:44},
series = {Leibniz International Proceedings in Informatics (LIPIcs)},
ISBN = {978-3-95977-134-4},
ISSN = {1868-8969},
year = {2020},
volume = {151},
editor = {Thomas Vidick},
publisher = {Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik},
address = {Dagstuhl, Germany},
URL = {https://drops.dagstuhl.de/opus/volltexte/2020/11729},
URN = {urn:nbn:de:0030-drops-117295},
doi = {10.4230/LIPIcs.ITCS.2020.44},
annote = {Keywords: Decision trees, Influence of variables, Analysis of boolean functions, Learning theory, Top-down decision tree heuristics}
}
Keywords: |
|
Decision trees, Influence of variables, Analysis of boolean functions, Learning theory, Top-down decision tree heuristics |
Collection: |
|
11th Innovations in Theoretical Computer Science Conference (ITCS 2020) |
Issue Date: |
|
2020 |
Date of publication: |
|
06.01.2020 |