2020
DOI: 10.48550/arxiv.2007.12652
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MurTree: Optimal Classification Trees via Dynamic Programming and Search

Abstract: Decision tree learning is a widely used approach in machine learning, favoured in applications that require concise and interpretable models. Heuristic methods are traditionally used to quickly produce models with reasonably high accuracy. A commonly criticised point, however, is that the resulting trees may not necessarily be the best representation of the data in terms of accuracy, size, and other considerations such as fairness. In recent years, this motivated the development of optimal classification tree … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 10 publications
(26 citation statements)
references
References 11 publications
0
26
0
Order By: Relevance
“…Benchmarks and experimental setting. We considered 75 binary classification datasets used in previous works (Verwer and Zhang 2019;Aglin, Nijssen, and Schaus 2020;Demirović et al 2020;Narodytska et al 2018;Hu, Rudin, and Seltzer 2019). The experiments were run one at a time on an Intel i7-3612QM@2.10 GHz with 8 GB RAM.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Benchmarks and experimental setting. We considered 75 binary classification datasets used in previous works (Verwer and Zhang 2019;Aglin, Nijssen, and Schaus 2020;Demirović et al 2020;Narodytska et al 2018;Hu, Rudin, and Seltzer 2019). The experiments were run one at a time on an Intel i7-3612QM@2.10 GHz with 8 GB RAM.…”
Section: Methodsmentioning
confidence: 99%
“…We describe previous work on dynamic programming and search for constructing decision trees with minimal misclassifications in more detail (Aglin, Nijssen, and Schaus 2020;Nijssen and Fromont 2007;Demirović et al 2020), as we generalise these methods.…”
Section: Dynamic Programming and Search For Treesmentioning
confidence: 99%
See 1 more Smart Citation
“…This is not the case of globally optimal trees, which are, on this respect, less interpretable (Bertsimas et al 2019a). Secondly, despite a vibrant corpus of works on formulating decision trees as mixed-integer optimization problems and solving them as such to provable optimality (Bertsimas and Dunn 2017, Günlük et al 2018, Zantedeschi et al 2020, Aglin et al 2020, Demirović et al 2020, Lin et al 2020, greedy procedures remain the gold-standard for balancing accuracy, interpretability, and scalability. Thirdly, recursive procedures can handle complex non-convex criteria like the Q-statistics.…”
Section: A Recursive Partitioning Proceduresmentioning
confidence: 99%
“…In [16] a column generation heuristic is described to build univariate binary classification trees for larger datasets. A dynamic programming and search algorithm is presented in [17] for constructing optimal univariate decision trees. In [18] the authors propose a flow-based MILO formulation for optimal univariate classification trees where they exploit the problem structure and max-flow/min-cut duality to derive a Benders' decomposition method for handling large datasets.…”
Section: Introductionmentioning
confidence: 99%