Improving Decision Trees through the Lens of Parameterized Local Search
By: Juha Harviainen, Frank Sommer, Manuel Sorge
Potential Business Impact:
Makes computer learning faster by finding better rules.
Algorithms for learning decision trees often include heuristic local-search operations such as (1) adjusting the threshold of a cut or (2) also exchanging the feature of that cut. We study minimizing the number of classification errors by performing a fixed number of a single type of these operations. Although we discover that the corresponding problems are NP-complete in general, we provide a comprehensive parameterized-complexity analysis with the aim of determining those properties of the problems that explain the hardness and those that make the problems tractable. For instance, we show that the problems remain hard for a small number $d$ of features or small domain size $D$ but the combination of both yields fixed-parameter tractability. That is, the problems are solvable in $(D + 1)^{2d} \cdot |I|^{O(1)}$ time, where $|I|$ is the size of the input. We also provide a proof-of-concept implementation of this algorithm and report on empirical results.
Similar Papers
A Parameterized-Complexity Framework for Finding Local Optima
Computational Complexity
Finds better solutions by showing how to get there.
Optimal Decision Tree Pruning Revisited: Algorithms and Complexity
Machine Learning (CS)
Makes smart computer programs learn faster.
Complexity of Local Search for CSPs Parameterized by Constraint Difference
Data Structures and Algorithms
Finds good solutions by fixing a few bad ones.