2021
DOI: 10.1109/access.2021.3070575
|View full text |Cite
|
Sign up to set email alerts
|

SDTR: Soft Decision Tree Regressor for Tabular Data

Abstract: Deep neural networks have been proved a success in multiple fields. However, researchers still favor traditional approaches to obtain more interpretable models, such as Bayesian methods and decision trees when processing heterogeneous tabular data. Such models are hard to differentiate, thus inconvenient to be integrated into end-to-end settings. On the other hand, traditional neural networks are differentiable but perform poorly on tabular data. We propose a hierarchical differentiable neural regression model… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 36 publications
(17 citation statements)
references
References 27 publications
0
15
0
Order By: Relevance
“…Since the input of the auxiliary models in the second and third stages of the TMSCNet was the numerical data rather than the image, the classical machine learning regression algorithm and fully connected network (FCN) were considered. There were many classical regression models in machine learning, such as Linear Regression ( Austin and Steyerberg, 2015 ), Support Vector Regressor ( Cortes and Vapnik, 1995 ), K-Nearest Neighbors Regressor ( Song et al, 2017 ), Decision Tree Regressor ( Luo et al, 2021 ), Random Forest Regressor ( Ding and Bar-Joseph, 2017 ), AdaBoost Regressor ( Chen et al, 2019 ), and Bagging Regressor ( Dal Molin Ribeiro and Coelho, 2020 ). For selecting more suitable self-correcting models, we first conducted comparative experiments on the classical machine learning regression algorithm to select the best performing method, and then implement comparative experiments on the optimized classical machine learning algorithm and fully connected network to determine the method of the auxiliary model in the second and third stages of the TMSCNet.…”
Section: Discussionmentioning
confidence: 99%
“…Since the input of the auxiliary models in the second and third stages of the TMSCNet was the numerical data rather than the image, the classical machine learning regression algorithm and fully connected network (FCN) were considered. There were many classical regression models in machine learning, such as Linear Regression ( Austin and Steyerberg, 2015 ), Support Vector Regressor ( Cortes and Vapnik, 1995 ), K-Nearest Neighbors Regressor ( Song et al, 2017 ), Decision Tree Regressor ( Luo et al, 2021 ), Random Forest Regressor ( Ding and Bar-Joseph, 2017 ), AdaBoost Regressor ( Chen et al, 2019 ), and Bagging Regressor ( Dal Molin Ribeiro and Coelho, 2020 ). For selecting more suitable self-correcting models, we first conducted comparative experiments on the classical machine learning regression algorithm to select the best performing method, and then implement comparative experiments on the optimized classical machine learning algorithm and fully connected network to determine the method of the auxiliary model in the second and third stages of the TMSCNet.…”
Section: Discussionmentioning
confidence: 99%
“…This paper by [21] (Ali et al 2021) uses Decision tree regressor with neural networks for performance optimization and has an accuracy of 90%. (Luo et al 2021)) [22] proposed a Soft Decision Tree Regressor (SDTR),a differentiable hierarchical neural regression model SDTR is a differentiable neural network that mimics a binary decision tree and is suitable for ensemble techniques such as bagging and boosting, as well as archiving the results of 95.34%.…”
Section: Discussionmentioning
confidence: 99%
“…CatBoost's implementation of gradient boosted decision trees has a proven track record in the context of several different and similarly sized problems (Bentéjac et al, 2021; Luo et al, 2021). Since the proposed solution is novel the explainability of the underlying decision trees can help in investigating performance and identifying what variables play the most significant role in location predictions.…”
Section: Methodsmentioning
confidence: 99%
“…This allows for more complex models to potentially learn policy and construction year patterns. Recent work by Luo et al (2021) addresses the favorability of traditional machine learning approaches over complex deep learning methods in the context of supervised regression with heterogeneous tabular input data. Furthermore, recent work conducting comprehensive comparisons of state‐of‐the‐art interpretable and uninterpretable models on several datasets reveals that CatBoost's implementation of an ensemble algorithm for gradient boosting on decision trees outperforms all other models on average (Bentéjac et al, 2021).…”
Section: Related Workmentioning
confidence: 99%