Proceedings of the 18th ACM Conference on Information and Knowledge Management 2009
DOI: 10.1145/1645953.1646301
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic gradient boosted distributed decision trees

Abstract: Stochastic Gradient Boosted Decision Trees (GBDT) is one of the most widely used learning algorithms in machine learning today. It is adaptable, easy to interpret, and produces highly accurate models. However, most implementations today are computationally expensive and require all training data to be in main memory. As training data becomes ever larger, there is motivation for us to parallelize the GBDT algorithm. Parallelizing decision tree training is intuitive and various approaches have been explored in e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
125
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 265 publications
(131 citation statements)
references
References 10 publications
0
125
0
Order By: Relevance
“…We test the performance of the model over a dataset containing eight million question instances, obtained after filtering out questions whose askers' demographics information is missing. We train our machine learning model using gradient boosted decision trees [10,27] and test on our data via 10-fold cross-validation. 15 The 10 most important features, as provided by the learning tool, are shown in Table 9.…”
Section: Attitude Predictionmentioning
confidence: 99%
“…We test the performance of the model over a dataset containing eight million question instances, obtained after filtering out questions whose askers' demographics information is missing. We train our machine learning model using gradient boosted decision trees [10,27] and test on our data via 10-fold cross-validation. 15 The 10 most important features, as provided by the learning tool, are shown in Table 9.…”
Section: Attitude Predictionmentioning
confidence: 99%
“…Overfitting normally does not happen on the training exam the notion of margins in the case of boosting. The margin ( ) ( , ) x y is based on the votes ( ) t h x along with t  denoting the weights for all hypotheses [29].…”
Section: Combining Hypothesis With Boostingmentioning
confidence: 99%
“…The effectiveness of tree-based ensembles for learning to rank has been widely demonstrated: an example is the family of gradient-boosted regression trees (GBRTs) [9,3,10,2]. In this context, our work uses LambdaMART [3], which is the combination of LambdaRank [11] and MART [12]-a class of boosting algorithms that performs gradient descent using regression trees.…”
Section: Lambdamartmentioning
confidence: 99%