2022
DOI: 10.48550/arxiv.2203.05556
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Embeddings for Numerical Features in Tabular Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(14 citation statements)
references
References 0 publications
0
14
0
Order By: Relevance
“…Note also that our observation could also explain the benefits of the ExU activation used in the Neural-GAM paper [Agarwal et al, 2021], and of the embeddings used in Gorishniy et al [2022]: the periodic embedding might help the model to learn the high-frequency part of the target function, and the target-aware binning might make the target function smoother.…”
Section: Resultsmentioning
confidence: 70%
See 1 more Smart Citation
“…Note also that our observation could also explain the benefits of the ExU activation used in the Neural-GAM paper [Agarwal et al, 2021], and of the embeddings used in Gorishniy et al [2022]: the periodic embedding might help the model to learn the high-frequency part of the target function, and the target-aware binning might make the target function smoother.…”
Section: Resultsmentioning
confidence: 70%
“…Our findings shed light on the results of Somepalli et al [2021] and Gorishniy et al [2022], which add an embedding layer, even for numerical features, before MLP or Transformer models. Indeed, this layer breaks rotation invariance.…”
Section: Finding 2: Uninformative Features Affect More Mlp-like Nnsmentioning
confidence: 70%
“…Although numerous models have been proposed based on using differentiable ensembles 45,46,47,48,49 , leveraging attention-based transformer neural networks 35,50,51,52,53,54 , as well as other approaches 55,56,57,58,59,60 , recent work on systematic evaluation of deep tabular models 35,44 shows that there is no universally best model capable of consistently outperforming GBDT. Transformer-based models have been shown to be the strongest competitor of GBDT 35,50,54,61,62 , especially when coupled with a powerful hyperparameter tuning toolkit 35,63 .…”
Section: Methodsmentioning
confidence: 99%
“…Tabular transformer model. We employ the recent transformer-based tabular deep learning method FT-Transformer proposed by Gorishniy et al 35 which has been shown to be the strongest neural network approach in the tabular data domain 35,61 . Additionally, we compare the performance of our model with the gradient boosted decision trees, and we use the popular CatBoost 36 and XGBoost 37 packages.…”
Section: Iib Transformer-based Tabular Deep Learning Modelmentioning
confidence: 99%
“…As reported in [10], an efficient transformation of categorical data for training DNNs is still a significant challenge. Furthermore, a work [11] shows that the embeddings (transformations) for numerical features can be also beneficial for DNNs.…”
Section: Introductionmentioning
confidence: 99%