“…Deep learning for tabular data As described by Borisov et al [2021] in their review of the field, there have been various attempts to make deep learning work on tabular data: data encoding techniques to make tabular data better suited for deep learning [Hancock andKhoshgoftaar, 2020, Yoon et al, 2020], "hybrid methods" to benefit from the flexibility of NNs while keeping the inductive biases of other algorithms like tree-based models [Lay et al, 2018, Popov et al, 2020, Abutbul et al, 2020, Hehn et al, 2019, Tanno et al, 2019, Chen, 2020, Kontschieder et al, 2015, Rodriguez et al, 2019, Popov et al, 2020, Lay et al, 2018 or Factorization Machines Guo et al [2017], tabularspecific transformers architectures Somepalli et al [2021], Kossen et al [2021], Arik and Pfister [2019], Huang et al [2020], and various regularization techniques to adapt classical architectures to tabular data [Lounici et al, 2021, Shavitt and Segal, 2018, Kadra et al, 2021a, Fiedler, 2021. In this paper, we focus on architectures directly inspired by classic deep learning models, in particular Transformers and Multi-Layer-Perceptrons (MLPs).…”