2023
DOI: 10.1162/tacl_a_00544
|View full text |Cite
|
Sign up to set email alerts
|

Transformers for Tabular Data Representation: A Survey of Models and Applications

Abstract: In the last few years, the natural language processing community has witnessed advances in neural representations of free texts with transformer-based language models (LMs). Given the importance of knowledge available in tabular data, recent research efforts extend LMs by developing neural representations for structured data. In this article, we present a survey that analyzes these efforts. We first abstract the different systems according to a traditional machine learning pipeline in terms of training data, i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 15 publications
(4 citation statements)
references
References 69 publications
0
3
0
Order By: Relevance
“…More generally, limited data availability constraints leverage more advanced deep learning architectures in rational nanozyme design, while there are many examples of their successful applications. 56,57 Another challenge is data preprocessing, which has only been possible so far with rigorous analysis of scienti c publications by domain experts and manual data extraction. With the appearance of LLMs now available on a commercial (e.g., GPT-4 34 ) but also on an opensource (e.g., Llama-2 58 ) basis, we are presented with an opportunity to automate or largely simplify this process.…”
Section: Prediction Of Multiple Catalytic Activitiesmentioning
confidence: 99%
“…More generally, limited data availability constraints leverage more advanced deep learning architectures in rational nanozyme design, while there are many examples of their successful applications. 56,57 Another challenge is data preprocessing, which has only been possible so far with rigorous analysis of scienti c publications by domain experts and manual data extraction. With the appearance of LLMs now available on a commercial (e.g., GPT-4 34 ) but also on an opensource (e.g., Llama-2 58 ) basis, we are presented with an opportunity to automate or largely simplify this process.…”
Section: Prediction Of Multiple Catalytic Activitiesmentioning
confidence: 99%
“…However, DLbased methods are primarily adapted to transfer learning because they are easier to pre-train than tree-based methods. Among others, self-supervised learning with Transformers is the most common pre-training approach of DL-based methods (see, e.g., Badaro et al (2023) for the survey).…”
Section: Related Workmentioning
confidence: 99%
“…Transformer [17] is an attention-based structure originally proposed as a sequence-tosequence model for machine translation tasks. In recent years, by virtue of its outstanding results in the field of Natural Language Processing (NLP) [18][19][20][21], it has attracted a wide range of attention from researchers in the field of computer vision [22], and more and more researchers are migrating its application to computer vision tasks such as target detection, video processing, image processing.…”
Section: Transformermentioning
confidence: 99%