2022
DOI: 10.1007/978-3-030-67024-5_14
|View full text |Cite
|
Sign up to set email alerts
|

Automating Data Science

Abstract: It has been observed that, in data science, a great part of the effort usually goes into various preparatory steps that precede model-building. The aim of this chapter is to focus on some of these steps. A comprehensive description of a given task to be resolved is usually supplied by the domain expert. Techniques exist that can process natural language description to obtain task descriptors (e.g., keywords), determine the task type, the domain, and the goals. This in turn can be used to search for the require… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 7 publications
(15 citation statements)
references
References 31 publications
(18 reference statements)
0
6
0
Order By: Relevance
“…For instance, does it make sense to smooth learning curves for model selection or curve extrapolation? Can meta-features [2], like the number of instances, be used to reliably predict (i) whether curves intersect, (ii) if they are monotone or convex, (iii) or which curve model will be accurate? Which nonparametric extrapolation techniques work best and in what case?…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…For instance, does it make sense to smooth learning curves for model selection or curve extrapolation? Can meta-features [2], like the number of instances, be used to reliably predict (i) whether curves intersect, (ii) if they are monotone or convex, (iii) or which curve model will be accurate? Which nonparametric extrapolation techniques work best and in what case?…”
Section: Discussionmentioning
confidence: 99%
“…For example, they can be extrapolated to determine the value of gathering more data or can be used to speed up training by selecting a smaller dataset size that still reaches sufficient accuracy. In addition, learning curves can provide useful information for model selection [2,26]. Particularly important questions concern the performance in the limit and the training set size at which the learning curves of two algorithms cross, as this can tell us when one learning algorithm should be preferred over the other.…”
Section: Introductionmentioning
confidence: 99%
“…A good alternative to select the best ML algorithm for a new dataset is to use previous knowledge regarding the performance of a set of algorithms in previous learning experiences. This is the idea behind a particular approach for metalearning, defined in [ 36 ] as learning to learn. According to the authors, metelarning is a research area that investigates how to recommend the most suitable algorithm, or set of algorithms, for a new task.…”
Section: Metalearningmentioning
confidence: 99%
“…To enable knowledge sharing across data sets, the scientific community has developed methods commonly referred to as meta-learning. 19 Whereas traditional machine learning models typically require an abundance of labeled data, meta-learning attempts to address this issue by asking how to learn to learn tasks? For this, meta-learning borrows intuition from how humans learn and solve problems.…”
Section: ■ Introductionmentioning
confidence: 99%
“…Instead of learning each task independently and anew, humans approach each challenge with prior knowledge. 19,20 With the success of transfer learning techniques in natural language processing or image analysis, its potential use in QSAR modeling has been recognized. 21,22 We believe the use of these techniques could be beneficial in utilizing and predicting the many low-resource tasks inherent to aquatic toxicity.…”
Section: ■ Introductionmentioning
confidence: 99%