2020
DOI: 10.26686/wgtn.12493820.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Genetic Programming for Evolving a Front of Interpretable Models for Data Visualization

Abstract: Data visualization is a key tool in data mining for understanding big datasets. Many visualization methods have been proposed, including the well-regarded state-of-the-art method t-distributed stochastic neighbor embedding. However, the most powerful visualization methods have a significant limitation: the manner in which they create their visualization from the original features of the dataset is completely opaque. Many domains require an understanding of the data in terms of the original features; there is h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 9 publications
(15 citation statements)
references
References 26 publications
(37 reference statements)
0
15
0
Order By: Relevance
“…In fact, we agree that is simplistic. However, we believe that minimizing represents one of the first baselines to compare against (and it was the only one we found being used to specifically promote interpretability [22]), and that designing a competitive baseline is non-trivial. We will investigate this further in future work.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In fact, we agree that is simplistic. However, we believe that minimizing represents one of the first baselines to compare against (and it was the only one we found being used to specifically promote interpretability [22]), and that designing a competitive baseline is non-trivial. We will investigate this further in future work.…”
Section: Discussionmentioning
confidence: 99%
“…The authors of [43] study whether modern model-based GP can be useful when particularly compact symbolic regression solutions are sought, to allow interpretability. A very different take to enable or improve interpretability is taken in [22,41,45], where interpretability is sought by means of feature construction and dimensionality reduction. In [22] in particular, MOGP is used, with solution size as a simple PHI.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…For decision trees and decision rules, besides reducing the number of features, approaches exist to restrict model size, prune unnecessary parts [5,26], aggregate local models in a hierarchy [48], or promote a trade-off between accuracy and complexity by means of loss functions [30,52] or prior distributions [33,60,61]. Regarding GP (and close relatives like grammatical evolution), perhaps the most simple and popular strategy to favor interpretability is to restrain the number of model components [16,31,57], sometimes in elaborate ways or particular settings [6,32,40,49,56]. Another strategy consists of penalizing models according to a weighted sum of the components that they include, after having pre-determined a weighing scheme [23,36].…”
Section: Related Workmentioning
confidence: 99%
“…Recently there has been new application perspectives for Genetic Programming (GP) regarding the increasing need for interpretable results. GP was for example used to provide interpretable policies in reinforcement learning [14], to learn manifolds [25], to create visualizations [26] or to explain complex deep learning models [10]. For dimensionality reduction tasks, GP has also been used a lot as a feature construction method [35].…”
Section: Grammar-guided Genetic Programmingmentioning
confidence: 99%