2021
DOI: 10.48550/arxiv.2104.08291
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Better Latent Spaces for Better Autoencoders

Barry M. Dillon,
Tilman Plehn,
Christof Sauer
et al.

Abstract: Autoencoders as tools behind anomaly searches at the LHC have the structural problem that they only work in one direction, extracting jets with higher complexity but not the other way around. To address this, we derive classifiers from the latent space of (variational) autoencoders, specifically in Gaussian mixture and Dirichlet latent spaces. In particular, the Dirichlet setup solves the problem and improves both the performance and the interpretability of the networks.

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
11
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6

Relationship

2
4

Authors

Journals

citations
Cited by 7 publications
(11 citation statements)
references
References 52 publications
0
11
0
Order By: Relevance
“…Generic taggers for multi-pronged jets can also be built by using representation learning, e.g. with an autoencoder [14][15][16][17][18]. Without the need of any signal assumption, but only using background (pseudo-)data, an unsupervised tagger can learn the background features in order to pinpoint outliers, i.e.…”
mentioning
confidence: 99%
“…Generic taggers for multi-pronged jets can also be built by using representation learning, e.g. with an autoencoder [14][15][16][17][18]. Without the need of any signal assumption, but only using background (pseudo-)data, an unsupervised tagger can learn the background features in order to pinpoint outliers, i.e.…”
mentioning
confidence: 99%
“…[50]). Some methods that search for outliers rely on abstract representations to try to characterize the event space, such as the latent space of an autoencoder [20,48]. Others give the event space itself a geometric interpretation in terms of distances [55][56][57].…”
mentioning
confidence: 99%
“…To do so, the autoencoder must establish a delicate balance in achieving a reconstruction fidelity which is accurate, but not too accurate. There are several cases where this is especially difficult, such as when the signal-to-background ratio S/B is small, when the dataset has certain topological properties [18], or when innate characteristics of the samples make the signal sample simpler than the background sample to reconstruct [46,48].…”
mentioning
confidence: 99%
See 2 more Smart Citations