2022
DOI: 10.48550/arxiv.2201.13094
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Designing Universal Causal Deep Learning Models: The Geometric (Hyper)Transformer

Abstract: A. We introduce a universal class of geometric deep learning models, called metric hypertransformers (MHTs), capable of approximating any adapted map F : X Z → Y Z with approximable complexity, where X ⊆ R d and Y is any suitable metric space, and X Z (resp. Y Z ) capture all discrete-time paths on X (resp. Y ). Suitable spaces Y include various (adapted) Wasserstein spaces, all Fréchet spaces admitting a Schauder basis, and a variety of Riemannian manifolds arising from information geometry. Even in the stati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 81 publications
0
3
0
Order By: Relevance
“…We set W i o = 0 in the attention block, so that the input directly goes into the pointwise feed-forward block. From the previous discussion, the pointwise feedforward network can be constructed to give an output y (1) : R → R 2τ+1 , such that y (1) (s) = c s [ φ 0,s (x(s)), . .…”
Section: Attention-based Architecturesmentioning
confidence: 99%
See 2 more Smart Citations
“…We set W i o = 0 in the attention block, so that the input directly goes into the pointwise feed-forward block. From the previous discussion, the pointwise feedforward network can be constructed to give an output y (1) : R → R 2τ+1 , such that y (1) (s) = c s [ φ 0,s (x(s)), . .…”
Section: Attention-based Architecturesmentioning
confidence: 99%
“…then we have Attn(y (1) )(1) = ∑ τ s=1 y (1) (s). Hence, the final output after the feed-forward network with linear readout c ⊤ = (1, .…”
Section: Attention-based Architecturesmentioning
confidence: 99%
See 1 more Smart Citation