Modern State Development, Capacity, and Institutions 2017
DOI: 10.2307/j.ctv1nzg214.5
|View full text |Cite
|
Sign up to set email alerts
|

The State, State Capabilities and Non-State Actors:

Abstract: Attention-based models such as Transformers and recurrent models like state space models (SSMs) have emerged as successful methods for autoregressive sequence modeling. Although both enable parallel training, none enable parallel generation due to their autoregressiveness. We propose the variational SSM (VSSM), a variational autoencoder (VAE) where both the encoder and decoder are SSMs. Since sampling the latent variables and decoding them with the SSM can be parallelized, both training and generation can be c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 4 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?