Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 1: Long Papers) 2017
DOI: 10.18653/v1/p17-1029
|View full text |Cite
|
Sign up to set email alerts
|

Multi-space Variational Encoder-Decoders for Semi-supervised Labeled Sequence Transduction

Abstract: Labeled sequence transduction is a task of transforming one sequence into another sequence that satisfies desiderata specified by a set of labels. In this paper we propose multi-space variational encoderdecoders, a new model for labeled sequence transduction with semi-supervised learning. The generative model can use neural networks to handle both discrete and continuous latent variables to exploit various features of data. Experiments show that our model provides not only a powerful supervised framework but a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
62
1

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 50 publications
(63 citation statements)
references
References 20 publications
0
62
1
Order By: Relevance
“…In this work, we further examine the method proposed in (Zhou and Neubig, 2017) for the shared task of SIGMORPHON 2017 on 52 languages and demonstrate the effectiveness of this approach. We will further improve our model's sophistication by investigating strategies for choosing appropriate semi-supervised data, and examining the model's performance on languages with a high inflection level.…”
Section: Discussionmentioning
confidence: 99%
“…In this work, we further examine the method proposed in (Zhou and Neubig, 2017) for the shared task of SIGMORPHON 2017 on 52 languages and demonstrate the effectiveness of this approach. We will further improve our model's sophistication by investigating strategies for choosing appropriate semi-supervised data, and examining the model's performance on languages with a high inflection level.…”
Section: Discussionmentioning
confidence: 99%
“…, y |Σy| }, respectively. In tasks where the tag is provided, i.e., labeled transduction (Zhou and Neubig, 2017), we denote the tag as an ordered set t ∈ Σ * t with a finite tag vocabulary Σ t = {t 1 , . .…”
Section: Preliminarymentioning
confidence: 99%
“…com/sigmorphon/conll2017/tree/master/ evaluation Zhou and Neubig (2017), where even after 150k unlabeled examples, performance still appears to be increasing.) After controlling for the amount of additional data, we see only a small benefit from autoencoding corpus words (AE-CW) rather than random strings (AE- Figure 2: The accuracy of our best systems on all languages.…”
Section: Multilingual Trainingmentioning
confidence: 99%
“…We train our system using Stochastic Gradient Descent. Our system is implemented using the Dynet toolkit (Neubig et al, 2017) 4 and our code is freely available. 5 There are three hyper-parameters in our system: the character embedding dimension, the size of the hidden layer of the LSTM models and the size of the hidden layer of the attention network.…”
Section: Rnn Encoder-decoder With Attentionmentioning
confidence: 99%
See 1 more Smart Citation