Proceedings of the Second Workshop on Subword/Character LEvel Models 2018
DOI: 10.18653/v1/w18-1205
|View full text |Cite
|
Sign up to set email alerts
|

Subword-level Composition Functions for Learning Word Embeddings

Abstract: Subword-level information is crucial for capturing the meaning and morphology of words, especially for out-of-vocabulary entries. We propose CNN-and RNN-based subword-level composition functions for learning word embeddings, and systematically compare them with popular word-level and subword-level models (Skip-Gram and FastText). Additionally, we propose a hybrid training scheme in which a pure subword-level model is trained jointly with a conventional word-level embedding model based on lookup-tables. This in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
24
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 25 publications
(26 citation statements)
references
References 34 publications
(28 reference statements)
1
24
0
Order By: Relevance
“…Finally, the subword embedding sequence is passed to a composition function, which computes the final word representation. Li et al (2018) and Zhu et al (2019) have empirically verified that composition by simple addition, among other more complex composition functions, is a robust choice. Therefore, we use addition in all our experiments.…”
Section: Methodsmentioning
confidence: 98%
“…Finally, the subword embedding sequence is passed to a composition function, which computes the final word representation. Li et al (2018) and Zhu et al (2019) have empirically verified that composition by simple addition, among other more complex composition functions, is a robust choice. Therefore, we use addition in all our experiments.…”
Section: Methodsmentioning
confidence: 98%
“…The proposed embeddings -Phonetic CNN subword embeddings -comprise of two distinct embeddings, concatenated together. First, each word w is processed using the CNN-based subword-level embedding model [10], which results in word embedding W for each word. Second, each word is also processed using the phonetic level model, as presented in [40], which gives the phonetic level embedding of the word P ho_w.…”
Section: Phonetic Cnn Subword Embeddingsmentioning
confidence: 99%
“…Li and others [10] proposed the CNN-based subword-level composition function for learning word embeddings. CNN-based subword-level embedding model (Fig.…”
Section: Cnn-based Subword-level Embedding Modelmentioning
confidence: 99%
See 2 more Smart Citations