Proceedings - Natural Language Processing in a Deep Learning World 2019
DOI: 10.26615/978-954-452-056-4_147
|View full text |Cite
|
Sign up to set email alerts
|

ETNLP: A Visual-Aided Systematic Approach to Select Pre-Trained Embeddings for a Downstream Task

Abstract: Given many recent advanced embedding models, selecting pre-trained word embedding (a.k.a., word representation) models best fit for a specific downstream task is non-trivial. In this paper, we propose a systematic approach, called ETNLP, for extracting, evaluating, and visualizing multiple sets of pre-trained word embeddings to determine which embeddings should be used in a downstream task.We demonstrate the effectiveness of the proposed approach on our pre-trained word embedding models in Vietnamese to select… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
17
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 18 publications
(19 citation statements)
references
References 19 publications
0
17
0
Order By: Relevance
“…For NER, PhoBERT large produces 1.1 points higher F 1 than PhoBERT base . In addition, PhoBERT base obtains 2+ points higher than the previous SOTA feature-and neural network-based models VnCoreNLP-NER and BiLSTM-CNN-CRF (Ma and Hovy, 2016) [ ] 88.6 BiLSTM-max (Conneau et al, 2018) 66.4 VNER (Nguyen et al, 2019b) 89.6 mBiLSTM (Artetxe and Schwenk, 2019) 72.0 BiLSTM-CNN-CRF + ETNLP [♠] 91.1 multilingual BERT (Devlin et al, 2019) [ ] 69.5 VnCoreNLP-NER + ETNLP [♠] 91.3 XLM MLM+TLM (Conneau and Lample, 2019) 76.6 XLM-R base (our result) 92.0 XLM-R base (Conneau et al, 2020) 75.4 XLM-R large (our result) 92.8 XLM-R large (Conneau et al, 2020) 79.7 PhoBERT base 93.6 PhoBERT base 78.5 PhoBERT large 94.7 PhoBERT large 80.0 are trained with the set of 15K BERT-based ETNLP word embeddings (Vu et al, 2019).…”
Section: Resultsmentioning
confidence: 89%
See 3 more Smart Citations
“…For NER, PhoBERT large produces 1.1 points higher F 1 than PhoBERT base . In addition, PhoBERT base obtains 2+ points higher than the previous SOTA feature-and neural network-based models VnCoreNLP-NER and BiLSTM-CNN-CRF (Ma and Hovy, 2016) [ ] 88.6 BiLSTM-max (Conneau et al, 2018) 66.4 VNER (Nguyen et al, 2019b) 89.6 mBiLSTM (Artetxe and Schwenk, 2019) 72.0 BiLSTM-CNN-CRF + ETNLP [♠] 91.1 multilingual BERT (Devlin et al, 2019) [ ] 69.5 VnCoreNLP-NER + ETNLP [♠] 91.3 XLM MLM+TLM (Conneau and Lample, 2019) 76.6 XLM-R base (our result) 92.0 XLM-R base (Conneau et al, 2020) 75.4 XLM-R large (our result) 92.8 XLM-R large (Conneau et al, 2020) 79.7 PhoBERT base 93.6 PhoBERT base 78.5 PhoBERT large 94.7 PhoBERT large 80.0 are trained with the set of 15K BERT-based ETNLP word embeddings (Vu et al, 2019).…”
Section: Resultsmentioning
confidence: 89%
“…The success of pre-trained BERT and its variants has largely been limited to the English language. For other languages, one could retrain a language-specific model using the BERT architecture (Cui et al, 2019;de Vries et al, 2019;Vu et al, 2019;Martin et al, 2020) or employ existing pre-trained multilingual BERT-based models (Devlin et al, 2019;Conneau and Lample, 2019;Conneau et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…In this model, we use external knowledge sources as word embeddings. To explore the effectiveness of word embeddings, we evaluate the performance of our proposed model on with several word embeddings including Word2vec [73], Word2vec and Character2vec [74], fastText [75], ELMo [72], BERT [49] and MULTI [76]. In particular, we use pre-trained embeddings on Vietnamese Wikipedia proposed by [76] for all experiments of our proposed method.…”
Section: External Knowledge Integrationmentioning
confidence: 99%