2020
DOI: 10.1609/aaai.v34i05.6351
|View full text |Cite
|
Sign up to set email alerts
|

Semi-Supervised Learning on Meta Structure: Multi-Task Tagging and Parsing in Low-Resource Scenarios

Abstract: Multi-view learning makes use of diverse models arising from multiple sources of input or different feature subsets for the same task. For example, a given natural language processing task can combine evidence from models arising from character, morpheme, lexical, or phrasal views. The most common strategy with multi-view learning, especially popular in the neural network community, is to unify multiple representations into one unified vector through concatenation, averaging, or pooling, and then build a singl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 20 publications
(22 citation statements)
references
References 19 publications
(12 reference statements)
0
22
0
Order By: Relevance
“…They include models trained on other grammar formalisms to improve dependency parsing on Twitter (Foster et al, 2011). Recently, this line of classics has been revisited (Ruder and Plank, 2018;Rotman and Reichart, 2019;Lim et al, 2020). For example, classic methods such as tri-training constitute a strong baseline for domain shift in neural times (Ruder and Plank, 2018).…”
Section: Pseudo-labelingmentioning
confidence: 99%
“…They include models trained on other grammar formalisms to improve dependency parsing on Twitter (Foster et al, 2011). Recently, this line of classics has been revisited (Ruder and Plank, 2018;Rotman and Reichart, 2019;Lim et al, 2020). For example, classic methods such as tri-training constitute a strong baseline for domain shift in neural times (Ruder and Plank, 2018).…”
Section: Pseudo-labelingmentioning
confidence: 99%
“…Compared with taggers that used BERT-like models, our join model shows slightly better performances. Even though Co-meta applied both the meta-LSTM and the sentence-based character model [15], our join model that applies two character models showed higher performances. However, it should be noted that the udfy model was first trained with 75 different languages and then tuned for English [11].…”
Section: Resultsmentioning
confidence: 85%
“…In POS tagging, multilingual BERT 6 that can handle 100 languages is applied to train a multilingual POS tagger, namely, udify [11] in 2019. More recently, Lim et al [15] proposed a Co-meta tagger that leverages its performance based on a semi-supervised learning approach with the multilingual BERT and BERT-base monolingual English model, and they achieved SOTA results. We compare the performance of our tagger with BERT-like models as existing SOTA systems, in particular, with multilingual BERT, BERT-base, and Roberta [35].…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations