Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1161
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Labeled Parsing with Deep Inside-Outside Recursive Autoencoders

Abstract: Understanding text often requires identifying meaningful constituent spans such as noun phrases and verb phrases. In this work, we show that we can effectively recover these types of labels using the learned phrase vectors from deep inside-outside recursive autoencoders (DIORA). Specifically, we cluster span representations to induce span labels. Additionally, we improve the model's labeling accuracy by integrating latent code learning into the training procedure. We evaluate this approach empirically through … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(22 citation statements)
references
References 21 publications
0
22
0
Order By: Relevance
“…Typically unsupervised constituency parsing is purely evaluated by its structure, although recent work fromDrozdov et al (2019b) shows that a simple approach to induce labels with DIORA can be done by clustering the inside and outside phrase vectors.…”
mentioning
confidence: 99%
“…Typically unsupervised constituency parsing is purely evaluated by its structure, although recent work fromDrozdov et al (2019b) shows that a simple approach to induce labels with DIORA can be done by clustering the inside and outside phrase vectors.…”
mentioning
confidence: 99%
“…Table 1 shows the unlabeled F 1 scores for our model compared to existing unsupervised parsers on PTB. The vanilla inside model is in itself competitive and is already in the range of previous best models like DIORA (Drozdov et al, 2019), Compound PCFG (Kim et al, 2019a). 4 See Appendix 3 https://nlp.cs.nyu.edu/evalb 4 We do not include the results of Shi et al (2021) in our analysis because their boost in the performance is contingent on the nature of the supervision data (especially the QA-SRL dataset) rather than on the actual learning process itself.…”
Section: Resultsmentioning
confidence: 99%
“…Following prior work (Kim et al, 2019a;Shen et al, 2018Shen et al, , 2019Cao et al, 2020), we remove punctuation and collapse unary chains before evaluation, and calculate F 1 ignoring trivial spans, i.e., single-word spans and whole-sentence spans, and we perform the averaging at sentence-level (macro average) rather than span-level (micro average), which means that we compute F 1 for each sentence and later average over all sentences. We also mention the oracle (Shen et al, 2019) 47.7 49.4 63.9 -Tree Transformer † (Wang et al, 2019) 50.5 52.0 66.2 -Neural PCFG † (Kim et al, 2019a) 50.8 52.6 64.6 -DIORA (Drozdov et al, 2019) -58.9 60.5 -Compound PCFG † (Kim et al, 2019a) 55.2 60.1 70.5 -S-DIORA † (Drozdov et al, 2020) 57.6 64.0 71.8 -Constituency Test (Cao et al, 2020) 62.8 65.9 (2019a) and take the baseline numbers of certain models from (Kim et al, 2019a;Cao et al, 2020). † denotes models trained without punctuation and denotes models trained on additional data.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations