2019
DOI: 10.48550/arxiv.1904.02142
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unsupervised Latent Tree Induction with Deep Inside-Outside Recursive Autoencoders

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…It is noteworthy that there are many other important miscellaneous works we do not mention in the previous sections. For example, numerous works have proposed to improve upon vanilla gradient-based methods [174,178,65]; linguistic rules such as negation, morphological inflection can be extracted by neural models [141,142,158]; probing tasks can used to explore linguistic properties of sentences [3,80,43,75,89,74,34]; the hidden state dynamics in recurrent nets are analysed to illuminate the learned long-range dependencies [73,96,67,179,94]; [169,166,168,101,57,167] studied the ability of neural sequence models to induce lexical, grammatical and syntactic structures; [91,90,12,136,159,24,151,85] modeled the reasoning process of the model to explain model behaviors; [157,139,28,163,219,170,180,137,106,58,162,81...…”
Section: Miscellaneousmentioning
confidence: 99%
“…It is noteworthy that there are many other important miscellaneous works we do not mention in the previous sections. For example, numerous works have proposed to improve upon vanilla gradient-based methods [174,178,65]; linguistic rules such as negation, morphological inflection can be extracted by neural models [141,142,158]; probing tasks can used to explore linguistic properties of sentences [3,80,43,75,89,74,34]; the hidden state dynamics in recurrent nets are analysed to illuminate the learned long-range dependencies [73,96,67,179,94]; [169,166,168,101,57,167] studied the ability of neural sequence models to induce lexical, grammatical and syntactic structures; [91,90,12,136,159,24,151,85] modeled the reasoning process of the model to explain model behaviors; [157,139,28,163,219,170,180,137,106,58,162,81...…”
Section: Miscellaneousmentioning
confidence: 99%
“…Shen et al (2019) designed a novel recurrent architecture to automatically capture the latent tree structure of an input sentence. Other studies learned syntactic parsers (Drozdov et al 2019;Htut, Cho, and Bowman 2019;Kitaev, Cao, and Klein 2019;Li, Mou, and Keller 2019;Li and Eisner 2019;Mrini et al 2019), although these methods pursued a high parsing accuracy, instead of explaining the DNN. Essentially, the learning of the syntactic parser aimed to make the parser fit syntactic structures defined by human experts.…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, explaining features encoded inside a DNN has become an emerging direction. Based on the inherent hierarchical structure of natural language, many methods use latent tree structures of language to guide the DNN to learn interpretable feature representations (Choi, Yoo, and Lee 2018;Drozdov et al 2019;Shen et al 2018Shen et al , 2019Shi et al 2018;Tai, Socher, and Manning 2015;Wang, Lee, and Chen 2019;Yogatama et al 2016). However, the interpretability usually conflicts with the discrimination power (Bau et al 2017).…”
Section: Introductionmentioning
confidence: 99%
“…However, these methods cannot learn simple grammar and meaningful semantics, though they perform well on NLI tasks [12]. Additionally, several approaches [13]- [15] aim to learn unsupervised parse trees; however, they perform poorly on end task. In this paper, we demonstrate that our approach can capture both grammar and semantics in the sentence, thus learning better parse trees and outperform RvNN-based model on several tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, as we will show in our experimental section, the grammar or meta-level association thus detected by their approach is relatively trivial. Recently, [13] proposed an unsupervised latent chart tree parsing algorithm, viz., DIORA, that uses the inside-outside algorithm for parsing and has an autoencoder-based neural network trained to reconstruct the input sentence. DIORA is trained end to end using masked language model via word prediction.…”
Section: Related Workmentioning
confidence: 99%