Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.511
|View full text |Cite
|
Sign up to set email alerts
|

A Novel Estimator of Mutual Information for Learning to Disentangle Textual Representations

Abstract: Learning disentangled representations of textual data is essential for many natural language tasks such as fair classification, style transfer and sentence generation, among others. The existent dominant approaches in the context of text data either rely on training an adversary (discriminator) that aims at making attribute values difficult to be inferred from the latent code or rely on minimising variational bounds of the mutual information between latent code and the value attribute. However, the available m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 16 publications
(10 citation statements)
references
References 43 publications
0
10
0
Order By: Relevance
“…Among available contrast measures, the Fisher-Rao distance is parameter-free and thus, it is easy to use in practice while the AB-Divergence achieves better results but requires to select α and β. Future work includes extending our metrics to new tasks such as SLU (Chapuis et al 2020(Chapuis et al , 2021Dinkar et al 2020;Colombo, Clavel, and Piantanida 2021), controlled sentence generation (Colombo et al 2019(Colombo et al , 2021b and multi-modal learning (Colombo et al 2021a;Garcia et al 2019).…”
Section: Summary and Concluding Remarksmentioning
confidence: 99%
“…Among available contrast measures, the Fisher-Rao distance is parameter-free and thus, it is easy to use in practice while the AB-Divergence achieves better results but requires to select α and β. Future work includes extending our metrics to new tasks such as SLU (Chapuis et al 2020(Chapuis et al , 2021Dinkar et al 2020;Colombo, Clavel, and Piantanida 2021), controlled sentence generation (Colombo et al 2019(Colombo et al , 2021b and multi-modal learning (Colombo et al 2021a;Garcia et al 2019).…”
Section: Summary and Concluding Remarksmentioning
confidence: 99%
“…. Therefore, this objective minimizes an upper-bound estimate of the MI between each pair of latent spaces, following (Cheng et al, 2020a,b;Colombo et al, 2021).…”
Section: Mutual-information Minimization (Min): a Natural Measure Of ...mentioning
confidence: 99%
“…Still, no previous work has tested whether negation, uncertainty, and content can be disentangled, as linguistic theory suggests, although previous works have disentangled attributes such as syntax, semantics, and style (Balasubramanian et al, 2021;John et al, 2019;Cheng et al, 2020b;Bao et al, 2019;Hu et al, 2017;Colombo et al, 2021). To fill this gap, we aim to answer the following research questions:…”
Section: Introductionmentioning
confidence: 99%
“…Disentangled VAEs in language Early approaches in text disentanglement use VAEs with multiple adversarial losses for style transfer (Hu et al, 2017;John et al, 2019). More recently, Cheng et al (2020) propose a style transfer method which minimizes the mutual information between the latent and the observed variable, while Colombo et al (2021) propose an upper bound of mutual information for fair text classification. Disentanglement of syntactic and semantic information on sentences is explored by , using multiple losses for word ordering and paraphrasing, and by Bao et al (2019) with linearized constituency tree losses.…”
Section: Related Workmentioning
confidence: 99%