Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.673
|View full text |Cite
|
Sign up to set email alerts
|

Improving Disentangled Text Representation Learning with Information-Theoretic Guidance

Abstract: Learning disentangled representations of natural language is essential for many NLP tasks, e.g., conditional text generation, style transfer, personalized dialogue systems, etc. Similar problems have been studied extensively for other forms of data, such as images and videos. However, the discrete nature of natural language makes the disentangling of textual representations more challenging (e.g., the manipulation over the data space cannot be easily achieved). Inspired by information theory, we propose a nove… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
53
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 45 publications
(53 citation statements)
references
References 21 publications
(21 reference statements)
0
53
0
Order By: Relevance
“…Because the exact value of Equation ( 8) is difficult to calculate in practice, we minimize its upperbound following Cheng et al (2020):…”
Section: Design Of Loss Functionmentioning
confidence: 99%
“…Because the exact value of Equation ( 8) is difficult to calculate in practice, we minimize its upperbound following Cheng et al (2020):…”
Section: Design Of Loss Functionmentioning
confidence: 99%
“…Third, it is possible that one feature detects several patterns (Jacovi et al, 2018) and it will be difficult to disable the feature if some of the detected patterns are useful while the others are harmful. Hence, FIND would be more effective when used together with disentangled text representations (Cheng et al, 2020).…”
Section: Limitationsmentioning
confidence: 99%
“…The sentiment information is contained in w t , while the content of the original sentence is represented by O t . To achieve styletransfer, one feeds the original sentence X with the target style label l to get the transferred sentence Y with style l. Following previous work (Hu et al, 2017;Yang et al, 2018;Cheng et al, 2020), we adopt a classifier as the discriminator and the soft-argmax approach (Kusner and Miguel, 2016) for the update of generator instead of policy gradient (Sutton and Barto, 1998).…”
Section: Extension To Non-parallel Text Style Transfermentioning
confidence: 99%