2017
DOI: 10.48550/arxiv.1707.07341
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Prediction-Constrained Training for Semi-Supervised Mixture and Topic Models

Abstract: Supervisory signals have the potential to make low-dimensional data representations, like those learned by mixture and topic models, more interpretable and useful. We propose a framework for training latent variable models that explicitly balances two goals: recovery of faithful generative explanations of high-dimensional data, and accurate prediction of associated semantic labels. Existing approaches fail to achieve these goals due to an incomplete treatment of a fundamental asymmetry: the intended applicatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(5 citation statements)
references
References 32 publications
0
5
0
Order By: Relevance
“…While this bound does explicitly demonstrate a generative-discriminative trade-off analogous to Hughes et al (2017), the bound can neither be optimized directly in a computationally efficient manner, nor does it imply that the joint likelihood in Eq. 1 captures this trade-off-indeed, it would be unfortunate if our results were because of our choice of approximate inference rather than the due to the model itself.…”
Section: Discussionmentioning
confidence: 98%
See 4 more Smart Citations
“…While this bound does explicitly demonstrate a generative-discriminative trade-off analogous to Hughes et al (2017), the bound can neither be optimized directly in a computationally efficient manner, nor does it imply that the joint likelihood in Eq. 1 captures this trade-off-indeed, it would be unfortunate if our results were because of our choice of approximate inference rather than the due to the model itself.…”
Section: Discussionmentioning
confidence: 98%
“…Baselines. We compare the generative and predictive performance of pf-GMM to (1) an approach that first learns a generative model using only the inputs and then trains a logistic regression classifier on the posterior beliefs p(Z|X) (2-Step-GMM), (2) a supervised Gaussian mixture model without the switch parameters (sup-GMM), (3) a logistic regression model, which is a discriminative approach that maximizes log p(Y |X) directly (LogReg), and (4) a prediction-constrained GMM (Hughes et al 2017) (pc-GMM). For the pf-HMM, we similarly compare to sup-HMM, 2-Step-HMM and LogReg.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations