2021
DOI: 10.1109/tpami.2021.3077397
|View full text |Cite
|
Sign up to set email alerts
|

Mutual Information Regularized Feature-level Frankenstein for Discriminative Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4
1

Relationship

6
4

Authors

Journals

citations
Cited by 21 publications
(9 citation statements)
references
References 64 publications
0
9
0
Order By: Relevance
“…Our classifier C and feature extractor f play an asymmetrical adversarial game to encourage that f eliminates the modality information. 27 Rather than maximizing the cross-entropy loss, f minimizes the KL-divergence of its softmax prediction and a uniform distribution. Specifically, we minimize the following loss: We note that the modality and position label of the sampled two patches are known, which can be used for supervised training.…”
Section: Our Proposed Networkmentioning
confidence: 99%
“…Our classifier C and feature extractor f play an asymmetrical adversarial game to encourage that f eliminates the modality information. 27 Rather than maximizing the cross-entropy loss, f minimizes the KL-divergence of its softmax prediction and a uniform distribution. Specifically, we minimize the following loss: We note that the modality and position label of the sampled two patches are known, which can be used for supervised training.…”
Section: Our Proposed Networkmentioning
confidence: 99%
“…A possible solution to avoid the largen scale labeling of the mobile captured video is using the unsupervised domain adaptation to transfer the knowledge from our dataset to the unlabeled mobile dataset [67], [68], [69]. In addition, it is promising to apply a face pose invariant or robust feature extractor as [35], [70], [71], [72].…”
Section: A Clinical Prospectsmentioning
confidence: 99%
“…L KL is only applied to the same utterance pairs. In parallel, s i is encouraged to inherit the subject-specific factors with an implicit complementary constraint [15,11]. By enforcing the information bottleneck, i.e., compact or low-dimensional latent feature [11], s i has to incorporate all the necessary complementary content (e.g., subject-specific style of the articulation) other than u i to achieve accurate reconstruction.…”
Section: Pair-wise Disentanglement Trainingmentioning
confidence: 99%