2010 IEEE International Conference on Acoustics, Speech and Signal Processing 2010
DOI: 10.1109/icassp.2010.5494947
|View full text |Cite
|
Sign up to set email alerts
|

A new topic-bridged model for transfer learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2010
2010
2021
2021

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 9 publications
0
6
0
Order By: Relevance
“…In order to fully utilize the knowledge of the source domain, additional penalty terms are added for must-link and cannot-link constraints. An extension of PLSA (a) Topic-bridged PLSA (Xue et al, 2008) and Topic-bridged LDA (Wu & Chien, 2010) (named Dual-PLSA (Yoo & Choi, 2009;Gao & Li, 2011) (Gao & Li, 2011) is proposed to mine the topics shared by two domains, where topics are defined as word-pair distributions rather than word distributions which models the cross-domain word co-occurrence relations. The aforementioned models all try to learn the shared topics between domains but the domain-specific properties are ignored, and irrelevant topics may degrade the task performance in the target domain.…”
Section: Share Topicmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to fully utilize the knowledge of the source domain, additional penalty terms are added for must-link and cannot-link constraints. An extension of PLSA (a) Topic-bridged PLSA (Xue et al, 2008) and Topic-bridged LDA (Wu & Chien, 2010) (named Dual-PLSA (Yoo & Choi, 2009;Gao & Li, 2011) (Gao & Li, 2011) is proposed to mine the topics shared by two domains, where topics are defined as word-pair distributions rather than word distributions which models the cross-domain word co-occurrence relations. The aforementioned models all try to learn the shared topics between domains but the domain-specific properties are ignored, and irrelevant topics may degrade the task performance in the target domain.…”
Section: Share Topicmentioning
confidence: 99%
“…For example, Short Text Similarity (STS) evaluates the semantic similarity between two short text snippets (target domain). Since they are short (only a few sentences, e.g., a tweet), standard statistical Bayesian linear and logistic regression (Friedman, Hastie, & Tibshirani, 2001) (Sultan et al, 2016) Probabilistic matrix factorization model (PMF) (Mnih & Salakhutdinov, 2008) (Jing et al, 2014) (Iwata & Koh, 2015) Flexible mixture model (Si & Jin, 2003) (Li et al, 2009) Polylingual topic models (Mimno, Wallach, Naradowsky, Smith, & McCallum, 2009) (Hu et al, 2014) Probabilistic latent semantic analysis (PLSA) (Hofmann, 1999) (Xue et al, 2008), (Gao & Li, 2011), (Zhuang et al, 2013), (Zhuang et al, 2010), (Zhuang et al, 2012), (Li et al, 2012), (Zhai et al, 2004(Zhai et al, ), et al, 2009 Latent Dirichlet allocation (LDA) (Blei et al, 2003) (Wu & Chien, 2010), (Jin et al, 2011), (Chen et al, 2015), (Yu & Aloimonos, 2010), (Yang et al, 2011), (Tang et al, 2012), (Phan et al, 2011) Probabilistic linear discriminant analysis (PLDA) (Prince & Elder, 2007) (Hong, Zhang, Li, Wan, & Tong, 2016) (López & Lleida, 2012) Conditional random field (CRF) (Lafferty et al, 2001) (Nallapati, Surdeanu, & Manning, 2010) (Finkel & Manning, 2009) (Arnold et al, 2008) Hierarchal latent Dirichlet allocation (hLDA)…”
Section: Natural Language Processingmentioning
confidence: 99%
“…Datasets structured like this were first used by Dai et al [13,14] to test two different approaches: CoCC [13], which co-clusters domains and words as a means to propagate the class structure from the source domain to the target domain; and TrAdaBoost [14], an extension of AdaBoost that implements transfer learning. Since then, many authors have adopted experimental settings with the same structure, in order to test transfer learning systems based on topic models (e.g., Topic-Bridged PLSA (TPLSA - [60]), Topic-Bridged LDA (TLDA - [55]), and Partially Supervised Cross-Collection LDA (PSCCLDA - [4])), non-negative matrix factorization (e.g., MTrick [65]), probabilistic models (e.g., Topic Correlation Analysis (TCA - [27])), and clustering techniques (e.g., Cross-Domain Spectral Classification (CDSC - [30])).…”
Section: Transductive Transfer Problemsmentioning
confidence: 99%
“…However, although these methods have been tested on transductive transfer problems (i.e., by having A T B T play the role of Ob U T and T e U T at the same time), not all of them are transductive transfer methods as defined in Section 2. Indeed, TrAdaBoost [14], TLDA [55], and TCA [27] are inductive transfer methods; i.e., when applied to a transductive problem, a "TTLP-via-ITLM approach" must be followed. When inductive transfer learning methods are tested on an inductive transfer learning problem, they are meant to be tested on a test set T e U T different from the unlabelled set T r U T on which they have been trained, in order to show that they generalize.…”
Section: Transductive Transfer Problemsmentioning
confidence: 99%
“…They make use of TF-IDF ranking technique in construction of the user profile, which they use for recommending other Twitter users to follow. c) Micro-post Classification: In [12], [26] the authors present LDA transfer learning. Transfer Learning is the process of generic learning in one domain and applying the model in a different domain.…”
Section: Tweet Message Classificationmentioning
confidence: 99%