TY - JOUR
T1 - Triplex transfer learning
T2 - Exploiting both shared and distinct concepts for text classification
AU - Zhuang, Fuzhen
AU - Luo, Ping
AU - Du, Changying
AU - He, Qing
AU - Shi, Zhongzhi
AU - Xiong, Hui
PY - 2014/7
Y1 - 2014/7
N2 - Transfer learning focuses on the learning scenarios when the test data from target domains and the training data from source domains are drawn from similar but different data distributions with respect to the raw features. Along this line, some recent studies revealed that the high-level concepts, such as word clusters, could help model the differences of data distributions, and thus are more appropriate for classification. In other words, these methods assume that all the data domains have the same set of shared concepts, which are used as the bridge for knowledge transfer. However, in addition to these shared concepts, each domain may have its own distinct concepts. In light of this, we systemically analyze the high-level concepts, and propose a general transfer learning framework based on nonnegative matrix trifactorization, which allows to explore both shared and distinct concepts among all the domains simultaneously. Since this model provides more flexibility in fitting the data, it can lead to better classification accuracy. Moreover, we propose to regularize the manifold structure in the target domains to improve the prediction performances. To solve the proposed optimization problem, we also develop an iterative algorithm and theoretically analyze its convergence properties. Finally, extensive experiments show that the proposed model can outperform the baseline methods with a significant margin. In particular, we show that our method works much better for the more challenging tasks when there are distinct concepts in the data.
AB - Transfer learning focuses on the learning scenarios when the test data from target domains and the training data from source domains are drawn from similar but different data distributions with respect to the raw features. Along this line, some recent studies revealed that the high-level concepts, such as word clusters, could help model the differences of data distributions, and thus are more appropriate for classification. In other words, these methods assume that all the data domains have the same set of shared concepts, which are used as the bridge for knowledge transfer. However, in addition to these shared concepts, each domain may have its own distinct concepts. In light of this, we systemically analyze the high-level concepts, and propose a general transfer learning framework based on nonnegative matrix trifactorization, which allows to explore both shared and distinct concepts among all the domains simultaneously. Since this model provides more flexibility in fitting the data, it can lead to better classification accuracy. Moreover, we propose to regularize the manifold structure in the target domains to improve the prediction performances. To solve the proposed optimization problem, we also develop an iterative algorithm and theoretically analyze its convergence properties. Finally, extensive experiments show that the proposed model can outperform the baseline methods with a significant margin. In particular, we show that our method works much better for the more challenging tasks when there are distinct concepts in the data.
KW - Common concept
KW - distinct concept
KW - distribution mismatch
KW - nonnegative matrix trifactorization
KW - triplex transfer learning
UR - https://www.scopus.com/pages/publications/84903132879
U2 - 10.1109/TCYB.2013.2281451
DO - 10.1109/TCYB.2013.2281451
M3 - 文章
C2 - 24108492
AN - SCOPUS:84903132879
SN - 2168-2267
VL - 44
SP - 1191
EP - 1203
JO - IEEE Transactions on Cybernetics
JF - IEEE Transactions on Cybernetics
IS - 7
M1 - 6606822
ER -