TY - JOUR
T1 - Relevant subtask learning by constrained mixture models
AU - Peltonen, Jaakko
AU - Yaslan, Yusuf
AU - Kaski, Samuel
PY - 2010
Y1 - 2010
N2 - We introduce relevant subtask learning, a new learning problem which is a variant of multi-task learning. The goal is to build a classifier for a task-of-interest for which we have too few training samples. We additionally have "supplementary data" collected from other tasks, but it is uncertain which of these other samples are relevant, that is, which samples are classified in the same way as in the task-of-interest. The research problem is how to use the "supplementary data" from the other tasks to improve the classifier in the task-of-interest. We show how to solve the problem, and demonstrate the solution with logistic regression classifiers. The key idea is to model all tasks as mixtures of relevant and irrelevant samples, and model the irrelevant part with a sufficiently flexible model such that it does not distort the model of relevant data. We give two learning algorithms for the method - a simple maximum likelihood optimization algorithm and a more advanced variational Bayes inference algorithm; in both cases we show that the method works better than a comparable multi-task learning model and naive methods.
AB - We introduce relevant subtask learning, a new learning problem which is a variant of multi-task learning. The goal is to build a classifier for a task-of-interest for which we have too few training samples. We additionally have "supplementary data" collected from other tasks, but it is uncertain which of these other samples are relevant, that is, which samples are classified in the same way as in the task-of-interest. The research problem is how to use the "supplementary data" from the other tasks to improve the classifier in the task-of-interest. We show how to solve the problem, and demonstrate the solution with logistic regression classifiers. The key idea is to model all tasks as mixtures of relevant and irrelevant samples, and model the irrelevant part with a sufficiently flexible model such that it does not distort the model of relevant data. We give two learning algorithms for the method - a simple maximum likelihood optimization algorithm and a more advanced variational Bayes inference algorithm; in both cases we show that the method works better than a comparable multi-task learning model and naive methods.
KW - Multi-task learning
KW - relevant subtask learning
UR - http://www.scopus.com/inward/record.url?scp=78650406464&partnerID=8YFLogxK
U2 - 10.3233/IDA-2010-0445
DO - 10.3233/IDA-2010-0445
M3 - Article
AN - SCOPUS:78650406464
SN - 1088-467X
VL - 14
SP - 641
EP - 662
JO - Intelligent Data Analysis
JF - Intelligent Data Analysis
IS - 6
ER -