TY - GEN
T1 - Learning incoherent sparse and low-rank patterns from multiple tasks
AU - Chen, Jianhui
AU - Ji, Liu
AU - Ye, Jieping
PY - 2010
Y1 - 2010
N2 - We consider the problem of learning incoherent sparse and low-rank patterns from multiple tasks. Our approach is based on a linear multi-task learning formulation, in which the sparse and low-rank patterns are induced by a cardinality regularization term and a low-rank constraint, respectively. This formulation is non-convex; we convert it into its convex surrogate, which can be routinely solved via semidefinite programming for small-size problems. We propose to employ the general projected gradient scheme to efficiently solve such a convex surrogate; however, in the optimization formulation, the objective function is non-differentiable and the feasible domain is non-trivial. We present the procedures for computing the projected gradient and ensuring the global convergence of the projected gradient scheme. The computation of projected gradient involves a constrained optimization problem; we show that the optimal solution to such a problem can be obtained via solving an unconstrained optimization subproblem and an Euclidean projection subproblem. In addition, we present two projected gradient algorithms and discuss their rates of convergence. Experimental results on benchmark data sets demonstrate the effectiveness of the proposed multi-task learning formulation and the efficiency of the proposed projected gradient algorithms.
AB - We consider the problem of learning incoherent sparse and low-rank patterns from multiple tasks. Our approach is based on a linear multi-task learning formulation, in which the sparse and low-rank patterns are induced by a cardinality regularization term and a low-rank constraint, respectively. This formulation is non-convex; we convert it into its convex surrogate, which can be routinely solved via semidefinite programming for small-size problems. We propose to employ the general projected gradient scheme to efficiently solve such a convex surrogate; however, in the optimization formulation, the objective function is non-differentiable and the feasible domain is non-trivial. We present the procedures for computing the projected gradient and ensuring the global convergence of the projected gradient scheme. The computation of projected gradient involves a constrained optimization problem; we show that the optimal solution to such a problem can be obtained via solving an unconstrained optimization subproblem and an Euclidean projection subproblem. In addition, we present two projected gradient algorithms and discuss their rates of convergence. Experimental results on benchmark data sets demonstrate the effectiveness of the proposed multi-task learning formulation and the efficiency of the proposed projected gradient algorithms.
KW - Multi-task learning
KW - Sparse and low-rank patterns
KW - Trace norm
UR - http://www.scopus.com/inward/record.url?scp=77956208061&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=77956208061&partnerID=8YFLogxK
U2 - 10.1145/1835804.1835952
DO - 10.1145/1835804.1835952
M3 - Conference contribution
AN - SCOPUS:77956208061
SN - 9781450300551
T3 - Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
SP - 1179
EP - 1187
BT - KDD'10 - Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data
T2 - 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD-2010
Y2 - 25 July 2010 through 28 July 2010
ER -