Least Square Regularized Regression for Multitask Learning
Yong-Li Xu,
Di-Rong Chen and
Han-Xiong Li
Abstract and Applied Analysis, 2013, vol. 2013, 1-7
Abstract:
The study of multitask learning algorithms is one of very important issues. This paper proposes a least-square regularized regression algorithm for multi-task learning with hypothesis space being the union of a sequence of Hilbert spaces. The algorithm consists of two steps of selecting the optimal Hilbert space and searching for the optimal function. We assume that the distributions of different tasks are related to a set of transformations under which any Hilbert space in the hypothesis space is norm invariant. We prove that under the above assumption the optimal prediction function of every task is in the same Hilbert space. Based on this result, a pivotal error decomposition is founded, which can use samples of related tasks to bound excess error of the target task. We obtain an upper bound for the sample error of related tasks, and based on this bound, potential faster learning rates are obtained compared to single-task learning algorithms.
Date: 2013
References: Add references at CitEc
Citations: View citations in EconPapers (1)
Downloads: (external link)
http://downloads.hindawi.com/journals/AAA/2013/715275.pdf (application/pdf)
http://downloads.hindawi.com/journals/AAA/2013/715275.xml (text/xml)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:hin:jnlaaa:715275
DOI: 10.1155/2013/715275
Access Statistics for this article
More articles in Abstract and Applied Analysis from Hindawi
Bibliographic data for series maintained by Mohamed Abdelhakeem ().