EconPapers    
Economics at your fingertips  
 

Error Bounds for -Norm Multiple Kernel Learning with Least Square Loss

Shao-Gao Lv and Jin- De Zhu

Abstract and Applied Analysis, 2012, vol. 2012, 1-18

Abstract:

The problem of learning the kernel function with linear combinations of multiple kernels has attracted considerable attention recently in machine learning. Specially, by imposing an -norm penalty on the kernel combination coefficient, multiple kernel learning (MKL) was proved useful and effective for theoretical analysis and practical applications (Kloft et al., 2009, 2011). In this paper, we present a theoretical analysis on the approximation error and learning ability of the -norm MKL. Our analysis shows explicit learning rates for -norm MKL and demonstrates some notable advantages compared with traditional kernel-based learning algorithms where the kernel is fixed.

Date: 2012
References: Add references at CitEc
Citations:

Downloads: (external link)
http://downloads.hindawi.com/journals/AAA/2012/915920.pdf (application/pdf)
http://downloads.hindawi.com/journals/AAA/2012/915920.xml (text/xml)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:hin:jnlaaa:915920

DOI: 10.1155/2012/915920

Access Statistics for this article

More articles in Abstract and Applied Analysis from Hindawi
Bibliographic data for series maintained by Mohamed Abdelhakeem ().

 
Page updated 2025-03-19
Handle: RePEc:hin:jnlaaa:915920