EconPapers    
Economics at your fingertips  
 

Learning Rates for l1‐Regularized Kernel Classifiers

Hongzhi Tong, Di-Rong Chen and Fenghong Yang

Journal of Applied Mathematics, 2013, vol. 2013, issue 1

Abstract: We consider a family of classification algorithms generated from a regularization kernel scheme associated with l1‐regularizer and convex loss function. Our main purpose is to provide an explicit convergence rate for the excess misclassification error of the produced classifiers. The error decomposition includes approximation error, hypothesis error, and sample error. We apply some novel techniques to estimate the hypothesis error and sample error. Learning rates are eventually derived under some assumptions on the kernel, the input space, the marginal distribution, and the approximation error.

Date: 2013
References: Add references at CitEc
Citations:

Downloads: (external link)
https://doi.org/10.1155/2013/496282

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:wly:jnljam:v:2013:y:2013:i:1:n:496282

Access Statistics for this article

More articles in Journal of Applied Mathematics from John Wiley & Sons
Bibliographic data for series maintained by Wiley Content Delivery ().

 
Page updated 2025-03-22
Handle: RePEc:wly:jnljam:v:2013:y:2013:i:1:n:496282