Abstract
A learning algorithm for regression is studied. It is a modified kernel projection machine (Blanchard et al., 2004 [2]) in the form of a least square regularization scheme withℓ1-regularizer in a data dependent hypothesis space based on empirical features (constructed by a reproducing kernel and the learning data). The algorithm has three advantages. First, it does not involve any optimization process. Second, it produces sparse representations with respect to empirical features under a mild condition, without assuming sparsity in terms of any basis or system. Third, the output function converges to the regression function in the reproducing kernel Hilbert space at a satisfactory rate. Our error analysis does not require any sparsity assumption about the underlying regression function.
| Original language | English |
|---|---|
| Pages (from-to) | 389-400 |
| Number of pages | 12 |
| Journal | Applied and Computational Harmonic Analysis |
| Volume | 32 |
| Issue number | 3 |
| DOIs | |
| Publication status | Published - 1 May 2012 |
| Externally published | Yes |
Keywords
- ℓ 1 -regularizer
- Empirical features
- Learning theory
- Reproducing kernel Hilbert space
- Sparsity
ASJC Scopus subject areas
- Applied Mathematics
Fingerprint
Dive into the research topics of 'An empirical feature-based learning algorithm producing sparse approximations'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver