2007 OptimalRatesfortheRegularizedLe

From GM-RKB
Jump to navigation Jump to search

Subject Headings: Regularized Least-Square Algorithm, Supervised Learning Algorithm, Vector-Valued Function.

Notes

Cited By

2008

Quotes

Abstract

We develop a theoretical analysis of the performance of the regularized least-square algorithm on a reproducing kernel Hilbert space in the supervised learning setting. The presented results hold in the general framework of vector-valued functions; therefore they can be applied to multitask problems. In particular, we observe that the concept of effective dimension plays a central role in the definition of a criterion for the choice of the regularization parameter as a function of the number of samples. Moreover, a complete minimax analysis of the problem is described, showing that the convergence rates obtained by regularized least-squares estimators are indeed optimal over a suitable class of priors defined by the considered kernel. Finally, we give an improved lower rate result describing worst asymptotic behavior on individual probability measures rather than over classes of priors.

Introduction

In this paper we investigate the estimation properties of the regularized least-squares (RLS) algorithm on a reproducing kernel Hilbert space (RKHS) in the regression setting. Following the general scheme of supervised statistical learning theory, the available input-output samples are assumed to be drawn i.i.d. according to an unknown probability distribution. The aim of a regression algorithm is estimating a particular invariant of the unknown distribution: the regression function, using only the available empirical samples.

References

;

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2007 OptimalRatesfortheRegularizedLeAndrea Caponnetto
Ernesto De Vito
Optimal Rates for the Regularized Least-Squares Algorithm10.1007/s10208-006-0196-82007