Statistical properties of the method of regularization with periodic
Gaussian reproducing kernel
release_jrkklbfbz5dbpktuhjbmlvomxm
by
Yi Lin, Lawrence D. Brown
2004
Abstract
The method of regularization with the Gaussian reproducing kernel is popular
in the machine learning literature and successful in many practical
applications.
In this paper we consider the periodic version of the Gaussian kernel
regularization.
We show in the white noise model setting, that in function spaces of very
smooth functions, such as the infinite-order Sobolev space and the space of
analytic functions, the method under consideration is asymptotically minimax;
in finite-order Sobolev spaces, the method is rate optimal, and the efficiency
in terms of constant when compared with the minimax estimator is reasonably
high. The smoothing parameters in the periodic Gaussian regularization can be
chosen adaptively without loss of asymptotic efficiency. The results derived in
this paper give a partial explanation of the success of the
Gaussian reproducing kernel in practice. Simulations are carried out to study
the finite sample properties of the periodic Gaussian regularization.
In text/plain
format
Archived Files and Locations
application/pdf 341.2 kB
file_krf4qjk7prg7zlm7bfaqkflm4e
|
archive.org (archive) |
access all versions, variants, and formats of this works (eg, pre-prints)