Error Analysis of Kernel Regularized Regression with Deterministic Spherical Scattered Data
2022-01-19LIJunyi李峻屹SHENGBaohuai盛宝怀
LI Junyi(李峻屹), SHENG Baohuai(盛宝怀)
(1.Department of Information Technology, Shanxi Police College, Xi’an 710021, China;2.Department of Applied Statistics, Shaoxing University, Shaoxing 312000, China)
Abstract: We investigate the error bounds of kernel regularized regression learning associated with deterministic data on the spherical cap (and the whole unit sphere) and the quadratic loss.We transform the learning error as the upper bound estimate for the numerical integration error and obtain the learning rates with the convergence rates of kernel-based quadrature.We express the learning rates with a modulus of smoothness which is equivalent to a K-functional in learning theory.The research results show that the learning rates are controlled by the mesh norm of the scattered data.
Key words: Number integration formula; Spherical cap; Worst-case error; Quadratic function loss; Kernel regularized regression; Learning rate
1.Introduction
It is known that learning theory is a new field of application and computation mathematics,many contributions have been done.[1−3]Among these researches,learning associating with spherical harmonics has been getting a lot of attentions.[4−5]Since kernel regularized approach associated with deterministic scattered data has found valuable applications in various fields of numerical analysis, including interpolation, meshless methods for solving partial differential equation, neural networks, machine learning, identification, recognition and control[6−9]and the kernel regularized framework has been established[10−11], it is necessary for us to give investigation on the performance of kernel regularized learning associating with determined spherical data.This paper provides a new method of deriving deterministic error bounds for the kernel regularized regression associated with kernel functionals on the unit sphere.We show how the error estimates of kernel-based quadrature in the deterministic setting associated with the reproducing kernel Hilbert space(RKHS)on the unit sphere are used to bound the learning rates and show that the learning rates are determined by the mesh norm of the scattered data.
Then,by Theorem 17.8 in[9],we know for given pairwise distinct pointsXN:={ξ1,ξ2,··· ,ξN}⊂Sq, the matrixAN:= (K(ξi, ξj))N×Nis positive definite.Therefore,K(x,y) is a Mercer kernel onSq×Sqwithelbeing its eigenvalues.
Define a nature spaceNφas
with the inner product
Define aK-functional
and a general modulus of smoothness as
Letz ∈Sqbe a given point,B(z;γ) ={x ∈Sq:x·z ≥cosγ}={x ∈Sq:d(x,z) =arccos(x·z)≤γ}be the spherical cap with centerz ∈Sqand radiusγ ∈(0,π).Herex·zdenotes the Euclidean inner product ofxandzin Rq+1.
We shall bound the learning rate of the kernel regularized regression on scattered data on the spherical capB(z;γ)and the whole unit sphereSqin Section 2 and Section 3 respectively.The discussions are along the following lines.We first provide the worst-case error bounds for kernel-based quadrature for kernel (1.2) with(see Proposition 2.1 and Proposition 3.1).Basing on these results, we establish the kernel regularized learning frameworks associated with scattered data onB(z;γ) andSqrespectively and then give the learning rates with the equivalent relation (1.4) and the worst-case error bounds.
2.Learning Rates on a Spherical Cap
Define the Sobolev spaceHK(s,q)=Hs,qas
ThenHs,qis a RKHS with an inner product defined as
i.e.,
Moreover, there holds (see (2.8) of [14])
Considerm-point rulesfor numerical integration over the spherical capB(z;γ),
with nodesxj ∈B(z;γ) and weightswj ∈R,j=1,2,··· ,m,for approximating the integral
of a continuous functionf,defined onSq,over the spherical capB(z;γ).The worst-case errorE() ofin Hilbert spaceHs,qis defined by
The worst-case error ofinHs,qsatisfies the estimate
where[a] denote the biggest integer≤a,|B(z;γ)|is the area of the spherical capB(z;γ), and
It follows
Theorem 2.1Letbe the solution of (2.8) andf∗∈L2(Sq).Then there exists a constantC=C(s,q,k)>0 such that
To show Theorem 2.1, we first give a lemma.
Lemma 2.1Letandbe defined as in (2.8) and (2.9) respectively.Then
Proof(2.11) can be obtained from (1.7) of [16] by takingy=f∗(x) andV(t) =−1 or it can be obtained from (5.14) in Chapter 5 of [3].
Proof of Theorem 2.1Simple computations yields
where
By the mean value theorem we have
where
and
It follows by (2.7), (2.16) and (2.11) that
(2.17),(2.14)together with(2.12)give(2.10).Since=0, takingλ=n−θ,0<θ
Corollary 2.1Letbe the solution of (2.8) andf∗∈L2(Sq).Then there holds almost everywhere the following convergence
Proof(2.19) can be obtained by (2.10) and the fact thatV(t)=−1→0+⇔|t|→0+.
3.Learning Rates on the Whole Unit Sphere
Considerm-point rulesQmfor numerical integration over the whole unit sphereSq,
with nodesxj ∈Sqand weightswj ∈R,j=1,2,··· ,m,for approximating the integral
of a continuous functionf, defined onSq.The worst-case errorE(Qm,Hs,q)ofQmin Hilbert spaceHs,qis defined by
Proposition 3.1[14]Forq ≥2 andthere exists a positive constantcsuch that for any positive weightm(n)-point numerical integration ruleQm(n), withQm(n)(p) =pfor all
The constantcdepends only onsandq.
We give a lemma to show that the conditions in Proposition 3.1 is satisfied.
Define the mesh normnd the separation diameterfor a data⊂Sqby
Theorem 3.1Letbe the solution of (3.7) andf∗∈L2(Sq).Then there exists a constantC=C(s,q,k)>0 such that
Proof(3.8) can be obtained by replacingB(z;γ) in (2.11) and (2.18) withSq.
Since the worst-case bounds (2.7) and (3.4) are determined by the mesh norm of the scattered datawe can say that the learning rates of the algorithms (2.8) and (3.7) are controlled by the mesh norm as well.
杂志排行
应用数学的其它文章
- Atomic Decomposition of Weighted Orlicz-Lorentz Martingale Spaces and Its Applications
- Solvability of Mixed Fractional Periodic Boundary Value Problem with p(t)-Laplacian Operator
- 一类含Hardy-Leray势的分数阶p-Laplacian方程解的单调性和对称性
- 一类非线性趋化方程的能控性及时间最优控制
- Existence of Positive Solutions for a Fractional Differential Equation with Multi-point Boundary Value Problems
- 奇异椭圆方程Robin问题多重正解的存在性