Next Article in Journal
Key Backup and Recovery for Resilient DID Environment
Previous Article in Journal
Consistency Improvement in the Analytic Hierarchy Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Application of Generalized Gaussian Radial Basis Function and Its Reproducing Kernel Theory

Department of Mathematics, The University of Texas at Tyler, Tyler, TX 75799, USA
Mathematics 2024, 12(6), 829; https://doi.org/10.3390/math12060829
Submission received: 17 February 2024 / Revised: 28 February 2024 / Accepted: 5 March 2024 / Published: 12 March 2024
(This article belongs to the Section Mathematics and Computer Science)

Abstract

:
Gaussian Radial Basis Function Kernels are the most-often-employed kernel function in artificial intelligence for providing the optimal results in contrast to their respective counterparts. However, our understanding surrounding the utilization of the Generalized Gaussian Radial Basis Function across different machine learning algorithms, such as kernel regression, support vector machines, and pattern recognition via neural networks is incomplete. The results delivered by the Generalized Gaussian Radial Basis Function Kernel in the previously mentioned applications remarkably outperforms those of the Gaussian Radial Basis Function Kernel, the Sigmoid function, and the ReLU function in terms of accuracy and misclassification. This article provides a concrete illustration of the utilization of the Generalized Gaussian Radial Basis Function Kernel as mentioned earlier. We also provide an explicit description of the reproducing kernel Hilbert space by embedding the Generalized Gaussian Radial Basis Function as an L 2 measure, which is utilized in implementing the analysis support vector machine. Finally, we provide the conclusion that we draw from the empirical experiments considered in the manuscript along with the possible future directions in terms of spectral decomposition of the Generalized Gaussian Radial Basis Function.

1. Introduction

Contemporary data science, which revolves around diverse architectural frameworks of machine learning, signifies crucial potential across a wide range of applications. These applications span from various fields such as physics and engineering [1] to areas like genomic-assisted prediction [2] and predictive modeling for cancer tumors [3]. The central task of (automated) machine learning is to devise an optimal activation function or optimal kernel function, depending upon the learning architecture, that is best-suited for the specific scientific model. Therefore, it is not hard to see that machine learning applications and many others across various scientific realms duly take advantage of various mathematical functions that arise in basic function theory and incorporate them along with their Hilbert space theories. This symbiotic relationship between advanced machine learning architecture avenues of artificial intelligence and the raw availability of various classes of mathematical functions is aptly captured by pictorial representations produced by the National Academy of Sciences in [4]. This does not only simply depict the importance of such functions but also questions and inquires about the possibility of the existence and applicability of other new, unexplored, better math functions, which can advance the horizons of artificial intelligence, for instance in [5].
Across the rich and diverse history of machine learning, numerous activation functions, along with their incorporation in a kernel sense, have been proposed. Certain important examples of activation functions are the Sigmoid function or S-Shaped curves in Figure 1, like the hyperbolic tangent function (given as 1 1 + e x cf. [6]). The class of Sigmoid functions is employed to classify data objects where the output is supposed to be constrained on a limited range (usually on [ 0 , 1 ] ).
The other class of important and popular activation functions comprises ReLU activation functions in Figure 2, defined by f ( x ) = x if x > 0 , which are employed in machine learning problems of reinforcement learning, quantification, and classification. If one adds a piece of function defined as f ( x ) = α x for x 0 , where α 0 to the actual definition of ReLU, we obtain the α ReLU activation function (see (29) in this article), which is often also referred to as LeakyReLU. It is essential to acknowledge that both ReLU and the α ReLU activation function have discontinuities at the origin; these lead to the failure of the gradient descent optimizer by producing undefined gradients [7].
Furthermore, one of the most important and interesting classes of mathematical functions that leverage crucial platforms of machine learning or deep learning, to name a few, are radial basis functions, defined as follows (cf. [8]).
Definition 1.
A function Φ : R d R is referred to as radial if there exists a univariate function ϕ : { 0 } R + R , such that
Φ x = ϕ ( r ) , w h e r e r = x .
Here, is some norm defined on R d , usually 2 , which is the Euclidean norm.
One such special type of mathematical radial basis function is the Gaussian Radial Basis Function (GRBF), which is given as:
g σ 2 ( r ) : = def exp σ 2 r 2 ,
where σ > 0 . The Gaussian Radial Basis Function is an activation function that possesses a bell-shaped curve (top in Figure 3); it has impact-full applications in modeling random variables, which are distributed according to the Gaussian probability distribution. For instance, a neural network architecture predicting the speed of a car might employ Gaussian kernel regression, since the speed of the car follows the Gaussian probability distribution.
Moreover, in recent years, learning algorithms of support vector machines (SVMs) have played a key role in data-mining methods; these are workhorse machine learning tools in industry and science. The crucial explanation for this aspect is the unique ability of SVMs to embed in higher-dimensional nonlinear spaces for nonlinear classifiers. The theoretical underpinnings of SVMs revolve around the kernel function, which essentially enables one to operate in a high-dimensional space without computing coordinates of data but by simply computing inner products between all pairs of data in the feature space (cf. page 197 in [9]). To that end, we define the most commonly used kernel function, that is, the Gaussian Radial Basis Function Kernel.
Let 2 be the usual Euclidean norm on R d ; then, the g ( 2 ) function is comfortably synonymous to its kernel notion, which is famously referred to as the Gaussian Radial Basis Function Kernel, given as K σ ( x , z ) : = def   g σ 2 ( x z 2 ) for x , z R d . Explicitly, that is
K σ ( x , z ) = exp σ 2 x z 2 2 .
The Gaussian Radial Basis Function Kernel (Figure 4) is a building block for various learning architectures, such as spatial statistics [10], dynamical system identification [11], Gaussian processes for machine learning [12], etc. Additionally, it is also employed for the classification of the existence of objects (cf. [13]). This particular study extends the idea of the Gaussian Radial Basis Function Kernel to what is referred to as the Generalized Gaussian Radial Basis Function Kernel (GGRBF Kernel in Figure 5), as introduced in [14], on the avenues of machine learning and deep learning.
Definition 2.
Let σ > 0 and σ 0 0 . Then, the Generalized Gaussian Radial Basis Function Kernel for x and z R d is defined as:
K σ , σ 0 x z : = d e f g σ 2 r exp g σ 0 2 ( r ) 1 ( x z 2 ) = g σ 2 ( x z 2 ) e g σ 0 2 ( x z 2 ) 1 = e σ 2 x z 2 2 e e σ 0 2 x z 2 2 1 .
Now that we have introduced the Generalized Gaussian Radial Basis Function Kernel, let us briefly discuss the motivation of the present article with respect to it.

1.1. Motivation

The applicability of the Generalized Gaussian Radial Basis Function was introduced in [15] to provide better results in contrast to the Gaussian Radial Basis Function results; here, we specifically refer to the results of convergence and stability for interpolation problems on Franke’s test function and Runge’s function or solving the system with Tikhonov regularization and Riley’s algorithm. Drawing significant inspiration from [15], the application of the unexplored Generalized Gaussian Radial Basis Function Kernel was documented in [14], specifically in the contexts of SVMs, kernel regression, and pattern recognition, through the activation function in a neural network. These state-of-the-art methods in learning architectures are leveraged by a peculiar topic from the Hilbert function space called the reproducing kernel Hilbert space (RKHS) [16]. The analysis from the reproducing kernel theory perspective for the Gaussian Radial Basis Function Kernel has already been established by [17]; in the paper, an investigation related to the norms, orthonormal basis, and feature space was presented. However, with the present empirical evidence supporting better results obtained by employing the Generalized Gaussian Radial Basis Function Kernel, it becomes important to perform the same investigation for the Generalized Gaussian Radial Basis Function Kernel.
Note that, if σ 0 = 0 , then we obtain the traditional Gaussian Radial Basis Function Kernel. This is demonstrated in Figure 6.

1.2. Contributions

This article is motivated by documenting the application of the Generalized Gaussian Radial Basis Function Kernel on various machine learning and deep learning architectures. In the process of executing the Generalized Gaussian Radial Basis Function Kernel over learning architectures, we discover that the results produced by the Generalized Gaussian Radial Basis Function Kernel excel in terms of comparative accuracy or misclassification; these are achieved in contrast to the results obtained from the existing counterparts. These intriguing and promising findings naturally demand an exploration of the Generalized Gaussian Radial Basis Function from the perspective of mathematical function theory, which starts with the identification of the Hilbert space, generated by it as an L 2 measure. Furthermore, with the established Hilbert function space theory, we are also able to address the corresponding reproducing kernel theory by virtue of Aronszajn’s theorem (see [16]).
Having aptly addressed the Hilbert function theory generated by the Generalized Gaussian Radial Basis Function, this study immediately provides its application in learning architectures. The detailed investigation on the reproducing kernel for the Hilbert space of the Generalized Gaussian Radial Basis Function is proved to be universal (see in Section 4) on the restrictions over R d × R d ; this makes it ready-to-use for learning architecture problems, such as standard binary classification through SVMs, function approximation through kernel regression, and pattern recognition through deep convolutional neural networks. In particular, we determine that the results related to the Hilbert function theory of the Generalized Gaussian Radial Basis Function Kernel can be leveraged for SVM applications by addressing the hinge loss function L and minimal L risk losses as well.
Finally, this article also provides promising future directions in light of the spectral decomposition of the Generalized Gaussian Radial Basis Function.

1.3. Plan of the Article

The present article is organized as follows: we provide the essential preliminaries of reproducing kernel theory in Section 2. Then, we give the results from the Hilbert space function theory, such as the orthonormal basis and the reproducing kernel of the Generalized Gaussian Radial Basis Function Kernel in Section 3. Then, we provide the functionality of the reproducing kernel in a real sense in Section 4; this is followed by compelling empirical comparison results in Section 5. Lastly, we provide our conclusions and results in Section 6 and describe our future research directions in Section 7.

2. Notation and Preliminaries

2.1. Hypergeometric Function Notation

We begin first by recalling the basic calculus results related to the Generalized Hypergeometric Function    p F q a 1 a 2 a p b 1 b 2 b q ; z [18].
Definition 3.
We define the Pochhammer symbol [19] (rising factorial notation) as
a k = Γ ( a + k ) Γ ( a ) = a ( a + 1 ) ( a + k 1 ) .
Then, we define Generalized Hypergeometric Function   p F q a 1 a 2 a p b 1 b 2 b q ; z , which is given as
  p F q a 1 a 2 a p b 1 b 2 b q ; z : = d e f l = 0 i = 1 p a i l i = 1 p b i l z l l ! .
Example 1.
The summation l = 0 1 l + x n + 1 1 l ! can be represented as:
l = 0 1 l + x n + 1 1 l ! = l = 0 Γ ( l + x ) Γ l + x + 1 n + 1 1 l ! = l = 0 x l Γ ( x ) x + 1 l Γ ( x + 1 ) n + 1 1 l ! = l = 0 x l Γ ( x ) x + 1 l x Γ ( x ) n + 1 1 l ! = 1 x n + 1 l = 0 x l n + 1 x + 1 l n + 1 1 l ! = 1 x n + 1 n + 1 F n + 1 x x + 1 ; 1 ( u s e ( 2 ) ) .
The example presented above will be useful in addressing further details such as the orthonormal basis in the present manuscript. To avoid heavy notation clutter, we write
  n + 1 F n + 1 x x + 1 ; 1 : = n o t a t i o n F n , x , 1
from now onward where necessary. Note that F n , , 1 = e .

2.2. Field and Space Notations

The set of natural numbers in union with 0 is denoted by W , that is W : = 0 , 1 , 2 , . We use Kronecker delta δ n m on non-negative integers n and m to show that δ n m = 1 ; this is the case whenever n = m and δ n m = 0 , if n m . We denote a complex number, z = x + i y , where x and y R . With that z, its conjugate part is given as z ¯ = x i y , along with its absolute value as | z | 2 = z · z ¯ = x 2 + y 2 . We reserve symbol K to treat with a choice of fields on which we will operate; in particular, K can either be R or C .
If X denotes the space of input values (sometimes a closed subset of R d ), then we write C ( X ) as the space of continuous functions f : X R ; on the other hand, C m ( X ) represents the space of the m times differentiable function, including m = .

2.3. Tensor Product Notation

We recall the tensor product between two functions; say that f 1 , f 2 : X K is given as f 1 f 2 : X × X K . Then, for all x , x X , the tensor product f 1 f 2 is defined as f 1 f 2 ( x , x ) : = f 1 ( x ) f 2 ( x ) .

2.4. Preliminaries

Definition 4.
Let X = ; then, a function k : X × X K is referred to as the kernel on X if there exists a K Hilbert space H , · , · H accompanied by a map Φ : X H such that x , x X , we have
k ( x , x ) = Φ ( x ) , Φ ( x ) H .
We regard Φ as the feature map and H as the feature space of k.
Now that we have introduced the basic notion from the kernel theory in the definition provided above, we can now comfortably define the building block of this article: reproducing kernel Hilbert space—RKHS.
Definition 5.
Let X = and H , · , · H be the Hilbert function space over X.
1. 
The space H is referred to as the reproducing kernel Hilbert space (RKHS) if x X , the evaluation functional ε x : H K defined as ε x ( f ) : = f ( x ) , f H is continuous.
Definition 6.
A function k : X × X K is called reproducing kernel of H if we have:
1. 
k ( · , x ) H x X , that is k ( · , x ) H < , and
2. 
k ( · , · ) has the reproducing property; that is
f ( x ) = f , k ( · , x ) H f H a n d x X .
It is worthwhile mentioning that the norm convergence yields the point-wise convergence inside RKHS. This fact can be readily learned due to the continuity of the evaluation function. This is demonstrated as follows for an arbitrary f H and f n n H with f f n H 0 as n ; then,
lim n f n ( x ) = lim n ε x f n = ( continuity of ε x ) ε x f = f ( x ) .
Now, we will state an important theorem from [16], which dictates the relationship between the reproducing kernel of the RKHS H and the orthonormal basis of it.
Theorem 1
(Aronszajn’s Theorem [16]). Let H be an RKHS over an nonempty set X; then, k : X × X K , defined as k ( x , x ) : = ε x , ε x H for x , x X is the only reproducing kernel of H. Additionally, for some index set, I , if we have e i i I as an orthonormal basis, then for all x , x X , we have
k ( x , x ) = i I e i ( x ) e i ( x ) ¯ ,
with an absolute convergence.

3. Function Space of Generalized Gaussian Radial Basis Measure

Let d N , σ > 0 and σ 0 0 and f be holomorphic functions f : C d C ; we first write the measure of our interest:
d μ σ , σ 0 , d ( z ) : = e σ 2 | z | 2 e e σ 0 2 | z | 2 1 d V C d ( z ) .
Here, ‘ d V C d ( z ) ’ is the usual Lebesgue measure on the entire C d . For d = 1 , we write simply d μ σ , σ 0 ( z ) to denote the typical Lebesgue area measure on C . We now provide the inner product associated with this measure as:
f , g σ , σ 0 , C d : = N σ , σ 0 , d C d f ( z ) g ( z ) ¯ d μ σ , σ 0 , d ( z ) .
Here, ‘ N σ , σ 0 , d ’ is the normalization constant whose value is explicitly given as e σ 2 2 π d . Once we have define the inner product for the space, the norm for holomorphic function f : C d C is:
f σ , σ 0 , C d 2 : = e σ 2 2 π d C d | f ( z ) | 2 d μ σ , σ 0 , d ( z ) .
We write the following for the collection of holomorphic functions f : C d C , whose σ , σ 0 , C d is finite, that is:
H σ , σ 0 , C d : = holomorphic function f : C d C : f σ , σ 0 , C d < .
Once we have defined norm in (8) and the associated Hilbert space in (9), we can provide the following formulation, which makes the Hilbert space H σ , σ 0 , C d as an RKHS.
Theorem 2.
For all σ > 0 , σ 0 0 and all compact sets K C d , there exists a positive constant c σ , σ 0 , d such that for all z K and f H σ , σ 0 , C d , we have
| f ( z ) | c σ , σ 0 , d f σ , σ 0 , C d .
Proof. 
Denote B ( 0 , 1 ) as the complex unit ball in C . Define
c σ , σ 0 , d : = sup z K + B ( 0 , 1 ) d e σ 2 | z | 2 e e σ 0 2 | z | 2 1 .
In the spirit of ([17], Lemma-3, Page 4639), we have
j = 1 d r j | f ( z ) | 2 1 ( 2 π ) d j = 1 d r j [ 0 , 2 π ] d f z 1 + r 1 e i θ 1 , , z 1 + r d e i θ d 2 d θ .
Integration of the above with respect to r 1 , , r d [ 0 , 1 ] d yields:
| f ( z ) | 2 1 ( 2 π ) d z + B ( 0 , 1 ) d | f z | 2 d V ( z ) c σ , σ 0 , d ( 2 π ) d z + B ( 0 , 1 ) d | f z | 2 e σ 2 | z | 2 e e σ 0 2 | z | 2 1 d V ( z ) c σ , σ 0 , d ( e σ 2 ) d f σ , σ 0 , d 2 .
In particular, we have c σ , σ 0 , d = c σ , σ 0 , d e σ 2 d . Hence, the result is established.    □
Establishing Theorem 2 immediately yields that H σ , σ 0 , C d is indeed an RKHS and we state here as an important corollary.
Corollary 1.
The Hilbert space H σ , σ 0 , C d identified by the inner product · , · σ , σ 0 , C d is an RKHS for all σ > 0 and σ 0 .

Orthonormal Basis

We will need following technical result to establish the orthonormal basis for the RKHS H σ , σ 0 , C d .
Theorem 3.
For every σ > 0 , σ 0 0 and n , m W , we have
C z n z m ¯ d μ σ , σ 0 ( z ) = 2 π n ! e σ 2 n + 2 F n , σ ^ , 1 2 π m ! e σ 2 m + 2 F m , σ ^ , 1 δ n m
where σ ^ = σ 0 2 σ 2 and F n , σ ^ , 1 are defined in (3).
Proof. 
Employ the polar coordinate of z = r e i θ to have:
C z n z m ¯ d μ σ , σ 0 ( z ) = 0 r n + m e σ 2 r 2 e e σ 0 2 r 2 1 r d r 0 2 π e i n m θ d θ .
The quantity 0 r n + m e σ 2 r 2 e e σ 0 2 r 2 1 r d r 0 2 π e i n m θ d θ is 0 when n m . Now, assume that n = m in (12), then:
C z n z m ¯ d μ σ , σ 0 ( z ) = 2 π 0 r 2 n e σ 2 r 2 e e σ 0 2 r 2 1 r d r = 2 π e 0 t n e σ 2 t e e σ 0 2 t d t ( put t = r 2 ) = 2 π e 0 s σ 2 n e s e e σ 0 2 / σ 2 s d s σ 2 ( put s = σ 2 t ) = 2 π e σ 2 n + 1 0 s n e s e e σ 0 2 / σ 2 s d s = 2 π e σ 2 n + 1 1 0 log 1 y n e y σ 0 2 / σ 2 ( d y ) ( put e s = y ) (13) = 2 π e σ 2 n + 1 0 1 log 1 y n l = 0 y σ 0 2 / σ 2 l l ! d y (14) = 2 π e σ 2 n + 1 l = 0 1 l ! 0 1 log 1 y n y l σ ^ d y ( use σ ^ = σ 0 2 σ 2 ) (15) = 2 π e σ 2 n + 1 l = 0 1 l ! Γ ( n + 1 ) ( l σ ^ + 1 ) n + 1 = 2 π n ! e σ 2 n + 2 F n , σ ^ , 1 ( use ( 3 ) ) .
We used the entry 4.272(6)), Page 551 in [20] for the result in (15). The summation over l in (13) is interchanged with integral with respect to d y for 0 y 1 by virtue of Fubini–Tonelli’s theorem as 0 1 log 1 y n e y σ 0 2 / σ 2 d y is finite which is demonstrated as:
0 1 log 1 y n e y σ 0 2 / σ 2 d y 0 1 log 1 y n e y σ 0 2 / σ 2 d y e 0 1 log 1 y n d y = e Γ ( n + 1 ) < .
Thus, the result prevails.    □
We will employ the notation σ ^ = σ 0 2 σ 2 for σ > 0 and σ 0 0 wherever it is required. In the light of Theorem 1, we have to determine the orthonormal basis of H σ , σ 0 , C d .
Theorem 4.
Let σ > 0 , σ 0 0 and
e n ( z ) : = σ 2 n n ! F n , σ ^ , 1 z n
for some z C . Then, the tensor–product system e n 1 e n d n 1 , , n d 0 forms the orthonormal basis of H σ , σ 0 , C d .
Proof. 
We establish our result for d = 1 for an initial basic understanding. For this, let us show that e n n W forms an orthonormal system. So, consider z C and let m , n W . Then,
e n , e m σ , σ 0 = e σ 2 2 π C e n ( z ) e m ( z ) ¯ d μ σ , σ 0 ( z ) = e σ 2 2 π σ 2 n n ! F n , σ ^ , 1 σ 2 m m ! F m , σ ^ , 1 C z n z m ¯ d μ σ , σ 0 ( z ) = 1 if n = m 0 otherwise ( use Theorem 3 ) .
We use the ‘polarplot’ function available in MATLAB R2023b to construct the polar plot of e n ( z ) . These results are given in Figure 7, where the numbers 1 , 2 , and 3 are the radius values in the polar plot.
The above result concludes that e n n W is actually an orthonormal system. To this end, we have to establish that it is also complete. So, for this, pick a holomorphic function f H σ , σ 0 , C with f ( z ) = l = 0 a l z l and observe that
f , e n σ , σ 0 = e σ 2 2 π C f ( z ) e n ( z ) ¯ d μ σ , σ 0 ( z ) = e σ 2 2 π l = 0 a l C z l e n ( z ) ¯ d μ σ , σ 0 ( z ) = e σ 2 2 π σ 2 n n ! F n , σ ^ , 1 l = 0 a l C z l z n ¯ d μ σ , σ 0 ( z ) = e σ 2 2 π σ 2 n n ! F n , σ ^ , 1 l = 0 a l 2 π l ! e σ 2 l + 2 F n , σ ^ , 1 2 π n ! e σ 2 n + 2 F n , σ ^ , 1 δ l n = σ 2 n n ! F n , σ ^ , 1 a n n ! F n , σ ^ , 1 σ 2 n = σ 2 n n ! F n , σ ^ , 1 1 a n .
Since the constant is σ 2 n n ! F n , σ ^ , 1 0 for any choice of n, the condition that f , e n = 0 for all n W yields that a n = 0 for all n W , which results in the conclusion that f 0 . Therefore, e n n W is complete. Now, we establish these results in a d dimensional situation by employing the tensor product notation of Section 2.3. To this end, we see that
e n 1 e n d , e m 1 e m d σ , σ 0 , d = j = 1 d e n j , e m j σ , σ 0 .
Hence, the orthonormality of e n 1 e n d n 1 , n d , W d is established due to the orthonormality of each e n j , e m j σ , σ 0 . We still need to ensure that this d dimensional orthonormal system is complete. Now, observe
f , e n 1 e n d σ , σ 0 , d = e σ 2 2 π d C d f ( z ) e n 1 e n d z ¯ d μ σ , σ 0 , C d z = e σ 2 2 π d l 1 , , l d a l 1 , , l d I l , d ,
where I l , d = C d z l e n 1 e n d z ¯ d μ σ , σ 0 , C d z . We further can simplify I d , as follows:
I l , d = C d z l e n 1 ( z 1 ¯ ) e n d ( z d ¯ ) d μ σ , σ 0 ( z 1 ) d μ σ , σ 0 ( z d ) = j = 1 d C z j l j e n j ( z j ¯ ) d μ σ , σ 0 ( z j ) = j = 1 d C z j l j z j ¯ n j d μ σ , σ 0 ( z j ) = j = 1 d 2 π l j ! e σ 2 l j + 2 F l j , σ ^ , 1 2 π n j ! e σ 2 n j + 2 F n j , σ ^ , 1 δ l j n j a l 1 , , l d .
Finally,
e σ 2 2 π d l 1 , , l d a l 1 , , l d I l , d = j = 1 d σ 2 n j n j ! F n j , σ ^ , 1 1 a n 1 , , n d .
The further result for completeness in d dimension follows a routine procedure from single-dimension case, as already discussed before.    □
The following theorem provides the reproducing kernel for the RKHS H σ , σ 0 , C d defined in (9).
Theorem 5.
For σ > 0 , σ 0 0 and σ ^ = σ 2 σ 0 2 , the reproducing kernel for the RKHS H σ , σ 0 , C d is indeed (because Theorem 1) given as
K z , w : = n 1 , , n d = 0 λ n z w ¯ n ,
where multi-index notation is employed: n = n 1 , , n d and λ n = i = 1 d σ 2 n i n i ! F n i , σ ^ , 1 .
Proof. 
We will demonstrate the desired proof as follows:
  • For w C d , we will show that K · , w σ , σ 0 , d < . For this, consider that
    K · , w σ , σ 0 , d 2 = e σ 2 2 π d C d | K z , w | 2 d μ σ , σ 0 , d z (18) = e σ 2 2 π d n 1 , , n d m 1 , , m d λ n λ m w n w ¯ m C d z n z ¯ m d μ σ , σ 0 , d z (19) = e σ 2 2 π d n 1 , , n d λ n 2 | w | 2 n n i = 1 d C z i n i z i ¯ m i d μ σ , σ 0 ( z i ) = e σ 2 2 π d n 1 , , n d λ n 2 | w | 2 n n i = 1 d 2 π n i ! σ 2 n i + 2 F n i , σ ^ , 1 = n 1 , , n d | w | 2 n i = 1 d n i ! σ 2 n i F n i , σ ^ , 1 .
    We used the result of Theorem 4 from (18) to (19). For all w C d , the quantity n 1 , , n d | w | 2 n i = 1 d n i ! σ 2 n i F n i , σ ^ , 1 achieves convergence. This implies that K · , w σ , σ 0 , d < for all w C d . Therefore, the kernel function is K ( · , w ) H σ , σ 0 , C d .
  • In order to establish the reproducing property of K ( · , w ) , pick an arbitrary f = n 1 , , n d a n 1 , , n d w n H σ , σ 0 , C d . Then, consider the inner product of f with K ( · , w ) , as follows:
    f , K ( · , w ) σ , σ 0 , C d = e σ 2 2 π d C d f ( z ) K z , w ¯ d μ σ , σ 0 , C d z = e σ 2 2 π d n 1 , , n d l 1 , , l d a n 1 , , n d λ l 1 , , l d w n i = 1 d 2 π e σ 2 σ 2 n i n i ! F n i , σ ^ , 1 = n 1 , , n d = 0 a n 1 , , n d w n = f w .
Hence, the desired result is achieved.    □
The proof in the preceding theorem to demonstrate reproducing kernel nature of K z , w of RKHS H σ , σ 0 , C d utilizes the basic machinery borrowed from the two-part definition for the reproducing kernel given in Definition 6. Now that we have determined the reproducing kernel theory for the Generalized Gaussian Radial Basis Function Kernel, we will investigate their functionality in terms of universality, which makes them applicable in addressing the learning architecture problems, such as standard binary classification through SVMs, function approximation through kernel regression, and pattern recognition through deep convolutional neural networks.

4. Restriction and Universality of Reproducing Kernel of RKHS H σ , σ 0 , C d

It is crucial to address the fact that the restriction of the reproducing kernel K z , w on R d × R d is also a kernel function, which eventually will assist us in establishing the universal nature of the K z , w .
Lemma 1
(Restriction of kernels [21]). Let k be a kernel on X, let X ˜ be a set, and let A : X ˜ X be a map. Then, k ˜ defined by k ˜ ( x , x ) : = k A ( x ) , A ( x ˜ ) , x , x ˜ X , which is a kernel on X ˜ . In particular, if X ˜ X , then k | X ˜ × X ˜ is a kernel.
In other words, the conclusion of the above lemma is given for all x , x R d ; if k ( x , x ) R , then k is also a kernel in the real sense (cf. [21]). This notion of restricting the kernel from C to R can allow the restriction of K z , w for the RKHS H σ , σ 0 , C d on R d × R d to have kernel in a real sense. Note that the λ n s present in the reproducing kernel K ( z , w ) in (17) are strictly positive real numbers for all n W ; also, observe that for all x , x R d , we have
K | R d × R d ( x , x ) = n 1 , , n d = 0 λ n x x ¯ n = n 1 , , n d = 0 λ n x x n R ,
which implies that the restriction of K ( · , · ) on R d × R d is a kernel. Now, we shall recall the notion of universal kernel.
Definition 7
(Universal Kernels [21]). A continuous kernel k on a compact metric space X is called universal if the RKHS H of k is dense in C ( X ) . This means that, for every function g C ( X ) and all ϵ > 0 , there exists f H , such that f g ϵ .
Now that we have defined the notion of universal kernel function in the above definition, we provide an example of the same.
Example 2
(Universal Taylor Kernels). For a fixed R ( 0 , ] and a C function f : ( R , R ) R that can be expanded into its Taylor series at 0, i.e.,
f ( t ) = n = 0 a n t n t ( R , R ) .
Let X x R d : x 2 < R . If we have a n > 0 for all n 0 in (21), then k given by
k ( x , x ) : = f x , x , x , x X ,
is a universal kernel on every compact subset of X.
With the help of universal Taylor kernels, as given above, we can see that the reproducing kernel of the RKHS H σ , σ 0 , C d is a universal Taylor kernel, followed by its restriction on R d × R d . This motivating result is captured in the following important theorem.
Theorem 6.
Let X be a compact subset of R d , σ > 0 and σ 0 . Then, the kernel K | R d × R d defined in (20) is universal.
Proof. 
Let σ > 0 and σ 0 0 . Recall a n from (21) and λ n from (17) and set both of them equal to each other, that is, a n = λ n for all n W . With these choices of a n for all n W , we now define a C function f : R , R R as
f ( t ) = n = 0 λ n t n t ( R , R ) .
We learn that, for x , x R d , whose Euclidean norm is bounded by R , we can express K | R d × R d x , x in terms of f as defined in (23) as:
K | R d × R d x , x = n = 0 λ n x x n = n = 0 λ n x , x n = f x , x
The establishment of K | R d × R d x , x for x , x R d in terms of C function f and in combination of λ n > 0 for all n W makes K | R d × R d a universal kernel by the virtue of Example 2.    □
The establishment of the universal nature of the reproducing kernel K ( · , · ) of the RKHS H σ , σ 0 , C d with restriction on R d × R d in Theorem 6 allows one to approximate arbitrary continuous functions on compact subsets of R d with those functions that can be represented against the reproducing kernel K ( · , · ) to the inner product, as defined in (7) or consequently whose norm given in (8) is finite.

5. Empirical Evidence and Results Comparison

We present the empirical evidence for the application of the Generalized Gaussian Radial Basis Function Kernel on various basic yet important learning architecture routines. The technical details of every experiment presented here are provided in Section 6.

5.1. Kernel Regression

We know that kernel regression aims to approximate or interpolate arbitrary continuous functions through the available choices of universal kernel functions on compact subsets of R d . Usually, these functions are radial basis functions that are employed in the kernel sense to achieve the approximation. In the current example given as follows, we leverage the universal nature of both Gaussian Radial Basis Function Kernel (cf. Corollary 4.58 in [21]) and Generalized Gaussian Radial Basis Function Kernel from Theorem 6.

5.1.1. Example 1

Kernel regression of f ( x ) = e ( 1 9 x 2 ) 4 + tan x + x 1 6 + sin x ϑ ( n ) is performed via both GRBF and (1) and the respective results are given below in Figure 8. Here, ϑ ( n ) is uniform random distributed number in ( 0 , 1 ) , x = n Z 100 .

5.1.2. Example 2

Kernel regression of f ( x ) = e sin x sin x 2 + 2 π | x + cos ϑ ( n ) | is performed via both GRBF and (1) and the respective results are given below in Figure 9. Here, ϑ ( n ) is uniform random distributed number in ( 0 , 1 ) , x = n Z 100 .

5.2. Support Vector Machine

Let us briefly address how the results derived in Section 3 can be used to understand the analysis of SVM. To that end, we define an important notion in SVMs referred to as hinge loss function, denoted by L ( · , · ) as:
L ( y , t ) : = max y Y : = 1 , 1 0 , 1 y t t R .
Let X R d be a nonempty open subset and P be a probability measure on X × Y ; then, the risk induced by the hinge loss function is L risk, which is defined with help of f : X R measurable function as:
R L , P f : = X × Y L y , f ( x ) d P ( x , y ) .
Additionally, the minimal L risk is given as
R L , P * : = inf f : X R R L , P f .
It is important to recall here that the RKHSs of universal kernels approximate the minimal L risk losses. It is performed in the following corollary (cf. Corollary 5.29 in [21]).
Corollary 2.
Let X be a compact metric space, H be the RKHS of a universal kernel on X, P be a distribution on X × Y , and L : X × Y × R [ 0 , ) be a continuous P integrable loss. Then, we have inf f H R L , P ( f ) = R L , P * .
Now, consider f H σ , σ 0 , C d ; then, let the restriction of f on R d be denoted by f | R d ; then, the aforementioned result forces us to establish the relation
inf f | R d H σ , σ 0 , C d R L , P ( f ) = R L , P * .
The setup presented in (28) enables to have consistent results concerning the classifiers of SVMs, which simply means that the classifiers that are based on minimization of a regularized risk can asymptotically learn in every classification task [22]. With this theory of SVM around the RKHS H σ , σ 0 , C d , we are now ready to ensure the application of the Generalized Gaussian Radial Basis Function Kernel in SVMs.
SVMs are implemented for the standard binary data classification (cf. [23]) data generated in top left of Figure 10; this is performed via a different choices of kernels; these kernels include the traditional Sigmoid function, the Gaussian Radial Basis Function Kernel, and the Generalized Gaussian Radial Basis Function Kernel in (1).
Of the available three kernels—the Generalized Gaussian Radial Basis Function in (1), the Sigmoid, and the Gaussian Radial Basis Function Kernel—the Generalized Gaussian Radial Basis Function Kernel yields the lowest misclassification for the data of 101 sampled points.

5.3. Neural Network

The work in this section revolves around the application of universal approximation theorem (adapted from [24]) of Cybenko-1989 [25], Hornik-1991 [26], and Pinkus-1999 [27], as follows:
Theorem 7
(Universal Approximation Theorem in Neural Network). Let ρ : R R be any continuous function. Let N d ρ represent the class of feed-forward neural networks with activation function ρ with d neurons in the input layer, one neuron in the output layer, and one hidden layer with an arbitrary number of neurons. Let X R d be compact. Then, N d ρ is dense in C ( X ) if and only if ρ is non-polynomial.
Now, under the application of the universal approximation theorem, one can have ρ ( r ) = g σ 2 ( r ) exp g σ 0 2 ( r ) 1 in Theorem 7; then, we realize that the Generalized Gaussian Radial Basis Function is an activation function, as it is a non-polynomial function. In the pursuit of an application of the Generalized Gaussian Radial Basis Function, we have following examples that provide the evidence of the optimal results that are yielded by the Generalized Gaussian Radial Basis Function in (1) employment both as in the usage of the activation function and the neural net layer in deep convolution neural nets (DCNNs).

5.3.1. Activation Function

Consider an α ReLU activation function, a simple modification of the powerful traditional ReLU activation function from NN (AF-NN), defined as
f ( x ) x if x > 0 α x if x 0 .
In the present experiment, two seven-layered NNs are constructed: one with the activation function defined by (29) and one with Generalized Gaussian Radial Basis Function in (1). Thereafter, the following performance Table 1 and Table 2 yield the respective results and comparison for the NNs.
Furthermore, the training progress data for the individual neural network architectures are presented next. The legends for accuracy and loss are same for both of the neural network architectures and their respective information is present in between the Figure 11 and Figure 12.

5.3.2. DCNN

The above experiments demonstrate the concrete functionality of the Generalized Gaussian Radial Basis Function in (1) as an activation function; this finding further motivates the performance of the deep convolution neural net. Therefore, in pursuit of this, a typical 7-layered deep convolution neural net is constructed; it has an activation function defined by (29) and one with (1). In the following experiment given in Figure 13 and Figure 14, the training data contain 1500– 28 × 28 gray-scale letter images of A, B, and C in a 4-D array.

6. Results and Technical Details

Now that we have successfully constructed the Hilbert function space theory around the Generalized Gaussian Radial Basis Function, when it is viewed as the L 2 measure, we immediately employ it in various basic artificial intelligence learning architectures. The learning architectures, namely kernel regression, SVM, and DCNN, that are considered here are quite basic yet well-versed in the field of machine learning and artificial intelligence. These can be regarded as the building blocks of several other artificial intelligence routines. Following Table 3 is essentially a compilation of the results documented for the various experiments we have discussed so far in comparison with the Generalized Gaussian Radial Basis Function Kernel as opposed to its available counterpart(s), where none of hyper-parameters were optimized for the available kernel functions.
The empirical evidence recorded here simply directs us to conclude that the results prompted by the Generalized Gaussian Radial Basis Function outperform every considered learning architecture routine. A detailed discussion of the empirical evidence collected is presented here:
  • We considered two experiments for the kernel regression of the functions mentioned in Figure 8 and Figure 9. These functions, as easily can be seen, are not well-behaved functions, since they involve transcendental behavior along with function discontinuities (due to tan x or x 1 / 6 , etc.) followed by uniform randomness due to the ϑ ( n ) of 101 random points. In both of the experiments of this nature, we record the minimum error obtained by employing either the Gaussian Radial Basis Function Kernel or the Generalized Gaussian Radial Basis Function Kernel. From this experiment, we learn that the minimum error recorded in both of them is remarkably small, in the order of ∼ 10 4 , if we use the Generalized Gaussian Radial Basis Function Kernel in contrast to the Gaussian Radial Basis Function Kernel, which provides a minimum error in the order of ∼ 10 3 .
  • Next, we consider the experiment of support vector machines for binary classification, in which we generated a random set of points (one can think of them as data) within the unit circle. We identified the random points spread across the circle with the positive class if it lies either in the first or the third quadrant; otherwise, we consider it to be within the negative class. After we generate and identify these datasets, we then identify the support vectors and the decision boundary by employing three (custom) kernel functions over MATLAB R2023b: Gaussian Radial Basis Function Kernel, Sigmoid Kernel, and Generalized Gaussian Radial Basis Function Kernel. We then register the out-of-sample misclassification rate by using 10-fold cross-validation from the aforementioned kernel functions. Here, we immediately learn that the misclassification yielded by Gaussian Radial Basis Function Kernel is comparatively high among all three employed kernel functions; the Generalized Gaussian Radial Basis Function Kernel yields the lowest misclassification rate: upward of 96 % .
  • For the activation function application in the domain of neural networks, we consider a practical example of pattern recognition for the synthetic hand-written English letter dataset in the form of a 4 D array (lettersTrainSet in MATLAB). As already mentioned in the respective subsection, two 7 layered neural nets were constructed whose layer names are given as follows:
    (a)
    Image input;
    (b)
    2D convolution;
    (c)
    Batch normalization;
    (d)
    (29) vs. Generalized Gaussian Radial Basis Function;
    (e)
    Ten fully connected layers;
    (f)
    Softmax;
    (g)
    Classification layer.
    It should be noted that, from the above 7 layered neural network constructed, both of the neural network training experiments came to a complete stop when the maximum epoch of 30 amounting to 330 iterations were finished. Additionally, the sequence of inserting the neural net layer was the same as that which was given above from (a)–(g). Once the desired neural net is set with both of the activation functions, we immediately learn that the neural net containing the activation function of the Generalized Gaussian Radial Basis Function Kernel as one of its layers simply outperforms its opponent, i.e., the neural net layer containing the activation function of (29).
  • With the establishment of the basics of the neural network architecture in the previous listing point, we train and execute a deep convolution neural net for the representative images of the alphabet characters A, B, and C. A list of the seven layers involved in the construction of a deep convolution neural network is presented as follows:
    (a)
    Image input;
    (b)
    2D convolution;
    (c)
    (29) vs. Generalized Gaussian Radial Basis Function;
    (d)
    A 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0];
    (e)
    Three fully connected layers;
    (f)
    Softmax;
    (g)
    Classification layer.
    Again, the sequence of inserting the neural net layers was the same as what is given above from (a)–(g) to construct the present DCNN. Similarly, the result for the successful classification performances by the DCNN which contains the activation layer of Generalized Gaussian Radial Basis Function Kernel achieves an accuracy upward of 96 % . However, we fail to see the match-able competitiveness by the DCNN, which contains the (29) as the activation layer.

7. Future Directions

Eigen Function Expansion of GGRBF

We recall the Mercer’s theorem from ([12], Theorem 4.2, Page 96).
Theorem 8
(Mercer’s Theorem). Let X , μ be a finite measure space and k L X 2 , μ 2 be a kernel, such that T k : L 2 X , μ L 2 X , μ is a definite positive. Let ϕ i i L 2 X , μ be the normalized eigenfunctions of T k associated with the eigenvalues Λ i i . Then,
1. 
the eigenvalues Λ i are absolutely summable
2. 
k x , x = i = 0 Λ i ϕ i x ϕ i x * holds μ 2 almost everywhere, where the series converges absolutely and uniformly μ 2 almost everywhere.
Example 3.
With the application of Theorem 8, we can provide the eigen function decomposition of K σ x , z = e σ 2 x z 2 for x , z R , that is:
e σ 2 x z 2 = i = 0 Λ i ϕ i x ϕ i z ,
Λ i = α σ 2 i α 2 2 1 + 1 + 2 σ α 2 + σ 2 2 i + 1 / 2 ,
ϕ i x = 1 + 1 + 2 σ α 2 8 2 i i ! e 1 + 2 σ α 2 1 α 2 x 2 2 H i 1 + 2 σ α 2 4 α x .
The expression H i i in (32) are the Hermite polynomials which are L 2 orthonormal against the weight α π e α 2 x 2 , that is:
R ϕ n ( x ) ϕ m ( x ) α π e α 2 x 2 d x = δ n m .
We have the graphical representation of first seven Hermite polynomials in Figure 15a,b.
Based on the Hermite polynomials [28] and as its application for the eigen decomposition analysis for the Gaussian Radial Basis Function Kernel, we have the following promising future research direction. In the spirit of the application of Mercer’s theorem, we know the eigen function decomposition of the Gaussian Radial Basis Function Kernel. However, presently, we are not fortunate enough to have such a decomposition for the Generalized Gaussian Radial Basis Function Kernel. A preliminary investigation towards the desired eigen function decomposition of the Generalized Gaussian Radial Basis Function Kernel (from [29,30,31]) directs us toward the incorporation of a new variety of functions, as defined in (34).
H n ( x ) : = ( 1 ) n e a x 2 e e b x 2 + 1 d n d x n e a x 2 e e b x 2 1
into our desired eigen function decomposition analysis. Here, a > 0 and b 0 ; therefore, if a = 1 & b = 0 , then H i = H i . The first two expressions for H n ( x ) are explicitly given as:
H 1 ( x ) = 2 a x + 2 b x e b x 2
H 2 ( x ) = 2 a + 4 b 2 x 2 e b x 2 2 b e b x 2 + 2 a x + 2 b x e b x 2 2 .
The real constants a and b present in (34) corresponds to the respective constants present in (1) as a = σ 2 and b = σ 0 2 . Next, see the graphical presentation of the first seven functions from the family defined in (34) in Figure 16 and Figure 17.
Future Direction 1.
We remain in a position lacking in knowledge of eigenvalues for the Generalized Gaussian Radial Basis Function Kernel. Additionally, we still need to investigate whether the function introduced in (34) is an orthonormal in the sense that the traditional Hermite polynomials are. Therefore, it will be interesting to understand the analysis of the function given in (34).

8. Discussion

In the present manuscript, we explore and introduce the Hilbert function space theory; in particular, the RKHS was generated by the Generalized Gaussian Radial Basis Function g σ 2 ( r ) exp g σ 0 2 ( r ) 1 as an L 2 measure, followed by an interesting application of it in the learning architectures of machine learning and deep learning. Throughout this article, we have learned from practical results that it is indeed beneficial to employ the Generalized Gaussian Radial Basis Function Kernel; here, one might achieve phenomenal results compared to those derived using its corresponding counterparts. Even though it is quite advantageous to use the Generalized Gaussian Radial Basis Function Kernel in learning architectures, we also encounter a few shortcomings in this endeavor. Oftentimes, the result yielded by the Generalized Gaussian Radial Basis Function is similar to the results that one might obtain using the preexisting and well-studied activation functions, such as Gaussian error linear unit (GELU) (cf. [32]), whose operation on x is given as
GELU ( x ) : = x 2 1 + erf x x 2 .
For instance, the comparison of the results based on choice of either Generalized Gaussian Radial Function Kernel or Gaussian error linear unit as an activation function is given in Table 4 and Table 5, respectively. We see that the results from both of the cases are comparable in terms of acquiring an accuracy rate. The key point to note here is that the Generalized Gaussian Radial Basis Function Kernel is a brand-new novel mathematical function to operate on; hence, it demands a detailed, in-depth investigation, which is initiated by its RKHS in this article, before it can be exploited to its full potential on the venues of artificial intelligence. Out of the other few major concerns in connection with the Generalized Gaussian Radial Basis Function Kernel is its mathematical structure, i.e., the presence of an extra exponential of an exponential of an argument. Hence, the simple presence of this can often slow the training process of learning architectures in which the Generalized Gaussian Radial Basis Function Kernel is employed; therefore, this might pose a challenge in attaining the desired results.

Funding

This research was funded by The Office of Dean, the Office of Research, Scholarship, and Sponsored Programs, Robert R. Muntz Library at The University of Texas at Tyler, TX-75799, USA.

Data Availability Statement

The coding framework presented in this article was executed on MATLAB R2023bat (https://www.mathworks.com/products/matlab.html, accessed on 15 February 2024).

Acknowledgments

The author would like to thank the reviewer for their precious time in reviewing the present manuscript. The author acknowledges the support of Drishty Singh, 4th Year MSc, Department of Mathematics, Babasaheb Bhimrao Ambedkar University, Lucknow, Uttar Pradesh, India. She provided the explicit expression of the first two functions in (35) and (36) from (34). Further analysis of function in (34) was extended due to these important results. The author wishes to thank the referees for their valuable comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Baddoo, P.J.; Herrmann, B.; McKeon, B.J.; Brunton, S.L. Kernel learning for robust Dynamic Mode Decomposition: Linear and Nonlinear disambiguation optimization. Proc. R. Soc. A 2022, 478, 20210830. [Google Scholar] [CrossRef] [PubMed]
  2. Gianola, D.; Van Kaam, J.B. Reproducing Kernel Hilbert spaces regression methods for genomic assisted prediction of quantitative traits. Genetics 2008, 178, 2289–2303. [Google Scholar] [CrossRef] [PubMed]
  3. Attia, N.; Akgül, A.; Seba, D.; Nour, A. Reproducing kernel Hilbert space method for the numerical solutions of fractional cancer tumor models. Math. Methods Appl. Sci. 2023, 46, 7632–7653. [Google Scholar] [CrossRef]
  4. Mathematical Functions Power Artificial Intelligence. Available online: https://nap.nationalacademies.org/resource/other/deps/illustrating-math/interactive/mathematical-functions-power-ai.html (accessed on 15 February 2024).
  5. Mathematics and Statistics of Weather Forecasting. Available online: https://nap.nationalacademies.org/resource/other/deps/illustrating-math/interactive/mathematics-and-statistics-of-weather-forecasting.html (accessed on 15 February 2024).
  6. Kalman, B.L.; Kwasny, S.C. Why tanh: Choosing a sigmoidal function. In Proceedings of the 1992 IJCNN International Joint Conference on Neural Networks, Baltimore, MD, USA, 7–11 June 1992; Volume 4, pp. 578–581. [Google Scholar]
  7. Lu, L.; Shin, Y.; Su, Y.; Karniadakis, G.E. Dying ReLU and initialization: Theory and numerical examples. arXiv 2019, arXiv:1903.06733. [Google Scholar] [CrossRef]
  8. Fasshauer, G.E. Meshfree Approximation Methods with MATLAB; World Scientific: Singapore, 2007; Volume 6. [Google Scholar]
  9. Brunton, S.L.; Kutz, J.N. Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control; Cambridge University Press: Cambridge, UK, 2022. [Google Scholar]
  10. Stein, M.L. Interpolation of Spatial Data: Some Theory for Kriging; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1999. [Google Scholar]
  11. Rosenfeld, J.A.; Russo, B.; Kamalapurkar, R.; Johnson, T.T. The occupation kernel method for nonlinear system identification. arXiv 2019, arXiv:1909.11792. [Google Scholar]
  12. Williams, C.K.; Rasmussen, C.E. Gaussian Processes for Machine Learning; MIT Press: Cambridge, MA, USA, 2006; Volume 2. [Google Scholar]
  13. Park, J.; Sandberg, I.W. Universal approximation using radial-basis-function networks. Neural Comput. 1991, 3, 246–257. [Google Scholar] [CrossRef] [PubMed]
  14. Singh, H. A new kernel function for better AI methods. In Proceedings of the 2023 Spring Eastern Sectional Meeting, Virtual, 1–2 April 2023. [Google Scholar]
  15. Karimi, N.; Kazem, S.; Ahmadian, D.; Adibi, H.; Ballestra, L. On a generalized Gaussian radial basis function: Analysis and applications. Eng. Anal. Bound. Elem. 2020, 112, 46–57. [Google Scholar] [CrossRef]
  16. Aronszajn, N. Theory of reproducing kernels. Trans. Am. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
  17. Steinwart, I.; Hush, D.; Scovel, C. An explicit description of the reproducing kernel Hilbert spaces of Gaussian RBF kernels. IEEE Trans. Inf. Theory 2006, 52, 4635–4643. [Google Scholar] [CrossRef]
  18. Barnes, E.W.V. The asymptotic expansion of integral functions defined by Taylor’s series. Philos. Trans. R. Soc. London. Ser. A Contain. Pap. A Math. Phys. Character 1906, 206, 249–297. [Google Scholar]
  19. Abramowitz, M.; Stegun, I.A. Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables; US Government Printing Office: Washington, DC, USA, 1968; Volume 55. [Google Scholar]
  20. Gradshteyn, I.S.; Ryzhik, I.M. Table of Integrals, Series, and Products, 7th ed.; Academic Press: Cambridgem MA, USA; Elsevier: Amsterdam, The Netherlands, 2007. [Google Scholar]
  21. Christmann, A.; Steinwart, I. Support Vector Machines; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  22. Steinwart, I. Consistency of support vector machines and other regularized kernel classifiers. IEEE Trans. Inf. Theory 2005, 51, 128–142. [Google Scholar] [CrossRef]
  23. Devroye, L.; Györfi, L.; Lugosi, G. A Probabilistic Theory of Pattern Recognition; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 31. [Google Scholar]
  24. Kidger, P.; Lyons, T. Universal approximation with deep narrow networks. In Proceedings of the Conference on Learning Theory, Graz, Austria, 9–12 July 2020; Proceedings of Machine Learning Research. pp. 2306–2327. [Google Scholar]
  25. Cybenko, G. Approximation by Superpositions of a Sigmoidal Function. Math. Control. Signals Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  26. Hornik, K. Approximation capabilities of multilayer feedforward networks. Neural Netw. 1991, 4, 251–257. [Google Scholar] [CrossRef]
  27. Pinkus, A. Approximation theory of the MLP model in neural networks. Acta Numer. 1999, 8, 143–195. [Google Scholar]
  28. Hermite, M. Sur un Nouveau Développement en Série des Fonctions; Imprimerie de Gauthier-Villars: 1864; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  29. Fasshauer, G.E.; McCourt, M.J. Stable evaluation of Gaussian radial basis function interpolants. Siam J. Sci. Comput. 2012, 34, A737–A762. [Google Scholar] [CrossRef]
  30. Rasmussen, C.E.; Williams, C.K. Gaussian Processes for Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; Volume 1. [Google Scholar]
  31. Zhu, H.; Williams, C.K.; Rohwer, R.; Morciniec, M. Gaussian Regression and Optimal Finite Dimensional Linear Models; Neural Networks and Machine Learning; Springer: Berlin/Heidelberg, Germany, 1997. [Google Scholar]
  32. Hendrycks, D.; Gimpel, K. Gaussian Error Linear Units (Gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
Figure 1. Graph of Sigmoid activation function.
Figure 1. Graph of Sigmoid activation function.
Mathematics 12 00829 g001
Figure 2. Graph of ReLU activation function.
Figure 2. Graph of ReLU activation function.
Mathematics 12 00829 g002
Figure 3. Graphs of radial basis functions of exponential types.
Figure 3. Graphs of radial basis functions of exponential types.
Mathematics 12 00829 g003
Figure 4. GRBF kernel.
Figure 4. GRBF kernel.
Mathematics 12 00829 g004
Figure 5. GGRBF kernel with σ = 1 2 and σ 0 = 1 0.98 .
Figure 5. GGRBF kernel with σ = 1 2 and σ 0 = 1 0.98 .
Mathematics 12 00829 g005
Figure 6. GGRBF kernel with σ = 1 2 + π and σ 0 = 0 .
Figure 6. GGRBF kernel with σ = 1 2 + π and σ 0 = 0 .
Mathematics 12 00829 g006
Figure 7. Polar plot of orthonormal basis e n ( z ) : = σ 2 n n ! F n , σ ^ , 1 z n given in (16).
Figure 7. Polar plot of orthonormal basis e n ( z ) : = σ 2 n n ! F n , σ ^ , 1 z n given in (16).
Mathematics 12 00829 g007
Figure 8. Kernel regression of f ( x ) = e ( 1 9 x 2 ) 4 + tan x + x 1 6 + sin x ϑ ( n ) with GRBF (L) and GGRBF (R).
Figure 8. Kernel regression of f ( x ) = e ( 1 9 x 2 ) 4 + tan x + x 1 6 + sin x ϑ ( n ) with GRBF (L) and GGRBF (R).
Mathematics 12 00829 g008
Figure 9. Kernel regression of f ( x ) = e sin x sin x 2 + 2 π | x + cos ϑ ( n ) | with GRBF (L) and GGRBF (R).
Figure 9. Kernel regression of f ( x ) = e sin x sin x 2 + 2 π | x + cos ϑ ( n ) | with GRBF (L) and GGRBF (R).
Mathematics 12 00829 g009
Figure 10. SVM classifier (standard binary classification) via different kernels, namely (respectively) GGRBF (in top right), Sigmoid (in bottom left) and GRBF (in bottom right).
Figure 10. SVM classifier (standard binary classification) via different kernels, namely (respectively) GGRBF (in top right), Sigmoid (in bottom left) and GRBF (in bottom right).
Mathematics 12 00829 g010
Figure 11. Training progress map for NN architecture with GGRBF as AF.
Figure 11. Training progress map for NN architecture with GGRBF as AF.
Mathematics 12 00829 g011
Figure 12. Training progress map for NN architecture with (29) as AF.
Figure 12. Training progress map for NN architecture with (29) as AF.
Mathematics 12 00829 g012
Figure 13. Accuracy of 96.33 % registered with GGRBF in (1) as a DCNN neural layer.
Figure 13. Accuracy of 96.33 % registered with GGRBF in (1) as a DCNN neural layer.
Mathematics 12 00829 g013
Figure 14. Accuracy of 91.87 % registered with (29) as a DCNN neural layer.
Figure 14. Accuracy of 91.87 % registered with (29) as a DCNN neural layer.
Mathematics 12 00829 g014
Figure 15. First seven Hermite polynomials. (a) Hermite polynomials H i ( x ) for i = 0 , 1 , 2 ; (b) Hermite polynomials H i ( x ) for i = 3 , 4 , 5 , 6 .
Figure 15. First seven Hermite polynomials. (a) Hermite polynomials H i ( x ) for i = 0 , 1 , 2 ; (b) Hermite polynomials H i ( x ) for i = 3 , 4 , 5 , 6 .
Mathematics 12 00829 g015
Figure 16. Graph of H n ( x ) in (34) for n = 0 , 1 , 2 with a = 0.091 & b = 0.81 .
Figure 16. Graph of H n ( x ) in (34) for n = 0 , 1 , 2 with a = 0.091 & b = 0.81 .
Mathematics 12 00829 g016
Figure 17. Graph of H n ( x ) in (34) for n = 3 , 4 , 5 , 6 with a = 0.091 & b = 0.81 .
Figure 17. Graph of H n ( x ) in (34) for n = 3 , 4 , 5 , 6 with a = 0.091 & b = 0.81 .
Mathematics 12 00829 g017
Table 1. Performance summary for GGRBF with an accuracy of 97.88 % .
Table 1. Performance summary for GGRBF with an accuracy of 97.88 % .
Training with GGRBF (on Single CPU)
EpochIterationRuntime (hh:mm:ss)Accuracy (%)Batch LossLearning Rate
1100:00:048.592.53290.0010
25000:00:1075.781.00870.0010
310000:00:1593.750.43770.0010
415000:00:2199.220.24520.0010
620000:00:2699.220.17900.0010
725000:00:31100.000.10280.0010
830000:00:36100.000.06810.0010
935000:00:41100.000.04580.0010
1039000:00:45100.000.03820.0010
Table 2. Performance summary for (29) with accuracy of 93.48 % .
Table 2. Performance summary for (29) with accuracy of 93.48 % .
Training with (29) (on Single CPU)
EpochIterationRuntime (hh:mm:ss)Accuracy (%)Batch LossLearning Rate
1100:00:088.592.99760.0010
25000:00:1578.910.60530.0010
310000:00:1986.720.41080.0010
415000:00:2590.620.31220.0010
620000:00:3096.880.16320.0010
725000:00:3699.220.06950.0010
830000:00:4199.220.07780.0010
935000:00:46100.000.03900.0010
1039000:00:49100.000.03550.0010
Table 3. Compilation of technical results.
Table 3. Compilation of technical results.
AI Learning ArchitectureActivation FunctionReferenceMin. ErrorMisclass. %Accuracy %Runtime (s)
Kernel RegressionGRBFFigure 80.00230---
Kernel RegressionGGRBFFigure 8 0.00097 ---
Kernel RegressionGRBFFigure 90.00100---
Kernel RegressionGGRBFFigure 9 0.00043 ---
SVMGRBFFigure 10-5.7594.25-
SVMSigmoidFigure 10-4.5095.50-
SVMGGRBFFigure 10- 3.75 96.25 -
AF in NNGGRBFTable 1- 2.12 97.88 0.43077
AF in NN α ReLUTable 2-6.5293.481.8935
DCNNGGRBFFigure 13- 3.67 96.33 -
DCNN α ReLUFigure 14-8.1391.87-
Table 4. Performance summary for GGRBF in comparison with (37) with accuracy of 98.18 % .
Table 4. Performance summary for GGRBF in comparison with (37) with accuracy of 98.18 % .
Training with GGRBF (on a Single CPU)
EpochIterationRuntime (hh:mm:ss)Accuracy (%)Batch LossLearning Rate
1100:00:0011.722.53290.0010
25000:00:0472.661.00870.0010
310000:00:0984.380.43770.0010
415000:00:1389.060.24520.0010
620000:00:1898.440.17900.0010
725000:00:23100.000.10280.0010
830000:00:27100.000.06810.0010
935000:00:32100.000.04580.0010
1140000:00:36100.000.03820.0010
1245000:00:41100.000.03820.0010
1350000:00:45100.000.03820.0010
1454600:00:50100.000.03820.0010
Table 5. Performance summary for (37) in comparison with GGRBF with accuracy of 98.06 % .
Table 5. Performance summary for (37) in comparison with GGRBF with accuracy of 98.06 % .
Training with (37) (on a Single CPU)
EpochIterationRuntime (hh:mm:ss)Accuracy (%)Batch LossLearning Rate
1100:00:0010.942.75560.0010
25000:00:0482.810.50650.0010
310000:00:0797.660.19230.0010
415000:00:11100.000.07380.0010
620000:00:14100.000.04460.0010
725000:00:28100.000.02620.0010
830000:00:21100.000.02110.0010
935000:00:25100.000.01730.0010
1140000:00:29100.000.00990.0010
1245000:00:32100.000.00850.0010
1350000:00:36100.000.00850.0010
1454600:00:39100.000.00680.0010
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Singh, H. Machine Learning Application of Generalized Gaussian Radial Basis Function and Its Reproducing Kernel Theory. Mathematics 2024, 12, 829. https://doi.org/10.3390/math12060829

AMA Style

Singh H. Machine Learning Application of Generalized Gaussian Radial Basis Function and Its Reproducing Kernel Theory. Mathematics. 2024; 12(6):829. https://doi.org/10.3390/math12060829

Chicago/Turabian Style

Singh, Himanshu. 2024. "Machine Learning Application of Generalized Gaussian Radial Basis Function and Its Reproducing Kernel Theory" Mathematics 12, no. 6: 829. https://doi.org/10.3390/math12060829

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop