Content deleted Content added
Mathreader17 (talk | contribs) →Connection between RKHS with ReLU function: Rewrite section to explain more clearly |
|||
Line 263:
== Connection between RKHS with ReLU function ==
The [[Rectifier (neural networks)|ReLU function]] is commonly defined as <math>f(x)=\max (0, x)</math> and is a mainstay in the architecture of neural networks, used as an activation function.
We show below how we can construct an RKHS using a ReLU-like nonlinear function. This illustrates the representation power of neural networks using ReLU activations.We will work with the Hilbert space <math> \mathcal{H}=C^{1}[0, \infty) </math> of absolutely continuous functions with inner product <math> \langle f,g \rangle_{\mathcal{H}} = \int_0^{\infty}f'(x)g'(x) dx </math>.
We start by introducing a function <math> G(x,t) = \begin{cases} 1, & \text{if }0\leq x<t \\ 0, & \text{
t, & \text{otherwise}
\end{cases}=\min(x, t)</math>.
x, & \text{if } x\geq 0\\
▲ x, & \text{if } x<0\\
0, & \text{otherwise}
\end{cases}=ReLU(x)</math>, and we obtain the ReLU function.
<math> \forall t\in [0,1] </math>, we have <math> f(t)=\int_{0}^{t} f'(x) dx=\int_{0}^{t} G(x,t)f'(x) dx = \langle K_t(x),f \rangle </math>. ▼
We can show that <math>K_t</math> is a kernel by noting that its corresponding matrix is positive definite. We can also show the reproducing property of <math>K_t</math> using [[Fundamental theorem of calculus|Fundamental Theorem of Calculus]]:
▲
Using this formulation, we can easily find the minimizing function in this Hilbert space, which corresponds to the optimum in training a neural network with ReLU activation.
== See also ==
*[[Positive definite kernel]]
|