Universal approximation theorem

This is an old revision of this page, as edited by Koertefa (talk | contribs) at 23:22, 11 February 2013 (there are many versions of the theorem + a paper written by Cybenko does not prove that it is also called "Cybenko theorem"). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In the mathematical theory of neural networks, the universal approximation theorem states[1] that the standard multilayer feed-forward network with a single hidden layer, which contains finite number of hidden neurons, is a universal approximator among continuous functions on compact subsets of Rn, under mild assumptions on the activation function.

One of the first versions of the theorem was proved by George Cybenko in 1989 for sigmoid activation functions.[2]

Kurt Hornik showed in 1991[3] that it is not the specific choice of the activation function, but rather the multilayer feedforward architecture itself which gives neural networks the potential of being universal approximators. The output units are always assumed to be linear. For notational convenience we shall explicitly formulate our results only for the case where there is only one output unit. (The general case can easily be deduced from the simple case.)

The theorem[2][3][4][5] in mathematical terms:

Formal statement

Let φ(·) be a nonconstant, bounded, and monotonically-increasing continuous function. Let Im denote the m-dimensional unit hypercube [0,1]m. The space of continuous functions on Im is denoted by C(Im). Then, given any function fC(Im) and є > 0, there exist an integer N and sets of real constants αi, biR, wiRm, where i = 1, ..., N such that we may define:

 

as an approximate realization of the function f; that is,

 

for all xIm.

References

  1. ^ Balázs Csanád Csáji. Approximation with Artificial Neural Networks; Faculty of Sciences; Eötvös Loránd University, Hungary
  2. ^ a b Cybenko., G. (1989) "Approximations by superpositions of sigmoidal functions", Mathematics of Control, Signals, and Systems, 2 (4), 303-314
  3. ^ a b Kurt Hornik (1991) "Approximation Capabilities of Multilayer Feedforward Networks", Neural Networks, 4(2), 251–257
  4. ^ Haykin, Simon (1998). Neural Networks: A Comprehensive Foundation, Volume 2, Prentice Hall. ISBN 0-13-273350-1.
  5. ^ Hassoun, M. (1995) Fundamentals of Artificial Neural Networks MIT Press, p. 48