Notice: This page requires JavaScript to function properly.
Please enable JavaScript in your browser settings or update your browser.
Impara The Universal Approximation Theorem | Approximation and Representational Power
Mathematical Foundations of Neural Networks

bookThe Universal Approximation Theorem

When you build neural networks, a central question arises: can these models represent any function you might encounter in practice? This question is crucial for understanding the power and limitations of neural networks. If a network can approximate any function to any desired level of accuracy, then, in theory, you can use it for a wide variety of tasks — such as regression, classification, or even more complex mappings. This idea motivates the study of function approximation in neural networks. The Universal Approximation Theorem addresses this very question, providing a foundational result about the representational abilities of neural networks.

Note
Definition

Universal Approximation Theorem (formal statement):
A feedforward neural network with a single hidden layer containing a finite number of neurons, using a nonconstant, bounded, and continuous activation function, can approximate any continuous function on compact subsets of ℝⁿ to any desired degree of accuracy, provided there are enough hidden units.
Main assumptions:

  • The activation function must be nonconstant, bounded, and continuous;
  • The function being approximated must be continuous;
  • Approximation is on compact subsets of ℝⁿ (closed and bounded sets).

The Universal Approximation Theorem has profound implications for neural network design. It guarantees that even a shallow network — with just one hidden layer — can, in principle, approximate any continuous function, as long as you choose a suitable activation function and provide enough neurons. However, this result crucially depends on the activation function being nonlinear. If you use only linear activations, the network collapses to a single linear transformation, which cannot capture complex, nonlinear relationships. This is why the introduction of nonlinearity, as discussed in the previous section, is essential for unlocking the full expressive power of neural networks. The theorem does not guarantee that learning such an approximation is practical or efficient, but it does assure you that the architecture has the capacity, under the right conditions, to represent a wide class of functions.

question mark

Which of the following conditions must be met for the Universal Approximation Theorem to apply to a neural network?

Select the correct answer

Tutto è chiaro?

Come possiamo migliorarlo?

Grazie per i tuoi commenti!

Sezione 2. Capitolo 1

Chieda ad AI

expand

Chieda ad AI

ChatGPT

Chieda pure quello che desidera o provi una delle domande suggerite per iniziare la nostra conversazione

bookThe Universal Approximation Theorem

Scorri per mostrare il menu

When you build neural networks, a central question arises: can these models represent any function you might encounter in practice? This question is crucial for understanding the power and limitations of neural networks. If a network can approximate any function to any desired level of accuracy, then, in theory, you can use it for a wide variety of tasks — such as regression, classification, or even more complex mappings. This idea motivates the study of function approximation in neural networks. The Universal Approximation Theorem addresses this very question, providing a foundational result about the representational abilities of neural networks.

Note
Definition

Universal Approximation Theorem (formal statement):
A feedforward neural network with a single hidden layer containing a finite number of neurons, using a nonconstant, bounded, and continuous activation function, can approximate any continuous function on compact subsets of ℝⁿ to any desired degree of accuracy, provided there are enough hidden units.
Main assumptions:

  • The activation function must be nonconstant, bounded, and continuous;
  • The function being approximated must be continuous;
  • Approximation is on compact subsets of ℝⁿ (closed and bounded sets).

The Universal Approximation Theorem has profound implications for neural network design. It guarantees that even a shallow network — with just one hidden layer — can, in principle, approximate any continuous function, as long as you choose a suitable activation function and provide enough neurons. However, this result crucially depends on the activation function being nonlinear. If you use only linear activations, the network collapses to a single linear transformation, which cannot capture complex, nonlinear relationships. This is why the introduction of nonlinearity, as discussed in the previous section, is essential for unlocking the full expressive power of neural networks. The theorem does not guarantee that learning such an approximation is practical or efficient, but it does assure you that the architecture has the capacity, under the right conditions, to represent a wide class of functions.

question mark

Which of the following conditions must be met for the Universal Approximation Theorem to apply to a neural network?

Select the correct answer

Tutto è chiaro?

Come possiamo migliorarlo?

Grazie per i tuoi commenti!

Sezione 2. Capitolo 1
some-alt