at least 1 hidden layer
Any continuous function can be approximated arbitrarily well by a
neural network with at least 1 hidden layer with a finite number of weights
One wide (latent) layer is enough, but it is just a memorizer (Cybenko ’89)

Seonglae Cho
Seonglae Cho