More precisely, they showed that piecewise linear networks (which can be obtained from rectifier nonlinearities or maxout units) can represent functions with a number of regions that is exponential in the depth of the network.
What does regions here referring to. How does this line helps us to conclude that deep networks require exponential units if we want to represent them using single hidden layer?