Computational capabilities of graph neural networks
IEEE Transactions on Neural Networks
Hi-index | 0.00 |
Given a multilayer perceptron (MLP) with a fixed architecture, there are functions that can be approximated up to any degree of accuracy, without having to increase the number of the hidden nodes. Those functions belong to the closure F¯ of the set F¯ of the maps realizable by the MLP. In this paper, we give a list of maps with this property. In particular, it is proven that: 1) rational functions belongs to F¯ for networks with inverse tangent activation function; and 2) products of polynomials and exponentials belongs to F¯ for networks with sigmoid activation function. Moreover, for a restricted class of MLPs, we prove that the list is complete and give an analytic definition of F¯