论文标题
Banach太空代表神经网络和山脊花纹定理
Banach Space Representer Theorems for Neural Networks and Ridge Splines
论文作者
论文摘要
我们开发了一个变分框架,以了解神经网络所学的功能适合数据的功能的属性。我们提出和研究连续域线性逆问题的家族,在ra域中,受数据拟合约束的ra域中的总变异样性。我们得出一个代表定理,表明有限的宽度,单隐层神经网络是解决这些反问题的解决方案。我们借鉴了各种样条理论的许多技术,因此我们提出了多项式脊样条的概念,该概念对应于具有截断功率函数作为激活函数的单隐层神经网络。代表定理让人联想到经典的繁殖内核希尔伯特太空代表定理,但我们表明神经网络问题是在非希尔伯特式的Banach Space上构成的。尽管学习问题是在连续域中构成的,但类似于内核方法,但这些问题可以作为有限维神经网络培训问题重新解决。这些神经网络训练问题具有与众所周知的重量衰减和路径正规化器有关的正规化器。因此,我们的结果洞悉了训练有素的神经网络的功能特征以及设计神经网络正则化器。我们还表明,这些正则化器促进具有理想概括特性的神经网络解决方案。
We develop a variational framework to understand the properties of the functions learned by neural networks fit to data. We propose and study a family of continuous-domain linear inverse problems with total variation-like regularization in the Radon domain subject to data fitting constraints. We derive a representer theorem showing that finite-width, single-hidden layer neural networks are solutions to these inverse problems. We draw on many techniques from variational spline theory and so we propose the notion of polynomial ridge splines, which correspond to single-hidden layer neural networks with truncated power functions as the activation function. The representer theorem is reminiscent of the classical reproducing kernel Hilbert space representer theorem, but we show that the neural network problem is posed over a non-Hilbertian Banach space. While the learning problems are posed in the continuous-domain, similar to kernel methods, the problems can be recast as finite-dimensional neural network training problems. These neural network training problems have regularizers which are related to the well-known weight decay and path-norm regularizers. Thus, our result gives insight into functional characteristics of trained neural networks and also into the design neural network regularizers. We also show that these regularizers promote neural network solutions with desirable generalization properties.