论文标题

由SGD和自适应学习规则学到的表示:在神经网络中改变稀疏性和选择性的条件

Representations learnt by SGD and Adaptive learning rules: Conditions that vary sparsity and selectivity in neural networks

论文作者

Park, Jin Hyun

论文摘要

从人脑的角度来看,持续学习可以执行各种任务而不会相互干扰。减少相互干扰的有效方法可以在神经元的稀疏性和选择性中找到。根据Aljundi等人的说法。 Hadsell等人在代表性水平上施加稀疏性对于持续学习是有利的,因为稀疏的神经元激活会使参数之间的重叠减少,从而减少了干扰。同样,高度选择性的神经网络可能会引起较少的干扰,因为神经元中的特定反应会减少与其他参数重叠的机会。考虑到人脑在整个生命周期内进行持续学习,因此发现自然产生稀疏性和选择性的条件可能会提供理解大脑功能的见解。本文研究了各种条件,这些条件自然会增加神经网络中的稀疏性和选择性。本文在MNIST分类任务中使用Hoyer的稀疏度指标和CCMAS选择性度量测试了不同的优化器。需要注意的是,直到今天,任何神经科学或机器学习的稀疏性和选择性的自然发生的调查都没有得到认可。本文发现,特定条件会提高稀疏性和选择性,例如应用较大的学习率并降低批量的大小。除了条件,稀疏性和选择性之间的关系外,将根据经验分析讨论以下内容:1。稀疏性与选择性之间的关系以及2。测试准确性,稀疏性和选择性之间的关系。

From the point of view of the human brain, continual learning can perform various tasks without mutual interference. An effective way to reduce mutual interference can be found in sparsity and selectivity of neurons. According to Aljundi et al. and Hadsell et al., imposing sparsity at the representational level is advantageous for continual learning because sparse neuronal activations encourage less overlap between parameters, resulting in less interference. Similarly, highly selective neural networks are likely to induce less interference since particular response in neurons will reduce the chance of overlap with other parameters. Considering that the human brain performs continual learning over the lifespan, finding conditions where sparsity and selectivity naturally arises may provide insight for understanding how the brain functions. This paper investigates various conditions that naturally increase sparsity and selectivity in a neural network. This paper tested different optimizers with Hoyer's sparsity metric and CCMAS selectivity metric in MNIST classification task. It is essential to note that investigations on the natural occurrence of sparsity and selectivity concerning various conditions have not been acknowledged in any sector of neuroscience nor machine learning until this day. This paper found that particular conditions increase sparsity and selectivity such as applying a large learning rate and lowering a batch size. In addition to the relationship between the condition, sparsity, and selectivity, the following will be discussed based on empirical analysis: 1. The relationship between sparsity and selectivity and 2. The relationship between test accuracy, sparsity, and selectivity.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源