Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations

作者:Guang Shi, Jiangshe Zhang, Huirong Li, Changpeng Wang

摘要

Deep neural networks (DNNs) are witnessing increasing attention in machine learning. However, the information propagation is becoming increasingly difficult as the networks get deeper, which makes the optimization of DNN extremely hard. One reason of this difficulty is saturation of hidden units. In this paper, we propose a novel methodology named RegA to decrease the influences of saturation on ReLU-DNNs (DNNs with ReLU). Instead of changing the activation functions or the initialization strategy, our methodology explicitly encourage the pre-activation to be out of the saturation region. Specifically, we add an auxiliary objective induced by L2-norm of the pre-activation values to the optimization problem. The auxiliary objective could help to active more units and promote effective information propagation in ReLU-DNNs. By conducting experiments on several large-scale real datasets, we demonstrate better representations could be learned by using RegA and the method help ReLU-DNNs get better performance on convergence and accuracy.

论文关键词:Neural networks, ReLU, Saturation phenomenon, L2 regularization

论文评审过程:

论文官网地址:https://doi.org/10.1007/s11063-018-9883-8