Oriented stochastic loss descent algorithm to train very deep multi-layer neural networks without vanishing gradients
作者:
Highlights:
• A new back-propagated gradient descent optimization algorithm without a vanishing problem is proposed.
• The proposed algorithm is called oriented stochastic loss descent (OSLD).
• OSLD updates in the opposite side of its gradient sign by a tuned ratio of random loss.
• OSLD is competitive to Adam and is compatible with most backpropagation architectures.
• OSLD is stable and opens more choices in front of very deep multi-layer neural networks.
摘要
•A new back-propagated gradient descent optimization algorithm without a vanishing problem is proposed.•The proposed algorithm is called oriented stochastic loss descent (OSLD).•OSLD updates in the opposite side of its gradient sign by a tuned ratio of random loss.•OSLD is competitive to Adam and is compatible with most backpropagation architectures.•OSLD is stable and opens more choices in front of very deep multi-layer neural networks.
论文关键词:Vanishing gradient problem,Deep learning,Backpropagation,Gradient descent algorithm,Optimization algorithm,Deep multi-layer neural network
论文评审过程:Received 27 November 2020, Revised 3 June 2021, Accepted 9 August 2021, Available online 11 August 2021, Version of Record 23 August 2021.
论文官网地址:https://doi.org/10.1016/j.knosys.2021.107391