ADMMiRNN: Training RNN with Stable Convergence via an Efficient ADMM Approach
It is hard to train Recurrent Neural Network (RNN) with stable convergence and avoid gradient vanishing and exploding, as the weights in the recurrent unit are repeated from iteration to iteration. Moreover, RNN is sensitive to the initialization of weights and bias, which brings difficulty in the t...
Saved in:
Published in | Machine Learning and Knowledge Discovery in Databases Vol. 12458; pp. 3 - 18 |
---|---|
Main Authors | , , , , , , |
Format | Book Chapter |
Language | English |
Published |
Switzerland
Springer International Publishing AG
2021
Springer International Publishing |
Series | Lecture Notes in Computer Science |
Online Access | Get full text |
Cover
Loading…
Be the first to leave a comment!