WingNews logo WingNews
top | new | best | ask | show | jobs
top | item 37065313

(no title)

pseudonom- | 2 years ago

There are other mechanisms for dealing with vanishing and exploding gradients. I (maybe wrongly?) think of batch normalization as being most distinctively about fighting internal covariate shift: https://machinelearning.wtf/terms/internal-covariate-shift/

discuss

order

No comments yet.

powered by hn/api // news.ycombinator.com