top | item 36065975

(no title)

vippy | 2 years ago

The math isn't super difficult. Some books will try to throw a mess of differential equations at you, but some simple calculus is all you need for backpropagation.

discuss

order

quickthrower2|2 years ago

I have been through the math thanks to the youtube videos by A. Karpathy. Deriving some of the differentials, e.g. for batchnorm seems fairly hard (hard as in slogging through something with many steps where you can't make a mistake at any step). But the principles are quite simple - I think by design. If they were hard to compute or reason about then the neural net wouldn't work very well!

sva_|2 years ago

Doing the compute efficiently, especially from Python, is the tricky part.