Learning Rate Back propagation calculus NN Learning
Gradient descent Vs Mini batch Stochastic Gradient descent Randomly shuffle your training data and divide it into a bunch of mini-batches Why are there different optimizers?
Oct 06, 20251 min read
Learning Rate Back propagation calculus NN Learning
Gradient descent Vs Mini batch Stochastic Gradient descent Randomly shuffle your training data and divide it into a bunch of mini-batches Why are there different optimizers?