Closed traveling-desi closed 7 years ago
It's a mix of both SGD and mini-batches. With normal SGD, you'd randomly grab one input at a time and train the network on each input. Mini-batches take a batch of non-random data and train on that. Here we're using a mini-batch of random data, so it's still SGD just with multiple data points.
Hello!
There have been recent changes in the notebook for the first project: https://github.com/udacity/deep-learning/blob/master/first-neural-network/Your_first_neural_network.ipynb
In particular, the weight updates are:
The instructions still refer to this as:
This is not SGD, but mini batching. in SGD we update the weights before we start with a new example. In this notebook, we accumulate the changes and then update the weights after the batch is completed. This is minibatching.
Refrence: https://en.wikipedia.org/wiki/Stochastic_gradient_descent#Iterative_method
Please update the instructions in case you agree with this.