mini batch gradient descent - AshokBhat/ml GitHub Wiki
- A type of gradient descent
- During each pass, process N random examples, with N << all the training examples
- What is it?
- How is it different from others?
- What are the pros and cons?
- Where is it used?