stochastic gradient descent - AshokBhat/ml GitHub Wiki
- A type of gradient descent
-
How is it different from other forms of gradient descent?
- Using a batch size of 1 random sample.
-
What is the advantage?
- Computationally less intensive.
-
Does it work?
- Yes, it does work. The path to minima will be staggered.