The Stochastic Gradient Descent (SGD) & Learning Rate

When I finished the article on gradient descent, I realized that there were two important points missing. The first concerns the stochastic approach when we have too large data sets, the second being to see very concretely what happens when we poorly choose the value of the learning rate. I will therefore take advantage of this article to finally continue the previous article šŸ˜‰

Read More

The Gradient Descent

How to talk about Machine Learning or even Deep Learning without addressing the – famous – gradient descent? There are many articles on this subject of course, but often you have to read several in order to fully understand all the mechanisms. Often too mathematical or not enough, I will try especially here to explain its operation smoothly and step by step in order to try to demystify the subject.

Read More

Fork me on GitHub