In this article, I introduce four of the most important optimization algorithms in Deep Learning. These algorithms allow neural networks to be trained faster while achieving better performance. These optimization...
Activation Functions in Deep Learning: Sigmoid, tanh, ReLU
In this guide, I will introduce you to four of the most important activation functions used in Deep Learning: Sigmoid, Tanh, ReLU & Leaky ReLU. Specifically, this guide will cover...