MLP, Gradient Descent, Activation Functions
MLP, Gradient Descent, Activation Functions
MLP is
composed of multiple layers, including an input layer, hidden layers,
and an output layer, where each layer contains a set of perception
elements known as neurons. Fig. 1 illustrates an MLP with two hidden
layers, an input and output layer.
4. Leaky ReLU
○ Definition:
● Local Minima and Saddle Points: Can get stuck in local minima or
saddle points where learning stops prematurely.
● Vanishing and Exploding Gradients: In deep networks, gradients
can become too small (vanishing) or too large (exploding),
hindering learning.
Key Concepts: