4 4 Choosing The Right Activation Function For Neural Networks
4 4 Choosing The Right Activation Function For Neural Networks
Introduction
The input layer just holds the input data and no calculation is
performed. Therefore, no activation function is used there.
Sigmoid activation function (Image by author, made with latex editor and matplotlib)
Key features:
Usage:
Drawbacks:
Tanh activation function (Image by author, made with latex editor and matplotlib)
Key features:
Usage:
Drawbacks:
ReLU activation function (Image by author, made with latex editor and matplotlib)
Key features:
Usage:
Drawbacks:
• The value of the positive side can go very high. That may
lead to a computational issue during the training.
Leaky ReLU activation function (Image by author, made with latex editor and matplotlib)
Key features:
• The leaky ReLU activation function is a modified version
of the default ReLU function.
Usage:
• The same usage of the ReLU function is also valid for the
leaky ReLU function.
Key features:
ReLU6 activation function (Image by author, made with latex editor and matplotlib)
Key features:
Key features:
Usage:
Key features:
Usage:
Identity activation function (Image by author, made with latex editor and matplotlib)
Key features:
Usage:
Swish activation function (Image by author, made with latex editor and matplotlib)
Key features:
Usage:
• This function is only used in the hidden layers.
Drawbacks:
Hard Swish (H-Swish) activation function (Image by author, made with latex editor and
matplotlib)
Key features:
• The graph is identical to the graph of the Swish function.
Usage:
Summary
By Rukshan Pramoditha