Sigmoid Deep Learning
Sigmoid Deep Learning
The sigmoid neuron is essentially the building block of the deep neural
networks. These sigmoid neurons are like perceptron, but they happen to
be slightly modified so that the output from the sigmoid neuron is far
smoother than the step functional output from perceptron.
The inputs to the sigmoid neuron can be real numbers unlike the Boolean
inputs in MP Neuron. Even the output will be a real number between 0–1.
In the sigmoid neuron, you are trying to regress the relationship between
X and Y in terms of probability. Even though the output is between 0 and
1, you can still make use of the sigmoid function for binary classification
tasks by selecting a threshold.
Mathematical Formulation
σ(x) = 1 / (1 + exp(-x))
Where:
Sigmoid graph
If the outcome of the sigmoid function is greater than 0.5 then you would
classify that label to be class 1 or positive class and if it is less than 0.5
then you would classify it to be a negative class or label it as class 0.
Basically, the function determines which value to pass as output and what
not to pass as output.
There are 7 types of activation functions that are used in machine learning
and deep learning.
There are several types of sigmoid functions available. Here are three of
the most common types of sigmoid functions.
Where,
Example problem 1:
How many years will it take for a bacteria population to reach 9000, if its
growth is modelled by
here, t in years?
Solution:
-0.12(t-20)=ln(0.111)
t = -ln(0.111)/0.12 + 20
On simplifying,
t=38.31 years
3. Arctangent Function
4. Guddermannian Function
5. Error Function
The error function (Gauss error function) is used in probability theory and
statistics for describing the probability distribution of a Gaussian random
variable.
6. Smoothstep Function
1972 - The biologists Hugh Wilson and Jack Cowan at the University of
Chicago were trying to model biological neurons computationally and
ended up publishing the Wilson–Cowan model, in which a neuron sends a
signal to another neuron if it receives a signal greater than an activation
potential. Wilson and Cowan employed the logistic sigmoid function to
model the activation of a neuron as a function of a stimulus.
Recently, ANNs have shifted away from sigmoid functions towards the
ReLU function, because all the variants of the sigmoid function are
computationally intensive to calculate, and the ReLU offers the required
nonlinearity to take advantage of the depth of the network, while also
being very fast to compute.