Unit 3_IntrotoNN (1)
Unit 3_IntrotoNN (1)
➢ 1960s-1970s: Perceptrons
This era is defined by the work of Rosenblatt on perceptrons. Perceptrons are single-layer
networks whose applicability was limited to linearly separable problems.
Neural networks are complex systems that mimic some features of the functioning of the human
brain. It is composed of an input layer, one or more hidden layers, and an output layer.
Network components include neurons, connections, weights, biases, propagation functions, and a
learning rule. Neurons receive inputs, governed by thresholds and activation functions.
Connections involve weights and biases regulating information transfer.
Bias: It is the constant which is added to the product of features and weights. It is used to offset
the result. It helps the models to shift the activation function towards the positive or negative side
by a constant value.
Activation Functions: It decides whether a neuron should be activated or not by calculating the
weighted sum and adding bias to it. The purpose of the activation function is to introduce non-
linearity into the output of a neuron. This nonlinearity allows neural networks to develop
complex representations and functions based on the inputs that would not be possible with a
simple linear regression model. Some types of activation functions are:
Learning Rule: The learning rules updates the weights and bias levels of a network when a
network simulates in a specific data environment. Applying learning rule is an iterative process.
It helps a neural network to learn from the existing conditions and improve its performance.
Example: Hebbian Rule, Delta Rule, Perceptron Rule etc.
Backpropagation
1. Loss Calculation: The network’s output is evaluated against the real goal values, and a
loss function is used to compute the difference.
2. Loss Reduction: A method like dradient descent is then used by the network to reduce the
loss. To lower the inaccuracy, weights are changed based on the derivative of the loss
with respect to each weight.
3. Adjusting weights: The weights are adjusted at each connection by applying this iterative
process, or backpropagation, backward across the network.
4. Training: During training with different data samples, the entire process of forward
propagation, loss calculation, and backpropagation is done iteratively, enabling the
network to adapt and learn patterns from the data.
5. Activation Functions: Model non-linearity is introduced by activation functions like the
rectified linear unit (ReLU) or sigmoid. Their decision on whether to “fire” a neuron is
based on the whole weighted input.
3.2 Linear-separability?
Linear Separability refers to the data points in binary classification problems which can be
separated using linear decision boundary. If the data points can be separated using a line, linear
function, or flat hyperplane, the problem is considered to be linearly separable.
Implementation of Linearly separable Boolean Functions (AND, OR, NOT, NAND, NOR):