What Is A Neural Network
What Is A Neural Network
A neural network is a machine learning (ML) model designed to mimic the function and
structure of the human brain. Neural networks are intricate networks of interconnected
nodes, or neurons, that collaborate to tackle complicated problems.
Also referred to as artificial neural networks (ANNs) or deep neural networks, neural
networks represent a type of deep learning technology that's classified under the broader
field of artificial intelligence (AI).
Each processing node has its own small sphere of knowledge, including what it has seen
and any rules it was originally programmed with or developed for itself. The tiers are
highly interconnected, which means each node in Tier N will be connected to many nodes
in Tier N-1 -- its inputs -- and in Tier N+1, which provides input data for those nodes.
There could be one or more nodes in the output layer, from which the answer it produces
can be read.
Artificial neural networks are noted for being adaptive, which means they modify
themselves as they learn from initial training and subsequent runs provide more
information about the world. The most basic learning model is centered on weighting the
input streams, which is how each node measures the importance of input data from each
of its predecessors. Inputs that contribute to getting the right answers are weighted
higher.
● Chatbots.
● Social media.
● Personal assistants.
Prime uses involve any process that operates according to strict rules or patterns and has
large amounts of data. If the data involved is too large for a human to make sense of in a
reasonable amount of time, the process is likely a prime candidate for automation through
artificial neural networks.
For example, if nodes David, Dianne and Dakota tell node Ernie the current input image
is a picture of Brad Pitt, but node Durango says it's George Clooney, and the training
program confirms it's Pitt, Ernie decreases the weight it assigns to Durango's input and
increase the weight it gives to David, Dianne and Dakota.
In defining the rules and making determinations -- the decisions of each node on what to
send to the next tier based on inputs from the previous tier -- neural networks use several
principles. These include gradient-based training, fuzzy logic, genetic algorithms and
Bayesian methods. They might be given some basic rules about object relationships in the
data being modeled.
For example, a facial recognition system might be instructed, "Eyebrows are found above
eyes," or, "Moustaches are below a nose. Moustaches are above and/or beside a mouth."
Preloading rules can make training faster and the model more powerful faster. But it also
includes assumptions about the nature of the problem, which could prove to be either
irrelevant and unhelpful or incorrect and counterproductive, making the decision about
what, if any, rules to build in important.
Further, the assumptions people make when training algorithms cause neural networks to
amplify cultural biases. Biased data sets are an ongoing challenge in training systems that
find answers on their own through pattern recognition in data. If the data feeding the
algorithm isn't neutral -- and almost no data is -- the machine propagates bias.
This type of ANN computational model is used in technologies such as facial recognition
and computer vision.
This neural network starts with the same front propagation as a feed-forward network but
then goes on to remember all processed information to reuse it in the future. If the
network's prediction is incorrect, then the system self-learns and continues working
toward the correct prediction during backpropagation.
The CNN model is particularly popular in the realm of image recognition. It has been
used in many of the most advanced applications of AI, including facial recognition, text
digitization and NLP. Other use cases include paraphrase detection, signal processing and
image classification.
● Parallel processing abilities. ANNs have parallel processing abilities, which means
the network can perform more than one job at a time.
● Information storage. ANNs store information on the entire network, not just in a
database. This ensures that even if a small amount of data disappears from one
location, the entire network continues to operate.
● Non-linearity. The ability to learn and model nonlinear, complex relationships helps
model the real-world relationships between input and output.
● Fault tolerance. ANNs come with fault tolerance, which means the corruption or
fault of one or more cells of the ANN won't stop the generation of output.
● Gradual corruption. This means the network slowly degrades over time instead of
degrading instantly when a problem occurs.
● Unrestricted input variables. No restrictions are placed on the input variables, such
as how they should be distributed.
● Obsevation-based decisions. Machine learning means the ANN can learn from
events and make decisions based on the observations.
● Unorganized data processing. Artificial neural networks are exceptionally good at
organizing large amounts of data by processing, sorting and categorizing it.
● Ability to learn hidden relationships. ANNs can learn the hidden relationships in
data without commanding any fixed relationship. This means ANNs can better model
highly volatile data and non-constant variance.
● Ability to generalize data. The ability to generalize and infer unseen relationships on
unseen data means ANNs can predict the output of unseen data.
● Lack of rules. The lack of rules for determining the proper network structure means
the appropriate artificial neural network architecture can only be found through trial,
error and experience.
● Hardware dependency. The requirement of processors with parallel processing
abilities makes neural networks dependent on hardware.
● Numerical translation. The network works with numerical information, meaning all
problems must be translated into numerical values before they can be presented to the
ANN.
● Lack of trust. The lack of explanation behind probing solutions is one of the biggest
disadvantages of ANNs. The inability to explain the why or how behind the solution
generates a lack of trust in the network.
● Inaccurate results. If not trained properly, ANNs can often produce incomplete or
inaccurate results.
● Black box nature. Because of their black box AI model, it can be challenging to
grasp how neural networks make their predictions or categorize data.
● 1940s. In 1943, mathematicians Warren McCulloch and Walter Pitts built a circuitry
system that ran simple algorithms and was intended to approximate the functioning of
the human brain.
● 1950s. In 1958, Frank Rosenblatt, an American psychologist who's also considered
the father of deep learning, created the perceptron, a form of artificial neural network
capable of learning and making judgments by modifying its weights. The perceptron
featured a single layer of computing units and could handle problems that were
linearly separate.
● 1970s. Paul Werbos, an American scientist, developed the backpropagation method,
which facilitated the training of multilayer neural networks. It made deep
learning possible by enabling weights to be adjusted across the network based on the
error calculated at the output layer.
● 1980s. Cognitive psychologist and computer scientist Geoffrey Hinton, along with
computer scientist Yann LeCun, and a group of fellow researchers began
investigating the concept of connectionism, which emphasizes the idea that cognitive
processes emerge through interconnected networks of simple processing units. This
period paved the way for modern neural networks and deep learning.
● 1990s. Jürgen Schmidhuberand Sepp Hochreiter, both computer scientists from
Germany, proposed the Long Short-Term Memory recurrent neural network
framework in 1997.
● 2000s. Geoffrey Hinton and his colleagues pioneered RBMs, a sort of generative
artificial neural network that enables unsupervised learning. RBMs opened the path
for deep belief networks and deep learning algorithms.
It wasn't until around 2010 that research in neural networks picked up great speed.
The big data trend, where companies amass vast troves of data and parallel computing
gave data scientists the training data and computing resources needed to run complex
artificial neural networks. In 2012, a neural network named AlexNet won the ImageNet
Large Scale Visual Recognition competition, an image classification challenge. Since
then, interest in artificial neural networks has soared and technology has continued to
improve.