0% found this document useful (0 votes)
103 views17 pages

An Ingression Into Deep Learning - FP

The document provides an overview of deep learning and neural networks. It defines key concepts like artificial intelligence, machine learning, deep learning, and neural network architecture. Deep learning involves complex algorithms and deep neural networks to train models in a way that mimics the human brain. It can process massive amounts of structured and unstructured data, perform complex operations, and maintain performance as more data is available. Neural networks work using principles of forward and backward propagation to update weights and biases in the network.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
103 views17 pages

An Ingression Into Deep Learning - FP

The document provides an overview of deep learning and neural networks. It defines key concepts like artificial intelligence, machine learning, deep learning, and neural network architecture. Deep learning involves complex algorithms and deep neural networks to train models in a way that mimics the human brain. It can process massive amounts of structured and unstructured data, perform complex operations, and maintain performance as more data is available. Neural networks work using principles of forward and backward propagation to update weights and biases in the network.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
You are on page 1/ 17

Digital : An Ingression into Deep Learning_FP

In ____________, information must flow from input to output only in one direction.

Feed-Forward artificial

Hiden

H2O.ai offers built-in integration tools for which platform?

All

tool supports the use of GPU.

turi

Machine Learning is inspired by the structure of the brain.

false

Which of the following is a type of Recurrent Neural Network?

All

Autoencoders are trained without supervision.

true

As the amount of data increases, the performance of Machine Learning algorithm

deIncrease

The following are tools offered by deepnet platforms, except

H2o

Every neuron in the Input Layer represents a/an _______________ variable that influences the
output.

Indepent

There is no exact formula for calculating the number of hidden layers, as well as the number of
neurons in each hidden layer.

True
introduces the concept of memory cell which can process data within time lags.

Autoend

____________ is a technique which helps machines to mimic human behavior.

Deep learning

Machine Learning is a subset o

Deep

Prelude

Have you ever wondered

How google translates an entire web page to a different language in a matter of seconds? Or,
how your phone gallery groups images into different folders based on their locations?

All these are innovations of Deep Learning! But, what exactly is Deep Learning? Explore to
know more!

Deep Learning

The Big Picture!


Deep Learning is a subset of Machine Learning, which in turn, is a subset of Artificial
Intelligence.

1. Artificial Intelligence

Artificial Intelligence (AI) is a technique that helps machines to mimic human behavior.

2. Machine Learning

Machine Learning is an application of AI that allows the system to learn and improve from
experience automatically.

3. Deep Learning

Deep Learning is a type of Machine Learning that is inspired by the structure of the brain. It
is also known as Artificial Neural Network (ANN). It uses complex algorithms and deep
neural networks to train models.

What is Deep Learning?

Definition
Deep Learning involves networks which are capable of learning from data and functions
similar to the human brain.

Why Deep Learning?

Let us explore the reasons which make Deep Learning shine in the industry.

 Processes massive amount of data

Deep Learning can process an enormous amount of both Structured and Unstructured
data.

 Performs Complex Operations

Deep Learning algorithms are capable enough to perform complex operations when
compared to the Machine Learning algorithms.

Why Deep Learning?

 Achieves Best Performance


o As the amount of data increases, the performance of Machine Learning
algorithms decreases.
o On the other hand, Deep Learning maintains the performance of the model.
 Feature Extraction
o Machine Learning algorithms extract patterns from labeled sample data,
while Deep Learning algorithms take large volumes of data as input, analyze
them to extract the features on its own.

Deep Learning vs. Machine Learning

Say, for instance, we develop a machine that differentiates between cherries and tomatoes.

Machine Learning
If done through Machine Learning, we need to specify the features based on which the two
can be differentiated like size and stem, in this case.

Deep Learning

In Deep Learning, the features are picked by the Neural Network without any human
intervention. But, that kind of independence can be achieved by a higher volume of data in
training the machine.

A Curtain Raiser

Watch this video to know more about Deep Learning

 The human brain contains billions of cells called Neurons. The structure of a neuron
is depicted in the above image.
 Neural Networks is a set of algorithms designed to learn the way our brain works.

 The biological neurons inspire the structure and the functions of the neural networks.

Biological and Artificial Neurons - Terminologies

Biological Neuron Artificial Neuron


Dendrites Inputs
Nucleus Nodes
Synapse Weights
Axon Output

A Node is also called a Neuron or Perceptron.

Structure of an Artificial Neural Network


The basic structure of an Artificial Neural Network (ANN) consists of artificial neuron that
are grouped into 3 different layers namely:

There are three different layers in a neural network, namely:

1. Input Layer
2. Hidden Layer
3. Output Layer

Input Layer

 The input layer communicates with the external environment and presents a
pattern to the neural network.
 Every neuron in the input layer represents an independent variable that influences
the output.

Structure of an Artificial Neural Network


Hidden Layer

 A Neural Network consists of several hidden layers, each consisting of a collection


of neurons.
 The hidden layer is an intermediate layer found between the input layer and the
output layer.
 This layer is responsible forextracting the features required from the input.
 There is no exact formula for calculating the number of the hidden layers as well as the
number of neurons in each hidden layer.

Output Layer

 The output layer of the neural network collects and transmits information in the
desired format.

Structure of an Artificial Neural Network

Every single layer is made up of one or more number of neurons. A simple model of an
Artificial neuron is shown in the figure.
Parameters represented are:

1. X1,X2,...,XmX_1,X_2,...,X_mX1,X2,...,Xm are the input signals.


2. W1,W2,...WmW_1,W_2,...W_mW1,W2,...Wm are the weights.
3. ∑\sum∑ is the Summation function
4. σ\sigmaσ is the Activation function.
5. bkb_kbk represents the bias value

Artificial Neural Network - Process Flow

The process flow depicted in the image is explained as follows:

 A node is where the computations happen


 A node combines the Input data X1,X2,...,XmX_1,X_2,...,X_mX1,X2,...,Xm with a set of
Weights W1,W2,...,WmW_1,W_2,...,W_mW1,W2,...,Wm which in turn aids in amplifying
or diminishing the input with respect to the polarity attached to it.
 On multiplying the inputs with weights, a specific significance is assigned to the inputs with
regard to the task the algorithm is learning. It helps in answering the question Which
input is the most helpful in learning data without error?

 The input-weight products are summed up using the Summing function, and the output
yielded is passed to the Activation function to decide whether and to what extent the
signal should progress to have an effect in the final output.

Summation and Activation Functions

Summation Function

 Various inputs are multiplied with their respective connection weights and summed up
together with the bias value.

∑j=1NXjWj+bk\sum_{j=1}^{N}X_jW_j+b_k∑j=1NXjWj+bk

where:

XjX_jXj represents the inputs

WjW_jWj represents the weights


Summation and Activation Functions

Activation Function

 Activation Function aids in deriving the output.


 It is also known as the Transfer Function.

 It maps the resulting values between 0 to 1 or -1 to 1.

 There are two types of Activation function, namely:

1. Linear Activation Function


2. Non-Linear Activation Function
 Non-Linear Activation Function is the most commonly used Activation function
in Neural Networks.

Non-Linear Activation Functions

Watch this video to know some of the most commonly used Activation Functions in Neural
Networks.

The default Activation Function is ReLU.

Working of a Neural Network

Watch this video to know the working of a Neural Network.

In January 2018, Google's CEO Sundar Pichai claimed that Artificial intelligence (AI) would be more
transformative to humanity than electricity.

Prelude
Curious about how weight and bias values are decided? This section will help you in
understanding the concepts of the learning process in Neural Network.

Learning Process of a Neural Network

The learning process of a Neural Network includes updating the network architecture and
connecting the weights for the network to perform efficiently.

Designing a Learning Process involves the following:

1. Learning Paradigm

Having a model from the environment in which Neural Network works.

2. Learning Rules

Figuring out the rules that aid in updating the weights.

3. Learning Algorithms

Identifying the procedure to update the weights according to the learning rules.

Learning Paradigm

The following are the various Learning Paradigms in Neural Networks:

 Supervised
 Unsupervised
 Reinforcement

Learning Rules

The four basic types of Learning Rules in Neural Network are:

 Error Correction Rules


 Hebbian
 Boltzmann
 Competitive Learning

Learning Algorithms

Following are a few Deep Learning algorithms:

 Adaptive Resonance Theory


 Kochen Self Organization Map
 ADALINE
 Perceptron
 Backpropagation

DRC-Hubo

View the following video on DRC-Hubo developed by Team KAIST from South Korea which
impresses us by remarkably precise and deliberate motion.

Principles

Principles

A Neural Network works based on two principles, namely:

1. Forward Propagation
2. Backward Propagation

Forward Propagation in Neural Network

Explore this video to know about the Forward Propagation.

Backward Propagation in Neural Network


Explore this video to know about the Backward Propagation.

Prelude

In this section, let us discover the Network Architecture that prevails in Neural Network.

Neural Network Architecture

The architecture of a Neural Network can be broadly classified into two, namely:

1. Feed Forward Artificial Neural Network


2. Recurrent Neural Network

Feed Forward Artificial Neural Network

 The information must flow from input to output only in one direction.
 No Feedback loops must be present.

A few Feed Forward Artificial Neural Networks are:

 Single-layer Feed Forward Network


 Multi-layer Feed Forward Network

Recurrent Neural Network

 Information can be transmitted in both directions.


 Feedback loops are allowed.
A few Recurrent Neural Networks are:

 Hopfield Network
 Elman Network
 Jordan Network

Explore More!

The following video will help you know more about the Feed Forward Networks and
Recurrent Neural Networks in mathematical terms.

MELTANT-α\alphaα is an avatar robot with the world’s most powerful and flexible robotic hand
which directly imitates human hand movements.

Prelude

Neural Networks are employed in various fields. Following are a few types of Neural
Networks that we will explore in this section:

 Radial Basis Function


 Long/Short Term Memory
 Gated Recurrent Unit
 Autoencoder
 Convolutional Neural Networks

Radial Basis Function

Radial Basis Function is similar to Feed Forward Network but uses Logistic function as the
Activation Function.

Check out this video to know more!

Long/Short Term Memory


 Long/Short Term Memory (LSTM) networks are used to classify, process, and make
predictions based on time series data.
 LSTM networks can predict the action in a specific video frame by keeping in mind the
action that occurred in the earlier frames.
 The applications of LSTM include writing, speech recognition, and so on.

Gated Recurrent Unit

 Gated Recurrent Unit (GRU) is LSTM with a forget gate.


 It is used in sound, speech synthesis, and so on.

More on LSTM and GRU

Watch the following video to know more about LSTM and GRU Neural Networks.

Autoencoder

 Autoencoder (AE) is used for classification, clustering, and feature compression.


 It is trained without supervision.

Convolutional Neural Networks

The following video will help you understand the basic concepts of Convolutional Neural
Networks (CNN).

Applications

Image Classification
 Image classification is one of the common applications of deep learning.
 A convolutional neural network can be used to recognize images and label them
automatically.

Object Recognition

Deepnets (an optimized version of Deep Neural Networks) can be trained to recognize different
objects within the same image.

Video Recognition

 Deepnets can also be trained for real-time object recognition with respect to time
and space.
 Some of the famous use cases are driverless cars, robots, and theft detection.

Sentiment Analysis

 Using sentiment analysis, the underlying intent of the text can be extracted.
 With social media channels, it is possible to automate and measure the feelings of the
public on a given news story, topic, brand, or product.

 Positive sentiment can be identified, thereby allowing the marketing of a product, or


understanding which elements of a business strategy are working.

Medical Applications

 Deepnets can be trained to detect cancerous cells, benign and malignant tumors from
MRI and CT scans.
 They are also applied in drug discovery by training nets with molecular structure and
chemical compositions.

Tools

Deepnet Platform
 A deepnet platform is a service that allows you to incorporate deepnet in your
applications without building one from scratch.
 This platform provides a set of tools and interfaces to create a custom deepnet.

Deepnet platforms are of two types, namely:

 Software platform: This platform is available as downloadable packages that need to be


deployed on your hardware.
 Full platform: It is available as an online interactive UI to build and deploy models
without any coding experience.

More on Deepnet Platform

The following video explains what exactly is a deepnet platform.

Deepnet Platforms - Tools

The following are the tools offered by the deepnet platforms:

 Deepnet capability
 Data Munging
 UI/Model Management
 Infrastructure

H2O.ai

 H2O.ai is an open-source machine learning platform.


 Along with many machine learning algorithms, this platform currently provides a
deepnet capability known as Multilayer Perceptron (MLP).

Features

 H2O.aiH_2O.aiH2O.ai supports MLP for a deep learning model, which uses the L-
BFGS algorithm for backpropagation.
 H2O.aiH_2O.aiH2O.ai offers built-in integration tools for platforms like HDFS,
Amazon S3, SQL, and NoSQL.
 Its intuitive UI can be accessed by programming environments like R, Python, and
JSON.
 You can model and analyze data with Tableau, Microsoft Excel, and RStudio.
 Since the package needs to be downloaded and deployed in the hardware,
H2O.aiH_2O.aiH2O.ai comes with an in-memory map-reduce, and columnar
compression to accelerate the training of models.

Turi - A Graph-based Tool

Turi is a software platform that provides two deepnet capabilities (convolutional network
and MLP) along with machine learning and graph algorithms.

Features

 Turi provides built-in support for integration of Amazon S3, SQL DB, HDFS, Spark RDD, and
Pandas data frames.
 It offers a UI interface for model management and also includes a visualization tool called
GraphLab Canvas for visualizing the model results.

 To work with large data, Turi comes with built-in storage support like SFrame, SArray, and
SGraph.
 Unlike H2O.aiH_2O.aiH2O.ai, Turi supports the use of GPUs.

Turi

Watch this video to know more about the Turi platform.

Course Summary

While we've seen great strides in robotic capability recently, there's still a vast gap between
human and robot motor skills. Even at simple motor skills like gripping, machine still have a
very long way to go to suit human ability.

By linking learning with continuous feedback and control, however, we may begin to bridge
that gap, enabling robots to manage the complexities of the real world intelligently and
reliably.

Replicating the human brain at entirety will not just be a Science fiction episode anymore.
The future indeed will be full of surprises. Keep learning!

You might also like