MobilenetV2 (Quantization)
MobilenetV2 (Quantization)
It is Neural Network Architecture Which is specially Optimized for the Edge devices
like Mobile.The architecture delivers high accuracy results while keeping the
parameters and mathematical operations as low as possible to bring deep neural
networks to mobile devices.
Earlier their was older version called MobilenetV1 which was designed for the same
purpose of edge device computing.
The main goal of these two Architecture was to Reduce two factors listed below
Architecture of MobilenetV1 :
The main difference of this architecture compared to Basic Convolution was The
Concept of Depthwise Separable Convolution.
It consisted two Sub Blocks
And After each convolution Batch Normalization and a non linear Activation
function Relu was applied.
However MobileNetV2 an improved version of Mobilenetv1 is released which is
much faster in comparison to MobilenetV1.It uses 2 times fewer operations, has
higher accuracy, needs 30 percent fewer parameters and is about 30-40 percent
faster.
Also the non-linearities present in mobilenetv1 are removed.
Architecture of MobileNetV2 :
The intuition is that the bottlenecks encode the model’s intermediate inputs and
outputs while the inner layer encapsulates the model’s ability to transform from
lower-level concepts such as pixels to higher level descriptors such as image
categories. Finally, as with traditional residual connections, shortcuts enable faster
training and better accuracy.
In MobileNetV2, there are two types of blocks. One is residual block with stride of
1 another one is block with stride of 2 for downsizing.
But their are three unique layers for both blocks
1. The first layer is 1×1 convolution with ReLU6.
3. The third layer is another 1×1 convolution but without any non-linearity
MobileNetV1
MobileNetV2