0% found this document useful (0 votes)
35 views

IP Report Final

This provisional patent application proposes integrating a computer vision and machine learning model into a painting production line to improve productivity. Currently a prototype model trained on the YOLOv5 algorithm can identify components with 68% accuracy. The proposed system aims to develop a model that can recognize components with higher accuracy and ensure 16,000 m2 of surface area is painted per month. It describes training a model to classify images into two categories and generating performance metrics like loss, accuracy, and confusion matrices over multiple epochs to evaluate the model.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
35 views

IP Report Final

This provisional patent application proposes integrating a computer vision and machine learning model into a painting production line to improve productivity. Currently a prototype model trained on the YOLOv5 algorithm can identify components with 68% accuracy. The proposed system aims to develop a model that can recognize components with higher accuracy and ensure 16,000 m2 of surface area is painted per month. It describes training a model to classify images into two categories and generating performance metrics like loss, accuracy, and confusion matrices over multiple epochs to evaluate the model.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 20

PROVISIONAL PATENT APPLICATION

1. TITLE: Monitoring Painting production line using Computer Vision

2. TECHNICAL FIELD: Computer Vision and Machine Learning

3. BACKGROUND:

Field of the Invention:

● An object recognition system that will be implemented in the painting production


line to improve productivity.

● Propel Industries, who are constructing a brand-new manufacturing facility, is the


target of the current invention. Their project includes a paint shop production line.
Our technology will be integrated into the painting assembly line, which will
identify the objects and ensure that 16,000 m2 surface area is painted each month.

Description of the Related Art:

● Currently, a prototype for the component recognition stage has been created, for
which a custom dataset is trained using the yolov5 algorithm for five components.
The model is able to provide findings with a 68% accuracy for epochs between 10
and 15, along with details in the form of graphs.

4. OBJECTS OF THE INVENTION:

● The technique employed will guarantee that the issue statement's needs are met,
and the outcome guarantees a prototype that will satisfy the advancement of
computer vision in the field of component recognition while also delivering higher
production and an optimized cycle time.

5. SUMMARY:

● Objectives of the invention:


➢ A machine learning model that will be able to recognize the components
with higher accuracy.
➢ To make sure that the designed system ensures that a surface area of 16,000
m2 production every month.

1
6. BRIEF DESCRIPTION OF THE DRAWINGS:

2
7. DETAILED DESCRIPTION:

● As the output is discrete, the deep learning model used in this case is a classification
problem. The trained model generates a confusion matrix based on the number of
classes when tested because the output will likewise be discrete. This kind of
learning is therefore monitored. YOLOv5 is a type of algorithm that makes use of
a unique coco dataset to identify objects in real time.

● Establishing what is truly being forecasted is crucial to comprehend the YOLO


algorithm. In the end, our goal is to forecast an item's class and the bounding box
that specifies where the object is located. Four descriptors can be used to describe
each bounding box:
➢ center of a bounding box (bxby)
➢ width (bw)
➢ height (bh)
➢ value cis corresponding to a class of an object

● Shortly after YOLOv4, the most popular object identification algorithm's fifth
iteration was made public, but this time by Glenn Jocher. YOLO applied the
PyTorch deep learning framework for the first time ever.

● The working of YOLO algorithm: So, what it does is to train the model, it takes
input as the image that is the X train and the labels of the image (after creating the
bounding box) that is the Y train. Labels means the descriptors of the bounding box
as discussed above in text in csv format. After this the algorithm start its training,
as the name says “You Only Look Once”, it will make the grid in the image to be
tested and each grid it will look only once and it will check whether the center point
of the required Object’s bounding box centroid is there in that particular grid or not,
if its there it will make another label having the same four descriptors or else it will
do nothing like that it will run and that’s how it train the model by looking through
each grid only once.

3
● As we can see in Fig. 1, the first grid has no centroid of any bonding box, therefore
its labels are NULL, for the fifth grid, centroid of dog’s bounding box is there, so
label for that grid has values according to the descriptors and similarly for human’s
centroid grid. And after this all this data is trained using CNN as we can see in Fig.
2 and we get the final output when we input data for testing

● Creating a dataset: A dataset is created for two classes. For thirty epochs the model
will be able to deliver results with the accuracy of 70%.

● The dataset is created in such a way that it has two folders, namely labels and
images.

4
5
● The labels for the given images are created with makesenseai.com

● Makesense.ai is a free-to-use online tool for labeling photos

● Labels are created for the different classes.

6
● The labels of the images are in YOLO format i.e txt format.

● The Dataset is trained for 30 epochs.

7
● Once it is trained for the given epochs it will generate the weights after passing
through the neural network. The weights will be used for the identification of the
object for the given input image.

● The output predicted will be stored in the folder runs/train.

8
● Identification for a set of different images is also done which are used for validation.

● Here zero refers to Mesh clamp and one refers to Feed Chute.

● For the given input image, the predicted output will be shown with the confidence level
of the prediction.

9
● F1 Score: The F1 score is defined as the harmonic mean of precision and recall. An F1
score is considered perfect when it's 1, while the model is a total failure when it's 0.

10
● No model can be perfect, for the components taken the f1 curve is generated. This curve
combines precision (PPV) and Recall (TPR) in a single visualization. The higher the y-axis
the model performs the better.

● It is seen from the graph that for feed chute it is higher so the model performs better in
identifying feed chute well.

● Confusion Matrix: It is a type of a table used to measure the performance of a


classification algorithm. Usually confusion matrix will be like:

11
● For the problem the classification is done for two classes namely: Mesh clamp and Feed
Chute. A confusion matrix is generated as:

12
● Precision-Confidence curve: The Precision is calculated as the ratio between the
number of positive samples correctly classified to the total number of samples
classified as positive (either correctly or incorrectly)

13
● Recall-Confidence Curve: The recall is calculated as the ratio between the number of
positive samples correctly classified as positive to the total number of positive samples.

14
● Precision-Recall Curve: A precision-recall (PR) curve shows the trade-off between the
precision and recall values for different thresholds.

15
● Labels: For the classification of two components the labels are made for a certain set of
images. This labels are marked in graphs for the specific instances.
● It is seen that Mesh clamp occurs for 25 instances and Feed Chute for 15 instances.
● In the below graph it is seen that all those instances are spread around it then it is classified
separately into two classes.

16
● Results:

● Box_loss: The box loss represents how well the algorithm can locate the center of an object
and how well the predicted bounding box covers an object. The graph shows the box loss
that happened for 30 epochs.

● Object_loss: The train loss represents the confidence of object presence. The graph shows
the object loss that happened for 30 epochs.

● Class_loss: The train loss represents the confidence of the different classes present in the
model. The graph shows the class loss that happened for 30 epochs.

● The losses mentioned above are represented in the form graph for the images of both train
and val.

● Precision is a metric that quantifies the number of correct positive predictions made.

● Recall is a metric that quantifies the number of correct positive predictions made out of all
positive predictions that could have been made.

17
● Mean average precision(mAP): AP (Average precision) is a popular metric in measuring
the accuracy of object detectors like Faster R-CNN, SSD, etc. Average precision computes
the average precision value for recall value over 0 to 1.

Teachable machine:

● The Google Teachable Machine is an online open-source environment which is used to


develop and train machine learning and deep learning supervised models without using any
programming language.
● Teachable machine was used to gain perspective on how a classification model should
work and to compare the desired output a model should have.
● It was used for research purposes to compare our model with the pre-existing model
developed by google.

● Here images of two different classes are fed into the machine.

● Once the model is trained for a given set of epochs the model will generate a desired output.

● For the given model there were no false positives given as input so the accuracy was high
and was able to deliver the desired output.

● Accuracy per class and per epoch is generated and a confusion matrix is formed.

18
● Loss per epoch is also calculated.

8. CLAIMS:

● An object recognition technology which when integrated along with the painting
manufacturing line will ultimately boost the production.

19
9. ABSTRACT:

● Recognition of the component plays an important role in the painting production


for an automated production line. Traditional object recognition adopts manual
recognition or primitive machine vision technology which has the issues of high
recognition error leading to low production rates. In order to address the problem
of the low recognition recall and accuracy we propose a novel method of
component recognition based on machine vision combining with deep learning.
This technique will ensure the needs for the problem statement and the result
promises that a prototype which will fulfill the advancement of computer vision in
the field of component recognition and providing higher production and an
optimized cycle time

20

You might also like