AD-17 documentation report projects
AD-17 documentation report projects
By
Dr P. Rambabu
Professor
School of Engineering
MALLA REDDY UNIVERSITY
Maisammaguda, Dulapally, Hyderabad, Telangana 500100
2024-25
Department of Computer Science and Engineering
CERTIFICATE
This is to certify that the project report entitled “Vehicle Monitoring System In A Gated
Community”, submitted by Sura Ashwanth Reddy(2111CS010073), Bhavana Anusha
(2111CS010052),Mohammad Arbaz ali(2111CS010057),Shaik Azeed(2111CS010077),
Damacharla Aravind(2111CS010056) towards the partial fulfillment for the award of
Bachelor’s Degree in Technology from the Department of Computer Science and
Engineering, Malla Reddy University, Hyderabad, is a record of bonafide work done by
him/ her. The results embodied in the work are not submitted to any other University or
Institute for award of any degree or diploma.
External Examiner
i
DECLARATION
We hereby declare that the project report entitled “Vehicle Monitoring System In A
Gated Community.” Has been carried out by us and this work has been submitted to the
Department of Computer Science and Engineering, Malla Reddy University, Hyderabad in
partial fulfillment of the requirements for the award of degree of Bachelor of Technology. We
further declare that this project work has not been submitted in full or part for the award of any
other degree in any other educational institutions.
Place: Hyderabad
Date:
ii
ACKNOWLEDGEMENT
We extend our sincere gratitude to all those who have contributed to the completion of this
project report. Firstly, we would like to extend our gratitude to Dr. V. S. K Reddy, Vice-
Chancellor, for his visionary leadership and unwavering commitment to academic excellence.
We would also like to express my deepest appreciation to our project guide Dr P.Rambabu,
Professor, whose invaluable guidance, insightful feedback, and unwavering support have been
instrumental throughout the course of this project for successful outcomes.
We are also grateful to Dr. Shaik Meeravali, Head of the Department of Computer Science and
Engineering, for providing us with the necessary resources and facilities to carry out this project.
We are deeply indebted to all of them for their support, encouragement, and guidance, without
which this project would not have been possible.
iii
ABSTRACT
This project aims to develop a deep learning-based system for monitoring vehicle movement in
gated communities to enhance security and operational efficiency. Utilizing the YOLO (You
Only Look Once) algorithm for real-time object detection and tracking, the system can
accurately identify and monitor vehicles at entry and exit points, as well as within community
spaces. The implementation leverages tools like TensorFlow and OpenCV for model training
and image processing, with an interactive Streamlit interface for a seamless monitoring
experience.
Key features include object detection to identify vehicles, image classification to categorize
them (e.g., cars, bikes, trucks), and vehicle tracking to count entries and exits. This data
provides valuable insights for applications such as security surveillance, traffic management,
and resource optimization. By integrating YOLO's high-speed detection capabilities with a
user-friendly Streamlit dashboard, the system enables rapid response to potential security
threats and contributes to safer residential environments.
By combining the efficiency of YOLO’s detection algorithms with a streamlined interface, the
system offers a high level of responsiveness and accuracy. This solution is ideal for
applications in community security, parking management, and emergency response,
empowering gated communities to proactively manage potential security threats and ensure a
safer, well-regulated environment for residents.
iv
CONTENTS
DESCRIPTION PAGE NO
CERTIFICATE i
DECLARATION ii
ACKNOWLEDGEMENTS iii
ABSTRACT iv
1.1 Introduction 1
1.2 Objective 1
1.3 Overview 2
4.1 Database 9
v
4.2.3 Use Case Diagram 12
4.3 Modules 15
4.5 Testing 18
4.5.1 Unit Testing 18
REFERENCES 34
vi
LIST OF FIGURES
vii
CHAPTER – 1
INTRODUCTION
1.1 Introduction
In modern gated communities, ensuring the security of residents and managing the flow of
vehicles are top priorities. Traditional manual or semi-automated monitoring methods often
struggle to provide accurate and real-time tracking, making it difficult to respond to security
threats or optimize traffic management. This project, "Vehicle Movement Monitoring in a
Gated Community Using Deep Learning," leverages the power of machine learning and real-
time data processing to provide an efficient and highly accurate vehicle surveillance system.
Utilizing the YOLO (You Only Look Once) algorithm, renowned for its high-speed object
detection capabilities, the system detects and tracks vehicles as they enter, exit, or move within
the community. The project employs a deep learning-based approach, with YOLO responsible
for vehicle detection and classification based on type, including cars, bikes, and trucks. The
system processes video feeds from cameras strategically positioned at key points within the
community and displays real-time results on an intuitive Streamlit dashboard, enabling quick
access to vehicle logs and detection alerts.
By automating vehicle monitoring, this system not only enhances security but also provides
valuable insights into vehicle flow, traffic patterns, and resource usage, supporting both
immediate response to security incidents and data-driven decision-making for long-term
community management.
1.2 Objective
Develop an automated vehicle monitoring system that ensures the security and operational
efficiency of gated communities.
Accurately detect and classify vehicles in real time as they move within the community,
using the YOLO algorithm for high-speed and reliable object detection.
Track entry and exit times and log vehicle movement to identify any unusual patterns or
unauthorized entries.
Provide a user-friendly, real-time interface via Streamlit, displaying logs, vehicle categories,
and alerts for quick and effective monitoring.
Collect and analyze data on vehicle movement patterns to assist community managers with
traffic management, resource optimization, and security insights.
1
1.3 Overview
This project, "Vehicle Movement Monitoring in a Gated Community Using Deep Learning,"
offers a sophisticated, automated solution for managing and analyzing vehicle movements
within residential communities. By combining YOLO-based real-time detection with a user-
friendly interface on Streamlit, the system brings both security and data insights to community
administrators, enhancing the quality of monitoring while minimizing manual intervention.
The system operates through a series of interdependent components, each optimized to process
data swiftly and accurately.
Here’s a high-level breakdown of the system’s components and their functions:
1.Data Collection: Video feeds from surveillance cameras capture the movement of vehicles
within the community. These feeds are continuously processed to detect and identify vehicles
in real time.
2.Vehicle Detection and Classification: Using YOLO, the system quickly identifies and
classifies different types of vehicles (cars, bikes, trucks, etc.), logging essential details such as
entry and exit times.
3.Data Processing and Storage: Each detection event is processed and stored, building a
database of vehicle movements that community managers can reference for tracking trends,
generating reports, and implementing policies.
5.Analysis and Insights: The stored data allows for comprehensive analysis, uncovering
valuable insights such as peak traffic times, vehicle density, and frequency of specific vehicle
types, which can aid in predictive modeling for security and operational needs.
2
CHAPTER–2
LITERATURE SURVEY
In recent years, gated communities and urban residential complexes have increasingly adopted
technological advancements for enhanced security and operational efficiency. Deep learning,
computer vision, and real-time data processing have proven to be effective in addressing the
challenges of vehicle monitoring in such environments. This literature survey covers key
studies, methodologies, and technologies that have informed the development of automated
vehicle monitoring systems, with a focus on object detection and tracking techniques, security
applications, and real-time implementation.
Object detection has progressed significantly with advancements in deep learning models.
Early methods relied on traditional computer vision techniques, such as feature matching and
edge detection (Viola-Jones, 2001; HOG-SVM, 2005), which were effective but limited in their
ability to generalize across complex environments. However, with the advent of deep
learning, algorithms such as CNNs and YOLO have improved detection accuracy and
processing speed, making them ideal for real-time applications.
YOLO (You Only Look Once): The YOLO family of algorithms (YOLOv1 to YOLOv5)
represents a significant leap in object detection. Developed by Redmon et al., YOLO uses a
single neural network to process an entire image, making detections in one forward pass.
Studies demonstrate that YOLOv5, in particular, offers a superior balance between detection
accuracy and speed, making it highly suitable for vehicle detection in real-time surveillance
(Redmon & Farhadi, 2018).
Application to Vehicle Detection: Several research works have validated the effectiveness of
YOLO for vehicle detection in live video feeds. For instance, Krishna et al. (2020) successfully
employed YOLOv3 to detect and classify vehicles in highway surveillance footage. The study
highlights YOLO’s ability to handle high frame rates while accurately identifying multiple
classes of vehicles, which is highly relevant to gated community
3
security where both accuracy and real-time performance are critical.
While YOLO performs well in detection tasks, tracking objects across frames requires
specialized algorithms. DeepSORT (Simple Online and Realtime Tracking) has gained
popularity for its robustness in tracking multiple objects in dynamic environments. DeepSORT
combines Kalman filtering with YOLO-based detection to maintain an ID for each detected
object, enabling continuous tracking as vehicles move across camera views. Research by
Bewley et al. (2016) demonstrated DeepSORT’s effectiveness in urban traffic monitoring,
where it successfully maintained object identities over time, even in complex scenarios with
overlapping vehicles.
A comparative study by Wu et al. (2021) evaluated the accuracy and speed of YOLOv5 against
Faster R-CNN, SSD, and RetinaNet in vehicle detection tasks. The study found that YOLOv5
outperformed these models in real-time applications, achieving higher accuracy at faster
processing speeds. This supports the choice of YOLOv5 for gated community monitoring,
where real-time performance is essential for immediate security response.
The application of computer vision in gated communities has evolved from basic motion
detection to sophisticated tracking and classification of vehicles. Studies like those by Zhang et
al. (2019) emphasize that real-time monitoring systems based on deep learning can
significantly reduce security risks. By analyzing entry and exit patterns, these systems provide
insights into unusual or unauthorized vehicle movement, which is crucial for enhancing
security.
Research in vehicle classification has shown that identifying vehicle types can aid in predictive
analytics for security (Lee et al., 2020). For instance, the system could flag a large delivery
truck entering a residential area during late hours as suspicious. Additionally, analysis of
vehicle movement patterns allows administrators to detect
4
abnormal behavior, such as repeated entries within a short period, which could indicate a
potential security threat. Studies by Wang et al. (2020) further show that deep learning-based
monitoring systems improve responsiveness to security incidents by automatically alerting
security personnel to irregularities.
OpenCV is widely adopted in surveillance projects for its versatility in handling video
processing tasks and its compatibility with various deep learning frameworks. By integrating
YOLO with OpenCV, research has shown that real-time performance can be significantly
enhanced. For example, a study by Reddy et al. (2019) on automated toll gate systems
demonstrated that the YOLO-OpenCV combination could effectively detect and log vehicle
details in real time, achieving minimal latency even on mid-range hardware.
Streamlit has emerged as a popular tool for developing quick, interactive, and user- friendly
dashboards in machine learning projects. For example, Kim et al. (2021) successfully deployed
a Streamlit-based interface for real-time monitoring of traffic in smart cities. The study
highlights Streamlit’s ease of use for developing interfaces that allow administrators to view
logs, receive alerts, and generate reports without extensive technical knowledge. This aligns
well with the requirements of gated communities, where security personnel can monitor vehicle
activity without specialized training.
While significant progress has been made in vehicle detection and tracking, several challenges
remain. Existing systems often struggle with:
Occlusion and Crowded Scenes: Detecting vehicles in densely populated areas where objects
partially overlap is still a challenge, though techniques like DeepSORT help address this to
some extent (Zhu et al., 2020).
5
Environmental Conditions: Weather conditions like rain, fog, and low lighting impact the
accuracy of vehicle detection, a challenge that most gated community solutions need to
overcome. Current research suggests that multi-sensor data fusion could improve accuracy
under adverse conditions.
Scalability and Data Privacy: Scaling the system for larger communities or multi-gate setups
remains challenging, especially when data privacy and security concerns are paramount.
Studies by Gupta et al. (2022) suggest cloud deployment with end-to-end encryption to address
scalability and privacy, though this approach requires further testing in residential applications.
The literature supports the use of YOLO for vehicle detection due to its high accuracy and real-
time capabilities, making it suitable for gated community monitoring. The combination of
YOLO with DeepSORT for tracking, OpenCV for efficient processing, and Streamlit for an
accessible interface aligns with the needs of this project, as validated by numerous studies.
Integrating these technologies provides a scalable, real-time solution that enhances security,
optimizes traffic flow, and supports decision-making with detailed data analytics.
This project builds upon these findings, leveraging the strengths of deep learning-based object
detection and tracking while addressing challenges such as occlusion and interface accessibility
through a tailored Streamlit dashboard. By implementing insights from prior research and using
YOLO and Streamlit as foundational tools, the project aims to provide a holistic, secure, and
efficient vehicle monitoring system for gated communities.
6
CHAPTER–3
SYSTEM ANALYSIS
Another commonly used solution is the CCTV-based monitoring system, where cameras are
installed at strategic points within the community, mainly at entry and exit gates. This setup
allows for basic record-keeping through video footage, which can be reviewed if an incident
occurs. However, CCTV systems still require human monitoring to detect suspicious activity in
real time, leading to potential oversight and delayed response. While CCTV provides some
levelof security deterrence, it does not provide data analytics, vehicle classification, or tracking
capabilities. Additionally, video footage review is often time-intensive and may not provide the
immediate insights needed for proactive security management.
In recent years, some communities have adopted Automatic Number Plate Recognition
(ANPR) systems with limited AI capabilities to improve security. ANPR systems combine
license plate recognition with basic AI algorithms to identify vehicle types and colors,
providing some level of tracking and classification. Although more advanced than traditional
LPR, ANPR systems still focus primarily on entry and exit points, lacking the capability to
monitor vehicle movement patterns within the community itself. They are also costly and may
require additional hardware and regular updates, making scalability an issue for larger gated
communities.
7
3.2 Proposed System
The proposed system for monitoring vehicle movement in a gated community combines
advanced deep learning algorithms with real-time data processing to enhance security and
operational efficiency. By leveraging the YOLO (You Only Look Once) object detection
algorithm, the system can accurately and quickly detect vehicles entering and exiting the
community. YOLO’s real-time detection capabilities allow for instant identification of vehicle
types, such as cars, bikes, and trucks, as well as detailed tracking of their movement within the
premises. The use of deep learning ensures high accuracy, even in challenging conditions like
varying lighting or partial obstructions. Vehicle movement is tracked using the DeepSORT
(Simple Online and Realtime Tracking) algorithm, which assigns unique identifiers to vehicles,
enabling continuous tracking across multiple camera views.
For the frontend, the system integrates a user-friendly interface built using Streamlit, which
provides real-time visualization and analytics of vehicle movements. Security personnel or
community administrators can view entry/exit logs, receive alerts about unusual behavior, and
monitor vehicle patterns through a comprehensive dashboard. This interface allows for quick
decision-making and enhances the system's operational efficiency by providing actionable
insights into peak traffic times, security threats, and overall community traffic management.
The proposed system is scalable, capable of being adapted for different community sizes, and
can be integrated with existing security measures for a more holistic approach to community
safety.
The system uses YOLO for vehicle detection and DeepSORT for tracking, offering efficient
and real-time monitoring. By utilizing Streamlit, the system provides an intuitive interface for
security teams to visualize and analyze data.
This deep learning-based solution offers accurate vehicle categorization, automatic tracking,
and real-time alerts.
It can efficiently monitor large communities with multiple entry points.The system’s
scalability and adaptability make it ideal for diverse gated communities, ensuring enhanced
security and operational oversight with minimal human intervention.
8
CHAPTER 4
4.1 Database
For this project, PostgreSQL is an ideal database choice due to its ability to handle structured
data efficiently and support complex queries. The database will consist of key tables such as
vehicle_ counts, which tracks the number of vehicles entering and exiting the gated
community by vehicle type, and vehicle_entry_exit_logs, which records detailed logs of
each vehicle's entry and exit events, including timestamps. PostgreSQL’s ACID compliance
ensures that transactions, such as vehicle count updates and logging of vehicle movements, are
completed reliably, preventing datainconsistencies and ensuring accurate real-time reporting.
4.2 Design
9
1. Input Source-Video Capture:
3. Vehicle Classification:
Vehicle Detection Model: Identifies different types of vehicles from the video stream.
10
4.2.2 Data Flow Diagram
11
4.2.4 Use Case Diagram
12
4.2.5 Sequence Diagram
13
4.2.6 Activity Diagram
14
4.3 Modules
2. Vehicle Detection and Classification Module: Using the YOLO object detection
algorithm, this module identifies and classifies vehicles in each frame, categorizing detected
objects (e.g., car, truck, motorcycle) and passing data to the tracking module. Tools: YOLO,
TensorFlow or PyTorch.
3. Vehicle Tracking and Counting Module: This module tracks vehicles across frames,
assigns unique identifiers, monitors vehicle movement through entry and exit points, and
counts each vehicle type based on these events. Tools: YOLO-based tracking, OpenCV.
4. Data Storage Module: This module stores real-time data on vehicle counts, types, and
timestamps of entry and exit events in PostgreSQL, ensuring efficient data retrieval and
updates for new entries. Tools: PostgreSQL.
6. Reporting and Analytics Module: Generates periodic reports and analyzes vehicle
movement data for security and management, identifying patterns, peak times, and trends
for future planning. Tools: SQL queries, Pandas.
7. Security and Access Control Module: Ensures secure access to the monitoring system
by authenticating users, authorizing access, and logging user interactions to maintain secure
data access. Tools: Streamlit or Flask authentication, database access control.
Wide Field of View: Cameras with at least a 90° field of view to cover broad entry
andexit zones.
2. Processing Hardware
15
High-Performance Computer (Edge Server): A local server or powerful
computer isrequired for real-time video processing and object detection tasks.
GPU: Dedicated GPU for running YOLO-based object detection models (e.g.,
NVIDIAGTX 1080 or higher, or NVIDIA Jetson devices for smaller installations).
Storage: SSD (512GB or higher) for fast read/write speeds and handling large
video data files, plus additional HDD storage for long-term data (1TB or more).
Access Control Devices: Card readers or RFID devices at entry/exit points for
securitypersonnel or authorized individuals.
OpenCV: For capturing, preprocessing, and handling video frames in real time.
16
3.Vehicle Tracking and Counting
17
4.5 Testing
Testing is essential to ensure that the Vehicle Movement Monitoring System functions
correctly, providing accurate detection, tracking, and data logging for vehicles within the
gated community. The testing phase is divided into three primary levels: Unit Testing,
Integration Testing, and System Testing.
Focus Areas:
o YOLO Vehicle Detection: Verify the YOLO model’s ability to detect vehicles
in each frame accurately, and validate that the classification logic works
across various vehicle types.
Objective: To test the interaction between multiple modules to ensure they work
together as a cohesive system.
Focus Areas:
o Vehicle Detection and Tracking: Validate that the detection and tracking
modules work together smoothly, accurately detecting and assigning unique
IDs to vehicles and maintaining their continuity across frames.
o Data Flow from Detection to Database: Test the flow of data from the
detection/tracking modules to the database, ensuring that vehicle counts and
logsare recorded accurately in real-time.
18
maintaining consistent information flow between components.
4.5.3 System Testing
Focus Areas:
o Performance under Load: Test the system’s response and accuracy under high
traffic scenarios, where multiple vehicles are entering and exiting simultaneously,
to ensure stability and speed.
o Data Accuracy and Real-Time Display: Ensure the Streamlit frontend accurately
reflects vehicle counts and logs in real time, providing users with a reliable
monitoring experience.
19
CHAPTER 5
SOURCE CODE
YOLO5:
YOLO-specific modules
Usage:
"""
#Importing libraries
import argparse
import contextlib
import os
import platform
import sys
import torch
import torch.nn as nn
# Set up paths
FILE = Path(__file__).resolve()
20
sys.path.append(str(ROOT)) # add ROOT to PATH
if platform.system() != 'Windows':
select_device, time_sync)
try:
except ImportError:
thop = None
class Detect(nn.Module):
super().__init__()
21
self.nl = len(anchors) # number of detection layers
z = [] # inference output
for i in range(self.nl):
wh = (wh.sigmoid() * 2) ** 2 * self.anchor_grid[i]
22
xy, wh, conf = x[i].sigmoid().split((2, 2, self.nc + 1), 4)
wh = (wh * 2) ** 2 * self.anchor_grid[i]
d = self.anchors[i].device
t = self.anchors[i].dtype
class Segment(Detect):
23
self.m = nn.ModuleList([nn.Conv2d(x, self.no * self.na, 1) for x in ch]) # output conv
p = self.proto(x[0])
x = self.detect(self, x)
return (x, p)
class BaseModel(nn.Module):
y, dt = [], [] # outputs
for m in self.model:
if m.f != -1:
if profile:
self._profile_one_layer(m, x, dt)
x = m(x)
if visualize:
return x
24
def fuse(self):
for m in self.model.modules():
delattr(m, 'bn')
m.forward = m.forward_fuse
return self
self = super()._apply(fn)
m = self.model[-1]
m.stride = fn(m.stride)
if isinstance(m.anchor_grid, list):
return self
25
Demo.py:
import os
import cv2
import torch
import tempfile
import streamlit as st
from track import * # Assuming `track.py` contains the necessary tracking functions
def display_footer():
"""Displays a footer with project details."""
footer_text = """
<div style="background-color:#F5F5DC; padding: 10px; border-radius: 5px;">
<center>
<h2>Vehicle Monitoring System in a Community</h2><br>
<h3>Done By:</h3><br>
<h3>Azeed.Sk</h3><br>
<h3>Anusha.B</h3><br>
<h3>Arbaz Ali.Mohammad</h3><br>
<h3>Ashwanth Reddy.S</h3><br>
<h3>Aravind.D</h3><br>
<h3>Guided By: DR. Ramababu.P Prof. Mruh</h3><br>
<h3>All Copyrights @ Reserved 2024</h3>
</center>
</div>
"""
st.markdown("<hr>", unsafe_allow_html=True)
st.markdown(footer_text, unsafe_allow_html=True)
def main():
st.title('Vehicle Monitoring System in a Gated Community')
st.markdown('<h3 style="color: red">with YOLOv5 and Deep Learning</h3>',
unsafe_allow_html=True)
# Sidebar settings
st.sidebar.markdown('---')
st.sidebar.title('Settings')
if custom_class:
assigned_class_id = []
selected_classes = st.sidebar.multiselect('Select custom classes', names)
for cls in selected_classes:
assigned_class_id.append(names.index(cls))
# Hyperparameter settings
confidence = st.sidebar.slider('Confidence', min_value=0.0, max_value=1.0, value=0.5)
line = st.sidebar.number_input('Line position', min_value=0.0, max_value=1.0, value=0.6,
step=0.1)
st.sidebar.markdown('---')
# Status display
status = st.empty()
stframe = st.empty()
if video_file_buffer is None:
status.markdown('<font size= "4">**Status:** Waiting for input</font>',
unsafe_allow_html=True)
else:
status.markdown('<font size= "4">**Status:** Ready</font>', unsafe_allow_html=True)
# FPS display
fps = st.columns(4)[0]
with fps:
st.markdown('**FPS**')
fps_text = st.markdown('__')
27
reset()
opt = parse_opt()
opt.conf_thres = confidence
opt.source = f'videos/{video_file_buffer.name}'
# Display footer
display_footer()
if __name__ == '__main__':
main()
import os
import sys
import cv2
import time
import torch
import shutil
import platform
import argparse
import streamlit as st
from pathlib import Path
from yolov5.utils.general import (
LOGGER, check_img_size, non_max_suppression, scale_boxes,
increment_path, check_imshow, xyxy2xywh
)
from yolov5.utils.torch_utils import select_device, time_sync
from yolov5.utils.plots import Annotator, colors
from yolov5.models.common import DetectMultiBackend
from yolov5.utils.dataloaders import LoadImages, LoadStreams
from deep_sort.deep_sort import DeepSort
from deep_sort.utils.parser import get_config
28
# Global Variables for Counting Vehicles
data_car, data_bus, data_truck, data_motor = [], [], [], []
already = []
line_pos = 0.6
# Initialize DeepSORT
cfg = get_config()
cfg.merge_from_file(opt.config_deepsort)
deepsort = DeepSort(deep_sort_model, max_dist=cfg.DEEPSORT.MAX_DIST,
max_iou_distance=cfg.DEEPSORT.MAX_IOU_DISTANCE,
max_age=cfg.DEEPSORT.MAX_AGE, n_init=cfg.DEEPSORT.N_INIT,
nn_budget=cfg.DEEPSORT.NN_BUDGET, use_cuda=True)
if pt:
model.model.half() if half else model.model.float()
# Load data
dataset = LoadStreams(source, img_size=imgsz, stride=stride, auto=pt and not jit) if
webcam else LoadImages(source, img_size=imgsz, stride=stride, auto=pt and not jit)
bs = len(dataset) if webcam else 1
vid_path, vid_writer = [None] * bs, [None] * bs
29
txt_file_name = source.split('/')[-1].split('.')[0]
txt_path = str(Path(save_dir)) + '/' + txt_file_name + '.txt'
30
global data_car, data_bus, data_truck, data_motor, already
center_y = int(box[1] + (box[3] - box[1]) / 2)
if center_y > (h * line_pos) and id not in already:
already.append(id)
if label == 'car': data_car.append(id)
elif label == 'bus': data_bus.append(id)
elif label == 'truck': data_truck.append(id)
elif label == 'motorcycle': data_motor.append(id)
# Main Execution
if __name__ == '__main__':
opt = parse_opt()
with torch.no_grad():
detect(opt)
31
CHAPTER-6
EXPERIMENTAL RESULTS
32
CHAPTER-7
Conclusions
Future Enhancements
Here are five focused future enhancements for your vehicle movement monitoring
system:
Multi-camera Integration
Expanding the system to include multiple cameras at different entry and exit
points or other critical areas would improve the accuracy of vehicle tracking and provide a
more comprehensive view of vehicle movements.
These targeted improvements would enhance the system's overall accuracy, responsiveness,
and usability in a gated community.
33
REFERENCES
1.S. Ren, K. He, R. Girshick, and C. Sun, "Faster R-CNN: Towards Real-Time
Object Detection with Region Proposal Networks," IEEE Transactions on Pattern Analysis
and Machine Intelligence, vol. 39, no. 6, pp. 1137-1149, June 2017.
2.J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You Only Look Once:
Unified, Real- Time Object Detection," IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), pp. 779-788, 2016.
4.G. Bradski and A. Kaehler, Learning OpenCV: Computer Vision with the
OpenCV Library, O'Reilly Media, 2008.
5.Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C. Y., & Berg, A. C.
"SSD: Single Shot Multibox Detector," European Conference on Computer Vision (ECCV),
pp. 21-37, 2016.
6.Ge, Z., Liu, S., Wang, F., Li, Z., & Sun, J. "YOLOX: Exceeding YOLO Series in
2021,"
arXiv preprint arXiv:2107.08430, 2021.
7.Zhu, X., Lyu, S., Wang, X., & Zhao, Q. "TMM: Deep Structured Models for Multi-
Object Tracking," IEEE Transactions on Multimedia, vol. 19, no. 4, pp. 750-762, 2017.
8.Bojarski, M., Testa, D. D., Dworakowski, D., Firner, B., Flepp, B., Goyal, P., ... &
Zhang,
X. "End to End Learning for Self-Driving Cars," arXiv preprint arXiv:1604.07316, 2016.
9.Goodfellow, I., Bengio, Y., & Courville, A. Deep Learning, MIT Press, 2016.
34