Smart Attendance System
Smart Attendance System
PROJECT SYNOPSIS
OF MAJOR PROJECT
BACHELOR OF TECHNOLOGY
SUBMITTED BY
Ayushi Singh – 2101331540028
I hereby certify that the work which is being presented in the major project
report entitled " REAL TIME OBJECT DETECTION " in fulfilment of the
requirement for the award of the Degree of Bachelor of Technology in
Department of CSE(DATA SCIENCE) of Noida Institute Of Engineering and
Technology, Greater Noida, U.P. is an authentic record of my own work carried
out during VIIIth semester.
Signature of:
AYUSHI SINGH
NAME OF THE STUDENT
FIGURE INDEX
1.3 Scope:
This project has significant potential in fields like:
● Surveillance systems – Essential for real-time threat detection, intrusion alerts, and
enhanced public safety.
● Smart cities – Valuable for traffic management, violation detection, and urban
planning enhancements.
2.1 Introduction:
2.1.1 Purpose:
The primary objective of this project is to design and implement a real-time object detection
system capable of identifying, localizing, and tracking multiple objects from live video feeds
or camera inputs. This technology aims to enhance automation, safety, and operational
efficiency in various domains. The system uses advanced machine learning and computer
vision techniques to process visual data instantly and provide meaningful insights.
● Automated Detection: Detect and classify objects (e.g., people, vehicles) in real time
without human input.
● Better Decision-Making: Provide real-time data for faster, smarter decisions in areas
like traffic and retail.
● Flexible & Scalable: Adapt to different environments and support multiple camera
inputs.
● Improved Safety: Monitor industrial areas to prevent accidents and automate checks.
● Smart Applications: Power systems like self-driving cars, smart homes, and
automated retail.
● Live Object Detection: Detect and classify objects (e.g., people, vehicles) in real
time from video feeds.
● Real-Time Alerts: Trigger alerts or notifications when specific objects or events are
detected.
● Edge Device Processing: Run detection models efficiently on edge devices (e.g.,
Jetson Nano, Raspberry Pi).
● Database Logging: Store detection results with timestamps and metadata in a
structured database.
● Visualization Dashboard: Display live video with detection overlays and object info
on a user interface.
.
● YOLO Model:The core deep learning model used to detect and classify objects in
real-time from video streams.
● Edge Computing Device:Hardware (e.g., NVIDIA Jetson Nano) used for local
processing of detection tasks, ensuring low-latency performance.
● Video Capture System:Cameras (IP cams, surveillance cameras) that provide
real-time video feeds for object detection.
● Database for Data Logging: Stores detection results, including object type,
confidence scores, and timestamps for later analysis.
● User Interface (UI) & Visualization:A dashboard that displays live video feeds,
detected objects with their locations, and real-time alerts
Developer Handles backend, model updates & Full system and backend
system integration access.
● Hardware Requirements:
o Computer/Laptop – Windows based system.
o Camera – Captures real-time video (e.g., USB, IP, or CSI cameras).
o Edge Device – Runs YOLO locally (e.g., Jetson Nano, Raspberry Pi).
o GPU/CPU – High-performance processing (e.g., NVIDIA RTX GPU or
desktop CPU).
o Storage – Saves detection data and logs (SSD/HDD).
● Software Requirements:
o Python 3.x – The programming language used.
o Deep Learning Framework – TensorFlow or PyTorch for YOLO model.
2.2.6 Constraints:
The project must ensure real-time processing with minimal latency (under 100ms/frame) and
maintain high detection accuracy (80-90%) while handling scalable data storage and secure
user authentication. It should also meet performance, security, and availability requirements,
including efficient resource usage, data retention, and backup mechanisms
These constraints can be categorized into:
I. Hardware Constraints:
o GPU/CPU Usage: Must support GPU acceleration for real-time
detection, with CPU fallback for edge devices.
o Fast Storage: Requires high-performance storage (e.g., SSD) for quick
data handling.
o Power Efficiency: Needs optimization for low power consumption,
especially in battery-powered deployments.
o System Requirements:
Actor Description
Admin Manages users, devices, system settings, and reports.
Fig: 2.1
2.2.8 Assumptions and Dependencies The Real Time Object Detection System relies
on several assumptions and dependencies that must be met for the system to function
correctly.
❖ Assumptions:
● Hardware Assumptions –
o Dedicated GPU (e.g., NVIDIA RTX) for real-time inference.
o High-resolution camera (1080p or above) for input.
o Minimum 8GB–16GB RAM for smooth processing.
o SSD storage for model files and runtime data.
● Software Assumptions –
o The system will be installed on Windows.
o Python (3.x) is installed and properly configured.
o Required libraries like PyTorch, OpenCV, and YOLO are available.
o OS supports GPU drivers (e.g., CUDA for NVIDIA GPUs).
o Internet access is available for initial model downloads and
updates.
● User Assumptions –
o Users have basic knowledge of running Python scripts and handling
command-line tools.
o Users can install necessary dependencies and configure the
environment.
o Users understand how to operate the camera and interpret detection
results.
● Security & Privacy Assumptions –
o User data and images stored in the database will be kept secure.
o The system assumes video data is processed locally to maintain user
privacy.
o It is assumed that access to the system is restricted to authorized
users only.
❖ Dependencies:
● Hardware Dependencies –
o CUDA-compatible GPU for fast inference
o Camera (USB/IP/CSI) for live video input
o Multi-core CPU for general processing
o 8GB–16GB RAM for smooth performance
o SSD for fast data access
o Stable power supply for continuous operation
● Software Dependencies –
o Python (3.x) – programming language for development.
o PyTorch – framework for running YOLO models.
o OpenCV – for video capture and image processing.
o YOLO (v5/v8) – object detection model.
o CUDA & cuDNN – GPU acceleration (for NVIDIA hardware).
● Database Dependencies –
o MySQL Database
ii. Non-Functional Features:The system ensures high performance with low latency,
scalability, and reliability, providing seamless real-time object detection across
various environments.
a. Performance & Speed Optimization –
✔ GPU Acceleration: Uses CUDA-compatible GPUs for faster inference.
✔ Model Optimization: Reduces model size with pruning and
quantization for improved speed.
✔ Parallel Processing: Utilizes multi-threading and batch processing for
faster frame handling.
b. Security & Data Privacy –
✔ Local Processing: Data is processed locally to ensure privacy.
❖ Internal Communication:
o Camera → Detector: Capture
o Detector → Post-process: Inference
o Post-process → UI/Alerts/Logs: Dispatch
✔ FPS (Frames Per Second) – Measures real-time processing speed. Higher is better.
✔ Encrypted Communication – Use HTTPS, SSL/TLS for video streams and API
calls.
✔ Secure Data Storage – Encrypt stored files, logs, and detection data; use secure
cloud storage.
Fig: 3.1
3.2 ER Diagram:
Fig: 3.2
Fig: 3.4
Fig: 3.6.1
Fig: 3.6.2
Fig: 3.6.3
3.7 Activity Diagram:
Fig: 3.7
Fig: 3.8
3.9 Sequence Diagram:
Fig: 3.8
Chapter 4 – System Implementation
4.1 Coding:
4.1.1 Import:
4.3 Snapshot:
Chapter 5 – Conclusion and future Scope
5.1 Conclusion:
The implementation of real-time object detection using YOLO (You Only Look Once) and
OpenCV has demonstrated the capability to accurately and efficiently detect multiple objects
within video streams. YOLO's fast inference time, combined with OpenCV’s flexible image
processing capabilities, enabled a seamless and responsive detection system suitable for
real-world applications. This project highlights the practicality of deep learning-based object
detection for use cases such as surveillance, autonomous systems, and smart vision
applications. Future improvements may include model optimization, deployment on edge
devices, and integration with other AI-based decision-making systems.
The project demonstrated high efficiency by achieving real-time object detection with
minimal latency and accurate results. Leveraging YOLO's speed and OpenCV’s processing
capabilities, it maintained a high frame rate while efficiently utilizing system resources. The
use of lightweight model variants ensured low memory consumption, making the system
suitable even for hardware-constrained environments. Overall, the architecture proved
scalable, responsive, and reliable for practical deployment.