Group_21-Report
Group_21-Report
1.1 Overview
● https://www.techtarget.com/whatis/definition/deepfake
Provides an overview of deepfake technology, its
underlying AI methods, and its applications.
● https://www.gao.gov/assets/gao-20-379sp.pdf This report
discusses the dangers of deepfake technology, including
misinformation, identity theft, and its impact on national
security.
● https://link.springer.com/article/10.1007/s10462-024-
10810-6 explores how deepfakes contribute to financial
fraud, cybercrime, and the challenges of detection.
● https://www.ft.com/content/61e4d68a-c7e4-4419-a5fe-
c8a72c9cb7c6 Examines how deepfake technology is
misused in cyberbullying, defamation, and privacy
violations.
● https://wires.onlinelibrary.wiley.com/doi/10.1002/
widm.1520 Reviews AI-driven detection methods,
including CNNs and RNNs, used to identify deepfake
content.
● https://ar5iv.labs.arxiv.org/html/2003.09234 analyzes
forensic techniques used to detect deepfake
inconsistencies, such as unnatural facial expressions and
lighting anomalies.
● https://behavioralsignals.com/the-duality-of-ai-and-the-
growing-challenge-of-deepfake-detection Discusses the
role of blockchain in verifying the authenticity of digital
media and preventing deepfake manipulation.
● https://www.trendmicro.com/vinfo/vn/security/news/
cyber-attacks/unusual-ceo-fraud-via-deepfake-audio-
steals-us-243-000-from-u-k-company an instance of
deepfake
https://www.fortinet.com/resources/cyberglossary/deepfake
https://www.fortinet.com/resources/cyberglossary/deepfake
spoken words.
Audio Analysis :
lip-sync errors.
https://ccoe.dsci.in/blog/deepfake-detection
1. FaceForensics++:
FaceForensics++ is a high-quality dataset designed for
evaluating deepfake detection methods. It includes video
sequences manipulated using various deepfake generation
techniques, such as FaceSwap, Deepfakes, and F2F. The
dataset contains more than 1,000 videos of real and
manipulated faces with high resolution and diverse scenes,
allowing models to generalize well across different settings.
https://github.com/ondyari/FaceForensics
2. DFDC (Deepfake Detection Challenge):
The DFDC dataset, created for the Deepfake Detection
Challenge by Facebook, includes a large set of videos and
images that cover a wide variety of actors and deepfake
techniques. The dataset is divided into training, validation, and
testing sets and includes both synthetic and real media with
extensive metadata, making it a valuable resource for
evaluating detection models.
https://ai.meta.com/datasets/dfdc
3. Other Datasets:
In addition to FaceForensics++ and DFDC, other datasets
such as Celeb-DF, DeepFake-TIMIT, and VGGFace2 can also be
used. These datasets contain a mix of real and manipulated
images or videos across a wide range of human faces and
expressions.
2. Video Processing:
- For video data, frames are extracted at consistent intervals.
These frames are then processed similarly to image data for
input into the model.
- Temporal features can be extracted through techniques like
optical flow or by leveraging 3D convolutions, which capture
the temporal relationship between frames.
5. Transformer-based Models:
Transformer models, such as Vision Transformers (ViTs) and
Spatio-Temporal Transformers, have gained attention for their
ability to model complex relationships in both space and time.
These models have been increasingly applied in deepfake
detection tasks, achieving state-of-the-art results.
6. Pretrained Models:
Using pre-trained models like EfficientNet, ResNet50, or
InceptionV3 (fine-tuned for the specific deepfake detection
task) can accelerate model training and improve detection
accuracy, especially when training data is limited.
2. Model Training:
- The selected model is trained using a supervised learning
approach, with labeled real and fake images or videos. During
training, the model learns to minimize a loss function (e.g.,
binary cross-entropy for binary classification tasks).
- Optimizers like Adam or SGD (Stochastic Gradient Descent)
are used to update the model's weights.
- Early stopping and dropout techniques may be employed to
prevent overfitting.
https://github.com/iperov/DeepFaceLab
3. Hyperparameter Tuning:
The model's hyperparameters (e.g., learning rate, batch size,
number of layers, etc.) are fine-tuned using grid search or
random search techniques. Hyperparameter optimization can
be done using cross-validation on the validation set to achieve
the best model performance.
After the model is trained and tested, the following metrics are
typically used to evaluate its performance:
1. Accuracy
The overall proportion of correctly classified samples (both
real and fake) relative to the total number of samples.
3. F1-Score:
The harmonic mean of precision and recall. This metric is
especially useful in imbalanced datasets, where one class (real
or fake) may be more prevalent than the other.
4. ROC-AUC Curve:
The Receiver Operating Characteristic curve plots the true
positive rate against the false positive rate at various threshold
settings. The Area Under the Curve (AUC) quantifies the
model’s ability to distinguish between real and fake samples.
5. Confusion Matrix:
A confusion matrix can be used to visualize the model's
predictions, providing insights into false positives, false
negatives, true positives, and true negatives.
Existing Tools
Microsoft Video Authenticator:
Microsoft’s Video Authenticator analyzes photos and videos
to determine the likelihood that the media has been artificially
manipulated. It provides a percentage score to help users
understand the authenticity of media content.
- [Microsoft Video Authenticator]
(https://www.microsoft.com/en-us/ai/ai-lab-video-authenticator)
Deepwater Scanner:
This app is designed to identify deepfakes in videos. It scans
media and checks for signs of manipulation like unnatural
facial expressions or mismatched lighting.
- [Deepware Scanner App](https://www.deepware.ai/)
---
International Cooperation
International bodies, such as the United Nations, have also
begun discussions on regulating the development and spread
of deepfake technology. Global agreements may be needed to
standardize the legal approach to deepfakes, considering their
cross-border nature.
---
6.3 User Awareness and Education
- AI Literacy Initiatives:
AI literacy is becoming a key area of focus in digital
education, helping users understand the basics of AI and how
technologies like deepfakes work. Platforms like Coursera and
edX offer courses that teach people about deepfake
technology and detection.
- [Coursera AI Courses](https://www.coursera.org/courses?
query=artificial%20intelligence)
- [edX AI Courses](https://www.edx.org/learn/artificial-
intelligence)
Debunking Deepfakes
- Fact-Checking Organizations:
Fact-checking organizations such as PolitiFact and
FactCheck.org have started to include deepfake detection in
their toolkits, helping users verify the media they encounter
online.
- [PolitiFact](https://www.politifact.com)
- [FactCheck.org](https://www.factcheck.org)