forensic tools are no longer effective (Carlini & Farid,
2020). Deepfake detectors based on convolutional
neural networks (CNNs), recurrent neural networks
(RNNs), and transformers have shown potential in
identifying tampered content, though they still face
limitations in adaptability, scalability, and robustness
(de Lima et al., 2020; Hussain et al., 2020). Moreover,
adversarial attacks continue to expose vulnerabilities
in many existing detection systems, making it
essential to design adaptive AI-powered solutions
capable of detecting deepfakes in real time and across
multiple platforms.
There is a pressing need to update detection
frameworks continuously and incorporate
sophisticated models that can learn from evolving
data patterns. This work seeks to explore the
effectiveness of hybrid AI approaches in building a
scalable and resilient deepfake detection system
aimed at preserving digital authenticity and
countering the malicious misuse of AI (Yang et al.,
2019; Hussain et al., 2020).
Problem Statement.
Deepfake technology has emerged as a serious threat
to media legitimacy, privacy, and, over time, digital
security (Rossler et al., 2019). Potential for
considerable improvements exists over existing
machine learning RM-based deepfake detection
techniques that face numerous challenges, ranging
from poor training datasets (Li et al., 2019; Zi et al.,
2020) to adversarial attacks to computationally
heavy deployments (Du et al., 2019). Moreover,
owing to the fast-changing nature of deepfake
generating technologies, detection systems must also
be constantly updated to face new threats. In this
work, we examine various transformer-based models,
RNNs and CNNs (Chollet, 2017; Huang et al., 2017)
to advance the detection of manipulated media
content by improving the effectiveness, scalability
and also robustness with the ultimate goal to develop
a state-of-the-art deepfake detector identifier. The
ultimate goal, however, is to develop AI-powered
solutions that successfully identify deepfakes, protect
digital authenticity, and avert the malicious misuse of
AI technology.
2 LITERATURE REVIEW
In November 2017, the term "deepfake" first
appeared in reference to the dissemination of explicit
content in which the faces of celebrities were
superimposed on original videos. By January 2018, a
number of websites supported by private sponsors
had introduced services that made it easier to create
deepfakes. However, because of the possible dangers
and privacy issues with deepfakes, these services
were banned within a month by websites such as
Twitter. The academic community quickly increased
its research into deepfake detection after realizing the
growing threats. FaceForensics, a comprehensive
video dataset created to train media forensic and
deepfake detection tools, was unveiled by Rössler et
al. in March 2018.
Next month, researchers at Stanford University
introduced "Deep Video Portraits," a technique that
allows for photorealistic re-animation of portrait
videos; at the same time, researchers at UC Berkeley
created a technique that transfers a person's body
movements to another person in a video; NVIDIA
advanced synthetic image generation by introducing
a style-based generator architecture for GANs; the
spread of deepfake content became apparent as search
engines indexed a large number of related web pages;
the top 10 adult platforms contained roughly 1,790
deepfake videos; adult websites contained 6,174
deepfake videos; and three new platforms were
created specifically for the purpose of deepfake
content.
The research community became very interested
in these developments, with 902 articles about GANs
published in 2018. Twelve of these papers, out of the
25 that addressed deepfake topics, were funded by
DARPA. Deepfakes have been used maliciously for
things like political instability, misinformation
campaigns, and cybercrimes in addition to explicit
content. Many detection methods have been
developed as a result of the substantial attention that
the deepfake detection field has received. A thorough
review covering every facet of deepfake research,
including available datasets, is still lacking, despite
the fact that some surveys have concentrated on
particular detection techniques or performance
evaluations.
This paper aims to fill this gap by providing a
systematic literature review (SLR) on deepfake
detection. As the security threats related to AI-
generated synthetic media become greater, deepfake
detection has generated an area of significant
importance. The advance of deepfake technology
raises serious concerns about disinformation, digital
fraud and violations of privacy because the line
between manipulated and real content continues to
blur. When it comes to drumming up authentic-
sounding voices, images, and videos that can be used
to fool people, shift public sentiment, and even
commit fraud, deepfakes have been effective.