1.2 Background
Traditional methods of scam detection often rely on
manual review or rule-based systems, making them
less adaptive to evolving scam tactics. In recent years,
with the rapid advancements in artificial intelligence
and machine learning technologies, deep learning
models have emerged as powerful tools for predictive
analytics and pattern recognition. The ANN and
CNN-LSTM models stand out as popular
architectures known for their effectiveness in
handling complex data patterns and sequential
information. The motivation behind this project stems
from the growing interest in exploring the capabilities
of different deep learning architectures for predictive
tasks. While ANN models have been extensively used
in various machine learning applications, the
integration of CNNs and LSTMs in the form of CNN-
LSTM models has shown promising results in
processing sequential data, such as time series, audio
signals, and natural language text.
1.3 Need
In real-world predictive tasks, paving the way for
informed decision-making and model selection in
future endeavours.
Understanding the strengths and weaknesses of
ANN and CNN-LSTM models is crucial for selecting
the most appropriate architecture for a given
predictive task. ANN models excel at capturing
complex relationships in structured data but may
struggle with sequential information due to the lack
of temporal context. On the other hand, CNN-LSTM
models leverage the spatial hierarchies learned by
CNNs and the long-term dependencies captured by
LSTMs, making them well-suited for tasks where
both spatial and temporal features are essential. By
conducting a comparative analysis of ANN and CNN-
LSTM models in this project, we aim to gain valuable
insights into their performance characteristics,
predictive accuracy, and computational efficiency.
This comparative study will provide a deeper
understanding of how these architectures process and
learn from data, ultimately guiding us in selecting the
most efficient and effective model for our specific
predictive task. The findings from this project have
the potential to advance the field of deep learning and
predictive modelling by offering empirical evidence
and practical guidance on choosing the optimal
architecture for similar tasks in the future. Through
this exploration, we aim to contribute to the ongoing
research efforts aimed at enhancing the capabilities of
deep learning models and their applications in diverse
domains.
1.4 Technology
In real-world predictive tasks, paving the way for
informed decision-making and model selection in
future endeavours.
Technology Stack for the Project: The project,
encompassing the implementation of Artificial
Neural Network (ANN) and Convolutional Neural
Network Long Short-Term Memory (CNN-LSTM)
models for predictive tasks, was executed within the
Anaconda environment, which provides a
comprehensive platform for managing Python
packages and environments. The following
technologies and libraries were utilized for the
successful development and analysis of the models:
• Python Programming Language: Python
played a central role in coding the machine
learning models, data manipulation, and analysis
tasks due to its simplicity and extensive libraries
for deep learning.
• NumPy and Pandas: NumPy and Pandas were
utilized for efficient numerical computations,
array operations, and data manipulation tasks,
including data preprocessing and cleaning.
• Matplotlib and Seaborn: Matplotlib and
Seaborn were employed for data visualization,
enabling the creation of informative plots and
graphs to analyse model performance and results
effectively.
• Scikit-Learn: Scikit-Learn was used for model
evaluation and metrics calculation, including
accuracy scores, classification reports, and
confusion matrices, providing valuable insights
into the model performance.
• Jupyter Notebooks Jupyter Notebooks served as
the interactive development environment for
coding, experimenting with different models, and
documenting the project workflow. The notebook
format facilitated seam- less integration of code,
visualizations, and explanatory text.
• Seaborn and Matplotlib: Seaborn and
Matplotlib libraries were used for data
visualization, aiding in the analysis and
interpretation of model results through various
plots and charts.
• Anaconda Environment: The project was
executed within the Anaconda environment,
which offers a comprehensive suite of tools for
data science, machine learning, and deep learning
tasks, streamlining package management and
environment setup.
• TensorFlow: TensorFlow, an open-source deep
learning library, was utilized for implementing
and training the neural network models. Its
flexibility and scalability made it suitable for