
vides standard and easily accessible training, testing,
and validation images. This study focuses on training
a CNN model using the MNIST dataset and assess-
ing its performance on two distinct datasets(USPS
and EMNIST), progressively diverging in similarity
from MNIST. These datasets serve as unseen data for
the model, facilitating a comprehensive evaluation of
its ability to generalize beyond the familiar MNIST
dataset.
The fundamental components of the CNN model
consist of convolutional layers, pooling layers, and
dense layers(Dai, 2021). Additionally, the paper ex-
amines how the model’s accuracy fluctuates with al-
terations in CNN design, particularly in modifying the
convolutional layers. Furthermore, the study investi-
gates the impact on accuracy when adjusting various
learning parameters, elucidating how these modifica-
tions influence the overall performance of the CNN
model for an unseen dataset.
This literature follows the introduction in section
1. Section 2 presents comprehensives the related
work on blind validation. Section 3 describes the
methodology and datasets. Results and discussions
are detailed in section 4 and conclude with section 5.
2 LITERATURE SURVEY
A research paper was conducted on the Convo-
lutional Neural Network (CNN) model for recogniz-
ing handwritten numbers. The model was trained us-
ing the well-known MNIST dataset, which consists of
grayscale handwritten digit photographs. The CNN
model achieved an impressive validation accuracy of
98.45% on the MNIST dataset. To test the model’s
ability to handle unknown data, the researchers ran it
through a series of random photos with handwritten
and printed digits. The model achieved a reasonable
accuracy of 68.57% on this new dataset. However,
it showed limitations in recognizing numbers not part
of the training data, particularly those in non-standard
formats.(Garg et al., 2019)
The paper delves deeper into the model’s archi-
tecture, revealing that it includes four convolutional
layers, ReLU activation, and max-pooling layers - a
standard arrangement for picture classification tasks.
This study highlights that CNNs are highly effective
at recognizing handwritten digits and can generalize
to previously unexplored data. However, the model’s
performance deteriorates when it encounters data that
considerably differs from the training set.
The paper explores EEG-based emotion recog-
nition, lever- aging Convolutional Neural Network
(CNN) architectures to enhance subject-independent
accuracy. Unlike conventional methods relying on
spectral band power features, raw EEG data is utilized
after windowing, pre-adjustments, and normalization,
removing manual feature extraction and harnessing
CNN’s capacity to uncover hidden features(Cimtay
and Ekmekcioglu, 2020). A median filter further im-
proves classification accuracy. The approach achieves
mean cross-subject accuracies of 86.56 and 78.34 on
the SEED dataset for two and three emotion classes,
re- respectively. Testing the SEED-trained model on
the DEAP dataset yields a mean accuracy of 58.1.
The paper extensively evaluates CNN models
in AI-assisted COVID-19 diagnostics, spotlighting
ResNet-50 as the top performer. Through itera-
tive rounds of training and testing across diverse
datasets, the study underscores the critical impor-
tance of achieving subject-independent accuracy and
the potential of enriching training datasets to bolster
model performance. Leveraging heatmaps and ac-
tivation features provides deeper insights into CNN
model learning dynamics, guiding future advance-
ments in COVID-19 and pneumonia detection diag-
nostic systems. During the initial evaluation round,
CNN models exhibited high accuracy rates of 95.2 to
99.2 for the Level 1 testing dataset, sourced from the
same clinic but designated solely for testing. How-
ever, model performance declined significantly with
the Level 3 dataset, characterized by outlier images,
reducing mean sensitivity from 99 to 36. These find-
ings emphasize the challenges outlier data poses and
the need for strategies to mitigate their impact on di-
agnostic model performance(Talaat et al., 2023).
This research gives a method for detecting biases
in picture attribute estimations learned by convolu-
tional neural networks (CNNs)(Zhang et al., 2018).
Even with great overall accuracy, these biases might
lead to erroneous findings. The method examines
CNN’s internal representation of characteristics to de-
tect probable blind spots (missing associations) and
failure modes (incorrect relationships) induced by bi-
ases in the training data. It does not require ad-
ditional labeled data and provides a more thorough
analysis than standard approaches. Experiments show
that the strategy successfully detects bias and outper-
forms other ways of identifying problems with CNN’s
learned representations.
3 COPYRIGHT FORM
Three datasets are chosen for experimentation,
consisting of grayscale images depicting handwritten
digits and characters, with dimensions of 28x28x1.
Only images containing digits 0-9 have been selected
INCOFT 2025 - International Conference on Futuristic Technology
296