of high quality. CNNs are among the deep learning
models trained on large quantity of labeled data for
learning of advanced properties and characteristics.
Though, it is not easy to gather this data, more often
this is a challenge, especially in certain niche such as
healthcare and security, specialized knowledge is
vital while labeling the data (Li, 2022). For example,
assigning diagnosis for particular diseases like cancer,
neurological diseases, etc. for prognosis from the
medical images is requires annotations on the data
and is usually accomplished by a radiologist which
not only increases cost but also time (Razzak et al.,
2018). Also, another problem that emerges is the data
imbalance. In many datasets, there is a prevalence of
a particular class or category, which introduces bias
in the models they provide, especially when
confronted with underrepresented data (Abdar et al.,
2021).
To overcome these limitations, the following
strategies have been used Namely, such techniques as
Generative Adverserial Networks (GANs), generate
artificial data to support training exercise. The third
way of creating an artificial increase in the size of the
dataset is data augmentation where these images can
be rotated, flipped or scaled to improve on the
outcome of the model (Hemanth & Estrela, 2017).
Nonetheless, transfer learning has been named as one
of the most effective strategies for coping with the
challenges arising from low data availability. To
facilitate this in transfer learning, models using large
datasets such as ImageNet are tweaked on a limited
data to enable the classifiers to perform other tasks as
desired despite limited data for labeling (Shafiq & Gu,
2022). However, the problems of finding diverse data
sets with annotations are still a major roadblock to the
expansion of deep learning in image recognition.
3.2 Computational Complexity and
Resource Demands
Learning deep neural networks particularly in image
recognition task requires huge computing power.
There is no doubt that everyone can develop a deep
learning model with millions of parameters, it could
take weeks or even days to train such models given
the layers and weights within the network architecture
(Najafabadi, Villanueva & Măruşter, 2015). The
training process involves the use of hardware such as
GPU and TPU with a view of optimising the training
process as well as improving the efficiency of the
models (Zhang et al., 2019). For example,
contemporary deep learning architectures such as
ResNet and EfficientNet use a huge amount of
computational resources, and the training processes
of such architectures on average hardware
instruments might be time-consuming experiences
(Shafiq & Gu, 2022).
Also, the electrical power being used to train such
models is also rising, which is not desirable given that
sustainability in AI is now becoming trendy. While it
is a fact that deep learning possesses a “carbon
footprint,” there are some questions about AI contact
with the environment, and scientists have urged to
train better models and algorithms (Abdar et al.,
2021). Techniques that have been proposed here
include the model pruning whereby one gets rid of
model parameters that are relatively irrelevant and
Quantization which simply cuts down the precision of
model weight. In addition, the new architectures
developed from the ground up, such as TPUs and
neuromorphic chips, pushed the deep learning
methods forward, and the issues of speed versus
accuracy were still an issue (Jacob & Darney, 2021).
3.3 Interpretability and Trust Issues
Learning deep neural networks particularly in image
recognition task requires huge computing power.
There is no doubt that everyone can develop a deep
learning model with millions of parameters, it could
take weeks or even days to train such models given
the layers and weights within the network architecture
(Najafabadi, Villanueva & Măruşter, 2015). The
training process involves the utilization of the
hardware such as the GPU and the TPU, in a way that
makes the training process more efficient, in addition
to boosting the effectiveness of the models (Zhang et
al., 2019). For example, contemporary deep learning
architectures such as ResNet and EfficientNet use a
huge amount of computational resources, and the
training processes of such architectures on average
hardware instruments might be time-consuming
experiences (Shafiq & Gu, 2022).
Also, the electrical power being used to train such
models is also rising, which is not desirable given that
sustainability in AI is now becoming trendy. While it
is a fact that deep learning possesses a “carbon
footprint,” there are some questions about AI contact
with the environment, and scientists have urged to
train better models and algorithms (Abdar et al.,
2021). Techniques that have been proposed here
include the model pruning whereby one gets rid of
model parameters that are relatively irrelevant and
Quantization which simply cuts down the precision of
model weight. Furthermore, new holistic
architectures including TPUs and neuromorphic
chips introduced deep learning methods, while the