3 DISCUSSIONS
Admittedly, in past research and experiments the
Federated Learning have presented its superiority in
privacy and heterogeneous data processing and have
maintained a relatively high level of accuracy in most
of its application circumstances. However, its
potential drawbacks and challenges have also been
discovered in these experiments, by either certain
features of the outcome data or the mathematical
structure of its foundational workflow. For some of
the illustrated issues, there has been modification or
expansion for the Federated Learning algorithms
towards their corresponding resolution; for some
others, however, it can only be expected to be solved
through the further development of other deep
learning algorithms.
One of the most common and directly meet
problems is that the models trained by the Federated
Learning algorithms lack interpretability. It is
admittedly that the Federated Learning trained
models can often attain great performance in the
prediction accuracy, but the researchers can hardly
comprehend how and why this is managed to happen
since the complexity of the neural network, the
commonly used architecture for the models trained by
the Federated Learning algorithms, makes it
challenge to be understood by human researchers.
Such drawbacks can be non-problematic in the result-
orientated applications; for instance, it is not very
necessary for human researchers and programmers to
fully understand why and how an image-recognition
model can tell all the cat images from the mixed
image dataset, as long as it has achieved its task by
attaining acceptable accuracy. Nevertheless, the
researches in medic and biology often acquires the
understanding of the mechanism behind the diagnosis,
no matter if it is aimed for adding the credibility for
the diagnosis or provoking further and more subtle
researches towards these newly discovered
mechanism. In this instance, unfortunately, the
Federated Learning algorithm can offer little aid for
the potential further researches.
One possible method in adding more
interpretability to the Federated Learning algorithm is
fusing it with SHapley Addictive exPlanation (SHAP)
algorithm. In the study of Lundberg et al, it is firstly
mentioned that Shapley values can be introduced in
the field of machine learning (Lundberg et al., 2017).
The SHAP algorithm can give out a contribution
value (SHAP value) of each attribution for the
eventual outcome of model prediction accuracy,
which can be both positive for the contribution of
reaching higher model accuracy and negative for the
opposite situation. If the researchers are able to
combine Federated Learning with SHAP, it is
possible for the researchers to achieve model
explainability by analyzing the SHAP values of all
the attributes, considering the fact that there have
been cases of SHAP algorithm being applied
separately in the domain of medical diagnosis.
Another existing drawback of the Federated
Learning that is due to its workflow is the lack of
applicability in all circumstances, a concession made
for its universality. What has caused this issue is the
difference in data distribution between heterogeneous
datasets: the decentralized characteristic of its
training process allows the Federated Learning
algorithm to process datasets with heterogeneous data
types, but the performance of the trained model can
also be impaired by their difference, making the
prediction accuracy of the aggregated model in an
epoch lower than the accuracy of the locally trained
model in the application of that particular dataset.
Some of researchers who have discovered this
phenomenon in their experiments also call it
“forgetting” (Darzi et al., 2024), stating that when
receiving the heterogeneous data from a new dataset
for further training, the model “forgets” some of the
information from the previous old dataset and thus
experience a decrease in its predication accuracy for
the old dataset.
In response to this problem, the solution proposed
by Yu et al in their research is the Dynamic Weighting
Translation Transfer Learning (DTTL), which builds
up a dynamic translation between imbalanced classes
in two domains and minimizes the cross-entropy
between the domains to reduce domain difference
(Yu et al., 2024).
Although frequently being described as a type of
decentralized deep learning algorithm, as a matter of
fact the Federated Learning algorithm is only a type
of partially distributed and coordinated algorithm,
which still requires a centre server as a coordinator
for model updates. Though still safer than the full
centralized algorithm, the remain centralized
architecture of the Federated Learning can lead to
potential safety loopholes, especially in the
transmission from local devices to the centre server.
If being intercepted and decrypted, the parameters of
the locally trained model from a specific dataset are
exposed to the hacker; though it is not the raw data
that is disclosed, the hackers might still be able to
learn the features of data distribution in this dataset
and manage to guess out a portion of the raw data
based on the parameters disclosed.
Unless being replaced by the development of fully
distributed algorithms, this potential safety issue can
always be problematic and hardly be removed, for it
relays on the architecture of Federated Learning’s
workflow. A possible improvement for this issue is
adopting more complex and secure encryption for the
transmission; however, it would surely increase the