intensity of contrastive learning during training,
adapting to the varying needs at different stages and
thereby optimizing performance. This integration
helps achieve more precise training results and
enhances the model's robustness and effectiveness in
practical applications.
However, the study is limited by the use of only
the CIFAR-10 dataset, which consists of small-sized
images (32 × 32) and a limited number of images
(60,000). This may not fully capture the complexity
and diversity of real-world image data. For future
research, this paper suggests expanding the
evaluation on more diverse and more complex
datasets. For example, ImageNet has over 1,000
categories and millions of high-resolution pictures.
Testing on such a dataset will help validate the
model's generalization capability and robustness in
different scenarios.
5 CONCLUSIONS
The temperature parameter τ is a super important
hyperparameter in federated contrastive learning. It
directly determines whether the model performs well
or not. τ controls how smooth and distinct the
similarity distribution is in the contrastive loss
function. If it can tune τ just right, it can boost the
model's training efficiency and ability to generalize.
By finding the perfect step sizes for the model, it can
make the training process smoother and give the
model more confidence when it encounters new
datasets. Conversely, improper settings may result in
unstable training, overfitting, or suboptimal
performance.
This paper demonstrates through experiments that
employing a simulated annealing algorithm to
dynamically adjust τ markedly improves the model's
adaptability across different training stages. This
dynamic adjustment enhances training stability and
final accuracy, allowing the model to flexibly balance
the learning of local and global information. It
accelerates convergence in the early stages of training
and mitigates overfitting in later stages.
Moreover, this paper optimizes the neural
network model by integrating CNN and Transformer
architectures. This integration enables the model to
more effectively capture both local and global
features of images, thereby achieving higher
performance and generalization ability.
In summary, the MOON federated learning
algorithm proposed in this paper addresses the non-
IID data problem in federated learning through
dynamic temperature adjustment and network
structure optimization. Future research can further
explore the adaptability of temperature parameters
and neural network structures under different datasets
and tasks. Additionally, future work can investigate
how to more efficiently leverage the privacy-
preserving features within the federated learning
framework to promote the application development
of this field.
AUTHORS CONTRIBUTION
All the authors contributed equally and their names
were listed in alphabetical order.
REFERENCES
Brendan, M. H., Moore, E., Ramage, D., Hampson, S., &
Arcas, Blaise Agüera y. 2016. Communication-
Efficient Learning of Deep Networks from
Decentralized Data. ArXiv.org.
Chen, D.-J., Lee, C.-Y., Park, C.-H., & Mendes, P. 2007.
Parallelizing simulated annealing algorithms based on
high-performance computer. Journal of Global
Optimization, 39(2), 261–289.
Chen, T., Kornblith, S., Norouzi, M., & Hinton, G. 2020. A
Simple Framework for Contrastive Learning of Visual
Representations. ArXiv:2002.05709.
Jafarigol, E., Trafalis, T., Razzaghi, T., & Zamankhani, M.
2023. Exploring Machine Learning Models for
Federated Learning: A Review of Approaches,
Performance, and Limitations. ArXiv.org.
Janiesch, C., Zschech, P., & Heinrich, K. 2021. Machine
learning and deep learning. Electronic Markets, 31(31),
685–695. Springer.
Kirkpatrick, S., Gelatt, C. D., & Vecchi, M. P. 1983.
Optimization by Simulated Annealing. Science,
220(4598), 671–680.
Li, Q., He, B., & Song, D. 2021. Model-Contrastive
Federated Learning. ArXiv.org.
Li, Y., Wang, C., Gao, L., Song, Y., & Li, X. 2020. An
improved simulated annealing algorithm based on
residual network for permutation flow shop scheduling.
Complex & Intelligent Systems, 7(3), 1173–1183.
Voigt, P., & Bussche, A. von dem. 2017. The EU General
Data Protection Regulation (GDPR). Springer
International Publishing.
World Innovation and Change Management Insititute.
2024. When Artificial intelligence becomes a part of
life? – World Innovation and Change Management
Institute. Wicmi.ch. https://wicmi.ch/techdata/when-
artificial-intelligence-becomes-a-part-of-life-2/