
handle multimodal fake news, such as misinformation
spread through images or videos. Additionally chal-
lenges arise in deploying the model at scale for mul-
tilingual datasets or adapting it to highly specialized
domains as it requires further refinement to maintain
optimal performance. These limitations underscore
the need for continued research to improve the sys-
tem’s versatility.
5 Conclusion and Future Work
This research developed a hybrid fake news detection
system by integrating BERT embeddings with en-
semble machine learning models. The system effec-
tively captured the semantic meaning of news content,
achieving improved accuracy and reliability through
Voting, Unweighted and Weighted Averaging tech-
niques. Weighted Averaging proved to be the most
reliable, leveraging the strengths of diverse models
and mitigating the impact of outliers using normal-
ized weights for consistent performance. Further-
more, the system demonstrated scalability and adapt-
ability across different datasets, making it suitable for
real-world applications. By combining the power of
deep learning with traditional classifiers, it addresses
key challenges such as overfitting and model inter-
pretability. The integration of these techniques lays
the foundation for building a robust and efficient fake
news detection system. Additionally, the approach’s
transparency helps enhance trust and accountability in
automated decision-making.
The proposed approach contributes to future ad-
vancements in fake news detection by enhancing
accuracy through weighted averaging in ensemble
learning, making it a scalable and adaptable frame-
work. News verification systems can leverage the
model to assist journalists and media organizations in
assessing the credibility of articles before publication.
Search engines can incorporate the model to filter out
misleading content, enhancing the integrity of online
information. The model can be enhanced to prevent
market manipulation through fake financial news and
also detect false health claims, medical misinforma-
tion, and prevent public health crises.
Future enhancements include exploring diverse
data types, fine-tuning BERT for domain-specific ap-
plications and enabling real-time detection capabili-
ties. Expanding support for multiple languages and
utilizing larger datasets will further improve system
performance. Additionally, incorporating explainable
AI and robust defenses against fake content can en-
hance transparency and reliability in detection.
REFERENCES
Bengio, Y. et al. (2012). Practical recommendations for
gradient-based training of deep architectures. Neural
Networks: Tricks of the Trade, pages 437–478.
Breiman, L. (2001). Random forests. Machine Learning,
45(1):5–32.
Caruana, R. and Niculescu-Mizil, A. (2006). An empirical
comparison of supervised learning algorithms. Pro-
ceedings of the 23rd international conference on Ma-
chine learning, pages 161–168.
Castillo, C. et al. (2011). Information credibility on twitter.
In Proceedings of the 22nd International Conference
on World Wide Web, pages 675–684. ACM.
Chen, T. et al. (2016). Xgboost: A scalable tree boosting
system. ACM SIGKDD, pages 785–794.
Czapla, P., Gugger, S., Howard, J., and Kardas, M. (2019).
Universal language model fine-tuning for polish hate
speech detection. In Proceedings of the PolEval2019
Workshop, page 149.
Devlin, J. et al. (2018). Bert: Pre-training of deep bidirec-
tional transformers for language understanding. arXiv
preprint arXiv:1810.04805.
Freund, Y. and Schapire, R. E. (1997). A decision-theoretic
generalization of on-line learning and an application
to boosting. Journal of Computer and System Sci-
ences, 55(1):119–139.
Friedman, J. H. (2001). Greedy function approximation: A
gradient boosting machine. The Annals of Statistics,
29(5):1189–1232.
Gilpin, L. H. et al. (2018). Explaining explanations: An
overview of interpretability of machine learning. ACM
Computing Surveys (CSUR), 51(5):93.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural Computation, 9(8):1735–1780.
Jiang, T., Yu, X., Li, C., Song, Y., and Zhan, Y. (2021). A
novel stacking approach for accurate detection of fake
news. IEEE Access, 9:22626–22639.
Jin, Y. et al. (2022). Towards fine-grained reasoning for
fake news detection. In Proceedings of the AAAI Con-
ference on Artificial Intelligence, volume 36, pages
6339–6346.
Joachims, T. (1998). Text categorization with support vec-
tor machines: Learning with many relevant features.
European Conference on Machine Learning (ECML).
Kim, Y. (2014). Convolutional neural networks for sentence
classification. In Proceedings of the 2014 Conference
on Empirical Methods in Natural Language Process-
ing (EMNLP).
Lifferth, W. (2018). Fake news. https://kaggle.com/
competitions/fake-news. Kaggle.
Marco Tulio Ribeiro, Sameer Singh, C. G. (2016). ”why
should i trust you?”: Explaining the predictions of any
classifier. Proceedings of the 22nd ACM SIGKDD In-
ternational Conference on Knowledge Discovery and
Data Mining, pages 1135–1144.
McCallum, A. and Nigam, K. (1998). A comparison of
event models for naive bayes text classification. In
Proceedings of the AAAI-98 Workshop on Learning
for Text Categorization, pages 41–48. AAAI Press.
Weighted Ensemble Model for Tackling Fake News
883