addressing critical gaps in real-time responsiveness,
energy efficiency, and environmental generalization.
Unlike existing models that are either too
computationally heavy or narrowly scoped to
simulated environments, the proposed framework
integrates a lightweight yet powerful RL architecture
capable of operating under diverse network
conditions, including dense urban deployments,
intelligent reflecting surfaces, and non-terrestrial
links.
The development of a flexible reward function,
combined with a multi-agent and federated learning
approach, has enabled spectrum decisions that are not
only optimal in terms of throughput and latency but
also considerate of energy consumption and resource
constraints. Extensive evaluations confirm that the
model offers significant gains in performance metrics
such as spectral efficiency, decision latency, and
convergence speed, while remaining scalable and
practical for deployment in future communication
infrastructures.
Beyond its immediate applications in 5G
networks, the framework is inherently forward-
compatible with the architectural needs and
operational philosophies of 6G, including support for
AI-native networking, edge intelligence, and
sustainable design principles. In doing so, this work
contributes not just a technological advancement, but
also a strategic foundation for how intelligent systems
can manage increasingly complex and dynamic
wireless ecosystems.
Ultimately, this research demonstrates that with
the right integration of AI and domain-specific
optimization, spectrum allocation can evolve from a
rigid, rule-based process to a self-optimizing,
context-aware systemcapable of empowering the next
generation of ultra-connected, intelligent digital
environments.
REFERENCES
AlSobhi, W., & Aghvami, A. H. (2019). QoS-Aware
resource allocation of two-tier HetNet: A Q-learning
approach. In 26th International Conference on
Telecommunications (ICT) (pp. 330–334). IEEE.
Springe rLink
Ansarifard, M., Mokari, N., Javan, M., Saeedi, H., &
Jorswieck, E. A. (2023). AI-based radio and computing
resource allocation and path planning in NOMA NTNs:
AoI minimization under CSI uncertainty. arXiv preprint
arXiv:2305.00780.
Cai, F., Gao, Y., Cheng, L., Sang, L., & Yang, D. (2016).
Spectrum sharing for LTE and WiFi coexistence using
decision tree and game theory. In Proceedings of the
2016 IEEE Wireless Communications and Networking
Conference.
Cao, J., Zou, X., Xie, R., & Li, Y. (2022). Cellular network
power allocation algorithm based on deep
reinforcement learning and artificial intelligence.
Computational Intelligence and Neuroscience, 2022,
Article 9456611. Wiley Online Library
Chen, X., Wu, C., & Chen, T. (2020). Age of information
aware radio resource management in vehicular
networks: A proactive deep reinforcement learning
perspective. IEEE Transactions on Wireless
Communications,19(4),2268–2281.SpringerLink
Grandl, R., Ananthanarayanan, G., & Kandula, S. (2014).
Multi-resource packing for cluster schedulers. ACM
SIGCOMM Computer Communication Review, 44(4),
455–466.SpringerLink
Hassan, S. S., Park, Y. M., Tun, Y. K., Saad, W., Han, Z.,
& Hong, C. S. (2024). Enhancing spectrum efficiency
in 6G satellite networks: A GAIL-powered policy
learning via asynchronous federated inverse
reinforcement learning.arXivpreprint arXiv:2409.187
18.arXiv
Haykin, S. (2005). Cognitive radio: Brain-empowered
wireless communications. IEEE Journal on Selected
Areas inCommunications,23,201–220.MDPI
Kasgari, A. T. Z., & Saad, W. (2019). Model-free ultra-
reliable low latency communication (URLLC): A deep
reinforcement learning framework. In ICC 2019–2019
IEEE International Conference on Communications
(ICC) (pp. 1–6). IEEE. Springer Link
Khadem, M., Zeinali, F., Mokari, N., & Saeedi, H. (2024).
AI-enabled priority and auction-based spectrum
managementfor6G.arXivpreprintarXiv:2401. 06484.ar
Xiv
Khalifa, N. B., Assaad, M., & Debbah, M. (2019). Risk-
sensitive reinforcement learning for URLLC traffic in
wireless networks. In 2019 IEEE Wireless
Communications and Networking Conference
(WCNC) (pp. 1–7). IEEE. Springe rLink
Lei, W., Wang, X., & Wang, J. (2020). Deep reinforcement
learning based spectrum allocation in integrated
accessandbackhaulnetworks. arXivpreprintarXiv:2004.
13133.arXiv
Liu, S., Wang, T., Pan, C., Zhang, C., Yang, F., & Song, J.
(2021). Deep reinforcement learning for spectrum
sharing in future mobile communication system. In
Proceedings of the 2021 IEEE International
Symposium on BroadbandMultimediaSystemsandBr
oadcasting(BMSB).MDPI
Mnih, V., Kavukcuoglu, K., Silver, D., et al. (2013).
Playing Atari with deep reinforcement learning. arXiv
preprint arXiv:1312. 5602.MDPI
Naparstek, O., & Cohen, K. (2013). Deep multi-user
reinforcement learning for distributed dynamic
spectrum access. IEEE Transactions on Wireless
Communications, 18, 310–323.MDPI
Nasir, Y. S., & Guo, D. (2020). Deep reinforcement
learning for joint spectrum and power allocation in
cellular networks. arXivpreprintarXiv:2012.
10682.arXiv