Deep Reinforcement Learning Oriented Queuing And Swarm Intelligence Framework For Adaptive Task Scheduling In End Edge Cloud Computing Environments
Keywords:
Deep reinforcement learning, cloud task scheduling, queuing theoryAbstract
The unprecedented growth of data intensive applications, cyber physical systems, and latency sensitive services has fundamentally reshaped the operational expectations of cloud, edge, and fog computing ecosystems. Modern computing environments are no longer limited to static cloud data centers but have evolved into complex end edge cloud architectures in which computation is dynamically distributed across geographically dispersed and resource heterogeneous nodes. Within this paradigm, task scheduling emerges as a central and persistent challenge, as the allocation of computational workloads to available resources directly determines system throughput, energy efficiency, delay, fairness, and economic sustainability. Traditional heuristic and metaheuristic approaches, while historically successful, increasingly struggle to cope with the stochasticity, scale, and heterogeneity that define contemporary distributed computing infrastructures. Consequently, there is a growing scholarly consensus that adaptive learning-based scheduling mechanisms are required to enable real time decision making under uncertainty.
This study develops and theorizes a novel reinforcement learning driven queuing and optimization framework for dynamic task scheduling across end edge cloud systems. Drawing upon foundational work in reinforcement learning theory, evolutionary computation, and queuing based performance modeling, the research constructs an integrated conceptual architecture that fuses deep Q learning with optimal queuing dynamics and swarm intelligence. In particular, the scheduling logic is informed by the theoretical insights of Kanikanti et al. (2025), who demonstrate that deep Q learning combined with optimal queuing theory can substantially improve dynamic task allocation in cloud computing environments. Their work provides a critical empirical and theoretical anchor for this study, which extends the paradigm into a broader end edge cloud orchestration context.
The article systematically analyzes how reinforcement learning agents can learn optimal task placement policies by observing system states such as queue length, processing capacity, latency, and energy consumption. These agents are embedded within a hybrid optimization framework that incorporates swarm-based search processes, genetic adaptation, and market-oriented scheduling strategies. The proposed model is not presented as a mathematical algorithmic artifact but as a deeply theorized scheduling ecosystem grounded in interdisciplinary research traditions. By integrating queuing theory with deep reinforcement learning, the framework accounts for both the temporal structure of task arrivals and the strategic adaptation of scheduling policies over time.
The results of this study, interpreted through extensive comparative reasoning against the existing literature, indicate that learning driven queuing-based schedulers can outperform static and metaheuristic approaches in terms of adaptability, delay minimization, and system robustness. Furthermore, the discussion situates these findings within broader debates about the future of intelligent computing infrastructures, arguing that end edge cloud systems represent not merely a technological shift but a paradigmatic transformation in how computational resources are conceptualized and governed. The article concludes by identifying key theoretical and practical implications for next generation scheduling systems and outlining a future research agenda centered on autonomous, self-optimizing distributed intelligence.
References
1. Zhou, C., Wu, W., He, H., Yang, P., Lyu, F., Cheng, N., Shen, X. Deep reinforcement learning for delay oriented iot task scheduling in sagin. IEEE Transactions on Wireless Communications 20(2), 911–925.
2. Huang, X., Li, C., Chen, H., An, D. Task scheduling in cloud computing using particle swarm optimization with time varying inertia weight strategies. Cluster Computing 23, 1137–1147.
3. Ren, J., Jiang, H., Shen, X., et al. Editorial of ccf transactions on networking: special issue on intelligence enabled end edge cloud orchestrated computing. CCF Transactions on Networking 3, 155–157.
4. Verma, A., Kaushal, S. A hybrid multi objective particle swarm optimization for scientific workflow scheduling. Parallel Computing 62, 1–19.
5. Khan, M.I., Xia, K., Ali, A., Aslam, N. Energy aware task scheduling by a true online reinforcement learning in wireless sensor networks. International Journal of Sensor Networks 25, 244–258.
6. Dorigo, M., Maniezzo, V., Colorni, A. Ant system: Optimization by a colony of cooperating agents. IEEE Transactions on Systems Man and Cybernetics 26, 29–41.
7. Pinciroli, R., Ali, A., Yan, F., Smirni, E. Cedule plus: Resource management for burstable cloud instances using predictive analytics. IEEE Transactions on Network and Service Management 18, 945–957.
8. Sutton, R.S., Barto, A.G. Reinforcement Learning: An Introduction. MIT Press.
9. Ren, J., Zhang, D., He, S., Zhang, Y., Li, T. A survey on end edge cloud orchestrated network computing paradigms: Transparent computing, mobile edge computing, fog computing, and cloudlet. ACM Computing Surveys 52(6).
10. Kanikanti, V. S. N., Tiwari, S. K., Nayan, V., Suryawanshi, S., Chauhan, R. Deep Q Learning Driven Dynamic Optimal Task Scheduling for Cloud Computing Using Optimal Queuing. Proceedings of the International Conference on Computational Intelligence and Knowledge Economy, 217–222.
11. Gabi, D., Ismail, A.S., Dankolo, N.M. Minimized makespan based improved cat swarm optimization for efficient task scheduling in cloud datacenter. Proceedings of the High Performance Computing and Cluster Technologies Conference, 16–20.
12. Wu, Z., Liu, X., Ni, Z., Yuan, D., Yang, Y. A market oriented hierarchical scheduling strategy in cloud workflow systems. Journal of Supercomputing 63, 256–293.
13. Wei, Z., Zhang, Y., Xu, X., Shi, L., Feng, L. A task scheduling algorithm based on Q learning and shared value function for wireless sensor networks. Computer Networks 126, 141–149.
14. Eberhart, R., Kennedy, J. A new optimizer using particle swarm theory. Proceedings of the International Symposium on Micro Machine and Human Science, 39–43.
15. Guevara, J.C., da Fonseca, N.L. Task scheduling in cloud fog computing systems. Peer to Peer Networking and Applications 14, 962–977.
16. Zhang, H., He, G. An adaptive task scheduling system based on real time clustering and NetFlow prediction. Proceedings of the International Conference on Computer and Communications, 77–80.
17. Holland, J.H. Adaptation in Natural and Artificial Systems. MIT Press.
18. Wei, Z., Liu, F., Zhang, Y., Xu, J., Ji, J., Lyu, Z. A Q learning algorithm for task scheduling based on improved support vector machine in wireless sensor networks. Computer Networks 161, 138–149.
19. Meshkati, J., Safi Esfahani, F. Energy aware resource utilization based on particle swarm optimization and artificial bee colony algorithms in cloud computing. Journal of Supercomputing 75, 2455–2496.
20. Shafi, U., Shah, M.A., Wahid, A., Abbasi, K., Javaid, Q., Asghar, M.N., Haider, M. A novel amended dynamic round robin scheduling algorithm for timeshared systems. International Arab Journal of Information Technology 17, 90–98.
21. Raju, R., Babukarthik, R.G., Chandramohan, D., Dhavachelvan, P., Vengattaraman, T. Minimizing the makespan using hybrid algorithm for cloud computing. Proceedings of the International Advance Computing Conference, 957–962.
22. Sulaiman, M., Halim, Z., Lebbah, M., Waqas, M., Tu, S. An evolutionary computing based efficient hybrid task scheduling approach for heterogeneous computing environment. Journal of Grid Computing 19, 11.
23. Duan, S., Wang, D., Ren, J., Lyu, F., Zhang, Y., Wu, H., Shen, X. Distributed artificial intelligence empowered by end edge cloud computing: A survey. IEEE Communications Surveys and Tutorials 25(1), 591–624.
24. Zhou, Z., Li, F., Zhu, H., Xie, H., Abawajy, J.H., Chowdhury, M.U. An improved genetic algorithm using greedy strategy toward task scheduling optimization in cloud environments. Neural Computing and Applications 32, 1531–1541.
25. Song, A., Chen, W.N., Luo, X., Zhan, Z.H., Zhang, J. Scheduling workflows with composite tasks: A nested particle swarm optimization approach. IEEE Transactions on Services Computing 15, 1074–1088.
26. Watkins, C.J., Dayan, P. Q learning. Machine Learning 8, 279–292.
27. Kaelbling, L.P., Littman, M.L., Moore, A.W. Reinforcement learning: A survey. Journal of Artificial Intelligence Research 4, 237–285.
28. Sharma, M., Garg, R. HIGA: Harmony inspired genetic algorithm for rack aware energy efficient task scheduling in cloud data centers. Engineering Science and Technology International Journal 23, 211–224.
29. Jamil, B., Shojafar, M., Ahmed, I., Ullah, A., Munir, K., Ijaz, H. A job scheduling algorithm for delay and performance optimization in fog computing. Concurrency and Computation Practice and Experience 32, e5581.
30. Jiang, M., Wu, T., Wang, Z., Gong, Y., Zhang, L., Liu, R.P. A multi intersection vehicular cooperative control based on end edge cloud computing. IEEE Transactions on Vehicular Technology 71(3), 2459–2471.
Downloads
Published
Issue
Section
License
Copyright (c) 2026 Dr. Leonhard Weiss

This work is licensed under a Creative Commons Attribution 4.0 International License.