Structure-Aware Deep Reinforcement Learning for Latency-Minimal Scheduling of Edge AI Inference on Heterogeneous Cores
DOI:
https://doi.org/10.70393/6a696574.343033ARK:
https://n2t.net/ark:/40704/JIET.v1n1a06Disciplines:
Intelligent SystemsSubjects:
OtherReferences:
25Keywords:
Edge Intelligence, Deep Reinforcement Learning, Heterogeneous Multi-Core Systems, Deep Learning Inference, Structure-Aware Learning, Combinatorial OptimizationAbstract
The exponential surge in latency-sensitive deep learning workloads at the network edge demands the implementation of highly stringent resource scheduling mechanisms to mitigate this situation. Yet, achieving this is impeded by the intrinsic computational asymmetry of contemporary edge processors, which imposes severe constraints on task scheduling.While heuristic algorithms have traditionally been employed to manage these tasks which are intrinsically modeled as Directed Acyclic Graphs they frequently stagnate at local optima due to an inability to dynamically adapt to complex, non-linear topological dependencies. Furthermore, existing learning-based solutions often fail to generalize across diverse neural network architectures, as they typically rely on scalar feature inputs that cannot capture the rich structural priors of the computation graph. To bridge this gap, this paper proposes a structure-aware scheduling framework that synergizes Graph Neural Networks with Deep Reinforcement Learning. By leveraging a Graph Isomorphism Network to extract high-dimensional topological embeddings of task dependencies, our agent learns to map computational layers to heterogeneous cores with the primary objective of minimizing end-to-end inference latency. Extensive experiments on real-world Deep Neural Network workloads, including ResNet and Inception architectures, demonstrate that the proposed approach not only surpasses classic heuristics such as HEFT in terms of makespan reduction but also achieves more robust convergence compared to structure-agnostic learning baselines. These findings suggest that explicitly encoding structural priors into learning algorithms offers a promising approach to effectively address the NP-hard scheduling problems inherent in next-generation edge artificial intelligence systems.
References
[1] Jayanth, R., Gupta, N., & Prasanna, V. (2024, September). Benchmarking Edge AI Platforms for High-Performance ML Inference. In 2024 IEEE High Performance Extreme Computing Conference (HPEC) (pp. 1-7). IEEE.
[2] Luo, M., Zhang, W., Song, T., Li, K., Zhu, H., Du, B., & Wen, H. (2021, January). Rebalancing expanding EV sharing systems with deep reinforcement learning. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence (pp. 1338-1344).
[3] Luo, M., Du, B., Zhang, W., Song, T., Li, K., Zhu, H., ... & Wen, H. (2023). Fleet rebalancing for expanding shared e-mobility systems: A multi-agent deep reinforcement learning approach. IEEE Transactions on Intelligent Transportation Systems, 24(4), 3868-3881.
[4] Zhu, H., Luo, Y., Liu, Q., Fan, H., Song, T., Yu, C. W., & Du, B. (2019). Multistep flow prediction on car-sharing systems: A multi-graph convolutional neural network with attention mechanism. International Journal of Software Engineering and Knowledge Engineering, 29(11n12), 1727-1740.
[5] Liu, W. (2025). Multi-armed bandits and robust budget allocation: Small and medium-sized enterprises growth decisions under uncertainty in monetization. European Journal of AI, Computing & Informatics, 1(4), 89–97.
[6] Liu, W. (2025). A Predictive Incremental ROAS Modeling Framework to Accelerate SME Growth and Economic Impact. Journal of Economic Theory and Business Management, 2(6), 25–30.
[7] Yu, C., Wang, H., Chen, J., Wang, Z., Deng, B., Hao, Z., ... & Song, Y. (2026). When Rules Fall Short: Agent-Driven Discovery of Emerging Content Issues in Short Video Platforms. arXiv preprint arXiv:2601.11634.
[8] Wang, H., Li, Q., & Liu, Y. (2023). Adaptive supervised learning on data streams in reproducing kernel Hilbert spaces with data sparsity constraint. Stat, 12(1), e514.
[9] Wang, P., Wang, H., Li, Q., Shen, D., & Liu, Y. (2024). Joint and individual component regression. Journal of Computational and Graphical Statistics, 33(3), 763-773.
[10] Wang, H., Sun, W., & Liu, Y. (2022). Prioritizing autism risk genes using personalized graphical models estimated from single-cell rna-seq data. Journal of the American Statistical Association, 117(537), 38-51.
[11] Liu, W. (2025). Few-Shot and Domain Adaptation Modeling for Evaluating Growth Strategies in Long-Tail Small and Medium-sized Enterprises. Journal of Industrial Engineering and Applied Science, 3(6), 30–35.
[12] Yu, C., Wu, H., Ding, J., Deng, B., & Xiong, H. (2025, September). Unified Survey Modeling to Limit Negative User Experiences in Recommendation Systems. In Proceedings of the Nineteenth ACM Conference on Recommender Systems (pp. 1104-1107).
[13] Yu, C., Li, P., Wu, H., Wen, Y., Deng, B., & Xiong, H. (2024). USM: Unbiased Survey Modeling for Limiting Negative User Experiences in Recommendation Systems. arXiv preprint arXiv:2412.10674.
[14] Raveendran Nair, G., Jiang, F., Zhang, J., & Cao, Y. (2024, August). A 16nm heterogeneous accelerator for energy-efficient sparse and dense ai computing. In Proceedings of the 29th ACM/IEEE International Symposium on Low Power Electronics and Design (pp. 1-6).
[15] Morabito, R., & Chiang, M. (2024). Exploring edge AI inference in heterogeneous environments: requirements, challenges, and solutions. In IoT Edge Intelligence (pp. 37-66). Cham: Springer Nature Switzerland.
[16] Kohli, P., Jayanth, R., Gupta, N., Fan, H., & Prasanna, V. (2025, September). Performance-Energy Characterization of ML Inference on Heterogeneous Edge AI Platforms. In 2025 IEEE High Performance Extreme Computing Conference (HPEC) (pp. 1-7). IEEE.
[17] Ngo, D., Park, H. C., & Kang, B. (2025). Edge intelligence: A review of deep neural network inference in resource-limited environments. Electronics, 14(12), 2495.
[18] Wu, Y. (2026). Research on Dynamic Prediction Model of Brand Marketing Content ROI Based on Machine Learning. International Journal of Advance in Applied Science Research, 5(2), 31-38.
[19] Wang, C. (2026). A Study on Data-Driven Budget Optimization for US Enterprises’ Cross-Border Marketing. Frontiers in Management Science, 5(1), 41-46.
[20] Tan, Z., Li, Z., Liu, T., Wang, H., Yun, H., Zeng, M., ... & Jiang, M. (2025). Aligning large language models with implicit preferences from user-generated content. arXiv preprint arXiv:2506.04463.
[21] Wang, J., Tim, K. T., Li, S., Chan, T. K., & Fung, J. C. (2023). A systematic comparison of the wind profile codifications in the Western Pacific Region. Wind & structures, 37(2), 105-115.
[22] Wang, J., Chang, Y., Cao, S., Dong, Y., Li, S., Jia, L., & Li, W. (2025). Explanatory framework of typhoon extreme wind speed predictions integrating the effects of climate changes. Climate Dynamics, 63(3), 142.
[23] Zhang, Z., Li, S., Zhang, Z., Liu, X., Jiang, H., Tang, X., ... & Jiang, M. (2025). IHEval: Evaluating language models on following the instruction hierarchy. arXiv preprint arXiv:2502.08745.
[24] Wu, Y. (2026). A Study on the Impact of Cross-Departmental Data Collaboration on Marketing Campaign Efficiency in Fast-Moving Consumer Goods E-commerce: The Case of PepsiCo (China)’s 7UP and Mirinda Project. Frontiers in Management Science, 5(1), 7-12.
[25] Lin, A. (2026). Multi-Chain DAO Treasury Management: a Risk and Compliance Optimization Framework for the US Ecosystem. Journal of Intelligence and Engineering Technology, 1(1), 11-18.
Downloads
Published
How to Cite
Issue
Section
ARK
License
Copyright (c) 2026 The author retains copyright and grants the journal the right of first publication.

This work is licensed under a Creative Commons Attribution 4.0 International License.