TDLB-DDPG——基于任务依赖和负载均衡的VEC计算卸载方案
TDLB-DDPG—Task Dependency and Load Balancing Based on Computation Offloading Scheme in VEC
摘要: 如何合理地对车辆上的计算任务进行卸载,以降低任务的响应时间、RSU的平均负载率,是车辆边缘计算(Vehicular Edge Computing, VEC)中的重要问题。深度强化学习是解决计算卸载问题的有力工具,但当前深度强化学习方法往往未考虑具有依赖关系的卸载任务之间是否可以合并以及路侧单元间的负载均衡。对此,文章提出一种基于任务依赖和负载均衡的深度确定性策略梯度(Deep Deterministic Policy Gradient, DDPG)计算卸载方案;通过任务分类合并,减小DDPG的动作空间及批量处理任务,降低卸载任务的平均响应时间;综合考虑任务群组的优先级、紧迫程度和关键路径长度来对任务群组进行排序,以提高任务的卸载成功率;设计一种RSU动态权重分配机制,辅助DDPG实现RSU之间的负载均衡。实验结果表明,相比已有基于DQN (Deep Q-Network)和DDPG的计算卸载方法,本文所提方案的平均响应时间分别降低了21.4%和22.5%,RSU平均负载率也显著低于DQN和DDPG。
Abstract: Efficiently offloading computational tasks from vehicles to reduce task response times and improve the average load rate of Roadside Units (RSUs) is a critical challenge in Vehicular Edge Computing (VEC). Deep Reinforcement Learning (DRL) has emerged as a powerful tool for addressing computational offloading problems. However, existing DRL-based approaches often fail to consider whether dependent offloading tasks can be merged and overlook the load balancing among roadside units. To address these limitations, this paper proposes a Deep Deterministic Policy Gradient (DDPG)-based computational offloading scheme that incorporates task dependency and load balancing. By categorizing and consolidating tasks, the action space of DDPG is reduced, and tasks are processed in batches, thereby decreasing the average response time for offloaded tasks. It further prioritizes task groups based on their priority, urgency, and critical path length to enhance offloading success rates. Additionally, a dynamic weight allocation mechanism for RSUs is designed to assist DDPG in achieving load balancing among RSUs. Experimental results demonstrate that, compared to existing Deep Q-Network (DQN) and DDPG-based offloading methods, the proposed scheme reduces the average response time by 21.4% and 22.5%, respectively, while significantly lowering the average RSU load rate.
文章引用:钱振. TDLB-DDPG——基于任务依赖和负载均衡的VEC计算卸载方案[J]. 建模与仿真, 2025, 14(5): 336-352. https://doi.org/10.12677/mos.2025.145398

参考文献

[1] Zheng, D., Wang, L., Kai, C. and Peng, M. (2023) Resource Optimization for Task Offloading with Real-Time Location Prediction in Pedestrian-Vehicle Interaction Scenarios. IEEE Transactions on Wireless Communications, 22, 7331-7344. [Google Scholar] [CrossRef
[2] Zabihi, Z., Eftekhari Moghadam, A.M. and Rezvani, M.H. (2023) Reinforcement Learning Methods for Computation Offloading: A Systematic Review. ACM Computing Surveys, 56, 1-41. [Google Scholar] [CrossRef
[3] Mao, M., Hu, T. and Zhao, W. (2023) Reliable Task Offloading Mechanism Based on Trusted Roadside Unit Service for Internet of Vehicles. Ad Hoc Networks, 139, Article 103045. [Google Scholar] [CrossRef
[4] Liu, L., Chen, C., Pei, Q., Maharjan, S. and Zhang, Y. (2020) Vehicular Edge Computing and Networking: A Survey. Mobile Networks and Applications, 26, 1145-1168. [Google Scholar] [CrossRef
[5] Huang, Z., Chen, Y. and Zhang, Y. (2024) Lyapunov-Guided Deep Reinforcement Learning for Vehicle Task Stable Offloading. 2024 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Tianjin, 8-10 May 2024, 833-838. [Google Scholar] [CrossRef
[6] Huang, B., Zhou, Y., Zhang, X., Chen, J. and Shang, L. (2024) Computation Offloading and Resource Allocation for Vehicle-Assisted Edge Computing Networks with Joint Access and Backhaul. IEEE Access, 12, 110248-110259. [Google Scholar] [CrossRef
[7] Zhang, X., Wu, W., Zhao, Z., Wang, J. and Liu, S. (2023) RMDDQN-Learning: Computation Offloading Algorithm Based on Dynamic Adaptive Multi-Objective Reinforcement Learning in Internet of Vehicles. IEEE Transactions on Vehicular Technology, 72, 11374-11388. [Google Scholar] [CrossRef
[8] Bi, X., Shi, J., Zhang, B., Lyu, Z. and Huang, L. (2023) An RSU-Crossed Dependent Task Offloading Scheme for Vehicular Edge Computing Based on Deep Reinforcement Learning. International Journal of Sensor Networks, 41, 244-256. [Google Scholar] [CrossRef
[9] Materwala, H., Ismail, L. and Hassanein, H.S. (2023) QoS-SLA-Aware Adaptive Genetic Algorithm for Multi-Request Offloading in Integrated Edge-Cloud Computing in Internet of Vehicles. Vehicular Communications, 43, Article 100654. [Google Scholar] [CrossRef
[10] Liu, J., Wang, Y., Pan, D. and Yuan, D. (2024) QoS-Aware Task Offloading and Resource Allocation Optimization in Vehicular Edge Computing Networks via MADDPG. Computer Networks, 242, Article 110282. [Google Scholar] [CrossRef
[11] Li, W., Sun, X., Wan, B., Liu, H., Fang, J. and Wen, Z. (2023) A Hybrid GA-PSO Strategy for Computing Task Offloading Towards MES Scenarios. PeerJ Computer Science, 9, e1273. [Google Scholar] [CrossRef] [PubMed]
[12] Zhou, W., Chen, L., Tang, S., Lai, L., Xia, J., Zhou, F., et al. (2021) Offloading Strategy with PSO for Mobile Edge Computing Based on Cache Mechanism. Cluster Computing, 25, 2389-2401. [Google Scholar] [CrossRef
[13] Liu, J., Ahmed, M., Mirza, M.A., Khan, W.U., Xu, D., Li, J., et al. (2022) RL/DRL Meets Vehicular Task Offloading Using Edge and Vehicular Cloudlet: A Survey. IEEE Internet of Things Journal, 9, 8315-8338. [Google Scholar] [CrossRef
[14] Zhu, Z. and Zhao, H. (2022) A Survey of Deep RL and IL for Autonomous Driving Policy Learning. IEEE Transactions on Intelligent Transportation Systems, 23, 14043-14065. [Google Scholar] [CrossRef
[15] Fofana, N., Letaifa, A.B. and Rachedi, A. (2025) Intelligent Task Offloading in Vehicular Networks: A Deep Reinforcement Learning Perspective. IEEE Transactions on Vehicular Technology, 74, 201-216. [Google Scholar] [CrossRef
[16] Walia, G.K. and Kumar, M. (2025) Computational Offloading and Resource Allocation for IoT Applications Using Decision Tree Based Reinforcement Learning. Ad Hoc Networks, 170, Article 103751. [Google Scholar] [CrossRef
[17] Pang, S., Hou, L., Gui, H., He, X., Wang, T. and Zhao, Y. (2024) Multi-Mobile Vehicles Task Offloading for Vehicle-Edge-Cloud Collaboration: A Dependency-Aware and Deep Reinforcement Learning Approach. Computer Communications, 213, 359-371. [Google Scholar] [CrossRef
[18] Zhao, L., Zhang, E., Wan, S., Hawbani, A., Al-Dubai, A.Y., Min, G., et al. (2024) MESON: A Mobility-Aware Dependent Task Offloading Scheme for Urban Vehicular Edge Computing. IEEE Transactions on Mobile Computing, 23, 4259-4272. [Google Scholar] [CrossRef
[19] Zhang, Y., He, X., Xing, J., Li, W. and Seah, W.K.G. (2024) Load-Balanced Offloading of Multiple Task Types for Mobile Edge Computing in IoT. Internet of Things, 28, Article 101385. [Google Scholar] [CrossRef
[20] Li, Z., Yu, K., Zhou, H. and Wu, X. (2023) DQN-Based Collaborative Computation Offloading for Edge Load Balancing. 2023 8th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC), Beijing, 3-5 November 2023, 1-6. [Google Scholar] [CrossRef
[21] Xie, J., Zheng, F., Wen, W. and Jia, Y. (2024) Price-Based Task Offloading for Load-Imbalance Vehicular Multi-Access Edge Computing. 2024 IEEE 99th Vehicular Technology Conference (VTC2024-Spring), Singapore, 24-27 June 2024, 1-6. [Google Scholar] [CrossRef
[22] OpenStreetMap.
https://openstreetmap.maps.arcgis.com
[23] SUMO (2024) SUMO—Simulation of Urban Mobility.
https://eclipse.dev/sumo/
[24] Lu, J., Li, Q., Guo, B., Li, J., Shen, Y., Li, G., et al. (2022) A Multi-Task Oriented Framework for Mobile Computation Offloading. IEEE Transactions on Cloud Computing, 10, 187-201. [Google Scholar] [CrossRef
[25] 王锦, 张新有. 基于DQN的无人驾驶任务卸载策略[J]. 计算机应用研究, 2022, 39(9): 2738-2744.
[26] Wang, Y., Fang, W., Ding, Y. and Xiong, N. (2021) Computation Offloading Optimization for UAV-Assisted Mobile Edge Computing: A Deep Deterministic Policy Gradient Approach. Wireless Networks, 27, 2991-3006. [Google Scholar] [CrossRef