Abstract
Collaborative edge computing (CEC) has emerged as a promising paradigm, enabling edge nodes to collaborate and execute tasks from end devices. Task offloading is a fundamental problem in CEC that decides when and where tasks are executed upon the arrival of tasks. However, the mobility of users often results in unstable connections, leading to network failures and resource underutilization. Existing works have not adequately addressed joint mobility-aware dependent task offloading and network flow scheduling, resulting in network congestion and suboptimal performance. To address this, we formulate an online joint mobility-aware dependent task offloading and bandwidth allocation problem, to improve the quality of service by reducing task completion time and energy consumption. We introduce a Mobility-aware Digital Twin-assisted Deep Reinforcement Learning (MDT-DRL) algorithm. Our digital twin model equips the reinforcement learning process by providing future states of mobile users, enabling efficient offloading plans for adapting to the mobile CEC system. Experimental results on real-world and synthetic datasets show that MDT-DRL surpasses state-of-the-art baselines on average task completion time and energy consumption.
| Original language | English |
|---|---|
| Pages (from-to) | 2979-2994 |
| Number of pages | 16 |
| Journal | IEEE Transactions on Mobile Computing |
| Volume | 24 |
| Issue number | 4 |
| DOIs | |
| Publication status | Published - 2025 |
Keywords
- Collaborative edge computing
- deep reinforcement learning
- digital twin
- mobility
- task offloading
ASJC Scopus subject areas
- Software
- Computer Networks and Communications
- Electrical and Electronic Engineering