| Literature DB >> 35684719 |
Abstract
Recently, intelligent IoT applications based on artificial intelligence (AI) have been deployed with mobile edge computing (MEC). Intelligent IoT applications demand more computing resources and lower service latencies for AI tasks in dynamic MEC environments. Thus, in this paper, considering the resource scalability and resource optimization of edge computing, an intelligent task dispatching model using a deep Q-network, which can efficiently use the computing resource of edge nodes is proposed to maximize the computation ability of the cluster edge system, which consists of multiple edge nodes. The cluster edge system can be implemented with the Kubernetes technology. The objective of the proposed model is to minimize the average response time of tasks offloaded to the edge computing system and optimize the resource allocation for computing the offloaded tasks. For this, we first formulate the optimization problem of resource allocation as a Markov decision process (MDP) and adopt a deep reinforcement learning technology to solve this problem. Thus, the proposed intelligent task dispatching model is designed based on a deep Q-network (DQN) algorithm to update the task dispatching policy. The simulation results show that the proposed model archives a better convergence performanc in terms of the average completion time of all offloaded tasks, than existing task dispatching methods, such as the Random Method, Least Load Method and Round-Robin Method, and has a better task completion rate than the existing task dispatching method when using the same resources as the cluster edge system.Entities:
Keywords: clustering; deep reinforcement learning; edge computing; task offloading
Year: 2022 PMID: 35684719 PMCID: PMC9185231 DOI: 10.3390/s22114098
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Comparison of relevant works.
| Work | Objective | Algorithm | Environments |
|---|---|---|---|
| Our work | Average task service delay | DQN | Static, Dynamic |
| [ | Average task service delay | SARSA | Static, Dynamic |
| [ | Task service delay for | DDPG | Dynamic |
| [ | Resource utilization for an | DQN | Static |
| [ | Average task service delay | MCTS | Static, Dynamic |
| [ | Task satisfaction degree for | Q-network | Static, Dynamic |
| [ | Service Migration | Multi-Agent DRL | Static, Dynamic |
| [ | Average task service delay | DRL | Dynamic |
| [ | Energy consumption for an | DRL | Dynamic |
Variables and notations used in our model.
| Notation | Definition |
|---|---|
|
| The |
|
| The edge controller in the cluster edge |
|
| |
|
| Collaborative core Cloud |
|
| Task of |
|
| The type of the task |
|
| The number of CPU cycles requested in the |
|
| The number of CPU cycles requested |
|
| The data size of the single task |
|
| The data size of |
|
| The task’s result deadline in the task required |
|
| The wireless link bandwidth between |
|
| The transmit power of |
|
| The channel gain of MN and |
|
| Task service delay of |
|
| |
|
| Bundle task service delay of |
|
| The task transmission delay |
|
| The task queuing delay |
|
| The task computation processing delay |
|
| The queuing delay in the task waiting queue of |
|
| The queuing delay in the task waiting queue of |
|
| The computation processing time of |
|
| The total computing resource of |
|
| The queuing delay in edge controller |
| The queuing delay in | |
|
| The average task service delay |
Figure 1Cluster edge system-based network model.
Figure 2DRL-based intelligent task dispatching method (DDM) using DQN in the cluster edge.
Figure 3Illustration of the proposed DQN for task dispatching policy.
Main simulation parameters for the environment.
| Parameters | Description | Value |
|---|---|---|
|
| The number of static nodes for | 50 |
|
| The number of mobile nodes |
|
|
| The number of sub-tasks in |
|
|
| The number of edge nodes in |
|
|
| The data size of the task |
|
|
| The total number of CPU |
|
|
| The tolerant service delay of |
|
The hyperparameters for DQN learning.
| Parameters | Description | Value |
|---|---|---|
|
| The number of iterations | 5000 |
|
| Learning rate | 0.005 |
|
| The size of experience replay | 10.000 |
|
| The number of mini-batches | 8 |
|
| The size of mini-batches | 32 |
|
| Factor discounting future | 0.9 |
|
| Step parameters | 1500 |
Figure 4Network model of simulation scenario with static nodes.
Figure 5The convergence performance of the proposed DDM model on: (a) average task service delay; (b) task completion rate with the number of edge node = 4.
Figure 6Average task service delay according to the computation capacity of edge nodes.
Figure 7Average task service delay according to the number of edge nodes.
Figure 8Network model of simulation scenario with mobile nodes.
Figure 9Average task service delay according to the number of mobile nodes.
Figure 10Task successful ratio according to the number of mobile nodes.