EDGE NETWORK COMPUTING SYSTEM WITH DEEP REINFORCEMENT LEARNING BASED TASK SCHEDULING

Information

  • Patent Application
  • 20230153124
  • Publication Number
    20230153124
  • Date Filed
    September 30, 2021
    3 years ago
  • Date Published
    May 18, 2023
    a year ago
Abstract
An edge network computing system includes: a plurality of terminal devices; a plurality of edge servers connected to the terminal device through an access network; and a plurality of cloud servers connected to the plurality of edge servers through a core network. Each edge server is configured to: receive a plurality of computing tasks originated from one of the plurality of terminal devices; use a deep Q-learning neural network (DQN) with experience replay to select one of the plurality of could servers to offload a portion of the plurality of computing tasks; and send the portion of the plurality of computing tasks to the selected cloud server and forward results of the portion of the plurality of computing tasks received from the selected cloud server to the originating terminal device.
Description
FIELD OF THE DISCLOSURE

The present disclosure generally relates to the field of network computing and, more particularly, relates to an edge network computing system with a deep reinforcement learning based task scheduling.


BACKGROUND

Computing at network edge is a new distributed computing paradigm which brings computation and data storage closer to the location where they are needed to improve response time and save bandwidth in a dynamic networking environment. Despite improvements in network technology, data centers are still unable to guarantee acceptable transfer rates and response time, which are critical requirements for most applications. An objective of the edge network computing is to move the computation and data storage away from the data centers towards the edge of a network, exploiting smart devices, mobile phones, computer terminals or network gateways to perform tasks and provide services on behalf of a cloud of the data centers.


The present disclosure provides a task scheduling/offloading method in the edge network to handle task distribution, offloading and management by applying a deep reinforcement learning model. Specifically, the task offloading problem is formulated as a multi-agent reinforcement learning problem. A decision-making process of each agent is modeled as a Markov decision process and a Q-learning approach configured to deal with a large scale of states and actions. The performance of the task offloading method is evaluated in a simulated environment of an edge computing network based on a simplified multi-armed bandit model. The task offloading method of the present disclosure outperforms the existing task offloading method and provides a lower latency for computation intensive tasks in the edge network.


BRIEF SUMMARY OF THE DISCLOSURE

One aspect or embodiment of the present disclosure includes an edge network computing system. The edge network computing system includes: a plurality of terminal devices, a plurality of edge servers connected to the terminal device through an access network, and a plurality of cloud servers connected to the plurality of edge servers through a core network. Each edge server is configured to: receive a plurality of computing tasks originated from one of the plurality of terminal devices; use a deep Q-learning neural network (DQN) with experience replay to select one of the plurality of could servers to offload a portion of the plurality of computing tasks; and send the portion of the plurality of computing tasks to the selected cloud server and forward results of the portion of the plurality of computing tasks received from the selected cloud server to the originating terminal device.


Another aspect or embodiment of the present disclosure includes a task offloading method for an edge server in an edge network computing system. The task offloading method includes: receiving a plurality of computing tasks from a terminal device, using a deep Q-learning neural network (DQN) with experience replay to select a could server to offload a portion of the plurality of computing tasks, and sending the portion of the plurality of computing tasks to the cloud server and forwarding results of the portion of the plurality of computing tasks received from the cloud server to the terminal device.


Another aspect or embodiment of the present disclosure includes computer-readable storage medium storing a computer program for task offloading. When being executed by an edge server, the computer program performs: receiving a plurality of computing tasks from a terminal device, using a deep Q-learning neural network (DQN) with experience replay to select a could server to offload a portion of the plurality of computing tasks, and sending the portion of the plurality of computing tasks to the cloud server and forwarding results of the portion of the plurality of computing tasks received from the cloud server to the terminal device.


Other aspects or embodiments of the present disclosure can be understood by those skilled in the art in light of the description, the claims, and the drawings of the present disclosure.





BRIEF DESCRIPTION OF THE DRAWINGS

The following drawings are merely examples for illustrative purposes according to various disclosed embodiments and are not intended to limit the scope of the present disclosure.



FIG. 1 depicts a schematic diagram illustrating an exemplary mobile edge computing system according to various disclosed embodiments of the present disclosure;



FIG. 2 depicts a schematic flowchart illustrating an exemplary task offloading method for an edge server in an edge network computing system according to various disclosed embodiments of the present disclosure; and



FIG. 3 depicts a schematic diagram illustrating an exemplary experience replay-based DQN training according to various disclosed embodiments of the present disclosure.





DETAILED DESCRIPTION

Reference will now be made in detail to exemplary embodiments of the disclosure, which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts. In the following description, reference is made to the accompanying drawings that form a part thereof, and in which is shown by way of illustration specific exemplary embodiments in which the disclosure may be practiced.


These embodiments are described in sufficient detail to enable those skilled in the art to practice the disclosure and it is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the disclosure. The following description is, therefore, merely exemplary.


Edge computing is distributed computing paradigm which brings computation and data storage closer to a location where they are needed to improve response time and save bandwidth. Despite improvements in network technology, data centers cannot always guarantee acceptable transfer rates and response time which could be critical requirements for most applications. The objective of the edge computing is to move the computation away from the data centers towards an edge of a cloud computing network, exploiting smart devices, mobile phones or network gateways to perform tasks and provide services on behalf of the cloud computing network. By moving services to the edge, it is possible to provide content caching, service delivery, storage and device management resulting in improved response time and transfer rates.


Mobile edge computing (MEC) is a new network paradigm that provides information technology services within a radio access network of mobile units (MUs). It is predicted that MEC will provide a new ecosystem to migrate intensive computing tasks of MUs. The MEC is located close to MUs and it is deployed within a Radio Access Network (RAN). Therefore, the MEC provides a higher bandwidth with a lower latency to improve Quality of Service (QoS) of the cloud computing network. In the 5G mobile network, the MEC plays an important role to satisfy the stringent delay requirement of the 5G standards.


However, the computation power of edge servers or edge devices may also be limited due to fast growing demand of network users and task loads. When tasks cannot be completely handled by the edge servers, the edge servers offload a portion of the task load to cloud servers at the data centers in the cloud computing network. In this case, it is important to reduce data transfer delays and data processing delays in the cloud computing network or backhaul network. The existing task offloading methods often select the cloud servers according to some static rules (e.g., minimum distance) and may cause substantial data transfer delays and/or data processing delays. On one hand, the static offloading rules are unable to adapt to time-varying network conditions. When data traffics and computation loads vary in the cloud computing network, any offload decision made without considering the time-varying network conditions may keep forwarding heavy data traffic to a node (i.e., a cloud server) that encounters congestion and may further aggravate the congestion at the node. The task offloading method consistent with the present disclosure senses the network conditions and actively avoids congested or busy cloud servers to reduce a task offloading cost. On the other hand, the existing task offloading methods often offload the tasks to a single cloud server. In this case, it is not possible to provide load balancing to the cloud servers. In other words, the edge server continues to offload the tasks to the same cloud server even if it already handles a heavy task load.


The present disclosure provides a mobile edge computing (MEC) server (hereinafter edge server) using deep reinforcement learning to help determining a target cloud server for tasks that need to be offloaded from an edge server. A multi-armed bandit model is used to simulate an MEC environment to evaluate task offloading performance. The simulation results indicate a QoS improvement in the MEC network.



FIG. 1 depicts a schematic diagram illustrating an exemplary edge network computing system according to various disclosed embodiments of the present disclosure. As shown in FIG. 1, the edge network computing system 100 includes a terminal device 110, an access network 120, a core network 130, an edge server 140, and a cloud server 150. Although only one terminal device 110, one edger server 140, and one cloud server 150 are shown in FIG. 1, a plurality of terminal devices, a plurality of edge servers, and a plurality of cloud servers may be present in the edge network computing system 100.


In some embodiments, the terminal device 110 originates and sends a task to the edge server 140 through the access network 120. The edge server 140 may process the task locally and return a result to the originating terminal device 110. In some other embodiments, depending on network conditions, the edge server 140 may offload the task to the cloud server 150 that is determined by the task offloading method. After completing the task, the cloud server 150 returns the result to the edge server 140 and the edge server 140 forwards the result to the originating terminal deice 110. The terminal device 110 communicates with the edge server 140 through the access network 120. The edge server 140 communicates with the cloud server 150 through the core network 130. In a process of processing the task, the cloud server 150 may communicate with the terminal device 110 directly or indirectly through the edge server 140.


The present disclosure also provides a task offloading method for an edge server in an edge network computing system. FIG. 2 depicts a schematic flowchart illustrating an exemplary task offloading method for an edge server in an edge network computing system according to various disclosed embodiments of the present disclosure. As shown in FIG. 2, the task offloading method includes the following processes.


At S210, a plurality of computing tasks is received by an edge server from a terminal device.


In some embodiments, the terminal device sends the plurality of computing tasks to an edge server through the access network connecting between the terminal device and the edge server. Generally, the edge server is located adjacent to the terminal device. The edge server may perform the computing tasks locally or may offload a portion of the plurality of computing tasks to a cloud server through the core network.


In some embodiments, the task offloading of the edge servers is formulated as a multi-agent reinforcement learning problem. Each edge server acts as an agent who makes a series of decisions on task offloading (specifically, cloud server selection) over time. The decision-making process of each agent is modeled as a Markov Decision Process (MDP).


In some embodiments, the edge network computing system includes K terminal devices indexed by k ∈ {1, . . . , K}, M edge servers indexed by m ∈ {1, . . . , M}, and J cloud servers indexed by j ∈ {1, . . . , J}. Each edge server includes an agent, and at each time slot, the edge server m handles Nm tasks indexed by i ∈ {1, . . . , Nm}. Each of Nm task has a size of input data sm,i and its complexity zm,i, defined by a number of CPU cycles required to execute one bit of the task i. The number of CPU cycles required to complete the task i is sm,izm,i, and the number of CPU cycles required to complete all the tasks at the edge server m is Σi=1Nm sm,izm,i.


To mitigate the computational load at the edge servers and guarantee timely processing of the plurality of computing tasks, a proportion of the plurality of computing tasks will be offloaded from the edge server m to the cloud server j. An association xm,j between the edge server m and the cloud server j is:







x

m
,
j


=

{




1
,




if


the


edge


server


m


offloads


to


the


cloud


server


j






0
,



otherwise








where m=1, M, and i=1, . . . , Nm.


In some embodiments, each edge server only selects at most one cloud server for task offloading, that is, Σj=1Jxm,j≤1. For example, the edge server m is associated with the cloud server j. The task assignment between the edge server m and the cloud servers is specified by the following:








a

m
,
i




is



a

m
,
j



=

{




1
,




offloaded


to


a


cloud


server






0
,




executed


by


the


edge


server


m









where m=1, M, and i=1, . . . , Nm.


The time required for locally executing the tasks at the edge server m is given by the following:








t

comp
,
m


(
E
)


=





i
=
1


N
m




(

1
-

a

m
,
i



)



s

m
,
i




z

m
,
i





c
m

(
E
)




,




where cm(E) is the computational capability of the edge server m measured in CPU cycles per second.


cj(C) is the computational capability of the cloud server j that is available for the edge servers and is measured in CPU cycles per second. In one example, cj(C) is a constant. In another example, cj(C) fluctuates according to a Markov process. cj(C) is allocated to the edge servers in a way such that all tasks offloaded from various edge servers associated with the cloud server j are completed at the same time.


The design rationale behind this allocation is that the computational capability is fully utilized. In other words, when the tasks that require less CPU cycles are completed sooner, the released computational capability will be used by other tasks that are still being executed. Under an optimal computational capability allocation that all tasks are completed at the same time, the computational capability allocated to a task is proportional to the number of CPU cycles required to execute the task. Note that the total number of CPU cycles for all tasks at the cloud server j is Σm=1M xm,jΣi=1Nm am,ism,izm,i, the computational capability allocated to the task i at the edge server m by the cloud server j is given by the following:







c

m
,
i


(
C
)


=




a

m
,
i




s

m
,
i




z

m
,
i







m
=
1

M



x

m
,
j







i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i










c
j

(
C
)


.






The execution time for all tasks offloaded to the cloud sever j is given by the following:







t

comp
,
j


(
C
)


=




a

m
,
i




s

m
,
i




z

m
,
i




c

m
,
j


(
C
)



=






m
=
1

M



x

m
,
j







i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i







c
j

(
C
)



.






The communication latency for offloading the tasks from the edge servers to the cloud servers includes an access latency and a data transmission latency. The access latency is a round-trip time for a packet to travel between the edge servers and the cloud servers, which is normally measured with “Ping” messages. The data transmission latency is a time spent on transmitting all the packets, which is determined by a size of transmitted data and a data rate of backhaul connection. Due to time-varying network dynamics, both the access latency and the data rate of backhaul connection fluctuate over time, and such fluctuation is expected to be memoryless. Note that, the access latency includes the uplink component (from the edge server to the cloud server) and the downlink component (from the cloud server to the edge server) of the access latency. In the edge network computing system, the two components are separated by a computing time at the cloud server, and the network dynamics may be changed during this period. Thus, the actual round-trip time could be different from the one measured by “Ping” messages. To obtain accurate round-trip time, the “Ping” message can be modified to record a total elapsed time from the beginning of task offloading to the completion of downloading outcome from the cloud server.


Then, the access latency is calculated by subtracting the data transmission time and the computing time from the total elapsed time. tacc,m,j(UL) and tacc,m,j(DL) are the time-varying uplink and downlink access latency of the backhaul connection between the edge server m and the cloud server j, respectively. Based on the memoryless property, tacc,m,j(UL) and tacc,m,j(DL)(m ∈ {1, . . . , M}, j ∈{1, . . . J}) follow a Markov process with a finite number of states. Each state corresponds to a certain (UL) (UL) range of tacc,m,j(UL). Take tacc,m,j(UL) as an example, a range of the possible values of tacc,m,j(UL) are divided into multiple intervals with an equal length of l, given by [tmin(UL), tmin(UL)+l], [tmin(UL), tmin(UL)+2l], . . . , [tmax(UL)−tmax(UL)], where tmin(UL) and tmax(UL) are the predefined minimum and maximum possible values of tacc,m,j(UL) respectively, which can be set according to historical data of tacc,m,j(UL) Given the interval length l, the total number of intervals is (tmax(UL)−tmin(UL))/l, which is also a number of states. tacc,m,j(UL) is in state s if tacc,m,j(UL) falls into the sth interval, i.e., tacc,m,j(UL) ∈ [tmin(B)+(s−1)l, tmin(B)+sl]. The Markov process for tacc,m,j(DL) is defined in the same way as tacc,m,j(UL). Note that the transition probabilities between different states are unknown to each edge server.


Similarly, rm,j(B) is the data rate of the backhaul connection between the edge server m and the cloud server j. rm,j(B) follows a Markov process with a finite number of states, given by [rmin(B), rmin(B)+l], rmin(B)+l′, [rmin(B)+2l′], . . . , [rmax(B)−l′, rmax(B)], where rmin(B) and rmax(B) are the predefined minimum and maximum possible values of rm,j(B), respectively, and l′ is the length of each interval. Same as tacc,m,j(B), rm,j(B) is in state s′ if rm,j(B) falls into the s′th interval, i.e., rm,j(B) ∈ [rmin(B)+(s′−1)l, rmin(B)+s′l].


With the data rate rm,j(B) and the size of the data to be transmitted Σi=1Nm am,jsm,j, the data transmission latency from the edge server m to the cloud server j is







t

trans
,
m
,
j


(
B
)


=






i
=
1


N
m




a

m
,
j




s

m
,
i





r

m
,
j


(
B
)



.





A size of output data of a task is small (e.g., a decision indicator), the latency for sending the output data (i.e., the outcome) of the task back to the edge server is negligible. In case such latency is not negligible, the time for downloading the task output data can be calculated in the same way as the uploading time.


Given the access latency, data transmission latency, and the computational latency, the total elapsed time for offloading the tasks of the edge server m to cloud server j and completing these tasks is toff,m,j(B)=tacc,m,j(UL)+ttrans,m,j(B)=tcomp,m,j(C)+tacc,m,j(DL). The time for executing the proportion of tasks at the edge server m is







t

comp
,
m


(
E
)


=






i
=
1


N
m




(

1
-

a

m
,
i



)



s

m
,
i




z

m
,
i





c
m

(
E
)



.


t

m
,
j







is the latency for completing all the tasks at the edge server m when associated with the cloud server j (i.e., xm,j=1). Given that a proportion of the tasks are offloaded to the cloud server j and executed there, tm,j is impacted by the dependency of the tasks. Specifically, if all the tasks are independent of each other, they can be executed in parallel. Then, tm,j is determined by a latest set of tasks completed between the edge server m and the cloud server j, which is given by the following:






t
m,j=max{toff,m,j(B),tcomp,m(E)}.


If the tasks are inter-dependent, they have to be executed following certain orders, making the calculation of tm,j more complicated. For example, the tasks of a training simulation system may include user device localization, map downloading and updating, image/video processing, trajectory prediction, and training outcome generation. The last task must rely on outcomes of previous tasks. A simple dependency pattern is considered, in which the tasks need to be executed sequentially. With such a dependency pattern, an output of one task is an input to a subsequent task. Based on the required order of execution, the set of tasks to be executed first (with proportion Σi=1Nm(1−am,i)sm,izm,i) are assigned to be executed by the edge server m, and the remaining tasks (with proportion Σi=1Nmam,ism,izm,i) are assigned to the cloud server j. Then, the output of the tasks executed by the edge server m is used as the input to start executing the tasks offloaded to the cloud server j. In this way, the computing at the edge server m and the offloading from the edge server m to the cloud server j can be performed in parallel, resulting in reduced latency. Under this setting, the time elapsed before the execution at the cloud server j is determined by the slower one between the computing at the edge server m and the offloading from the edge server m to the cloud server j, which is given by max{tcomp,m(E), tacc,m,j(UL)+ttrans,m,j(B)} Finally, tm,j is calculated by the following:






t
m,j=max{tcomp,m(E),tacc,m,j(UL)+ttrans,m,j(B)}+tacc,m,j(DL)+tcomp,m,j.


The optimal task partitioning is achieved when toff,mj(B)=tcomp,m(E) for independent tasks and tcomp,m(E)=tacc,m,j(UL)+ttrans,m,j(B) for sequentially dependent tasks.


In some embodiments, the task offloading of the edge servers is formulated as a multi-agent reinforcement learning problem. Each edge server acts as an agent who makes a series of decisions on task offloading (specifically, cloud server selection) over time. The decision-making process of each agent is modeled as a Markov Decision Process (MDP). For the edge server m, a system state is defined by the observed UL and DL access latencies tacc,m,j(UL) and tacc,m,j(DL) and the backhaul data date rm,j(B)observed at a current time slot. The system state observed by the edge server m at time t is specified by a 1×3 J vector Sm(t)=[sm,1(UL) (t), . . . , sm,j(UL)(t), sm,1(DL)(t), . . . , sm,j(DL)(t), s′m,1(t), . . . s′m,j(t)], where {sm,j(UL)(t)}, {sm,j(DL)(t)}, and {s′m,j(t)}(j ∈ {1, . . . , J}) are the states of tacc,m,j(UL), tacc,m,j(DL), and rm,j(B) at time t, respectively. The action space of the edge server m is specified by a combination of two variables {Am(t), Bm(t)}, which corresponds to the strategies of cloud server selection and task partitioning ratio of the edge server m. The association between the edge server m and various cloud servers is specified by a set of binary variables [xm,1(t), . . . xm,j(t)}, where xm,j=1 indicates that the edge server m offloads a proportion of tasks to the cloud server j and xm,j=0 indicates otherwise. Given the constraint that at most one cloud server will be selected (i.e., Σj=1Jxm,j≤1), the number of possible values of Am(t) is J+1 and can be indexed by the following:








A
m

(
t
)

=

{




j



if


the


cloud


server


j


is


selected





0



if


no


cloud


server


is


selected




,






where m=1, . . . , M, and j=1, . . . , J.


The proportion of tasks (measured in total required CPU cycles) to be offloaded to the cloud server j is given by











i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i





.




Considering that the edge server m may receive different kinds of tasks over time,










i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i








may take any value between 0 and 1. Thus, if the value of










i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i








is directly set as the variable for task partitioning in the action space







(


i
.
e
.

,




B
m

(
t
)

=





i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i







)

,




the dimension of action space would be extremely high. To reduce the complexity of training, Bm(t) is set to take a relatively small number of discrete values between 0 and 1, e.g., {0, 0.1, 0.2, . . . , 0.9, 1}. Once a certain value of Bm(t) is selected, the task assignment is set such that










i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i








is the closest to the selected value of Bm(t),







i
.
e
.

,



"\[LeftBracketingBar]"




B
m

(
t
)

-





i
=
1


N
m




a

m
,
i




s

m
,
i




z

m
,
i








i
=
1


N
m




s

m
,
i




z

m
,
i








"\[RightBracketingBar]"






is minimized. In this case, there is a tradeoff between performance and complexity when different resolution of Bm(t) are selected. The reward of the edge server m at each time slot is set to be−tm,j. The objective is to find the optimal policy that maximizes the expected long-term accumulated discounted reward.


Returning to FIG. 2, at S220, a deep Q-learning neural network (DQN) with experience replay is used for each edge server to select a cloud server to offload a portion of the plurality of computing tasks.


Each edge server includes a reinforcement learning (RL) agent aiming to learn from the environment and take action to maximize its long-term cumulative reward. The environment is modeled as an MDP with state space custom-character and an RL agent can take actions from space custom-character. The agent interacts with the environment by taking actions, observing the reward and system state transition, and updating its knowledge about the environment. The objective of an RL algorithm is to find the optimal policy, which determines the strategy of taking actions under certain system states. A policy π is specified by π(s|a)=P (At=a|St=s). In general, a policy is in a stochastic form to enable exploration over different actions. To find the optimal policy, the key component is to determine the value of each state-action function, also known as Q-function, which is defined by the following:







Q

(


s
t

,

a
t


)




Q

(


s
t

,

a
t


)

+


α
[


r

t
+
1


+

γ


max

a

t
+
1





Q

(


s

t
+
1


,

a

t
+
1



)


-

Q

(


s
t

,

a
t


)


]

.






Therefore, in large scale systems with large numbers of states and actions, the traditional Q-learning approach becomes infeasible since a table is required to store all the Q-values. In addition, traditional Q-learning needs to visit and evaluate every state-action pair, resulting in huge complexity and slow convergence. An effective approach to deal with such a challenge is to use a neural network (NN) to approximate the Q-values, given by Q(s,a,w)≈Qπ(s,a), where w are the weights of the NN. By training a NN with sampled data, the NN can map the inputs of state-action pairs to their corresponding Q-values.


At S230, the portion of plurality of computing tasks is sent to the cloud server and results of the portion of the plurality of computing tasks received by the edge server from the cloud server is forwarded by the edge server to the terminal device.


In some embodiments, each edge server corresponds to a DQN to be trained and the DQN is used to generate task offloading decisions once the training is completed. The input layer of the DQN is the current system state, which consists of 2 J neurons, each corresponds to an element of Sm(t). The output layer is set to generate the values of Q-functions when taking all the J+1 actions, given by Q (Sm(t), 1), Q (Sm(t), 2), . . . , Q (Sm(t), J+1).


The direct application of NN in Q-learning may be unstable or even diverge due to the correlations between observations and the correlations between Q-values and target values. To deal with this challenge, a mechanism called experience replay is applied for the training of the DQN.


As described in the background section, due to the interaction between edge servers that share the computational resource of the same cloud server, the task offloading strategies of different edge servers are coupled, resulting in a multi-agent RL problem. A simple solution to such a problem is independent Q-learning (IQL), where each agent (edge server) regards the activities of other agents as part of the environment and learns the optimal policy based on the received rewards over time. In some embodiments, the edge servers learn to optimize the task offloading policies via the history of observed latency, hence gradually learn to select the proper cloud server. However, with IQL, all edge servers are learning and adjusting their policies simultaneously, the environment from the perspective of each edge server is non-stationary and there may be a Ping-Pong effect. For example, two edge servers may select the same cloud server at a time slot and experience high computational latency. Then, they select another same cloud server at the next time slot and observe high computational latency again. Hence, the system may take an extremely long time to converge.


To break such correlations, experience replay-based deep Q-network (DQN) is considered. The idea is to “freeze” the agent's experience for a certain time and use it to train the DQN later. Specifically, the agent first explores the environment by randomly taking actions and stores the experience, et=(st, at, rt, st+1), in a target network. With the samples randomly drawn from the target network, the weights of the DQN is updated by minimizing the loss function given by the following:








L
i

(

w
i

)

=


E


(

s
,
a
,
r
,

s



)



U

(
D
)



[


(

r
+

γ


max

a





Q

(


s


,

a


,

w
i
-


)


-

Q

(

s
,
a
,

w
i


)


)

2

]





where wi and wiare weights of the DQN and the target network at iteration i, respectively.


The loss function is the mean square error between DQN and target network, which can be minimized through stochastic gradient descent. To reduce the correlation between DQN and target network, the target network is updated less frequently. After the training of DQN, the agent then takes action based on the estimated Q-values. The framework of the experience replay based DQN training is shown in FIG. 3.


With experience replay-based training that “freezes” each agent's experience, the edge servers are not interacting with each other at the same pace and unable to learn from instantaneous feedback. As a result, the experience replay may be unstable, causing the system fail to converge. To stabilize training and accelerate convergence, a key observation is that the environment observed by an single unit (SU) can be made stationary conditioned on the policies of other SUs. However, given that samples generated by experience replay are obsolete (i.e., cannot reflect current system dynamics), the edge servers may not be able disambiguate the age of the sampled data from the replay memory. Importance sampling with off-environment training is an effective approach to tackle this issue. The idea of importance sampling is assigning an importance ratio (or weight) to each sample in a minibatch. More recent experiences have greater weights and less recent experiences have smaller weights to naturally decay obsolete experiences.


In the embodiments of the present disclosure, the deep reinforcement learning algorithm is applied to the edge network computing system. Each edge server includes a reinforcement learning agent. The state space includes the access latency and the back haul data rate. The action space includes the strategies of cloud server selection and the task partitioning ratio. Each edge server corresponds to a DQN trained to generate task offloading decisions. Thus, the optimal policy is used to make task offloading decisions.


The performance of the task offloading method consistent with the present disclosure is evaluated through simulation. The model used for evaluation is called multiarmed bandit model. The objective of the model is to maximize the expected total reward over some time period. In a k-armed bandit problem, each of the k actions that has an expected or mean reward is given. The goal of the agent is to select a proper action form that maximizes the long-term expectation of the reward. In some embodiments, the objective turns to choose a proper target cloud server that minimize the latency for each arrived task. The evaluation includes the following assumptions. The edge server and the cloud servers are undistinguishable, i.e., the task offloading method considers that the edge sever and the cloud servers have the same distribution of the latency. The task offloading does not affect the distribution of the latency, i.e., the task offloading decision does not affect the environment. The latencies of the edge server and the cloud servers are uniformly distributed, as well as the randomness of latency for each of the edge server and the cloud servers. As the RL algorithms maximize the reward, the optimal policy for the task offloading in the edge network computing system is intended to achieve a minimum latency. The relation between the latency and the reward is R=e−T, where R is an immediate reward, and T is a latency in second.


To simulate the scenario where the distribution of the latency changes, two perturbations are added at t=300 and t=600, where, a congestion occurs at t=300 on the connection between the edge server and the cloud server which is the best initial task offloading target (or the computational resource in the edge server is exhausting, and the latency increases). At t=600, a small random perturbation in the MEC system occurs. The best action may or may not change at this time.


The strategy of the naïve algorithms is simply testing every action at the beginning of each run and selects the sever which has the shortest latency. The selected server is used throughout the simulation process. This strategy outperforms at the beginning of the simulation process but fails when the distribution of the latency changes due to lack of flexibility. Both sample average and constant step-size c-greedy algorithms are able to select the suboptimal sever when the best original link is congested or select new optimal sever when a small random perturbation occurs in the system. However, the constant step-size c-greedy algorithm is more resilient to both perturbations. As such, the constant step-size c-greedy algorithm is adaptive to a non-stationary environment.


According to the simulations, the constant step-size c-greedy algorithm shows good performance on adaptively selecting the best servers for incoming tasks even the distribution of the latency in the system changes. It is also observed that the RL method performs well for task offloading in the edge network computing environment. However, the task offloading decisions do not affect the latencies of future tasks. In some other embodiments, the multi-armed bandit model is replaced with deep Q-learning to explore removal of the assumptions and further performance improvement.


The present disclosure provides the task offloading method based on the deep reinforcement learning for the computation intensive tasks within the edge network computing environment. The detailed design and the model description are provided with mathematical analysis and modeling. The simulation environment is built to conduct the simplified validation and evaluation experiment to verify the performance of the task offloading method. Preliminary simulation results show that the task offloading method outperforms the existing task offloading strategies.


The present disclosure also provides a computer-readable storage medium. The computer-readable storage medium stores a computer program. When being executed by an edge sever, the computer program implements the embodiments of the task offloading method shown in FIG. 2. The description thereof is omitted.


The computer-readable storage medium may be an internal storage unit of the edge server described in any of the foregoing embodiments. For example, the computer-readable storage medium may be a hard disk or an internal memory of the device. The computer-readable storage medium may also be an external storage device of the device, such as a plug-in hard disk, a smart media card (SMC), a secure digital (SD) card, a flash card, etc. The computer-readable storage medium may also store the computer program, and other programs and data required by the edge server. The computer-readable storage medium may also temporarily store already outputted data or to-be-outputted data.


Those skilled in the art should understand that all or part of the processes in the foregoing method embodiments can be implemented by instructing relevant hardware through a computer program. The computer program may be stored in the computer-readable storage medium, and when being executed, the computer program implements the processes of the foregoing method embodiments. The storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), or a random-access memory (RAM).


Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.

Claims
  • 1. An edge network computing system, comprising: a plurality of terminal devices;a plurality of edge servers connected to the terminal device through an access network; anda plurality of cloud servers connected to the plurality of edge servers through a core network,wherein each edge server is configured to: receive a plurality of computing tasks originated from one of the plurality of terminal devices;use a deep Q-learning neural network (DQN) with experience replay to select one of the plurality of could servers to offload a portion of the plurality of computing tasks; andsend the portion of the plurality of computing tasks to the selected cloud server and forward results of the portion of the plurality of computing tasks received from the selected cloud server to the originating terminal device.
  • 2. The system according to claim 1, wherein: a first time is a time for completing a remaining portion of the plurality of computing tasks performed by the edge server;a second time is a time for offloading and completing the portion of the plurality of computing tasks offloaded to the cloud server, and includes an uplink access latency, a data transmission time for transmitting the portion of the plurality of computing tasks from the edge server to the cloud server, a computing time for completing the portion of the plurality of computing tasks by the cloud server, and a downlink access latency;when the plurality of computing tasks are independent of each other, a total time for completing the plurality of computing tasks is the greater of the first time and the second time, and an optimal task partitioning is achieved when the first time is equal to the second time; andwhen the plurality of computing tasks are required to be performed sequentially, the total time for completing the plurality of computing tasks includes the greater of the first time and a sum of the uplink access latency and the data transmission time, the computing time, and the downlink access latency, and the optimal task partitioning is achieved when the first time is equal to the sum of the uplink access latency and the data transmission time.
  • 3. The system according to claim 2, wherein: each edge server includes an agent of multi-agent reinforcement learning in a Markov decision process (MDP), and the agent learns from an environment, and takes action to maximize a long-term accumulative reward;a state space includes the uplink access latency, the downlink access latency, and a data rate of a backhaul link between the edge server and the cloud server, observed at a current time slot;an action space includes a plurality of cloud servers as candidates for task offloading, and a task partition ratio having a small number of discrete values; anda reward includes minus the time for completing the plurality of computing tasks.
  • 4. The system according to claim 3, wherein when using the deep Q-learning neural network (DQN) with experience replay to select the could server to offload the portion of the plurality of computing tasks to the cloud server, the edge server is configured to: each time after receiving a plurality of computing tasks, use the DQN to determine a cloud server and a task partition ratio to offload some of the plurality of computing tasks to the cloud server;store each experience including a current state, a current action, a current reward, and a next state in a replay memory;randomly select a minibatch of past experiences from the replay memory to input to the DQN and a target network;optimize the DQN with updated weights based on Q-values from the DQN and Q-values from the target network; andperiodically update weights of the target network with the weights of the DQN.
  • 5. The system according to claim 4, wherein: the DQN includes an input layer of neurons representing the state space and the action space, and an output layer of Q-values for the state space and the action space at the input layer;training the DQN includes mapping inputs of state-action pairs in the minibatch of past experiences to their corresponding Q-values by approximating Q-values and minimizing a loss function of mean square error between the Q-values outputted by the DQN and the Q-values outputted by the target network; andthe target network includes a structure same as the DQN and weights periodically updated by the weights of the DQN obtained from training.
  • 6. The system according to claim 4, wherein: each experience stored in the replay memory includes an importance weight to naturally decay obsolete data.
  • 7. The system according to claim 1, wherein: the edge network computing system is a mobile edge computing system;the plurality of terminal devices are mobile devices; andthe plurality of mobile devices are connected to the plurality of edge servers through a radio access network.
  • 8. A task offloading method for an edge server in an edge network computing system, comprising: receiving a plurality of computing tasks from a terminal device;using a deep Q-learning neural network (DQN) with experience replay to select a could server to offload a portion of the plurality of computing tasks; andsending the portion of the plurality of computing tasks to the cloud server and forwarding results of the portion of the plurality of computing tasks received from the cloud server to the terminal device.
  • 9. The method according to claim 8, wherein: a first time is a time for completing a remaining portion of the plurality of computing tasks performed by the edge server;a second time is a time for offloading and completing the portion of the plurality of computing tasks offloaded to the cloud server, and includes an uplink access latency, a data transmission time for transmitting the portion of the plurality of computing tasks from the edge server to the cloud server, a computing time for completing the portion of the plurality of computing tasks by the cloud server, and a downlink access latency;when the plurality of computing tasks are independent of each other, a total time for completing the plurality of computing tasks is the greater of the first time and the second time, and an optimal task partitioning is achieved when the first time is equal to the second time; andwhen the plurality of computing tasks are required to be performed sequentially, the total time for completing the plurality of computing tasks includes the greater of the first time and a sum of the uplink access latency and the data transmission time, the computing time, and the downlink access latency, and the optimal task partitioning is achieved when the first time is equal to the sum of the uplink access latency and the data transmission time.
  • 10. The method according to claim 9, wherein: each edge server includes an agent of multi-agent reinforcement learning in a Markov decision process (MDP), and the agent learns from an environment, and takes action to maximize a long-term accumulative reward;a state space includes the uplink access latency, the downlink access latency, and a data rate of a backhaul link between the edge server and the cloud server, observed at a current time slot;an action space includes a plurality of cloud servers as candidates for task offloading, and a task partition ratio having a small number of discrete values; anda reward includes minus the time for completing the plurality of computing tasks.
  • 11. The method according to claim 10, wherein using the deep Q-learning neural network (DQN) with experience replay to select the could server to offload the portion of the plurality of computing tasks to the cloud server includes: each time after receiving a plurality of computing tasks, using the DQN to determine a cloud server and a task partition ratio to offload some of the plurality of computing tasks to the cloud server;storing each experience including a current state, a current action, a current reward, and a next state in a replay memory;randomly selecting a minibatch of past experiences from the replay memory to input to the DQN and a target network;optimizing the DQN with updated weights based on Q-values from the DQN and Q-values from the target network; andperiodically updating weights of the target network with the weights of the DQN.
  • 12. The method according to claim 10, wherein: the DQN includes an input layer of neurons representing the state space and the action space, and an output layer of Q-values for the state space and the action space at the input layer;training the DQN includes mapping inputs of state-action pairs in the minibatch of past experiences to their corresponding Q-values by approximating Q-values and minimizing a loss function of mean square error between the Q-values outputted by the DQN and the Q-values outputted by the target network; andthe target network includes a structure same as the DQN and weights periodically updated by the weights of the DQN obtained from training.
  • 13. The method according to claim 10, wherein: each experience stored in the replay memory includes an importance weight to naturally decay obsolete data.
  • 14. The method according to claim 8, wherein: the edge network computing system includes a plurality of terminal devices, a plurality of edge servers, and a plurality of cloud server;the plurality of terminal devices are connected to the plurality of edge servers through an access network; andthe plurality of edge servers are connected to the plurality of cloud server through a core network.
  • 15. The method according to claim 14, wherein: the edge network computing system is a mobile edge computing system;the plurality of terminal devices are mobile devices; andthe plurality of mobile devices are connected to the plurality of edge servers through a radio access network.
  • 16. A computer-readable storage medium storing a computer program for task offloading, when being executed by an edge server, the computer program performing: receiving a plurality of computing tasks from a terminal device;using a deep Q-learning neural network (DQN) with experience replay to select the could server to offload a portion of the plurality of computing tasks; andsending the portion of the plurality of computing tasks to the cloud server and forwarding results of the portion of the plurality of computing tasks received from the cloud server to the terminal device.
  • 17. The computer-readable storage medium according to claim 16, wherein: a first time is a time for completing a remaining portion of the plurality of computing tasks performed by the edge server;a second time is a time for offloading and completing the portion of the plurality of computing tasks offloaded to the cloud server, and includes an uplink access latency, a data transmission time for transmitting the portion of the plurality of computing tasks from the edge server to the cloud server, a computing time for completing the portion of the plurality of computing tasks by the cloud server, and a downlink access latency;when the plurality of computing tasks are independent of each other, a total time for completing the plurality of computing tasks is the greater of the first time and the second time, and an optimal task partitioning is achieved when the first time is equal to the second time; andwhen the plurality of computing tasks are required to be performed sequentially, the total time for completing the plurality of computing tasks includes the greater of the first time and a sum of the uplink access latency and the data transmission time, the computing time, and the downlink access latency, and the optimal task partitioning is achieved when the first time is equal to the sum of the uplink access latency and the data transmission time.
  • 18. The computer-readable storage medium according to claim 17, wherein: each edge server includes an agent of multi-agent reinforcement learning in a Markov decision process (MDP), and the agent learns from an environment, and takes action to maximize a long-term accumulative reward;a state space at each edge server includes the uplink access latency, the downlink access latency, and a data rate of a backhaul link between the edge server and the cloud server, observed at a current time slot;an action space at each edge server includes a plurality of cloud servers as candidates for task offloading, and a task partition ratio having a small number of discrete values; anda reward includes minus the time for completing the plurality of computing tasks.
  • 19. The computer-readable storage medium according to claim 18, wherein using the deep Q-learning neural network (DQN) with experience replay to select the could server to offload the portion of the plurality of computing tasks to the cloud server includes: each time after receiving a plurality of computing tasks, using the DQN to determine a cloud server and a task partition ratio to offload some of the plurality of computing tasks to the cloud server;storing each experience including a current state, a current action, a current reward, and a next state in a replay memory;randomly selecting a minibatch of past experiences from the replay memory to input to the DQN and a target network;optimizing the DQN with updated weights based on Q-values from the DQN and Q-values from the target network; andperiodically updating weights of the target network with the weights of the DQN.
  • 20. The computer-readable storage medium according to claim 18, wherein: the DQN includes an input layer of neurons representing the state space and the action space, and an output layer of Q-values for the state space and the action space at the input layer;training the DQN includes mapping inputs of state-action pairs in the minibatch of past experiences to their corresponding Q-values by approximating Q-values and minimizing a loss function of mean square error between the Q-values outputted by the DQN and the Q-values outputted by the target network; andthe target network includes a structure same as the DQN and weights periodically updated by the weights of the DQN obtained from training.
GOVERNMENT RIGHTS

The present disclosure was made with Government support under Contract No. W912CG21P0001, awarded by United States Army. The U.S. Government has certain rights in the present disclosure.