Skip to content

It's a implementation about the paper Liang Huang, Suzhi Bi, and Ying-jun Angela Zhang, "Deep Reinforcement Learning for Online Computation Offloading in Wireless Powered Mobile-Edge Computing Networks", on https://ieeexplore.ieee.org/document/8771176

Notifications You must be signed in to change notification settings

yinrhh/RL-for-binary-computation-offloading-in-wireless-powered-MEC-networks

 
 

Repository files navigation

RL-for-binary-computation-offloading-in-wireless-powered-MEC-networks

I've modified some parameters and the number of neurons in the decoder network, and I switch the decoder method to 'K nearest neighbor'. This work considered a wireless powered mobile edge-computing network that follows binary offloading policy. That is, each computation is either computed on the energy harvesting WD or offloaded to an server(AP). The goal of DROO is to optimize the offloading decisions and allocate the resouces according to the channel conditions(channel gain).

Performance

  • The train loss and the test loss. In fact it's validation loss cause I calculate it during training phase. In addition, the decoder never seen these data.
  • train
  • The nomalized computation rate during training the nomalized computation rate is defined as follows

The blue curve denotes the moving average of \hat{Q} over the last 50 time frames, and the light blue shadow denotes the maximum and minimum of \hat{Q} in the last 50 frames.

  • train

I test the performance of the model every 100 time frames, we can see that \hat{Q} is almost 0.95 after 30000 time frames

  • train

Data

The folder includs all pre-generated training and testing data set, including:

  • data_#.mat: , where # = {10, 20, 30} is the number of WDs

Data samples are generated by enumerating all 2^N binary offloading actions for N <= 10 and by following the CD method presented in https://ieeexplore.ieee.org/document/8334188 for N = 20, 30. There are 30,000 (for N = 10, 20, 30) or 10,000 (otherwise) samples saved in each *.mat file. Where each data sample includes:

variable description
input_h The wireless channel gain between WDs and the AP
output_mode The optimal binary offloading action
output_a The optimal fraction of time that the AP broadcasts RF energy for the WDs to harvest
output_tau The optimal fraction of time allocated to WDs for task offloading
output_obj The optimal weighted sum computation rate

About

It's a implementation about the paper Liang Huang, Suzhi Bi, and Ying-jun Angela Zhang, "Deep Reinforcement Learning for Online Computation Offloading in Wireless Powered Mobile-Edge Computing Networks", on https://ieeexplore.ieee.org/document/8771176

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%