Deep reinforcement learning online offloading for SWIPT multiple access edge computing network

More computation-intensive and low latency applications are emerging recently, and they are constrained by the computing power and battery life of internet of things (IoT). Simultaneous wireless information and power transfer (SWIPT) with mobile-edge computing (MEC) can improve the data processing c...

Full description

Saved in:
Bibliographic Details
Main Authors: Teckchai, Tiong, Ismail Saad, Tze, Kenneth Kin Teo, Herwansyah Lago
Format: Proceedings
Language:English
English
Published: Institute of Electrical and Electronics Engineers 2021
Subjects:
Online Access:https://eprints.ums.edu.my/id/eprint/32718/1/Deep%20reinforcement%20learning%20online%20offloading%20for%20SWIPT%20multiple%20access%20edge%20computing%20network.ABSTRACT.pdf
https://eprints.ums.edu.my/id/eprint/32718/2/Deep%20Reinforcement%20Learning%20Online%20Offloading%20for%20SWIPT%20Multiple%20Access%20Edge%20Computing%20Network.pdf
https://eprints.ums.edu.my/id/eprint/32718/
https://ieeexplore.ieee.org/document/9612551
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Universiti Malaysia Sabah
Language: English
English
Description
Summary:More computation-intensive and low latency applications are emerging recently, and they are constrained by the computing power and battery life of internet of things (IoT). Simultaneous wireless information and power transfer (SWIPT) with mobile-edge computing (MEC) can improve the data processing capability of energy constrained networks. In this paper, a SWIPT-based MEC system is proposed, comprising a multi-antenna access point (AP), multiple single antenna low power IoT devices and a MEC server. The IoT devices exploit the harvested energy for either locally computing or offloading the tasks to the MEC server. Conventional numerical optimization methods are not able to solve combinatorial problems within the limit of the wireless channel coherence time. Thus, Online Offloading with Deep Reinforcement learning (OODRL) is proposed. The proposed algorithm jointly optimizes the offloading decisions, the time slots devoted to energy harvesting (EH), and local computation/offloading to maximize the MEC computation rate. Deep Q network (DQN) is used to learn the binary offloading decisions from the learning experience. This method no longer needs to solve combinatorial problems. Simulation results are presented to demonstrate that the proposed algorithm is able to approach near-optimal performance and superior in decreasing tasks computation time compared with existing optimization methods, enabling real time optimal resource allocation and offloading achievable in a fast-fading wireless environment.