Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system

With the rapid development of communication technology, various wireless terminals have been invented and applied, while the quality of communication services required by the terminals has gradually improved as well. NOMA (Non-orthogonal Multiple Access) technology was proposed to meet the chall...

Full description

Saved in:
Bibliographic Details
Main Author: Dong, Junyi
Other Authors: Teh Kah Chan
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2023
Subjects:
Online Access:https://hdl.handle.net/10356/164005
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-164005
record_format dspace
spelling sg-ntu-dr.10356-1640052023-01-04T02:55:47Z Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system Dong, Junyi Teh Kah Chan School of Electrical and Electronic Engineering EKCTeh@ntu.edu.sg Engineering::Electrical and electronic engineering::Wireless communication systems With the rapid development of communication technology, various wireless terminals have been invented and applied, while the quality of communication services required by the terminals has gradually improved as well. NOMA (Non-orthogonal Multiple Access) technology was proposed to meet the challenges posed by the rapid growth in the number of terminals and mobile data traffic, especially the scarcity of spectrum resources. This technology actively introduces interference in the power or code domain at the transmitter, allowing multiple users to share the same spectrum resource block. At the same time, SIC (Successive Interference Cancellation ) technology is used at the receiver to decode the correct user information. Therefore, NOMA can help increase the efficiency of spectrum utilization and system capacity. For energy and cost saving, the EE (Energy Efficiency) of NOMA systems is the focus of various studies and an important indicator of system performance. So, a reasonable allocation of channel resources while meeting the basic requirements of the NOMA system is one of the most critical problems to be solved. At the same time, ML (Machine Learning) technology has also witnessed many breakthroughs. DRL (Deep Reinforcement Learning), which is derived from RL (Reinforcement Learning) and DL (Deep Learning), has been applied in many fields with good results due to its self-learning and unsupervised advantages. This project focuses on whether the DRL method can be applied to NOMA systems for better EE. The DQN (Deep Q-learning Network)-DDPG (Deep Deterministic Policy Gradient) network is firstly constructed with subchannel assignment and transmission power allocation as the two decision variables. The DQN is responsible for deciding how the subchannels are allocated, while the DDPG generates the strategy of transmission power for each user on each subchannel. Then, the DDQN (Double Deep Q Networks)-TD3(Twin Delayed DDPG) algorithm is proposed as an optimized method which uses DDQN and TD3 instead of DQN and DDPG, respectively. It is demonstrated that this algorithm could achieve higher system EE after simulation. Besides, this project also explores the feasibility of an A3C-based resource allocation network. Master of Science (Communications Engineering) 2023-01-04T02:55:46Z 2023-01-04T02:55:46Z 2022 Thesis-Master by Coursework Dong, J. (2022). Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/164005 https://hdl.handle.net/10356/164005 en ISM-DISS-02775 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering::Wireless communication systems
spellingShingle Engineering::Electrical and electronic engineering::Wireless communication systems
Dong, Junyi
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
description With the rapid development of communication technology, various wireless terminals have been invented and applied, while the quality of communication services required by the terminals has gradually improved as well. NOMA (Non-orthogonal Multiple Access) technology was proposed to meet the challenges posed by the rapid growth in the number of terminals and mobile data traffic, especially the scarcity of spectrum resources. This technology actively introduces interference in the power or code domain at the transmitter, allowing multiple users to share the same spectrum resource block. At the same time, SIC (Successive Interference Cancellation ) technology is used at the receiver to decode the correct user information. Therefore, NOMA can help increase the efficiency of spectrum utilization and system capacity. For energy and cost saving, the EE (Energy Efficiency) of NOMA systems is the focus of various studies and an important indicator of system performance. So, a reasonable allocation of channel resources while meeting the basic requirements of the NOMA system is one of the most critical problems to be solved. At the same time, ML (Machine Learning) technology has also witnessed many breakthroughs. DRL (Deep Reinforcement Learning), which is derived from RL (Reinforcement Learning) and DL (Deep Learning), has been applied in many fields with good results due to its self-learning and unsupervised advantages. This project focuses on whether the DRL method can be applied to NOMA systems for better EE. The DQN (Deep Q-learning Network)-DDPG (Deep Deterministic Policy Gradient) network is firstly constructed with subchannel assignment and transmission power allocation as the two decision variables. The DQN is responsible for deciding how the subchannels are allocated, while the DDPG generates the strategy of transmission power for each user on each subchannel. Then, the DDQN (Double Deep Q Networks)-TD3(Twin Delayed DDPG) algorithm is proposed as an optimized method which uses DDQN and TD3 instead of DQN and DDPG, respectively. It is demonstrated that this algorithm could achieve higher system EE after simulation. Besides, this project also explores the feasibility of an A3C-based resource allocation network.
author2 Teh Kah Chan
author_facet Teh Kah Chan
Dong, Junyi
format Thesis-Master by Coursework
author Dong, Junyi
author_sort Dong, Junyi
title Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
title_short Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
title_full Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
title_fullStr Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
title_full_unstemmed Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
title_sort deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (noma) system
publisher Nanyang Technological University
publishDate 2023
url https://hdl.handle.net/10356/164005
_version_ 1754611258750926848