Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system
With the rapid development of communication technology, various wireless terminals have been invented and applied, while the quality of communication services required by the terminals has gradually improved as well. NOMA (Non-orthogonal Multiple Access) technology was proposed to meet the chall...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/164005 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-164005 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1640052023-01-04T02:55:47Z Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system Dong, Junyi Teh Kah Chan School of Electrical and Electronic Engineering EKCTeh@ntu.edu.sg Engineering::Electrical and electronic engineering::Wireless communication systems With the rapid development of communication technology, various wireless terminals have been invented and applied, while the quality of communication services required by the terminals has gradually improved as well. NOMA (Non-orthogonal Multiple Access) technology was proposed to meet the challenges posed by the rapid growth in the number of terminals and mobile data traffic, especially the scarcity of spectrum resources. This technology actively introduces interference in the power or code domain at the transmitter, allowing multiple users to share the same spectrum resource block. At the same time, SIC (Successive Interference Cancellation ) technology is used at the receiver to decode the correct user information. Therefore, NOMA can help increase the efficiency of spectrum utilization and system capacity. For energy and cost saving, the EE (Energy Efficiency) of NOMA systems is the focus of various studies and an important indicator of system performance. So, a reasonable allocation of channel resources while meeting the basic requirements of the NOMA system is one of the most critical problems to be solved. At the same time, ML (Machine Learning) technology has also witnessed many breakthroughs. DRL (Deep Reinforcement Learning), which is derived from RL (Reinforcement Learning) and DL (Deep Learning), has been applied in many fields with good results due to its self-learning and unsupervised advantages. This project focuses on whether the DRL method can be applied to NOMA systems for better EE. The DQN (Deep Q-learning Network)-DDPG (Deep Deterministic Policy Gradient) network is firstly constructed with subchannel assignment and transmission power allocation as the two decision variables. The DQN is responsible for deciding how the subchannels are allocated, while the DDPG generates the strategy of transmission power for each user on each subchannel. Then, the DDQN (Double Deep Q Networks)-TD3(Twin Delayed DDPG) algorithm is proposed as an optimized method which uses DDQN and TD3 instead of DQN and DDPG, respectively. It is demonstrated that this algorithm could achieve higher system EE after simulation. Besides, this project also explores the feasibility of an A3C-based resource allocation network. Master of Science (Communications Engineering) 2023-01-04T02:55:46Z 2023-01-04T02:55:46Z 2022 Thesis-Master by Coursework Dong, J. (2022). Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/164005 https://hdl.handle.net/10356/164005 en ISM-DISS-02775 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering::Wireless communication systems |
spellingShingle |
Engineering::Electrical and electronic engineering::Wireless communication systems Dong, Junyi Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
description |
With the rapid development of communication technology, various wireless
terminals have been invented and applied, while the quality of communication
services required by the terminals has gradually improved
as well. NOMA (Non-orthogonal Multiple Access) technology was proposed
to meet the challenges posed by the rapid growth in the number
of terminals and mobile data traffic, especially the scarcity of spectrum
resources. This technology actively introduces interference in the power
or code domain at the transmitter, allowing multiple users to share the
same spectrum resource block. At the same time, SIC (Successive Interference
Cancellation ) technology is used at the receiver to decode the
correct user information. Therefore, NOMA can help increase the efficiency
of spectrum utilization and system capacity. For energy and cost
saving, the EE (Energy Efficiency) of NOMA systems is the focus of
various studies and an important indicator of system performance. So,
a reasonable allocation of channel resources while meeting the basic requirements
of the NOMA system is one of the most critical problems
to be solved. At the same time, ML (Machine Learning) technology
has also witnessed many breakthroughs. DRL (Deep Reinforcement
Learning), which is derived from RL (Reinforcement Learning) and DL
(Deep Learning), has been applied in many fields with good results due
to its self-learning and unsupervised advantages. This project focuses on
whether the DRL method can be applied to NOMA systems for better
EE. The DQN (Deep Q-learning Network)-DDPG (Deep Deterministic
Policy Gradient) network is firstly constructed with subchannel assignment
and transmission power allocation as the two decision variables.
The DQN is responsible for deciding how the subchannels are
allocated, while the DDPG generates the strategy of transmission power
for each user on each subchannel. Then, the DDQN (Double Deep Q
Networks)-TD3(Twin Delayed DDPG) algorithm is proposed as an optimized
method which uses DDQN and TD3 instead of DQN and DDPG,
respectively. It is demonstrated that this algorithm could achieve higher
system EE after simulation. Besides, this project also explores the feasibility
of an A3C-based resource allocation network. |
author2 |
Teh Kah Chan |
author_facet |
Teh Kah Chan Dong, Junyi |
format |
Thesis-Master by Coursework |
author |
Dong, Junyi |
author_sort |
Dong, Junyi |
title |
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
title_short |
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
title_full |
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
title_fullStr |
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
title_full_unstemmed |
Deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (NOMA) system |
title_sort |
deep reinforcement learning for dynamic power allocation for non-orthogonal multiple-access (noma) system |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/164005 |
_version_ |
1754611258750926848 |