Machine Learning-Based NOMA for Multiuser MISO Broadcast Channels
- Authors
- Kang, Min Jeong; Lee, Jung Hoon; Ryu, Jong Yeol
- Issue Date
- Jan-2024
- Publisher
- Institute of Electrical and Electronics Engineers Inc.
- Keywords
- decoding order; deep neural network; machine learning; multiple-input single-output broadcast channel; Non-orthogonal multiple access
- Citation
- IEEE Communications Letters, v.28, no.1, pp 93 - 97
- Pages
- 5
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE Communications Letters
- Volume
- 28
- Number
- 1
- Start Page
- 93
- End Page
- 97
- URI
- https://scholarworks.gnu.ac.kr/handle/sw.gnu/69473
- DOI
- 10.1109/LCOMM.2023.3338475
- ISSN
- 1089-7798
1558-2558
- Abstract
- This letter proposes machine learning-based non-orthogonal multiple access (NOMA) for a multiuser multiple-input single-output (MISO) broadcast channel (BC), where a transmitter selects and serves multiple users with a fixed rate using NOMA. Unlike a single-input single-output NOMA, where the optimal decoding order is determined by channel gains, the optimal decoding order for a MISO NOMA should be found with a brute force approach. In this case, the transmitter needs to check the validity of all decoding orders under a transmit power budget, so should calculate power allocations for all decoding order candidates, whose complexity is prohibitive as the number of users increases. Our proposed scheme uses a machine learning model to directly predict the optimal decoding order based on channel gains and cross-channel correlations. Once the decoding order is determined, we only need to calculate the power allocation for that decoding order, greatly reducing complexity. Our results show that our machine learning model performs well in finding the optimal decoding order and achieves performance comparable to the optimal scheme. © 1997-2012 IEEE.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - 해양과학대학 > 지능형통신공학과 > Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.