Cited 9 time in
Multi-Agent Reinforcement Learning for a Random Access Game
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lee, Dongwoo | - |
| dc.contributor.author | Zhao, Yu | - |
| dc.contributor.author | Seo, Jun-Bae | - |
| dc.contributor.author | Lee, Joohyun | - |
| dc.date.accessioned | 2022-12-26T05:41:35Z | - |
| dc.date.available | 2022-12-26T05:41:35Z | - |
| dc.date.issued | 2022-08 | - |
| dc.identifier.issn | 0018-9545 | - |
| dc.identifier.issn | 1939-9359 | - |
| dc.identifier.uri | https://scholarworks.gnu.ac.kr/handle/sw.gnu/1011 | - |
| dc.description.abstract | This work investigates a random access (RA) game for a time-slotted RA system, where N players choose a set of slots of a frame and each frame consists of M multiple time slots. We obtain the pure strategy Nash equilibria (PNEs) of this RA game, where slots are fully utilized as in the centralized scheduling. As an algorithm to realize a PNE (Pure strategy Nash Equilibrium), we propose an Exponential-weight algorithm for Exploration and Exploitation (EXP3)-based multi-agent (MA) learning algorithm, which has the computational complexity of O(N (NmaxT)-T-2). EXP3 is a bandit algorithm designed to find an optimal strategy in a multi-armed bandit (MAB) problem that users do not know the expected payoff of each strategy. Our simulation results show that the proposed algorithm can achieve PNEs. Moreover, it can adapt to time-varying environments, where the number of players varies over time. | - |
| dc.format.extent | 6 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Institute of Electrical and Electronics Engineers | - |
| dc.title | Multi-Agent Reinforcement Learning for a Random Access Game | - |
| dc.type | Article | - |
| dc.publisher.location | 미국 | - |
| dc.identifier.doi | 10.1109/TVT.2022.3176722 | - |
| dc.identifier.scopusid | 2-s2.0-85130779348 | - |
| dc.identifier.wosid | 000846892800095 | - |
| dc.identifier.bibliographicCitation | IEEE Transactions on Vehicular Technology, v.71, no.8, pp 9119 - 9124 | - |
| dc.citation.title | IEEE Transactions on Vehicular Technology | - |
| dc.citation.volume | 71 | - |
| dc.citation.number | 8 | - |
| dc.citation.startPage | 9119 | - |
| dc.citation.endPage | 9124 | - |
| dc.type.docType | Article | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.relation.journalResearchArea | Engineering | - |
| dc.relation.journalResearchArea | Telecommunications | - |
| dc.relation.journalResearchArea | Transportation | - |
| dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
| dc.relation.journalWebOfScienceCategory | Telecommunications | - |
| dc.relation.journalWebOfScienceCategory | Transportation Science & Technology | - |
| dc.subject.keywordPlus | ALOHA | - |
| dc.subject.keywordAuthor | Multi-armed bandit | - |
| dc.subject.keywordAuthor | nash equilibrium | - |
| dc.subject.keywordAuthor | non-cooperative game | - |
| dc.subject.keywordAuthor | random access | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
Gyeongsang National University Central Library, 501, Jinju-daero, Jinju-si, Gyeongsangnam-do, 52828, Republic of Korea+82-55-772-0534
COPYRIGHT 2022 GYEONGSANG NATIONAL UNIVERSITY LIBRARY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
