DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Yi, Yung | - |
dc.contributor.advisor | 이융 | - |
dc.contributor.author | Son, Kyunghwan | - |
dc.date.accessioned | 2019-09-04T02:43:49Z | - |
dc.date.available | 2019-09-04T02:43:49Z | - |
dc.date.issued | 2019 | - |
dc.identifier.uri | http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=843399&flag=dissertation | en_US |
dc.identifier.uri | http://hdl.handle.net/10203/266898 | - |
dc.description | 학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2019.2,[iv, 21 p. :] | - |
dc.description.abstract | Multi-agent reinforcement learning tasks require that agents learn in a stable and scalable manner. To this end, we explore solutions in centralized training and decentralized execution (CTDE) regime popularized recently and focus on value-based methods. VDN and QMIX are representative examples employing centralized training to resolve instability and non-stationarity issues, and distributed execution to render the algorithm scalable. While appropriately factorizing the joint value functions into individual ones is key to distributed execution, we find that the existing methods of value function factorization address only a fraction of game-theoretically modelable MARL tasks. We propose QREG, which takes on a new approach to value function factorization: regularizing the joint value function. This approach translates to relaxing the previously assumed conditions placed on the nature of the value functions. Upon relaxing those assumptions, we illustrate that QREG covers every game satisfying a set of relatively mild conditions, enabling QREG to cover a wider class of games. Our simulations indicate superior performance in a variety of settings, with especially larger margins in games whose payoffs penalize non-cooperative behavior more harshly. | - |
dc.language | eng | - |
dc.publisher | 한국과학기술원 | - |
dc.subject | Machine learning▼adeep learning▼areinforcement learning▼amulti-agent reinforcement learning | - |
dc.subject | 기계 학습▼a딥 러닝▼a강화 학습▼a다중 에이전트 강화 학습 | - |
dc.title | Learning to factorize with regularization for cooperative multi-agent reinforcement learning | - |
dc.title.alternative | 협력을 위한 다중 에이전트 강화 학습에서의 정규화를 통한 분해 학습 | - |
dc.type | Thesis(Master) | - |
dc.identifier.CNRN | 325007 | - |
dc.description.department | 한국과학기술원 :전기및전자공학부, | - |
dc.contributor.alternativeauthor | 손경환 | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.