Algorithms for efficient offline reinforcement learning효율적인 오프라인 강화학습을 위한 알고리즘 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 147
  • Download : 0
Offline reinforcement learning (RL) aims to learn without additional interaction with the environment, based on the pre-collected dataset. It has recently gathered attention due to its promise for real-world applications. Unlike online RL where the agent's predictions can be further corrected through additional interactions, offline RL requires robust policy improvement under the potentially incorrect predictions. To do this, it is necessary to accurately measure the uncertainty of the implicitly or explicitly constructed environment model, and design an algorithm that can find a trade-off between the potential policy performance and the uncertainty in policy evaluation. In this thesis, we study offline RL algorithms about (1) finding a good trade-off using a validation split and (2) learning model that is more robust especially for offline RL.
Advisors
Kim, Kee-Eungresearcher김기응researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 전산학부, 2021.2,[iv, 61 p. :]

Keywords

Machine Learning▼aReinforcement Learning▼aOffline Reinforcement Learning▼aHypergradient▼aBalanced Representation; 기계학습▼a강화학습▼a오프라인 강화학습▼a하이퍼그래디언트▼a표현 밸런싱

URI
http://hdl.handle.net/10203/295721
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=956453&flag=dissertation
Appears in Collection
CS-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0