Value-instilled diffusion for offline goal-conditioned reinforcement learning오프라인 목표조건부 강화학습을 위한 가치관입 디퓨전

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor김기응-
dc.contributor.authorKim, Sungyoon-
dc.contributor.author김성윤-
dc.date.accessioned2024-07-25T19:30:43Z-
dc.date.available2024-07-25T19:30:43Z-
dc.date.issued2023-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045715&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/320527-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.8,[iii, 19 p. :]-
dc.description.abstractThis paper discusses an approach to offline Goal-Conditioned Reinforcement Learning(GCRL) using a diffusion model. GCRL is a problem of learning policies that depend on given goals, and it is important to extract useful information from sparse reward signals. To address this, the proposed methodology introduces a planning-based approach using a conditional diffusion model to generate trajectories for taking actions. Specifically, a goal relabeling method is proposed to overcome sparse rewards, and a conditional diffusion model is tasked to generate trajectories that satisfy certain action value level. The proposed methodology demonstrates high performance in various GCRL experimental environments, as demonstrated through qualitative and quantitative evaluations.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject오프라인 목표지향 강화학습▼a디퓨전 모델▼a플래닝-
dc.subjectOffline goal-conditioned reinforcement learning▼aDiffusion model▼aPlanning-
dc.titleValue-instilled diffusion for offline goal-conditioned reinforcement learning-
dc.title.alternative오프라인 목표조건부 강화학습을 위한 가치관입 디퓨전-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthorKim, Kee-Eung-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0