Value-instilled diffusion for offline goal-conditioned reinforcement learning오프라인 목표조건부 강화학습을 위한 가치관입 디퓨전

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 4
  • Download : 0
This paper discusses an approach to offline Goal-Conditioned Reinforcement Learning(GCRL) using a diffusion model. GCRL is a problem of learning policies that depend on given goals, and it is important to extract useful information from sparse reward signals. To address this, the proposed methodology introduces a planning-based approach using a conditional diffusion model to generate trajectories for taking actions. Specifically, a goal relabeling method is proposed to overcome sparse rewards, and a conditional diffusion model is tasked to generate trajectories that satisfy certain action value level. The proposed methodology demonstrates high performance in various GCRL experimental environments, as demonstrated through qualitative and quantitative evaluations.
Advisors
김기응researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.8,[iii, 19 p. :]

Keywords

오프라인 목표지향 강화학습▼a디퓨전 모델▼a플래닝; Offline goal-conditioned reinforcement learning▼aDiffusion model▼aPlanning

URI
http://hdl.handle.net/10203/320527
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045715&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0