Learning input-agnostic manipulation directions in styleGAN with text guidance텍스트를 이용한 StyleGAN의 Input-agnostic 방향 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 184
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorYang, Eunho-
dc.contributor.advisor양은호-
dc.contributor.authorKim, Yoonjeon-
dc.date.accessioned2023-06-22T19:31:15Z-
dc.date.available2023-06-22T19:31:15Z-
dc.date.issued2023-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1032320&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/308189-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.2,[iv, 33 p. :]-
dc.description.abstractWith the advantages of fast inference and human-friendly flexible manipulation, image-agnostic style manipulation via text guidance enables new applications that were not previously available. The state-of-the-art text-guided image-agnostic manipulation method embeds the representation of each channel of StyleGAN independently in the Contrastive Language-Image Pre-training (CLIP) space, and provides it in the form of a Dictionary to quickly find out the channel-wise manipulation direction during inference time. However, in this paper we argue that this dictionary which is constructed by controlling single channel individually is limited to accommodate the versatility of text guidance since the collective and interactive relation among multiple channels are not considered. Indeed, we show that it fails to discover a large portion of manipulation directions that can be found by existing methods, which manually manipulates latent space without texts. To alleviate this issue, we propose a novel method Multi2One that learns a Dictionary, whose entry corresponds to the representation of a single channel, by taking into account the manipulation effect coming from the interaction with multiple other channels. We demonstrate that our strategy resolves the inability of previous methods in finding diverse known directions from unsupervised methods and unknown directions from random text while maintaining the real-time inference speed and disentanglement ability.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectGenerative models▼aImage manipulation▼aText guidance-
dc.subject생성 모델▼a이미지 조작▼a텍스트 기반-
dc.titleLearning input-agnostic manipulation directions in styleGAN with text guidance-
dc.title.alternative텍스트를 이용한 StyleGAN의 Input-agnostic 방향 학습-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthor김윤전-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0