Manifold learning and stylization of human motion인간 행동의 매니폴드 학습 및 스타일화

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 274
  • Download : 0
In this thesis, we present two motion generative models for constructing motion manifold and motion stylization. First, we propose a novel recurrent neural network-based method to construct a latent motion manifold that can represent a wide range of human motions in a long sequence. Our model create compact and versatile motion manifold that allows for generating new motions by performing random sampling and algebraic operations, such as interpolation and analogy, in the latent motion manifold. Second, we propose the motion style transfer model that can control the motion style of individual body parts. Inputs to our framework are one source motion for content, and multiple target motions, each for stylizing a different body part, and its output is a content-preserving motion stylized by body part. The advantages of our framework are thus it can generate an extensive variety of motions by stylizing them by body part and can allows style interpolation by body part, which further enhances its usefulness. In addition, our framework can be easily integrated with other motion generator. We demonstrate a real-time motion style transfer by integrating with an existing motion controller.
Advisors
Lee, Sung-Heeresearcher이성희researcher
Description
한국과학기술원 :문화기술대학원,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 문화기술대학원, 2023.2,[v, 49 p. :]

Keywords

Motion manifold▼aMotion style transfer▼aCharacter animation▼aDeep learning; 행동 매니폴드▼a행동 스타일 전이▼a캐릭터 애니메이션▼a딥러닝

URI
http://hdl.handle.net/10203/307976
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1030384&flag=dissertation
Appears in Collection
GCT-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0