Temporal procrustes alignment framework for 3D human pose and shape estimation from video비디오를 통한 3D 인간 자세 및 형태 추정을 위한 시계열 프로크루스테스 정렬 프레임워크

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 149
  • Download : 0
This dissertation considers the task of model-based 3D human pose and shape estimation from 2D monocular RGB video. Human pose and shape estimation takes a monocular image containing a person as input and returns that person’s joint position and body shapes. To obtain model-based 3D results, depth estimation is added to the 2D joint location estimation, which is used with the body shapes to generate 3D human mesh model parameters. For the case where the input is a video, the above process is done for each image frame of a video. Although the estimated pose and shape of a person may seem accurate for each frame of image, simply listing the estimated results of the images does not give smooth results for they show jitter along the predictions due to the uncertainty of the joint positions. To solve this problem, previous methods have made progress by improving networks to consider the temporal consistency of human motions in sequential frames by supervising the average acceleration of joints for each frame. After maintaining temporal consistency, however, geometric misalignments within the sequence of joints are observed. Geometric misalignment refers to the steady deviation between the geometric path drawn by a sequence of predicted joints and that of ground-truth joints. To this end, we propose Temporal Procrustes Alignment (TPA) framework, which is a model-agnostic framework that mitigates geometric misalignments by performing group-wise sequential learning of every joint’s movement paths. While previous methods rely entirely on per-frame supervision for accuracy, our framework can supervise sequential accuracy by performing Procrustes Analysis to the sequence of predicted joints. Experiments show that TPA framework mitigates the misalignment of the results without damaging their temporal consistency, advancing the network to mostly exceed the previous state-of-the-art performances on benchmark datasets in both per-frame accuracy and video smoothness metric.
Advisors
Yoo, Chang Dongresearcher유창동researcher
Description
한국과학기술원 :로봇공학학제전공,
Publisher
한국과학기술원
Issue Date
2022
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 로봇공학학제전공, 2022.8,[v, 40 p. :]

Keywords

Human Pose Estimation from Video▼a3D Human Pose and Shape Estimation▼aProcrustes Analysis; 비디오를 통한 인간 자세 추정▼a3D 인간 자세 및 형태 추정▼a프로크루스테스 분석

URI
http://hdl.handle.net/10203/308273
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1008235&flag=dissertation
Appears in Collection
RE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0