Preserving Semantic and Temporal Consistency for Unpaired Video-to-Video Translation

Cited 20 time in webofscience Cited 13 time in scopus
  • Hit : 178
  • Download : 0
In this paper, we investigate the problem of unpaired video-to-video translation. Given a video in the source domain, we aim to learn the conditional distribution of the corresponding video in the target domain, without seeing any pairs of corresponding videos. While signiicant progress has been made in the unpaired translation of images, directly applying these methods to an input video leads to low visual quality due to the additional time dimension. In particular, previous methods sufer from semantic inconsistency (i.e., semantic label lipping) and temporal lickering artifacts. To alleviate these issues, we propose a new framework that is composed of carefully-designed generators and discriminators, coupled with two core objective functions: 1) content preserving loss and 2) temporal consistency loss. Extensive qualitative and quantitative evaluations demonstrate the superior performance of the proposed method against previous approaches. We further apply our framework to a domain adaptation task and achieve favorable results.
Publisher
ACM Multimedia
Issue Date
2019-10
Language
English
Citation

ACM International Conference on Multimedia, pp.1248 - 1257

DOI
10.1145/3343031.3350864
URI
http://hdl.handle.net/10203/269277
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 20 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0