1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Preserving Semantic and Temporal Consistency for Unpaired Video-to-Video Translation

      Preprint
      , , , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In this paper, we investigate the problem of unpaired video-to-video translation. Given a video in the source domain, we aim to learn the conditional distribution of the corresponding video in the target domain, without seeing any pairs of corresponding videos. While significant progress has been made in the unpaired translation of images, directly applying these methods to an input video leads to low visual quality due to the additional time dimension. In particular, previous methods suffer from semantic inconsistency (i.e., semantic label flipping) and temporal flickering artifacts. To alleviate these issues, we propose a new framework that is composed of carefully-designed generators and discriminators, coupled with two core objective functions: 1) content preserving loss and 2) temporal consistency loss. Extensive qualitative and quantitative evaluations demonstrate the superior performance of the proposed method against previous approaches. We further apply our framework to a domain adaptation task and achieve favorable results.

          Related collections

          Most cited references15

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Pyramid Scene Parsing Network

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs

              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation

                Bookmark

                Author and article information

                Journal
                20 August 2019
                Article
                10.1145/3343031.3350864
                1908.07683
                a9422042-6cc4-4040-9134-cdcd00832139

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                Accepted by ACM Multimedia(ACM MM) 2019
                cs.CV cs.MM

                Computer vision & Pattern recognition,Graphics & Multimedia design
                Computer vision & Pattern recognition, Graphics & Multimedia design

                Comments

                Comment on this article