Improving Shape Deformation in Unsupervised Image-to-Image Translation

Aaron Gokaslan, Vivek Ramanujan, Daniel Ritchie, Kwang In Kim, James Tompkin

Research output: Chapter or section in a book/report/conference proceedingChapter in a published conference proceeding

7 Citations (SciVal)

Abstract

Unsupervised image-to-image translation techniques are able to map local texture between two domains, but they are typically unsuccessful when the domains require larger shape change. Inspired by semantic segmentation, we introduce a discriminator with dilated convolutions that is able to use information from across the entire image to train a more context-aware generator. This is coupled with a multi-scale perceptual loss that is better able to represent error in the underlying shape of objects. We demonstrate that this design is more capable of representing shape deformation in a challenging toy dataset, plus in complex mappings with significant dataset variation between humans, dolls, and anime faces, and between cats and dogs.
Original languageEnglish
Title of host publicationProceedings of European Conference on Computer Vision (ECCV)
Subtitle of host publicationComputer Vision - ECCV 2018
EditorsVittorio Ferrari, Martial Hebert, Cristian Sminchisescu, Yair Weiss
PublisherSpringer Verlag
Pages662-678
Number of pages17
Volume11216
ISBN (Electronic)978-3-030-01258-8
ISBN (Print)978-3-030-01257-1
DOIs
Publication statusE-pub ahead of print - 6 Oct 2018

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume11216 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Keywords

  • Generative adversarial networks
  • Image translation

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Improving Shape Deformation in Unsupervised Image-to-Image Translation'. Together they form a unique fingerprint.

Cite this