IterGANs: Iterative GANs to learn and control 3D object transformation
作者:
Highlights:
•
摘要
We are interested in learning visual representations which allow for 3D manipulations of visual objects based on a single 2D image. We cast this into an image-to-image transformation task, and propose Iterative Generative Adversarial Networks (IterGANs) which iteratively transform an input image into an output image. Our models learn a visual representation that can be used for objects seen in training, but also for never seen objects. Since object manipulation requires a full understanding of the geometry and appearance of the object, our IterGANs learn an implicit 3D model and a full appearance model of the object, which are both inferred from a single (test) image. Two advantages of IterGANs are that the intermediate generated images can be used for an additional supervision signal, even in an unsupervised fashion, and that the number of iterations can be used as a control signal to steer the transformation. Experiments on rotated objects and scenes show how IterGANs help with the generation process.
论文关键词:
论文评审过程:Received 16 July 2018, Revised 14 June 2019, Accepted 21 August 2019, Available online 28 August 2019, Version of Record 1 November 2019.
论文官网地址:https://doi.org/10.1016/j.cviu.2019.102803