Cross-modality motion parameterization for fine-grained video prediction
作者:
Highlights:
•
摘要
While predicting video content is challenging given the huge unconstrained searching space, this work explores cross-modality constraints to safeguard the video generation process and seeks improved content prediction. By observing the underlying correspondence between the sound and the object movement, we propose a novel cross-modality video generation network. Via adversarial training, this network directly links sound with the movement parameters of the operated object and automatically outputs corresponding object motion according to the rhythm of the given audio signal. We experiment on both rigid object and non-rigid object motion prediction tasks and show that our method significantly reduces motion uncertainty for the generated video content, with the guidance of the associated audio information.
论文关键词:
论文评审过程:Received 15 June 2018, Revised 15 December 2018, Accepted 23 March 2019, Available online 3 April 2019, Version of Record 15 May 2019.
论文官网地址:https://doi.org/10.1016/j.cviu.2019.03.006