ORCID: https://orcid.org/0000-0003-0766-120X
(2021):
Understanding Object Dynamics for Interactive Image-to-Video Synthesis.
2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20-25 June 2021.
In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR),
New York: IEEE. pp. 5167-5177
Abstract
What would be the effect of locally poking a static scene? We present an approach that learns naturally-looking global articulations caused by a local manipulation at a pixel level. Training requires only videos of moving objects but no information of the underlying manipulation of the physical scene. Our generative model learns to infer natural object dynamics as a response to user interaction and learns about the interrelations between different object body regions. Given a static image of an object and a local poking of a pixel, the approach then predicts how the object would deform over time. In contrast to existing work on video prediction, we do not synthesize arbitrary realistic videos but enable local interactive control of the deformation. Our model is not restricted to particular object categories and can transfer dynamics onto novel unseen object instances. Extensive experiments on diverse objects demonstrate the effectiveness of our approach compared to common video prediction frameworks.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Faculties: | History and Art History > Department of Art History > Art History |
Subjects: | 000 Computer science, information and general works > 004 Data processing computer science 700 Arts and recreation > 700 Arts |
Place of Publication: | New York |
Language: | English |
Item ID: | 107296 |
Date Deposited: | 04. Oct 2023, 13:31 |
Last Modified: | 28. May 2024, 12:34 |