Abstract
We present an approach to modeling an image-space prior on scene dynamics. Our prior is learned from a collection of motion trajectories extracted from real video sequences containing natural, oscillating motion such as trees, flowers, candles, and clothes blowing in the wind. Given a single image, our trained model uses a frequency-coordinated diffusion sampling process to predict a per-pixel long-term motion representation in the Fourier domain, which we call a neural stochastic motion texture. This representation can be converted into dense motion trajectories that span an entire video. Along with an image-based rendering module, these trajectories can be used for a number of downstream applications, such as turning still images into seamlessly looping dynamic videos, or allowing users to realistically interact with objects in real pictures.
Community
Amazing!
interesting approach
Hi. Curious to know how the demo runs in real-time considering it is using multiple models?
Hi. Curious to know how the demo runs in real-time considering it is using multiple models?
The web demo uses a pregenerated neural stochastic motion texture, and for speed it uses a lower-quality rendering method (mesh warping in webgl) than the full rendering model in the paper.
Will code be published?
Good Job!
Will code be published?
This looks incredible, and would be a great asset to my video productions! Is this going to be released as an app or by code per chance?
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper