Home

Awesome

Owl-1: Omni World Model for Consistent Long Video Generation

Paper

Owl-1: Omni World Model for Consistent Long Video Generation

Yuanhui Huang, Wenzhao Zheng$\dagger$, Yuan Gao, Xin Tao, Pengfei Wan, Di Zhang, Jie Zhou, Jiwen Lu

$\dagger$ Project leader

Owl-1 approaches consistent long video generation with an omni world model, which models the evolution of the underlying world with latent state, explicit observation and world dynamics variables.

teaser

News.

Demo

Short Videos

demo

demo

Long Videos

demo

Overview

<p align = "center"> <img src="./assets/comparison.png" width="60%" /> </p>

We propose an Omni World modeL (Owl-1) to produce long-term coherent and comprehensive conditions for consistent long video generation. As videos are observations of the underlying evolving world, we propose to model the long-term developments in a latent space and use VGMs to film them into videos. Specifically, we represent the world with a latent state variable which can be decoded into explicit video observations. These observations serve as a basis for anticipating temporal dynamics which in turn update the state variable. The interaction between evolving dynamics and persistent state enhances the diversity and consistency of the long videos. Extensive experiments show that Owl-1 achieves comparable performance with SOTA methods on VBench-I2V and VBench-Long, validating its ability to generate high-quality video observations.

overview

Getting Started

We will release the code soon.

Citation

If you find this project helpful, please consider citing the following paper:

@article{huang2024owl1omniworldmodel,
    title={Owl-1: Omni World Model for Consistent Long Video Generation}, 
    author={Huang, Yuanhui and Zheng, Wenzhao and Gao, Yuan and Tao, Xin and Wan, Pengfei and Zhang, Di and Zhou, Jie and Lu, Jiwen},
    journal={arXiv preprint arXiv:2412.09600},
    year={2024},
}