Diffusers
Latte / README.md
maxin-cn's picture
Update README.md
ade76bf verified
|
raw
history blame
2.43 kB
metadata
license: mit

Latte: Latent Diffusion Transformer for Video Generation

This repo contains pre-trained weights for our paper exploring latent diffusion models with transformers (Latte). You can find more visualizations on our project page.

News

  • (πŸ”₯ New) May. 23, 2024. πŸ’₯ The updated LatteT2V model is released at here. If you want to use the updated model to generate images directly, please make sure video_length=1, enable_temporal_attentions=True and enable_vae_temporal_decoder=False in t2v_sample.yaml.

  • (πŸ”₯ New) Mar. 20, 2024. πŸ’₯ An updated LatteT2V model is coming soon, stay tuned!

  • (πŸ”₯ New) Feb. 24, 2024. πŸ’₯ We are very grateful that researchers and developers like our work. We will continue to update our LatteT2V model, hoping that our efforts can help the community develop. Our Latte discord channel is created for discussions. Coders are welcome to contribute.

  • (πŸ”₯ New) Jan. 9, 2024. πŸ’₯ An updated LatteT2V model initialized with the PixArt-Ξ± is released, the checkpoint can be found here.

  • (πŸ”₯ New) Oct. 31, 2023. πŸ’₯ The training and inference code is released. All checkpoints (including FaceForensics, SkyTimelapse, UCF101, and Taichi-HD) can be found here. In addition, the LatteT2V inference code is provided.

Contact Us

Yaohui Wang: wangyaohui@pjlab.org.cn Xin Ma: xin.ma1@monash.edu

Citation

If you find this work useful for your research, please consider citing it.

@article{ma2024latte,
  title={Latte: Latent Diffusion Transformer for Video Generation},
  author={Ma, Xin and Wang, Yaohui and Jia, Gengyun and Chen, Xinyuan and Liu, Ziwei and Li, Yuan-Fang and Chen, Cunjian and Qiao, Yu},
  journal={arXiv preprint arXiv:2401.03048},
  year={2024}
}

Acknowledgments

Latte has been greatly inspired by the following amazing works and teams: DiT and PixArt-Ξ±, we thank all the contributors for open-sourcing.