thanks to kabachuha ❤
Browse files- README.md +23 -0
- VQGAN_autoencoder.pth +3 -0
- configuration.json +35 -0
- open_clip_pytorch_model.bin +3 -0
- text2video_pytorch_model.pth +3 -0
README.md
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-4.0
|
3 |
+
task_categories:
|
4 |
+
- text-to-video
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
tags:
|
8 |
+
- anime
|
9 |
+
---
|
10 |
+
|
11 |
+
This is https://huggingface.co/datasets/strangeman3107/animov-0.1 model by strangeman3107 that was converted by [me](https://github.com/kabachuha) into the ModelScope original format using this script https://github.com/ExponentialML/Text-To-Video-Finetuning/pull/52.
|
12 |
+
|
13 |
+
Ready to use in Auto1111 webui with this extension https://github.com/deforum-art/sd-webui-text2video
|
14 |
+
|
15 |
+
---
|
16 |
+
|
17 |
+
Now, copyting info from the original page
|
18 |
+
|
19 |
+
This is a text2video model for diffusers, fine-tuned with a [modelscope](https://huggingface.co/damo-vilab/text-to-video-ms-1.7b) to have an anime-style appearance.
|
20 |
+
It was trained at 384x384 resolution.
|
21 |
+
It still generates unstable content often. The usage is the same as with the original modelscope model.
|
22 |
+
|
23 |
+
example images are [here](https://imgur.com/a/sCwmKG1).
|
VQGAN_autoencoder.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:930e9865584beae2405d29bc06a05db3bb6a5b34eedd40a7db29b9156ed7d098
|
3 |
+
size 2607657443
|
configuration.json
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{ "framework": "pytorch",
|
2 |
+
"task": "text-to-video-synthesis",
|
3 |
+
"model": {
|
4 |
+
"type": "latent-text-to-video-synthesis",
|
5 |
+
"model_args": {
|
6 |
+
"ckpt_clip": "open_clip_pytorch_model.bin",
|
7 |
+
"ckpt_unet": "text2video_pytorch_model.pth",
|
8 |
+
"ckpt_autoencoder": "VQGAN_autoencoder.pth",
|
9 |
+
"max_frames": 16,
|
10 |
+
"tiny_gpu": 1
|
11 |
+
},
|
12 |
+
"model_cfg": {
|
13 |
+
"unet_in_dim": 4,
|
14 |
+
"unet_dim": 320,
|
15 |
+
"unet_y_dim": 768,
|
16 |
+
"unet_context_dim": 1024,
|
17 |
+
"unet_out_dim": 4,
|
18 |
+
"unet_dim_mult": [1, 2, 4, 4],
|
19 |
+
"unet_num_heads": 8,
|
20 |
+
"unet_head_dim": 64,
|
21 |
+
"unet_res_blocks": 2,
|
22 |
+
"unet_attn_scales": [1, 0.5, 0.25],
|
23 |
+
"unet_dropout": 0.1,
|
24 |
+
"temporal_attention": "True",
|
25 |
+
"num_timesteps": 1000,
|
26 |
+
"mean_type": "eps",
|
27 |
+
"var_type": "fixed_small",
|
28 |
+
"loss_type": "mse"
|
29 |
+
}
|
30 |
+
},
|
31 |
+
"pipeline": {
|
32 |
+
"type": "latent-text-to-video-synthesis"
|
33 |
+
}
|
34 |
+
}
|
35 |
+
|
open_clip_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73c32c62eebf1112b0693ff9e3ecfa0573ba02cd279420ea4da4af1cbfb39e3b
|
3 |
+
size 1972451989
|
text2video_pytorch_model.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b710dd73d49c598b339ea9c76c78750fa38e6477793f6373be18087dbe9740c
|
3 |
+
size 2822972283
|