Translate model card
#1
by
osanseviero
- opened
README.md
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
3 |
---
|
4 |
|
5 |
The original repo is [here](https://modelscope.cn/models/damo/text-to-video-synthesis/summary).
|
@@ -9,3 +10,69 @@ We Are Hiring! (Based on Beijing / Hangzhou, China.)
|
|
9 |
If you're looking for an exciting challenge and the opportunity to work with cutting-edge technologies in AIGC and large-scale pretraining, then we are the place for you. We are looking for talented, motivated and creative individuals to join our team. If you are interested, please send your CV to us.
|
10 |
|
11 |
EMAIL: wangjiuniu.wjn@alibaba-inc.com
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
pipeline_tag: text-to-video
|
4 |
---
|
5 |
|
6 |
The original repo is [here](https://modelscope.cn/models/damo/text-to-video-synthesis/summary).
|
|
|
10 |
If you're looking for an exciting challenge and the opportunity to work with cutting-edge technologies in AIGC and large-scale pretraining, then we are the place for you. We are looking for talented, motivated and creative individuals to join our team. If you are interested, please send your CV to us.
|
11 |
|
12 |
EMAIL: wangjiuniu.wjn@alibaba-inc.com
|
13 |
+
|
14 |
+
This model is based on a multi-stage text-to-video generation diffusion model, which inputs a description text and returns a video that matches the text description. Only English input is supported.
|
15 |
+
|
16 |
+
## Model Description
|
17 |
+
|
18 |
+
The text-to-video generation diffusion model consists of three sub-networks: text feature extraction, text feature-to-video latent space diffusion model, and video latent space to video visual space. The overall model parameters are about 1.7 billion. Support English input. The diffusion model adopts the Unet3D structure, and realizes the function of video generation through the iterative denoising process from the pure Gaussian noise video.
|
19 |
+
|
20 |
+
**How to expect the model to be used and where it is applicable**
|
21 |
+
|
22 |
+
This model has a wide range of applications and can reason and generate videos based on arbitrary English text descriptions.
|
23 |
+
|
24 |
+
## How to use
|
25 |
+
|
26 |
+
Under the ModelScope framework, the current model can be used by calling a simple Pipeline, where the input must be in dictionary format, the legal key value is 'text', and the content is a short text. This model currently only supports inference on the GPU. Enter specific code examples as follows:
|
27 |
+
|
28 |
+
### Operating environment (Python Package)
|
29 |
+
|
30 |
+
```
|
31 |
+
pip install modelscope
|
32 |
+
pip install open_clip_torch
|
33 |
+
```
|
34 |
+
|
35 |
+
### Code example (Demo Code)
|
36 |
+
|
37 |
+
```python
|
38 |
+
from huggingface_hub import snapshot_download
|
39 |
+
|
40 |
+
from modelscope.pipelines import pipeline
|
41 |
+
from modelscope.outputs import OutputKeys
|
42 |
+
|
43 |
+
model_dir = pathlib.Path('weights')
|
44 |
+
snapshot_download('damo-vilab/modelscope-damo-text-to-video-synthesis',
|
45 |
+
repo_type='model', local_dir=model_dir)
|
46 |
+
|
47 |
+
pipe = pipeline('text-to-video-synthesis', model_dir.as_posix())
|
48 |
+
test_text = {
|
49 |
+
'text': 'A panda eating bamboo on a rock.',
|
50 |
+
}
|
51 |
+
output_video_path = pipe(test_text,)[OutputKeys.OUTPUT_VIDEO]
|
52 |
+
print('output_video_path:', output_video_path)
|
53 |
+
```
|
54 |
+
|
55 |
+
### View results
|
56 |
+
|
57 |
+
The above code will display the save path of the output video, and the current encoding format can be played normally with [VLC player](https://www.videolan.org/vlc/).
|
58 |
+
|
59 |
+
The output mp4 file can be viewed by [VLC media player](https://www.videolan.org/vlc/). Some other media players may not view it normally.
|
60 |
+
|
61 |
+
## Model limitations and biases
|
62 |
+
|
63 |
+
* The model is trained based on public data sets such as Webvid, and the generated results may have deviations related to the distribution of training data.
|
64 |
+
* This model cannot achieve perfect film and television quality generation.
|
65 |
+
* The model cannot generate clear text.
|
66 |
+
* The model is mainly trained with English corpus and does not support other languages at the moment**.
|
67 |
+
* The performance of this model needs to be improved on complex compositional generation tasks.
|
68 |
+
|
69 |
+
## Misuse, Malicious Use and Excessive Use
|
70 |
+
|
71 |
+
* The model was not trained to realistically represent people or events, so using it to generate such content is beyond the model's capabilities.
|
72 |
+
* It is prohibited to generate content that is demeaning or harmful to people or their environment, culture, religion, etc.
|
73 |
+
* Prohibited for pornographic, violent and bloody content generation.
|
74 |
+
* Prohibited for error and false information generation.
|
75 |
+
|
76 |
+
## Training data
|
77 |
+
|
78 |
+
The training data includes LAION5B, ImageNet, Webvid and other public datasets. Image and video filtering is performed after pre-training such as aesthetic score, watermark score, and deduplication.
|