Spaces:
Running
Running
<br/> | |
This system generates a talking face video based on the input text. | |
You can provide the input text in one of the four languages: Chinese (Mandarin), English, Japanese, and Korean. | |
You may also select the target language, the language of the output speech. | |
If the input text language and the target language are different, the input text will be translated to the target language using Google Translate API. | |
## Updates | |
(2022.06.17.) To maximize everyone's experience, we capped the length of the translated texts at: | |
- 200 characters for English | |
- 100 characters for Chinese, Japaense, and Korean. | |
Also, we found that there were a lot of inappropriate input texts. We decided to filter the inputs based on toxicity using [Perspective API @Google](https://developers.perspectiveapi.com/s/). Now, if you enter a possibily toxic text, the video generation will fail. We hope you understand. | |
(2022.06.05.) Due to the latency from HuggingFace Spaces and video rendering, it takes 15 ~ 30 seconds to get a video result. | |
<details> | |
<summary><i>Outdated updates</i></summary> | |
(2022.09.29.) ~~The core part of the demonstration has been working on the AWS instance of MINDsLab, and I found that it can't connect to the instance now. I want to fix this issue, but I'm sorry to say that I left the company last week. I've contacted the company, but it takes some time to restore the session. If you're in a hurry, please send the e-mail directly to MINDsLab (hello@mindslab.ai). | |
Whatever the reason, I'm sorry again. Hope you understand.~~ | |
</details> |