Spaces:
Runtime error
Runtime error
Bo-Kyeong Kim
commited on
Commit
路
327e673
1
Parent(s):
94b4e9d
Update docs/description.md
Browse files- docs/description.md +5 -2
docs/description.md
CHANGED
@@ -1,7 +1,10 @@
|
|
1 |
This demo showcases a lightweight Stable Diffusion model (SDM) for general-purpose text-to-image synthesis. Our model **BK-SDM-Small** achieves **36% reduced** parameters and latency. This model is bulit with (i) removing several residual and attention blocks from the U-Net of SDM-v1.4 and (ii) distillation pretraining on only 0.22M LAION pairs (fewer than 0.1% of the full training set). Despite very limited training resources, our model can imitate the original SDM by benefiting from transferred knowledge.
|
2 |
|
3 |
<center>
|
4 |
-
<img alt="U-Net architectures and KD-based pretraining" img src="https://huggingface.co/spaces/nota-ai/theme/resolve/3bb3eed8b911d0baf306767bb9548bf732052c53/docs/compressed_stable_diffusion/fig_model.png" width="
|
5 |
</center>
|
6 |
|
7 |
-
<br
|
|
|
|
|
|
|
|
1 |
This demo showcases a lightweight Stable Diffusion model (SDM) for general-purpose text-to-image synthesis. Our model **BK-SDM-Small** achieves **36% reduced** parameters and latency. This model is bulit with (i) removing several residual and attention blocks from the U-Net of SDM-v1.4 and (ii) distillation pretraining on only 0.22M LAION pairs (fewer than 0.1% of the full training set). Despite very limited training resources, our model can imitate the original SDM by benefiting from transferred knowledge.
|
2 |
|
3 |
<center>
|
4 |
+
<img alt="U-Net architectures and KD-based pretraining" img src="https://huggingface.co/spaces/nota-ai/theme/resolve/3bb3eed8b911d0baf306767bb9548bf732052c53/docs/compressed_stable_diffusion/fig_model.png" width="65%">
|
5 |
</center>
|
6 |
|
7 |
+
<br/>
|
8 |
+
|
9 |
+
### Updates
|
10 |
+
(May/31/2023) The demo is running on T4 small (4 vCPU 路 15 GB RAM 路 16GB VRAM). It takes 5~10 seconds for the original model to generate a 512脳512 image with 25 denoising steps. Our compressed model accelerates inference speed while preserving visually compelling results.
|