End of training
Browse files- .gitattributes +4 -0
- README.md +48 -0
- image_0.png +3 -0
- image_1.png +3 -0
- image_2.png +3 -0
- image_3.png +3 -0
- pytorch_lora_weights.safetensors +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
image_0.png filter=lfs diff=lfs merge=lfs -text
|
37 |
+
image_1.png filter=lfs diff=lfs merge=lfs -text
|
38 |
+
image_2.png filter=lfs diff=lfs merge=lfs -text
|
39 |
+
image_3.png filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: stabilityai/stable-diffusion-xl-base-1.0
|
3 |
+
library_name: diffusers
|
4 |
+
license: creativeml-openrail-m
|
5 |
+
tags:
|
6 |
+
- stable-diffusion-xl
|
7 |
+
- stable-diffusion-xl-diffusers
|
8 |
+
- text-to-image
|
9 |
+
- diffusers
|
10 |
+
- diffusers-training
|
11 |
+
- lora
|
12 |
+
inference: true
|
13 |
+
---
|
14 |
+
|
15 |
+
<!-- This model card has been generated automatically according to the information the training script had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
+
|
18 |
+
|
19 |
+
# LoRA text2image fine-tuning - Jayw1/Dunhuang_img
|
20 |
+
|
21 |
+
These are LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. The weights were fine-tuned on the /home/hjw/code/diffusers/Dunhuang_murals dataset. You can find some example images in the following.
|
22 |
+
|
23 |
+

|
24 |
+

|
25 |
+

|
26 |
+

|
27 |
+
|
28 |
+
|
29 |
+
LoRA for the text encoder was enabled: False.
|
30 |
+
|
31 |
+
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
|
32 |
+
|
33 |
+
|
34 |
+
## Intended uses & limitations
|
35 |
+
|
36 |
+
#### How to use
|
37 |
+
|
38 |
+
```python
|
39 |
+
# TODO: add an example code snippet for running this diffusion pipeline
|
40 |
+
```
|
41 |
+
|
42 |
+
#### Limitations and bias
|
43 |
+
|
44 |
+
[TODO: provide examples of latent issues and potential remediations]
|
45 |
+
|
46 |
+
## Training details
|
47 |
+
|
48 |
+
[TODO: describe the data used to train the model]
|
image_0.png
ADDED
![]() |
Git LFS Details
|
image_1.png
ADDED
![]() |
Git LFS Details
|
image_2.png
ADDED
![]() |
Git LFS Details
|
image_3.png
ADDED
![]() |
Git LFS Details
|
pytorch_lora_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4c6d40f8fdd4d33b0210805548961ffeb94236b2bdc3988beeaf0db985074ee
|
3 |
+
size 23390424
|