hasininawoda
commited on
Commit
•
4a1a100
1
Parent(s):
bc311dd
End of training
Browse files- README.md +39 -0
- logs/text2image-fine-tune/1719936659.6956227/events.out.tfevents.1719936659.ed7fc0fd9bec.5460.1 +3 -0
- logs/text2image-fine-tune/1719936659.697233/hparams.yml +50 -0
- logs/text2image-fine-tune/1719936894.6760967/events.out.tfevents.1719936894.ed7fc0fd9bec.6625.1 +3 -0
- logs/text2image-fine-tune/1719936894.6777575/hparams.yml +50 -0
- logs/text2image-fine-tune/1719937023.4383209/events.out.tfevents.1719937023.ed7fc0fd9bec.7261.1 +3 -0
- logs/text2image-fine-tune/1719937023.439974/hparams.yml +50 -0
- logs/text2image-fine-tune/events.out.tfevents.1719936659.ed7fc0fd9bec.5460.0 +3 -0
- logs/text2image-fine-tune/events.out.tfevents.1719936894.ed7fc0fd9bec.6625.0 +3 -0
- logs/text2image-fine-tune/events.out.tfevents.1719937023.ed7fc0fd9bec.7261.0 +3 -0
- pytorch_lora_weights.safetensors +3 -0
README.md
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: CompVis/stable-diffusion-v1-4
|
3 |
+
library_name: diffusers
|
4 |
+
license: creativeml-openrail-m
|
5 |
+
tags:
|
6 |
+
- stable-diffusion
|
7 |
+
- stable-diffusion-diffusers
|
8 |
+
- text-to-image
|
9 |
+
- diffusers
|
10 |
+
- diffusers-training
|
11 |
+
- lora
|
12 |
+
inference: true
|
13 |
+
---
|
14 |
+
|
15 |
+
<!-- This model card has been generated automatically according to the information the training script had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
+
|
18 |
+
|
19 |
+
# LoRA text2image fine-tuning - hasininawoda/output1
|
20 |
+
These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were fine-tuned on the None dataset. You can find some example images in the following.
|
21 |
+
|
22 |
+
|
23 |
+
|
24 |
+
|
25 |
+
## Intended uses & limitations
|
26 |
+
|
27 |
+
#### How to use
|
28 |
+
|
29 |
+
```python
|
30 |
+
# TODO: add an example code snippet for running this diffusion pipeline
|
31 |
+
```
|
32 |
+
|
33 |
+
#### Limitations and bias
|
34 |
+
|
35 |
+
[TODO: provide examples of latent issues and potential remediations]
|
36 |
+
|
37 |
+
## Training details
|
38 |
+
|
39 |
+
[TODO: describe the data used to train the model]
|
logs/text2image-fine-tune/1719936659.6956227/events.out.tfevents.1719936659.ed7fc0fd9bec.5460.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5a8bae368ab1dc388661c7b95cc1c1838636e4d9ff23ce97b0b74d24ca4d8a9
|
3 |
+
size 2204
|
logs/text2image-fine-tune/1719936659.697233/hparams.yml
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
adam_beta1: 0.9
|
2 |
+
adam_beta2: 0.999
|
3 |
+
adam_epsilon: 1.0e-08
|
4 |
+
adam_weight_decay: 0.01
|
5 |
+
allow_tf32: false
|
6 |
+
cache_dir: null
|
7 |
+
caption_column: caption
|
8 |
+
center_crop: true
|
9 |
+
checkpointing_steps: 500
|
10 |
+
checkpoints_total_limit: null
|
11 |
+
dataloader_num_workers: 8
|
12 |
+
dataset_config_name: null
|
13 |
+
dataset_name: images
|
14 |
+
enable_xformers_memory_efficient_attention: false
|
15 |
+
gradient_accumulation_steps: 4
|
16 |
+
gradient_checkpointing: false
|
17 |
+
hub_model_id: null
|
18 |
+
hub_token: null
|
19 |
+
image_column: image
|
20 |
+
learning_rate: 0.0001
|
21 |
+
local_rank: -1
|
22 |
+
logging_dir: logs
|
23 |
+
lr_scheduler: cosine
|
24 |
+
lr_warmup_steps: 0
|
25 |
+
max_grad_norm: 1.0
|
26 |
+
max_train_samples: null
|
27 |
+
max_train_steps: 15
|
28 |
+
mixed_precision: null
|
29 |
+
noise_offset: 0
|
30 |
+
num_train_epochs: 2
|
31 |
+
num_validation_images: 4
|
32 |
+
output_dir: output1
|
33 |
+
prediction_type: null
|
34 |
+
pretrained_model_name_or_path: CompVis/stable-diffusion-v1-4
|
35 |
+
push_to_hub: false
|
36 |
+
random_flip: true
|
37 |
+
rank: 4
|
38 |
+
report_to: tensorboard
|
39 |
+
resolution: 512
|
40 |
+
resume_from_checkpoint: null
|
41 |
+
revision: null
|
42 |
+
scale_lr: false
|
43 |
+
seed: 1337
|
44 |
+
snr_gamma: null
|
45 |
+
train_batch_size: 1
|
46 |
+
train_data_dir: null
|
47 |
+
use_8bit_adam: false
|
48 |
+
validation_epochs: 1
|
49 |
+
validation_prompt: A photo of TOK person
|
50 |
+
variant: null
|
logs/text2image-fine-tune/1719936894.6760967/events.out.tfevents.1719936894.ed7fc0fd9bec.6625.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:890b473a355b8fda6408a8803f91c88603f8f25f345a8050f9804e5b24b8e627
|
3 |
+
size 2204
|
logs/text2image-fine-tune/1719936894.6777575/hparams.yml
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
adam_beta1: 0.9
|
2 |
+
adam_beta2: 0.999
|
3 |
+
adam_epsilon: 1.0e-08
|
4 |
+
adam_weight_decay: 0.01
|
5 |
+
allow_tf32: false
|
6 |
+
cache_dir: null
|
7 |
+
caption_column: caption
|
8 |
+
center_crop: true
|
9 |
+
checkpointing_steps: 500
|
10 |
+
checkpoints_total_limit: null
|
11 |
+
dataloader_num_workers: 8
|
12 |
+
dataset_config_name: null
|
13 |
+
dataset_name: images
|
14 |
+
enable_xformers_memory_efficient_attention: false
|
15 |
+
gradient_accumulation_steps: 3
|
16 |
+
gradient_checkpointing: false
|
17 |
+
hub_model_id: null
|
18 |
+
hub_token: null
|
19 |
+
image_column: image
|
20 |
+
learning_rate: 0.0001
|
21 |
+
local_rank: -1
|
22 |
+
logging_dir: logs
|
23 |
+
lr_scheduler: cosine
|
24 |
+
lr_warmup_steps: 0
|
25 |
+
max_grad_norm: 1.0
|
26 |
+
max_train_samples: null
|
27 |
+
max_train_steps: 15
|
28 |
+
mixed_precision: null
|
29 |
+
noise_offset: 0
|
30 |
+
num_train_epochs: 2
|
31 |
+
num_validation_images: 4
|
32 |
+
output_dir: output1
|
33 |
+
prediction_type: null
|
34 |
+
pretrained_model_name_or_path: CompVis/stable-diffusion-v1-4
|
35 |
+
push_to_hub: false
|
36 |
+
random_flip: true
|
37 |
+
rank: 4
|
38 |
+
report_to: tensorboard
|
39 |
+
resolution: 512
|
40 |
+
resume_from_checkpoint: null
|
41 |
+
revision: null
|
42 |
+
scale_lr: false
|
43 |
+
seed: 1337
|
44 |
+
snr_gamma: null
|
45 |
+
train_batch_size: 1
|
46 |
+
train_data_dir: null
|
47 |
+
use_8bit_adam: true
|
48 |
+
validation_epochs: 1
|
49 |
+
validation_prompt: A photo of TOK person
|
50 |
+
variant: null
|
logs/text2image-fine-tune/1719937023.4383209/events.out.tfevents.1719937023.ed7fc0fd9bec.7261.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24f3afabf402215cb4c7c087c91ea8d7a8362ea0a3e1b56c3a23c0038e09a1da
|
3 |
+
size 2204
|
logs/text2image-fine-tune/1719937023.439974/hparams.yml
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
adam_beta1: 0.9
|
2 |
+
adam_beta2: 0.999
|
3 |
+
adam_epsilon: 1.0e-08
|
4 |
+
adam_weight_decay: 0.01
|
5 |
+
allow_tf32: false
|
6 |
+
cache_dir: null
|
7 |
+
caption_column: caption
|
8 |
+
center_crop: true
|
9 |
+
checkpointing_steps: 500
|
10 |
+
checkpoints_total_limit: null
|
11 |
+
dataloader_num_workers: 2
|
12 |
+
dataset_config_name: null
|
13 |
+
dataset_name: images
|
14 |
+
enable_xformers_memory_efficient_attention: false
|
15 |
+
gradient_accumulation_steps: 3
|
16 |
+
gradient_checkpointing: false
|
17 |
+
hub_model_id: null
|
18 |
+
hub_token: null
|
19 |
+
image_column: image
|
20 |
+
learning_rate: 0.0001
|
21 |
+
local_rank: -1
|
22 |
+
logging_dir: logs
|
23 |
+
lr_scheduler: cosine
|
24 |
+
lr_warmup_steps: 0
|
25 |
+
max_grad_norm: 1.0
|
26 |
+
max_train_samples: null
|
27 |
+
max_train_steps: 15
|
28 |
+
mixed_precision: null
|
29 |
+
noise_offset: 0
|
30 |
+
num_train_epochs: 2
|
31 |
+
num_validation_images: 4
|
32 |
+
output_dir: output1
|
33 |
+
prediction_type: null
|
34 |
+
pretrained_model_name_or_path: CompVis/stable-diffusion-v1-4
|
35 |
+
push_to_hub: false
|
36 |
+
random_flip: true
|
37 |
+
rank: 4
|
38 |
+
report_to: tensorboard
|
39 |
+
resolution: 512
|
40 |
+
resume_from_checkpoint: null
|
41 |
+
revision: null
|
42 |
+
scale_lr: false
|
43 |
+
seed: 1337
|
44 |
+
snr_gamma: null
|
45 |
+
train_batch_size: 1
|
46 |
+
train_data_dir: null
|
47 |
+
use_8bit_adam: true
|
48 |
+
validation_epochs: 1
|
49 |
+
validation_prompt: A photo of TOK person
|
50 |
+
variant: null
|
logs/text2image-fine-tune/events.out.tfevents.1719936659.ed7fc0fd9bec.5460.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:639243e657c7e4b9292c66bd043440f5b17b07e940abdc2a2656c4461754d51f
|
3 |
+
size 88
|
logs/text2image-fine-tune/events.out.tfevents.1719936894.ed7fc0fd9bec.6625.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e25aff651cdcc7dfb29671518d4031cc1dc4d9ba4c06e3e06381a4a5ccf76a0
|
3 |
+
size 88
|
logs/text2image-fine-tune/events.out.tfevents.1719937023.ed7fc0fd9bec.7261.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e74fc5ea8884884b5a7e39cb581ab26e055628b4b7936b1314b4aa08aa88eb3
|
3 |
+
size 5477758
|
pytorch_lora_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1875717ccf0c1cc55ab90c418879d336f001fa70fae57d77f49965970b23279a
|
3 |
+
size 3226184
|