End of training
Browse files- README.md +37 -0
- prompt.txt +1 -0
- pytorch_lora_weights.safetensors +3 -0
- pytorch_lora_weights_kohya.safetensors +3 -0
- training_params.json +1 -0
README.md
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- autotrain
|
4 |
+
- stable-diffusion
|
5 |
+
- stable-diffusion-diffusers
|
6 |
+
- text-to-image
|
7 |
+
- diffusers
|
8 |
+
- lora
|
9 |
+
- template:sd-lora
|
10 |
+
|
11 |
+
base_model: stable-diffusion-v1-5/stable-diffusion-v1-5
|
12 |
+
instance_prompt: photo of ambika0 man
|
13 |
+
license: openrail++
|
14 |
+
---
|
15 |
+
# ModelsLab LoRA DreamBooth Training - stablediffusionapi/my-stablediffusion-lora-6484
|
16 |
+
These are LoRA adaption weights for stable-diffusion-v1-5/stable-diffusion-v1-5. The weights were trained on photo of ambika0 man using [ModelsLab](https://modelslab.com).
|
17 |
+
LoRA for the text encoder was enabled: False.
|
18 |
+
|
19 |
+
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
|
20 |
+
```py
|
21 |
+
!pip install -q transformers accelerate peft diffusers
|
22 |
+
from diffusers import DiffusionPipeline
|
23 |
+
import torch
|
24 |
+
|
25 |
+
pipe_id = "Lykon/DreamShaper"
|
26 |
+
pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda")
|
27 |
+
pipe.load_lora_weights("stablediffusionapi/my-stablediffusion-lora-6484", weight_name="pytorch_lora_weights.safetensors", adapter_name="abc")
|
28 |
+
prompt = "abc of a hacker with a hoodie"
|
29 |
+
lora_scale = 0.9
|
30 |
+
image = pipe(
|
31 |
+
prompt,
|
32 |
+
num_inference_steps=30,
|
33 |
+
cross_attention_kwargs={"scale": 0.9},
|
34 |
+
generator=torch.manual_seed(0)
|
35 |
+
).images[0]
|
36 |
+
image
|
37 |
+
```
|
prompt.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
photo of ambika0 man
|
pytorch_lora_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e25f2713e510a93ab9bc67b24b64694fa4466186fbddb2a6747cb72a87a02df7
|
3 |
+
size 12792952
|
pytorch_lora_weights_kohya.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d30148d7ab47dcd7c6efb039339145c72ab3876de2bf12e1d849b73d9073692
|
3 |
+
size 12811664
|
training_params.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"model": "stable-diffusion-v1-5/stable-diffusion-v1-5", "vae_model": null, "revision": null, "tokenizer": null, "image_path": "my-stablediffusion-lora-6484/autotrain-data", "class_image_path": null, "prompt": "photo of ambika0 man", "class_prompt": null, "num_class_images": 100, "class_labels_conditioning": null, "prior_preservation": false, "prior_loss_weight": 1.0, "project_name": "my-stablediffusion-lora-6484", "seed": 42, "resolution": 1024, "center_crop": false, "train_text_encoder": false, "batch_size": 4, "sample_batch_size": 4, "epochs": 1, "num_steps": 5, "checkpointing_steps": 500, "resume_from_checkpoint": null, "gradient_accumulation": 4, "disable_gradient_checkpointing": false, "lr": 0.0005, "scale_lr": false, "scheduler": "constant", "warmup_steps": 0, "num_cycles": 1, "lr_power": 1.0, "dataloader_num_workers": 0, "use_8bit_adam": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_weight_decay": 0.01, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "allow_tf32": false, "prior_generation_precision": null, "local_rank": -1, "xformers": false, "pre_compute_text_embeddings": false, "tokenizer_max_length": null, "text_encoder_use_attention_mask": false, "rank": 16, "xl": false, "mixed_precision": "fp16", "push_to_hub": true, "username": "stablediffusionapi", "hf_private": false, "validation_prompt": null, "num_validation_images": 4, "validation_epochs": 50, "checkpoints_total_limit": null, "validation_images": null, "logging": false}
|