It is a repository for storing as many LECOs as I can think of, emphasizing quantity over quality.

Files will continue to be added as needed.

Because the guidance_scale parameter is somewhat excessive, these LECOs tend to be very sensitive and too effective; using a weight of -0.1 to -1 is appropriate in most cases.

All LECOs are trained with target eq positive, erase settings.

The target is a one of among danbooru's GENERAL tags what most frequently used in order from the top to the bottom, and sometimes I also add phrases that I have personally come up with.

- target: "$query"
  positive: "$query"
  unconditional: ""
  neutral: ""
  action: "erase"
  guidance_scale: 1.0
  resolution: 512
  batch_size: 4
prompts_file: prompts.yaml
pretrained_model:
  name_or_path: "/storage/model-1892-0000-0000.safetensors"
  v2: false
  v_pred: false
network:
  type: "lierla"
  rank: 4
  alpha: 1.0
  training_method: "full"
train:
  precision: "bfloat16"
  noise_scheduler: "ddim"
  iterations: 50
  lr: 1
  optimizer: "Prodigy"
  lr_scheduler: "cosine"
  max_denoising_steps: 50

save:
  name: "$query"
  path: "/stable-diffusion-webui/models/Lora/LECO/"
  per_steps: 50
  precision: "float16"

logging:
  use_wandb: false
  verbose: false

other:
  use_xformers: true
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.