This is my first time doing any sort of Stable Diffusion training so I went through a lot of trial and error. Here are my findings in case it helps anyone.
Training
All parameters are provided in the accompanying JSON files.
- Trained on 138 curated images, repeated 8 times (1104 total images / 3 batchsize = 368 iterations)
- I pruned most images with white backgrounds because I felt they might have been negatively impacting my results early on, but in hindsight I think that was bad training parameters instead.
- Dataset was tagged with WD1.4 interrogator. Shuffling was disabled.
mutsuki, blue archive
were added to the start of each caption.
- Two variants included; one trained at 512px max resolution, and another trained at 768px max resolution. All other params identical.
- Trained on RTX 4090 for about 2min30sec (512px variant) and 6min30sec (768px variant)
- I tried using higher batch sizes with the 512px variant for faster training, but the results seemed noticably worse.
- Small batch sizes seem to work better even when you have the VRAM for 10 or 12, so I instead put the VRAM towards training a higher resolution variant.
Usage
Mutsuki needs a few tags to be summoned reliably. Some common tags in her dataset:
1girl, halo, side ponytail, long hair, white hair, purple eyes, jacket, red skirt, light grin, small breasts
You can add or ignore mutsuki, blue archive
; while they were in her captions, they don't seem to be particularly strong for some reason.
You can use the 512px or 768px variants. I want to say the 768px one is better, but it's hard to say definitively. Give both a shot and post your findings.
Weight 0.80-1.05 should work well depending on model.