--- pipeline_tag: text-to-image --- # wdv-tests alright since i've been promising everyone something for ages and i don't have a clue when wdV actual will get done other than soon, here's something that's better than nothing ## **NOTE:** this is NOT a release, per se. it's a public test, if anything. this uses the TEs from Animagine 3.1 (with permission) and is not representative of the final model really. But it's something, and i rather like it. ### see below for setup instructions - you *need* to use "simple" or "ddim_uniform" or *maybe* "karras" schedule for the sampler - dpmpp_2s_ancestral works best. **SDE samplers will not work.** - 20 steps or so at 5.5-7.5 CFG ## Setup instructions ### Option 1: SwarmUI extension (recommended) 1. the SwarmUI extension repo is [here on GitHub](https://github.com/waifu-diffusion/SwarmWaifuDiffusionV) 2. Open that link above and read the instructions in the README 3. If you made it to this step, go back to step 2. ### Option 1: Raw Comfy node (not recommended) 1. get the node from [the WaifuNodes folder of the SwarmUI extension](https://github.com/waifu-diffusion/SwarmWaifuDiffusionV/blob/main/WaifuNodes/wdv_test_nodes.py) 2. put it between your model loader and sampler, 3. set sampling mode to "EDM" 4. set schedule to "XLTC", 5. set scaling 1.6 (these 3 should all be defaults) ## Model training info v0 is trained for five epochs or so on about 4.8 million assorted anime images and uses the text encoder from Animagine 3.1 (with permission from the creator). Actual release wdV models will not use this TE, to be quite honest i swapped it in for a test and forgot to remove it so here we are.