--- license: creativeml-openrail-m base_model: amused/amused-512 datasets: - lambdalabs/pokemon-blip-captions tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - amused inference: true --- # aMUSEd finetuning - suvadityamuk/amused-512-pokemon This pipeline was finetuned from **amused/amused-512** on the **lambdalabs/pokemon-blip-captions** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['a pokemon red mammoth with unicorn horns', 'a pokemon blue fish with golden scales', 'a pokemon green goblin with glasses, wearing black pants and red shirt', 'a pokemon golden unicorn with shiny black hair and deep blue horns', 'a pokemon drawing of a dragon with its mouth closed', 'a pokemon red and yellow phoenix with fire on its wings', 'a pokemon purple tree with white leaves and golden nectar flowing', 'a pokemon green caterpillar']: ![val_imgs_grid](./val_imgs_grid.png) ## Pipeline usage You can use the pipeline like so: ```python from diffusers import DiffusionPipeline import torch pipeline = DiffusionPipeline.from_pretrained("suvadityamuk/amused-512-pokemon", torch_dtype=torch.float16) prompt = "a pokemon red mammoth with unicorn horns" image = pipeline(prompt).images[0] image.save("my_image.png") ``` ## Training info These are the key hyperparameters used during training: * Train Steps: 750 * Learning rate: 5e-06 * Batch size: 8 * Gradient accumulation steps: 4 * Image resolution: 512 * Mixed-precision: bf16 More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/ml-colabs/fconn-amused/runs/0cljkrh9).