amused-512-pokemon / README.md
suvadityamuk's picture
End of training
ae2cc14 verified
|
raw
history blame
1.7 kB
metadata
license: creativeml-openrail-m
base_model: amused/amused-512
datasets:
  - lambdalabs/pokemon-blip-captions
tags:
  - stable-diffusion
  - stable-diffusion-diffusers
  - text-to-image
  - diffusers
  - amused
inference: true

aMUSEd finetuning - suvadityamuk/amused-512-pokemon

This pipeline was finetuned from amused/amused-512 on the lambdalabs/pokemon-blip-captions dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['a pokemon red mammoth with unicorn horns', 'a pokemon blue fish with golden scales', 'a pokemon green goblin with glasses, wearing black pants and red shirt', 'a pokemon golden unicorn with shiny black hair and deep blue horns', 'a pokemon drawing of a dragon with its mouth closed', 'a pokemon red and yellow phoenix with fire on its wings', 'a pokemon purple tree with white leaves and golden nectar flowing', 'a pokemon green caterpillar']:

val_imgs_grid

Pipeline usage

You can use the pipeline like so:

from diffusers import DiffusionPipeline
import torch

pipeline = DiffusionPipeline.from_pretrained("suvadityamuk/amused-512-pokemon", torch_dtype=torch.float16)
prompt = "a pokemon red mammoth with unicorn horns"
image = pipeline(prompt).images[0]
image.save("my_image.png")

Training info

These are the key hyperparameters used during training:

  • Training Steps: 750
  • Learning rate: 0.00025
  • Batch size: 8
  • Gradient accumulation steps: 4
  • Image resolution: 512
  • Mixed-precision: bf16

More information on all the CLI arguments and the environment are available on your wandb run page.