File size: 1,694 Bytes
ae2cc14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8eb0e39
 
ae2cc14
 
 
 
 
 
8eb0e39
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50

---
license: creativeml-openrail-m
base_model: amused/amused-512
datasets:
- lambdalabs/pokemon-blip-captions
tags:
- stable-diffusion
- stable-diffusion-diffusers
- text-to-image
- diffusers
- amused
inference: true
---
    
# aMUSEd finetuning - suvadityamuk/amused-512-pokemon

This pipeline was finetuned from **amused/amused-512** on the **lambdalabs/pokemon-blip-captions** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['a pokemon red mammoth with unicorn horns', 'a pokemon blue fish with golden scales', 'a pokemon green goblin with glasses, wearing black pants and red shirt', 'a pokemon golden unicorn with shiny black hair and deep blue horns', 'a pokemon drawing of a dragon with its mouth closed', 'a pokemon red and yellow phoenix with fire on its wings', 'a pokemon purple tree with white leaves and golden nectar flowing', 'a pokemon green caterpillar']: 

![val_imgs_grid](./val_imgs_grid.png)


## Pipeline usage

You can use the pipeline like so:

```python
from diffusers import DiffusionPipeline
import torch

pipeline = DiffusionPipeline.from_pretrained("suvadityamuk/amused-512-pokemon", torch_dtype=torch.float16)
prompt = "a pokemon red mammoth with unicorn horns"
image = pipeline(prompt).images[0]
image.save("my_image.png")
```

## Training info

These are the key hyperparameters used during training:

* Train Steps: 750
* Learning rate: 5e-06
* Batch size: 8
* Gradient accumulation steps: 4
* Image resolution: 512
* Mixed-precision: bf16


More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/ml-colabs/fconn-amused/runs/0cljkrh9).