Muhammadreza commited on
Commit
4dba445
1 Parent(s): 06fb064

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -2
README.md CHANGED
@@ -10,8 +10,49 @@ library_name: diffusers
10
 
11
  ## How to use the model
12
 
13
- _TBD_
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  ## Tips and Tricks
16
 
17
- _TBD_
 
 
 
10
 
11
  ## How to use the model
12
 
13
+ ### Install needed libraries
14
+
15
+ ```
16
+ pip install git+https://github.com/huggingface/diffusers.git transformers==4.42.4 accelerate xformers peft sentencepiece protobuf -q
17
+ ```
18
+
19
+ ### Execution code
20
+
21
+ ```python
22
+ import numpy as np
23
+ import random
24
+ import torch
25
+ from diffusers import DiffusionPipeline, FlowMatchEulerDiscreteScheduler, AutoencoderTiny, AutoencoderKL
26
+ from transformers import CLIPTextModel, CLIPTokenizer,T5EncoderModel, T5TokenizerFast
27
+
28
+ dtype = torch.bfloat16
29
+ device = "cuda" if torch.cuda.is_available() else "cpu"
30
+
31
+ taef1 = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=dtype).to(device)
32
+ pipe = DiffusionPipeline.from_pretrained("mann-e/mann-e_flux", torch_dtype=dtype, vae=taef1).to(device)
33
+ torch.cuda.empty_cache()
34
+
35
+ MAX_SEED = np.iinfo(np.int32).max
36
+ MAX_IMAGE_SIZE = 2048
37
+
38
+ seed = random.randint(0, MAX_SEED)
39
+ generator = torch.Generator().manual_seed(seed)
40
+
41
+ prompt = "an astronaut riding a horse"
42
+
43
+ pipe(
44
+ prompt=f"{prompt}",
45
+ guidance_scale=3.5,
46
+ num_inference_steps=10,
47
+ width=720,
48
+ height=1280,
49
+ generator=generator,
50
+ output_type="pil"
51
+ ).images[0].save("output.png")
52
+ ```
53
 
54
  ## Tips and Tricks
55
 
56
+ 1. Adding `mj-v6.1-style` to the prompts specially the cinematic and photo realistic prompts can make the result quality high as hell! Give it a try.
57
+ 2. The best `guidance_scale` is somewhere between 3.5 and 5.0
58
+ 3. Inference steps between 8 and 16 are working very well.