paulkm commited on
Commit
84d6f0b
1 Parent(s): 63d27f6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -0
README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: text-to-image
3
+ library_name: optimum_neuronx
4
+ ---
5
+
6
+ This is a model compiled for optimized inference on AWS inf2 instances. The overall performance can achieve 4it/s for a 1024x1024 picture.
7
+
8
+ NOTE: In order to load and infer on inf2.xlarge instances, a minimum of 8 GB SWAP memory is required. Other instance sizes do not require this.
9
+
10
+ Assuming you're using the official DLAMI or installed the drivers and libraries required for neuron devices, install the latest optimum_neuronx library:
11
+ ```
12
+ pip install -U optimum[neuronx] diffusers=0.20.0
13
+ ```
14
+
15
+ Then, enjoy the super-fast experience:
16
+ ```py
17
+ from optimum.neuron import NeuronStableDiffusionXLPipeline
18
+ pipe = NeuronStableDiffusionXLPipeline.from_pretrained("paulkm/sdxl_neuron_pipe", device_ids=[0,1])
19
+ img = pipe("a cute black cat").images[0]
20
+ ```