Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
pipeline_tag: text-to-image
|
3 |
+
library_name: optimum_neuronx
|
4 |
+
---
|
5 |
+
|
6 |
+
This is a model compiled for optimized inference on AWS inf2 instances. The overall performance can achieve 4it/s for a 1024x1024 picture.
|
7 |
+
|
8 |
+
NOTE: In order to load and infer on inf2.xlarge instances, a minimum of 8 GB SWAP memory is required. Other instance sizes do not require this.
|
9 |
+
|
10 |
+
Assuming you're using the official DLAMI or installed the drivers and libraries required for neuron devices, install the latest optimum_neuronx library:
|
11 |
+
```
|
12 |
+
pip install -U optimum[neuronx] diffusers=0.20.0
|
13 |
+
```
|
14 |
+
|
15 |
+
Then, enjoy the super-fast experience:
|
16 |
+
```py
|
17 |
+
from optimum.neuron import NeuronStableDiffusionXLPipeline
|
18 |
+
pipe = NeuronStableDiffusionXLPipeline.from_pretrained("paulkm/sdxl_neuron_pipe", device_ids=[0,1])
|
19 |
+
img = pipe("a cute black cat").images[0]
|
20 |
+
```
|