diffusers-qa-chatbot-artifacts / scrapped_outputs /041d6ec5bc898d377b96ad1c3e5ce22b.txt
sayakpaul's picture
sayakpaul HF staff
Upload folder using huggingface_hub
9b4884a verified
raw
history blame
749 Bytes
Overview Generating high-quality outputs is computationally intensive, especially during each iterative step where you go from a noisy output to a less noisy output. One of 🤗 Diffuser’s goals is to make this technology widely accessible to everyone, which includes enabling fast inference on consumer and specialized hardware. This section will cover tips and tricks - like half-precision weights and sliced attention - for optimizing inference speed and reducing memory-consumption. You’ll also learn how to speed up your PyTorch code with torch.compile or ONNX Runtime, and enable memory-efficient attention with xFormers. There are also guides for running inference on specific hardware like Apple Silicon, and Intel or Habana processors.