Pytorch int8 quantized version of gpt2-large
Usage
Download the .bin file locally. Load with:
Rest of the usage according to original instructions.
import torch
model = torch.load("path/to/pytorch_model_quantized.bin")
- Downloads last month
- 19
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.