Pytorch int8 quantized version of gpt2-large

Usage

Download the .bin file locally. Load with:

Rest of the usage according to original instructions.

import torch

model = torch.load("path/to/pytorch_model_quantized.bin")
Downloads last month
19
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.