Edit model card

This is an EXL2 6-bit quantization of Goliath-longLORA-120b-rope8-32k-fp16

Downloads last month
2
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.