Edit model card

This is an EXL2 4-bit quantization of lzlv-longLORA-70b-rope8-32k-fp16

Downloads last month
0
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.