Edit model card

Bagel DPO 57B

exllamav2 quant for TeeZee/2xbagel-dpo-34b-v0.2

Runs smoothly on single 3090 in webui with context length set to 4096, ExLlamav2_HF loader and cache_8bit=True

All comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel: Buy Me A Coffee

Downloads last month
1
Inference API
Input a message to start chatting with TeeZee/2xbagel-dpo-34b-v0.2-bpw3.0-h6-exl2.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Collection including TeeZee/2xbagel-dpo-34b-v0.2-bpw3.0-h6-exl2