Edit model card

Qwen1.5-32B-Chat-3.5bpw-exl2

This is a 3.5bpw quantized version of Qwen/Qwen1.5-32B-Chat made with exllamav2.

To run this, make sure you installed the up-to-date version of Exllamav2.

License

This project is distributed under the Tongyi Qianwen LICENSE AGREEMENT. See the LICENSE file for more information.

Feeling Generous? 😊

Eager to buy me a cup of 2$ coffe or iced tea?πŸ΅β˜• Sure, here is the link: https://ko-fi.com/drnicefellow. Please add a note on which one you want me to drink?

Downloads last month
0
Inference API
Input a message to start chatting with DrNicefellow/Qwen1.5-32B-Chat-3.5bpw-exl2.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Collection including DrNicefellow/Qwen1.5-32B-Chat-3.5bpw-exl2