原始模型:Sakura-13B-LNovel-v0.11pre1
4Bit AWQ量化,未测试,不建议使用。
采用未安装flash_attn的环境进行量化
Intel-XPU测试用,该量化模型可能不适合所有人。
- Downloads last month
- 7
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The HF Inference API does not support model that require custom code execution.