ddh0's picture
Update README.md
926cd48 verified
---
license: apache-2.0
library_name: transformers
pipeline_tag: text-generation
---
# Starling-LM-10.7B-beta
This is Starling-LM-10.7B-beta, a depth-upscaled version of [Nexusflow/Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta).
This model is intended to be used as a drop-in upgrade from the original 7 billion parameter model.
# GGUF quantizations (courtesy of bartowski)
See [bartowski/Starling-LM-10.7B-beta-GGUF](https://huggingface.co/bartowski/Starling-LM-10.7B-beta-GGUF)
# ExLlamaV2 quantizations (courtesy of [blockblockblock](https://huggingface.co/blockblockblock))
- [2.5 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw2.5)
- [3 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw3)
- [3.5 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw3.5)
- [3.7 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw3.7)
- [4 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw4)
- [4.4 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw4.4)
- [4.6 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw4.6)
- [4.8 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw4.8)
- [5 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw5)
- [5.5 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw5.5)
- [6 bpw](https://huggingface.co/blockblockblock/Starling-LM-10.7B-beta-bpw6)