File size: 1,100 Bytes
9b4d480
 
 
642a988
abfdf8f
099b02b
abfdf8f
9b7ef2f
abfdf8f
2d9173b
abfdf8f
099b02b
abfdf8f
860e903
099b02b
 
 
 
 
 
 
860e903
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
---
license: apache-2.0
---

**NOTE:** Tested with KoboldCpp. I can't promise that it will work with other frontends. Use at your own risk!

**2023-04-20:** Converted using ggerganov/ggml's "stablelm" conversion script and quantization code as of [commit 05f3079](https://github.com/ggerganov/ggml/tree/05f307971862b83df12fada0c42ee027ba5a82b5/examples/stablelm).

**2023-04-30:** Converted using ggerganov/ggml's "stablelm" conversion script and quantization code as of [commit 5dd92f4](https://github.com/ggerganov/ggml/tree/5dd92f421ee44f18b8fde0afbf5ca8fc7bf93841/examples/stablelm).

These conversions are based off of the updated Pythia Deduped checkpoints, not the original v0 trainings.

# RAM USAGE:
Model | Initial RAM usage
:--:|:--:
ggml-pythia-70m-deduped-q4_3.bin | 121.2 MiB
ggml-pythia-160m-deduped-q4_3.bin | 225.2 MiB
ggml-pythia-410m-deduped-q4_3.bin | 498.1 MiB
ggml-pythia-1b-deduped-q4_3.bin | 951.5 MiB
ggml-pythia-1.4b-deduped-q4_3.bin | 1.3 GiB
ggml-pythia-2.8b-deduped-q4_3.bin | 2.4 GiB
ggml-pythia-6.9b-deduped-q4_3.bin | 5.4 GiB
ggml-pythia-12b-deduped-q4_3.bin | 9.2 GiB