Update README.md
Browse files
README.md
CHANGED
@@ -73,6 +73,7 @@ Refer to the Provided Files table below to see what files use which methods, and
|
|
73 |
## Provided files
|
74 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
75 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
|
|
76 |
| [allenai_tulu-2-dpo-70b-q5_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q5_K_M.gguf) | Q5_K_M | 5 | 48.75 GB| 51.25 GB | large, very low quality loss - recommended |
|
77 |
|
78 |
**Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
|
|
|
73 |
## Provided files
|
74 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
75 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
76 |
+
| [allenai_tulu-2-dpo-70b-q4_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q4_K_M.gguf) | Q4_K_M | 4 | 41.42 GB| 43.92 GB | medium, balanced quality - recommended |
|
77 |
| [allenai_tulu-2-dpo-70b-q5_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q5_K_M.gguf) | Q5_K_M | 5 | 48.75 GB| 51.25 GB | large, very low quality loss - recommended |
|
78 |
|
79 |
**Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
|