radm
/

GGUF
English
radm commited on
Commit
4961b5a
1 Parent(s): 60d1f38

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -73,6 +73,7 @@ Refer to the Provided Files table below to see what files use which methods, and
73
  ## Provided files
74
  | Name | Quant method | Bits | Size | Max RAM required | Use case |
75
  | ---- | ---- | ---- | ---- | ---- | ----- |
 
76
  | [allenai_tulu-2-dpo-70b-q5_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q5_K_M.gguf) | Q5_K_M | 5 | 48.75 GB| 51.25 GB | large, very low quality loss - recommended |
77
 
78
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
 
73
  ## Provided files
74
  | Name | Quant method | Bits | Size | Max RAM required | Use case |
75
  | ---- | ---- | ---- | ---- | ---- | ----- |
76
+ | [allenai_tulu-2-dpo-70b-q4_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q4_K_M.gguf) | Q4_K_M | 4 | 41.42 GB| 43.92 GB | medium, balanced quality - recommended |
77
  | [allenai_tulu-2-dpo-70b-q5_K_M.gguf](https://huggingface.co/radm/allenai_tulu-2-dpo-70b-gguf/blob/main/allenai_tulu-2-dpo-70b-q5_K_M.gguf) | Q5_K_M | 5 | 48.75 GB| 51.25 GB | large, very low quality loss - recommended |
78
 
79
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.