UltraCM-13B-GGUF / README.md
alvarobartt's picture
alvarobartt HF staff
Update README.md
bec663b
|
raw
history blame
3.8 kB
metadata
license: mit
language:
  - en
datasets:
  - openbmb/UltraFeedback
model_creator: OpenBMB
model_name: UltraCM-13b
model_type: llama
base_model: openbmb/UltraCM-13b
library_name: transformers
pipeline_tag: text-generation
inference: false
tags:
  - dpo
  - rlaif
  - preference
  - ultrafeedback
quantized_by: alvarobartt

Model Card for UltraCM-13b-GGUF

UltraCM-13B is a fine-tuned LLM for completion-critique in order to evaluate LLM outputs on helpfulness, truthfulness, honesty, and to what extent the answer follows the given instructions.

UltraCM-13B is a 13b param LLM that was released by OpenBMB, as part of their paper UltraFeedback: Boosting Language Models with High-quality Feedback.

This model contains the quantized variants using the GGUF format, introduced by the llama.cpp team, and also heavily inspired by TheBloke work on quantizing most of the LLMs out there.

Model Details

Model Description

Model Files

Name Quant method Bits Size Max RAM required Use case
UltraCM-13b.q4_0.gguf Q4_0 4 3.83 GB 6.33 GB legacy; small, very high quality loss - prefer using Q3_K_M
UltraCM-13b.q4_k_s.gguf Q4_K_S 4 7.41 GB 9.91 GB small, greater quality loss
UltraCM-13b.q4_k_m.gguf Q4_K_M 4 7.87 GB 10.37 GB medium, balanced quality - recommended
UltraCM-13b.q5_0.gguf Q5_0 5 4.65 GB 7.15 GB legacy; medium, balanced quality - prefer using Q4_K_M
UltraCM-13b.q5_k_s.gguf Q5_K_S 5 8.97 GB 11.47 GB large, low quality loss - recommended
UltraCM-13b.q5_k_m.gguf Q5_K_M 5 9.23 GB 11.73 GB large, very low quality loss - recommended

Note: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.

For more information on quantization, I'd highly suggest anyone reading to go check TheBloke out, as well as joining their Discord server.

Uses

Direct Use

[More Information Needed]

Citation

Since this is only a GGUF-quantization of the original weights, please refer and cite the original authors instead.

@misc{cui2023ultrafeedback,
      title={UltraFeedback: Boosting Language Models with High-quality Feedback}, 
      author={Ganqu Cui and Lifan Yuan and Ning Ding and Guanming Yao and Wei Zhu and Yuan Ni and Guotong Xie and Zhiyuan Liu and Maosong Sun},
      year={2023},
      eprint={2310.01377},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}