feihu.hf commited on
Commit
12145bd
·
1 Parent(s): df5bf01

upload fp16 weights

Browse files
Files changed (2) hide show
  1. README.md +1 -1
  2. qwen2.5-0.5b-instruct-fp16.gguf +3 -0
README.md CHANGED
@@ -28,7 +28,7 @@ Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we rele
28
  - Number of Paramaters (Non-Embedding): 0.36B
29
  - Number of Layers: 24
30
  - Number of Attention Heads (GQA): 14 for Q and 2 for KV
31
- {{GGUF_LONG_SUMMARY}}
32
  - Quantization: q2_K, q3_K_M, q4_0, q4_K_M, q5_0, q5_K_M, q6_K, q8_0
33
 
34
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
 
28
  - Number of Paramaters (Non-Embedding): 0.36B
29
  - Number of Layers: 24
30
  - Number of Attention Heads (GQA): 14 for Q and 2 for KV
31
+ - Context Length: Full 32,768 tokens and generation 8192 tokens
32
  - Quantization: q2_K, q3_K_M, q4_0, q4_K_M, q5_0, q5_K_M, q6_K, q8_0
33
 
34
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
qwen2.5-0.5b-instruct-fp16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e0ae26000627ed62de0e78e41860af70094558b9d2913385c842a6aa06cf3fc
3
+ size 1266425696