Text Generation
Transformers
PyTorch
longllama
code
text-generation-inference
custom_code
Eval Results
syzymon commited on
Commit
3dea16c
1 Parent(s): 81abcb1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -129,7 +129,8 @@ with three layers used for context extension. **Crucially, LongLLaMA is able to
129
  |----------------|----------|----------|-----------|
130
  | Source model | [OpenLLaMA-3B](https://huggingface.co/openlm-research/open_llama_3b_easylm) | [OpenLLaMA-3Bv2](https://huggingface.co/openlm-research/open_llama_3b_v2_easylm) | [CodeLLaMA-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) |
131
  | Source model tokens | 1T | 1 T | 2T + 0.5 T |
132
- | Fine-tuning tokens | 10B | 5B | 35B | - |
 
133
  | Memory layers | 6, 12, 18 | 6, 12, 18 | 8, 16, 24 |
134
 
135
  </div>
 
129
  |----------------|----------|----------|-----------|
130
  | Source model | [OpenLLaMA-3B](https://huggingface.co/openlm-research/open_llama_3b_easylm) | [OpenLLaMA-3Bv2](https://huggingface.co/openlm-research/open_llama_3b_v2_easylm) | [CodeLLaMA-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) |
131
  | Source model tokens | 1T | 1 T | 2T + 0.5 T |
132
+ | Fine-tuning context | 8K | 32K | 32K |
133
+ | Fine-tuning tokens | 10B | 5B | 35B |
134
  | Memory layers | 6, 12, 18 | 6, 12, 18 | 8, 16, 24 |
135
 
136
  </div>