Update README.md
Browse files
README.md
CHANGED
@@ -129,7 +129,8 @@ with three layers used for context extension. **Crucially, LongLLaMA is able to
|
|
129 |
|----------------|----------|----------|-----------|
|
130 |
| Source model | [OpenLLaMA-3B](https://huggingface.co/openlm-research/open_llama_3b_easylm) | [OpenLLaMA-3Bv2](https://huggingface.co/openlm-research/open_llama_3b_v2_easylm) | [CodeLLaMA-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) |
|
131 |
| Source model tokens | 1T | 1 T | 2T + 0.5 T |
|
132 |
-
| Fine-tuning
|
|
|
133 |
| Memory layers | 6, 12, 18 | 6, 12, 18 | 8, 16, 24 |
|
134 |
|
135 |
</div>
|
|
|
129 |
|----------------|----------|----------|-----------|
|
130 |
| Source model | [OpenLLaMA-3B](https://huggingface.co/openlm-research/open_llama_3b_easylm) | [OpenLLaMA-3Bv2](https://huggingface.co/openlm-research/open_llama_3b_v2_easylm) | [CodeLLaMA-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) |
|
131 |
| Source model tokens | 1T | 1 T | 2T + 0.5 T |
|
132 |
+
| Fine-tuning context | 8K | 32K | 32K |
|
133 |
+
| Fine-tuning tokens | 10B | 5B | 35B |
|
134 |
| Memory layers | 6, 12, 18 | 6, 12, 18 | 8, 16, 24 |
|
135 |
|
136 |
</div>
|