Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ We finetuned this checkpoint on the German Instruction dataset from DiscoResearc
|
|
22 |
|
23 |
## How to use
|
24 |
Llama3_DiscoLeo_Instruct_8B_v0.1 uses the [Llama-3 chat template](https://github.com/meta-llama/llama3?tab=readme-ov-file#instruction-tuned-models), which can be easily used with [transformer's chat templating](https://huggingface.co/docs/transformers/main/en/chat_templating).
|
25 |
-
|
26 |
|
27 |
## Model Training and Hyperparameters
|
28 |
The model was full-fintuned with axolotl on the [hessian.Ai 42](hessian.ai) with 8192 context-length, learning rate 2e-5 and batch size of 16.
|
@@ -52,7 +52,7 @@ We release DiscoLeo-8B in the following configurations:
|
|
52 |
5. [Experimental `DARE-TIES` Merge with Llama3-Instruct](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_8B_DARE_Experimental)
|
53 |
6. [Collection of Quantized versions](https://huggingface.co/collections/DiscoResearch/discoleo-8b-quants-6651bcf8f72c9a37ce485d42)
|
54 |
|
55 |
-
##
|
56 |
Here's how to use the model with transformers:
|
57 |
```python
|
58 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
22 |
|
23 |
## How to use
|
24 |
Llama3_DiscoLeo_Instruct_8B_v0.1 uses the [Llama-3 chat template](https://github.com/meta-llama/llama3?tab=readme-ov-file#instruction-tuned-models), which can be easily used with [transformer's chat templating](https://huggingface.co/docs/transformers/main/en/chat_templating).
|
25 |
+
See [below](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_Instruct_8B_v0.1#usage-example) for a usage example.
|
26 |
|
27 |
## Model Training and Hyperparameters
|
28 |
The model was full-fintuned with axolotl on the [hessian.Ai 42](hessian.ai) with 8192 context-length, learning rate 2e-5 and batch size of 16.
|
|
|
52 |
5. [Experimental `DARE-TIES` Merge with Llama3-Instruct](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_8B_DARE_Experimental)
|
53 |
6. [Collection of Quantized versions](https://huggingface.co/collections/DiscoResearch/discoleo-8b-quants-6651bcf8f72c9a37ce485d42)
|
54 |
|
55 |
+
## Usage Example
|
56 |
Here's how to use the model with transformers:
|
57 |
```python
|
58 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|