mbrack commited on
Commit
eeafb78
1 Parent(s): de19672

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,7 +22,7 @@ We finetuned this checkpoint on the German Instruction dataset from DiscoResearc
22
 
23
  ## How to use
24
  Llama3_DiscoLeo_Instruct_8B_v0.1 uses the [Llama-3 chat template](https://github.com/meta-llama/llama3?tab=readme-ov-file#instruction-tuned-models), which can be easily used with [transformer's chat templating](https://huggingface.co/docs/transformers/main/en/chat_templating).
25
-
26
 
27
  ## Model Training and Hyperparameters
28
  The model was full-fintuned with axolotl on the [hessian.Ai 42](hessian.ai) with 8192 context-length, learning rate 2e-5 and batch size of 16.
@@ -52,7 +52,7 @@ We release DiscoLeo-8B in the following configurations:
52
  5. [Experimental `DARE-TIES` Merge with Llama3-Instruct](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_8B_DARE_Experimental)
53
  6. [Collection of Quantized versions](https://huggingface.co/collections/DiscoResearch/discoleo-8b-quants-6651bcf8f72c9a37ce485d42)
54
 
55
- ## How to use:
56
  Here's how to use the model with transformers:
57
  ```python
58
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
22
 
23
  ## How to use
24
  Llama3_DiscoLeo_Instruct_8B_v0.1 uses the [Llama-3 chat template](https://github.com/meta-llama/llama3?tab=readme-ov-file#instruction-tuned-models), which can be easily used with [transformer's chat templating](https://huggingface.co/docs/transformers/main/en/chat_templating).
25
+ See [below](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_Instruct_8B_v0.1#usage-example) for a usage example.
26
 
27
  ## Model Training and Hyperparameters
28
  The model was full-fintuned with axolotl on the [hessian.Ai 42](hessian.ai) with 8192 context-length, learning rate 2e-5 and batch size of 16.
 
52
  5. [Experimental `DARE-TIES` Merge with Llama3-Instruct](https://huggingface.co/DiscoResearch/Llama3_DiscoLeo_8B_DARE_Experimental)
53
  6. [Collection of Quantized versions](https://huggingface.co/collections/DiscoResearch/discoleo-8b-quants-6651bcf8f72c9a37ce485d42)
54
 
55
+ ## Usage Example
56
  Here's how to use the model with transformers:
57
  ```python
58
  from transformers import AutoModelForCausalLM, AutoTokenizer