Text Generation
scaling
GregorZiegltrumAA commited on
Commit
edc895f
1 Parent(s): a1c7328

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
16
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/671a0238b080a748c29b8fea/F1-zbAXF5LGvxpIRrYfU4.png)
17
 
18
 
19
- This Repository holds the model weights for the 7B u-μP models trained at Aleph Alpha Research for 72k steps (300B tokens). Please note, that the released checkpoints are not fully converged models and are intended for research use only.
20
 
21
  You can find all model weights and their corresponding safetensors conversions at the following links:
22
  - [umup-research-7b-bf16](https://huggingface.co/Aleph-Alpha/umup-research-7b-bf16)
@@ -42,7 +42,7 @@ ckpt_path = Path("<path_to_repo>/7B_umup_fp8")
42
 
43
  model = TransformerInferenceModule.from_checkpoint(ckpt_path)
44
 
45
- prompt = "Once upon a time"
46
 
47
  output = model.generate(max_tokens=100, input_text=prompt)
48
 
 
16
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/671a0238b080a748c29b8fea/F1-zbAXF5LGvxpIRrYfU4.png)
17
 
18
 
19
+ This Repository holds the model weights for the 7B u-μP models trained at Aleph Alpha Research, in collaboration with Graphcore, for 72k steps (300B tokens). Please note, that the released checkpoints are not fully converged models and are intended for research use only.
20
 
21
  You can find all model weights and their corresponding safetensors conversions at the following links:
22
  - [umup-research-7b-bf16](https://huggingface.co/Aleph-Alpha/umup-research-7b-bf16)
 
42
 
43
  model = TransformerInferenceModule.from_checkpoint(ckpt_path)
44
 
45
+ prompt = "Yesterday I dreamt of "
46
 
47
  output = model.generate(max_tokens=100, input_text=prompt)
48