GregorZiegltrumAA
commited on
Commit
•
edc895f
1
Parent(s):
a1c7328
Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
|
|
16 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/671a0238b080a748c29b8fea/F1-zbAXF5LGvxpIRrYfU4.png)
|
17 |
|
18 |
|
19 |
-
This Repository holds the model weights for the 7B u-μP models trained at Aleph Alpha Research for 72k steps (300B tokens). Please note, that the released checkpoints are not fully converged models and are intended for research use only.
|
20 |
|
21 |
You can find all model weights and their corresponding safetensors conversions at the following links:
|
22 |
- [umup-research-7b-bf16](https://huggingface.co/Aleph-Alpha/umup-research-7b-bf16)
|
@@ -42,7 +42,7 @@ ckpt_path = Path("<path_to_repo>/7B_umup_fp8")
|
|
42 |
|
43 |
model = TransformerInferenceModule.from_checkpoint(ckpt_path)
|
44 |
|
45 |
-
prompt = "
|
46 |
|
47 |
output = model.generate(max_tokens=100, input_text=prompt)
|
48 |
|
|
|
16 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/671a0238b080a748c29b8fea/F1-zbAXF5LGvxpIRrYfU4.png)
|
17 |
|
18 |
|
19 |
+
This Repository holds the model weights for the 7B u-μP models trained at Aleph Alpha Research, in collaboration with Graphcore, for 72k steps (300B tokens). Please note, that the released checkpoints are not fully converged models and are intended for research use only.
|
20 |
|
21 |
You can find all model weights and their corresponding safetensors conversions at the following links:
|
22 |
- [umup-research-7b-bf16](https://huggingface.co/Aleph-Alpha/umup-research-7b-bf16)
|
|
|
42 |
|
43 |
model = TransformerInferenceModule.from_checkpoint(ckpt_path)
|
44 |
|
45 |
+
prompt = "Yesterday I dreamt of "
|
46 |
|
47 |
output = model.generate(max_tokens=100, input_text=prompt)
|
48 |
|