kaiokendev
commited on
Commit
•
b5290ed
1
Parent(s):
41fef8d
Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ Trained against LLaMa 30B 4-bit for 3 epochs with cutoff length 1024, using a mi
|
|
23 |
- 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
|
24 |
|
25 |
### Compatibility
|
26 |
-
This LoRA is compatible with any 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
|
27 |
|
28 |
### Prompting
|
29 |
You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
|
@@ -51,8 +51,9 @@ Remember that with lower parameter sizes, the structure of the prompt becomes mo
|
|
51 |
|
52 |
### Coming Soon
|
53 |
- 2048 7B version
|
54 |
-
-
|
55 |
-
- merged ggml models
|
|
|
56 |
|
57 |
### Citations
|
58 |
Alpaca COT datasets
|
|
|
23 |
- 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
|
24 |
|
25 |
### Compatibility
|
26 |
+
This LoRA is compatible with any 7B, 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
|
27 |
|
28 |
### Prompting
|
29 |
You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
|
|
|
51 |
|
52 |
### Coming Soon
|
53 |
- 2048 7B version
|
54 |
+
- 512 variants of 13B and 7B
|
55 |
+
- merged ggml models for 13B and 7B
|
56 |
+
- Tweet fix for 13B and 7B
|
57 |
|
58 |
### Citations
|
59 |
Alpaca COT datasets
|