kaiokendev commited on
Commit
b5290ed
1 Parent(s): 41fef8d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -23,7 +23,7 @@ Trained against LLaMa 30B 4-bit for 3 epochs with cutoff length 1024, using a mi
23
  - 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
24
 
25
  ### Compatibility
26
- This LoRA is compatible with any 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
27
 
28
  ### Prompting
29
  You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
@@ -51,8 +51,9 @@ Remember that with lower parameter sizes, the structure of the prompt becomes mo
51
 
52
  ### Coming Soon
53
  - 2048 7B version
54
- - 1024 and 512 variants of 13B and 7B
55
- - merged ggml models
 
56
 
57
  ### Citations
58
  Alpaca COT datasets
 
23
  - 30B 4-bit 128g CUDA: [https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda](https://huggingface.co/tsumeone/llama-30b-supercot-4bit-128g-cuda)
24
 
25
  ### Compatibility
26
+ This LoRA is compatible with any 7B, 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
27
 
28
  ### Prompting
29
  You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
 
51
 
52
  ### Coming Soon
53
  - 2048 7B version
54
+ - 512 variants of 13B and 7B
55
+ - merged ggml models for 13B and 7B
56
+ - Tweet fix for 13B and 7B
57
 
58
  ### Citations
59
  Alpaca COT datasets