kaiokendev commited on
Commit
27a8de1
1 Parent(s): b71cff8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -9,6 +9,12 @@ Tests have shown that the model does indeed leverage the extended context at 8K.
9
 
10
  You will need to **use either the monkeypatch** or, if you are already using the monkeypatch, **change the scaling factor to 0.25 and the maximum sequence length to 8192**
11
 
 
 
 
 
 
 
12
  I trained the LoRA with the following configuration:
13
  - 1200 samples (~400 samples over 2048 sequence length)
14
  - learning rate of 3e-4
 
9
 
10
  You will need to **use either the monkeypatch** or, if you are already using the monkeypatch, **change the scaling factor to 0.25 and the maximum sequence length to 8192**
11
 
12
+ #### Looking for Merged & Quantized Models?
13
+ 30B 4-bit CUDA: [tmpupload/superhot-30b-8k-4bit-safetensors](https://huggingface.co/tmpupload/superhot-30b-8k-4bit-safetensors)
14
+ 30B 4-bit CUDA 128g: [tmpupload/superhot-30b-8k-4bit-128g-safetensors](https://huggingface.co/tmpupload/superhot-30b-8k-4bit-128g-safetensors)
15
+
16
+
17
+ #### Training Details
18
  I trained the LoRA with the following configuration:
19
  - 1200 samples (~400 samples over 2048 sequence length)
20
  - learning rate of 3e-4