Kooten commited on
Commit
6ba02d0
1 Parent(s): eed7585

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ language:
4
+ - en
5
+ ---
6
+
7
+ # Euryale-1.4-L2-70B IQ2-GGUF
8
+
9
+ ## Description
10
+ IQ2-GGUF quants of [Sao10K/Euryale-1.4-L2-70B](https://huggingface.co/Sao10K/Euryale-1.4-L2-70B)
11
+
12
+ Unlike regular GGUF quants this uses important matrix similar to Quip# to keep the quant from degrading too much even at 2bpw allowing you to run larger models on less powerful machines.
13
+
14
+ ***NOTE:*** As of uploading these this llamacpp can run these quants but i am unsure what guis like oobabooga / koboldcpp can run them.
15
+
16
+ [More info](https://github.com/ggerganov/llama.cpp/pull/4897)
17
+
18
+
19
+ # Models
20
+
21
+ Models: [IQ2-XS](https://huggingface.co/Kooten/Euryale-1.4-L2-70B-IQ2-GGUF/raw/main/Euryale-1.4-L2-70B-IQ2_XS.gguf), [IQ2-XSS](https://huggingface.co/Kooten/Euryale-1.4-L2-70B-IQ2-GGUF/raw/main/Euryale-1.4-L2-70B-IQ2_XXS.gguf)
22
+
23
+ Regular GGUF Quants: [Here](https://huggingface.co/Sao10K/Euryale-1.4-L2-70B-GGUF)
24
+
25
+ ## Prompt Format
26
+
27
+ ### Alpaca:
28
+ ```
29
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
30
+
31
+ ### Instruction:
32
+ {prompt}
33
+
34
+ ### Input:
35
+ {input}
36
+
37
+ ### Response:
38
+
39
+ ```
40
+
41
+ ## Contact
42
+ Kooten on discord