Update README.md
Browse files
README.md
CHANGED
@@ -72,8 +72,8 @@ PPL = Perplexity (lower is better) - https://huggingface.co/docs/transformers/pe
|
|
72 |
- You will likely face many bugs until text-generation-webui loads, ranging between missing PATH or env variables to having to manually pip uninstall/install packages.
|
73 |
- The notes below will likely become outdated once both text-generation-webui and GPTQ-for-LLaMa receive the appropriate bug fixes.
|
74 |
- If this model produces very slow answers (1 token/s), it means you are not using Cuda for bitsandbytes or that your hardware needs an upgrade.
|
75 |
-
- If this model produces answers with weird characters, it means you
|
76 |
-
- If this model produces answers that are out of topic or if it talks to itself, it means you
|
77 |
|
78 |
RECOMMENDED - Triton (Fast tokens/s) - Works on Windows with WSL (what I've used) or Linux:
|
79 |
```
|
@@ -87,7 +87,8 @@ mkdir repositories
|
|
87 |
cd repositories
|
88 |
git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git # -b cuda
|
89 |
cd GPTQ-for-LLaMa
|
90 |
-
git checkout 508de42 # Before qwopqwop200 broke everything... - https://github.com/qwopqwop200/GPTQ-for-LLaMa/issues/183
|
|
|
91 |
pip install -r requirements.txt
|
92 |
```
|
93 |
|
@@ -101,6 +102,7 @@ mkdir repositories
|
|
101 |
cd repositories
|
102 |
git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git -b cuda # Make sure you obtain the qwopqwop200 version, not the oobabooga one! (because "act-order: yes")
|
103 |
cd GPTQ-for-LLaMa
|
|
|
104 |
pip install -r requirements.txt
|
105 |
python setup_cuda.py install
|
106 |
```
|
|
|
72 |
- You will likely face many bugs until text-generation-webui loads, ranging between missing PATH or env variables to having to manually pip uninstall/install packages.
|
73 |
- The notes below will likely become outdated once both text-generation-webui and GPTQ-for-LLaMa receive the appropriate bug fixes.
|
74 |
- If this model produces very slow answers (1 token/s), it means you are not using Cuda for bitsandbytes or that your hardware needs an upgrade.
|
75 |
+
- If this model produces answers with weird characters, it means you a using a broken commit of qwopqwop200/GPTQ-for-LLaMa.
|
76 |
+
- If this model produces answers that are out of topic or if it talks to itself, it means you a using a broken commit of qwopqwop200/GPTQ-for-LLaMa.
|
77 |
|
78 |
RECOMMENDED - Triton (Fast tokens/s) - Works on Windows with WSL (what I've used) or Linux:
|
79 |
```
|
|
|
87 |
cd repositories
|
88 |
git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git # -b cuda
|
89 |
cd GPTQ-for-LLaMa
|
90 |
+
#git checkout 508de42 # Since been fixed # Before qwopqwop200 broke everything... - https://github.com/qwopqwop200/GPTQ-for-LLaMa/issues/183
|
91 |
+
git checkout 210c379 # Optional - This is a commit I have verified, you may want to try the latest commit instead, if the latest commit doesn't work revert to an older one such as this one
|
92 |
pip install -r requirements.txt
|
93 |
```
|
94 |
|
|
|
102 |
cd repositories
|
103 |
git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git -b cuda # Make sure you obtain the qwopqwop200 version, not the oobabooga one! (because "act-order: yes")
|
104 |
cd GPTQ-for-LLaMa
|
105 |
+
git checkout 505c2c7 # Optional - This is a commit I have verified, you may want to try the latest commit instead, if the latest commit doesn't work revert to an older one such as this one
|
106 |
pip install -r requirements.txt
|
107 |
python setup_cuda.py install
|
108 |
```
|