Thireus commited on
Commit
c669a07
1 Parent(s): a567ec5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -72,8 +72,8 @@ PPL = Perplexity (lower is better) - https://huggingface.co/docs/transformers/pe
72
  - You will likely face many bugs until text-generation-webui loads, ranging between missing PATH or env variables to having to manually pip uninstall/install packages.
73
  - The notes below will likely become outdated once both text-generation-webui and GPTQ-for-LLaMa receive the appropriate bug fixes.
74
  - If this model produces very slow answers (1 token/s), it means you are not using Cuda for bitsandbytes or that your hardware needs an upgrade.
75
- - If this model produces answers with weird characters, it means you are not using the correct version of qwopqwop200/GPTQ-for-LLaMa as mentioned below.
76
- - If this model produces answers that are out of topic or if it talks to itself, it means you are not using the correct checkout 508de42 of qwopqwop200/GPTQ-for-LLaMa as mentioned below.
77
 
78
  RECOMMENDED - Triton (Fast tokens/s) - Works on Windows with WSL (what I've used) or Linux:
79
  ```
@@ -87,7 +87,8 @@ mkdir repositories
87
  cd repositories
88
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git # -b cuda
89
  cd GPTQ-for-LLaMa
90
- git checkout 508de42 # Before qwopqwop200 broke everything... - https://github.com/qwopqwop200/GPTQ-for-LLaMa/issues/183
 
91
  pip install -r requirements.txt
92
  ```
93
 
@@ -101,6 +102,7 @@ mkdir repositories
101
  cd repositories
102
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git -b cuda # Make sure you obtain the qwopqwop200 version, not the oobabooga one! (because "act-order: yes")
103
  cd GPTQ-for-LLaMa
 
104
  pip install -r requirements.txt
105
  python setup_cuda.py install
106
  ```
 
72
  - You will likely face many bugs until text-generation-webui loads, ranging between missing PATH or env variables to having to manually pip uninstall/install packages.
73
  - The notes below will likely become outdated once both text-generation-webui and GPTQ-for-LLaMa receive the appropriate bug fixes.
74
  - If this model produces very slow answers (1 token/s), it means you are not using Cuda for bitsandbytes or that your hardware needs an upgrade.
75
+ - If this model produces answers with weird characters, it means you a using a broken commit of qwopqwop200/GPTQ-for-LLaMa.
76
+ - If this model produces answers that are out of topic or if it talks to itself, it means you a using a broken commit of qwopqwop200/GPTQ-for-LLaMa.
77
 
78
  RECOMMENDED - Triton (Fast tokens/s) - Works on Windows with WSL (what I've used) or Linux:
79
  ```
 
87
  cd repositories
88
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git # -b cuda
89
  cd GPTQ-for-LLaMa
90
+ #git checkout 508de42 # Since been fixed # Before qwopqwop200 broke everything... - https://github.com/qwopqwop200/GPTQ-for-LLaMa/issues/183
91
+ git checkout 210c379 # Optional - This is a commit I have verified, you may want to try the latest commit instead, if the latest commit doesn't work revert to an older one such as this one
92
  pip install -r requirements.txt
93
  ```
94
 
 
102
  cd repositories
103
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git -b cuda # Make sure you obtain the qwopqwop200 version, not the oobabooga one! (because "act-order: yes")
104
  cd GPTQ-for-LLaMa
105
+ git checkout 505c2c7 # Optional - This is a commit I have verified, you may want to try the latest commit instead, if the latest commit doesn't work revert to an older one such as this one
106
  pip install -r requirements.txt
107
  python setup_cuda.py install
108
  ```