TheBloke commited on
Commit
ee2d9ff
1 Parent(s): 4bf8036

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +24 -2
README.md CHANGED
@@ -2,6 +2,17 @@
2
  license: other
3
  inference: false
4
  ---
 
 
 
 
 
 
 
 
 
 
 
5
 
6
  # Alpaca LoRA 65B GPTQ 4bit
7
 
@@ -30,7 +41,7 @@ Three files are provided. **The second and third files will not work unless you
30
  Specifically, the last two files use `--act-order` for maximum quantisation quality and will not work with oobabooga's fork of GPTQ-for-LLaMa. Therefore at this time it will also not work with the CUDA branch of GPTQ-for-LLaMa, or `text-generation-webui` one-click installers.
31
 
32
  Unless you are able to use the latest Triton GPTQ-for-LLaMa code, please use `medalpaca-13B-GPTQ-4bit-128g.no-act-order.safetensors`
33
-
34
  * `alpaca-lora-65B-GPTQ-4bit-128g.no-act-order.safetensors`
35
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
36
  * Works with text-generation-webui one-click-installers
@@ -96,8 +107,19 @@ The above commands assume you have installed all dependencies for GPTQ-for-LLaMa
96
 
97
  If you can't update GPTQ-for-LLaMa to the latest Triton branch, or don't want to, you can use `alpaca-lora-65B-GPTQ-4bit-128g.no-act-order.safetensors` as mentioned above, which should work without any upgrades to text-generation-webui.
98
 
 
 
 
 
 
 
 
 
 
 
 
99
  # Original model card not provided
100
 
101
  No model card was provided in [changsung's original repository](https://huggingface.co/chansung/alpaca-lora-65b).
102
 
103
- Based on the name, I assume this is the result of fine tuning using the original GPT 3.5 Alpaca dataset. It is unknown as to whether the original Stanford data was used, or the [cleaned tloen/alpaca-lora variant](https://github.com/tloen/alpaca-lora).
 
2
  license: other
3
  inference: false
4
  ---
5
+ <div style="width: 100%;">
6
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
7
+ </div>
8
+ <div style="display: flex; justify-content: space-between; width: 100%;">
9
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
10
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
11
+ </div>
12
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
13
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
14
+ </div>
15
+ </div>
16
 
17
  # Alpaca LoRA 65B GPTQ 4bit
18
 
 
41
  Specifically, the last two files use `--act-order` for maximum quantisation quality and will not work with oobabooga's fork of GPTQ-for-LLaMa. Therefore at this time it will also not work with the CUDA branch of GPTQ-for-LLaMa, or `text-generation-webui` one-click installers.
42
 
43
  Unless you are able to use the latest Triton GPTQ-for-LLaMa code, please use `medalpaca-13B-GPTQ-4bit-128g.no-act-order.safetensors`
44
+
45
  * `alpaca-lora-65B-GPTQ-4bit-128g.no-act-order.safetensors`
46
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
47
  * Works with text-generation-webui one-click-installers
 
107
 
108
  If you can't update GPTQ-for-LLaMa to the latest Triton branch, or don't want to, you can use `alpaca-lora-65B-GPTQ-4bit-128g.no-act-order.safetensors` as mentioned above, which should work without any upgrades to text-generation-webui.
109
 
110
+ ## Want to support my work?
111
+
112
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
113
+
114
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
115
+
116
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
117
+
118
+ * Patreon: coming soon! (just awaiting approval)
119
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
120
+ * Discord: https://discord.gg/UBgz4VXf
121
  # Original model card not provided
122
 
123
  No model card was provided in [changsung's original repository](https://huggingface.co/chansung/alpaca-lora-65b).
124
 
125
+ Based on the name, I assume this is the result of fine tuning using the original GPT 3.5 Alpaca dataset. It is unknown as to whether the original Stanford data was used, or the [cleaned tloen/alpaca-lora variant](https://github.com/tloen/alpaca-lora).