big-thousand commited on
Commit
f54bc5d
1 Parent(s): b1f0a55
.gitignore CHANGED
@@ -6,8 +6,6 @@ extensions/elevenlabs_tts/outputs
6
  extensions/sd_api_pictures/outputs
7
  extensions/multimodal/pipelines
8
  logs
9
- loras
10
- models
11
  repositories
12
  softprompts
13
  torch-dumps
 
6
  extensions/sd_api_pictures/outputs
7
  extensions/multimodal/pipelines
8
  logs
 
 
9
  repositories
10
  softprompts
11
  torch-dumps
loras/place-your-loras-here.txt ADDED
File without changes
models/config.yaml ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ .*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b):
2
+ model_type: 'llama'
3
+ .*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
4
+ model_type: 'opt'
5
+ .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b):
6
+ model_type: 'gptj'
7
+ .*(4bit|int4):
8
+ wbits: 4
9
+ .*(3bit|int3):
10
+ wbits: 3
11
+ .*(-2bit|_2bit|int2-):
12
+ wbits: 2
13
+ .*(-1bit|_1bit|int1-):
14
+ wbits: 1
15
+ .*(8bit|int8):
16
+ wbits: 8
17
+ .*(-7bit|_7bit|int7-):
18
+ wbits: 7
19
+ .*(-6bit|_6bit|int6-):
20
+ wbits: 6
21
+ .*(-5bit|_5bit|int5-):
22
+ wbits: 5
23
+ .*(-gr32-|-32g-|groupsize32|-32g$):
24
+ groupsize: 32
25
+ .*(-gr64-|-64g-|groupsize64|-64g$):
26
+ groupsize: 64
27
+ .*(gr128|128g|groupsize128):
28
+ groupsize: 128
29
+ .*(gr1024|1024g|groupsize1024):
30
+ groupsize: 1024
31
+ .*gptq(?!u|arl|v2):
32
+ wbits: 4
33
+ groupsize: 128
34
+ .*gptqv2:
35
+ groupsize: 'None'
36
+ .*(oasst|stablelm-7b-sft-v7-epoch-3):
37
+ mode: 'instruct'
38
+ instruction_template: 'Open Assistant'
39
+ skip_special_tokens: false
40
+ (?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable).*vicuna:
41
+ mode: 'instruct'
42
+ instruction_template: 'Vicuna-v0'
43
+ .*vicuna.*v0:
44
+ mode: 'instruct'
45
+ instruction_template: 'Vicuna-v0'
46
+ .*vicuna.*(1.1|1_1):
47
+ mode: 'instruct'
48
+ instruction_template: 'Vicuna-v1.1'
49
+ .*wizard.*vicuna:
50
+ mode: 'instruct'
51
+ instruction_template: 'Vicuna-v1.1'
52
+ .*stable.*vicuna:
53
+ mode: 'instruct'
54
+ instruction_template: 'StableVicuna'
55
+ .*alpaca:
56
+ mode: 'instruct'
57
+ instruction_template: 'Alpaca'
58
+ .*alpaca-native-4bit:
59
+ mode: 'instruct'
60
+ instruction_template: 'Alpaca'
61
+ wbits: 4
62
+ groupsize: 128
63
+ .*galactica:
64
+ skip_special_tokens: false
65
+ .*dolly-v[0-9]-[0-9]*b:
66
+ mode: 'instruct'
67
+ instruction_template: 'Alpaca'
68
+ skip_special_tokens: false
69
+ custom_stopping_strings: '"### End"'
70
+ .*koala:
71
+ mode: 'instruct'
72
+ instruction_template: 'Koala'
73
+ .*chatglm:
74
+ mode: 'instruct'
75
+ instruction_template: 'ChatGLM'
76
+ .*metharme:
77
+ mode: 'instruct'
78
+ instruction_template: 'Metharme'
79
+ .*llava:
80
+ mode: 'instruct'
81
+ model_type: 'llama'
82
+ instruction_template: 'LLaVA'
83
+ custom_stopping_strings: '"\n###"'
84
+ .*raven:
85
+ mode: 'instruct'
86
+ instruction_template: 'RWKV-Raven'
87
+ .*moss-moon.*sft:
88
+ mode: 'instruct'
89
+ instruction_template: 'MOSS'
90
+ .*stablelm-tuned:
91
+ mode: 'instruct'
92
+ instruction_template: 'StableLM'
93
+ truncation_length: 4096
94
+ chat_prompt_size: 4096
95
+ chat_prompt_size_max: 4096
96
+ .*stablelm-base:
97
+ truncation_length: 4096
98
+ chat_prompt_size: 4096
99
+ chat_prompt_size_max: 4096
100
+ .*wizardlm:
101
+ mode: 'instruct'
102
+ model_type: 'llama'
103
+ instruction_template: 'WizardLM'
104
+ .*galactica.*finetuned:
105
+ mode: 'instruct'
106
+ instruction_template: 'Galactica Finetuned'
107
+ .*galactica.*-v2:
108
+ mode: 'instruct'
109
+ instruction_template: 'Galactica v2'
110
+ (?!.*finetuned)(?!.*-v2).*galactica:
111
+ mode: 'instruct'
112
+ instruction_template: 'Galactica'
113
+ .*guanaco:
114
+ mode: 'instruct'
115
+ instruction_template: 'Guanaco non-chat'
116
+ .*baize:
117
+ mode: 'instruct'
118
+ instruction_template: 'Baize'
119
+ .*mpt-.*instruct:
120
+ mode: 'instruct'
121
+ instruction_template: 'Alpaca'
122
+ .*mpt-.*chat:
123
+ mode: 'instruct'
124
+ instruction_template: 'MPT-Chat'
125
+ (?!.*-flan-)(?!.*-t5-).*lamini-:
126
+ mode: 'instruct'
127
+ instruction_template: 'Alpaca'
128
+ .*incite.*chat:
129
+ mode: 'instruct'
130
+ instruction_template: 'INCITE-Chat'
131
+ .*incite.*instruct:
132
+ mode: 'instruct'
133
+ instruction_template: 'INCITE-Instruct'
models/next-social_dcard-ggml/ggml-model-q4_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b722860f0572740e94a437b3f596979448be9f9e68d769bf5a8113dc5a50525
3
+ size 4212859520
models/next-social_dcard-ggml/huggingface-metadata.txt ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ url: https://huggingface.co/next-social/dcard-int4-ggml
2
+ branch: main
3
+ download date: 2023-05-15 15:21:54
4
+ sha256sum:
5
+ 5b722860f0572740e94a437b3f596979448be9f9e68d769bf5a8113dc5a50525 ggml-model-q4_0.bin
models/place-your-models-here.txt ADDED
File without changes