Spaces:
Running
Running
Merge branch 'main' into link_requests_and_results
Browse files- app.py +13 -2
- src/assets/text_content.py +6 -5
- src/auto_leaderboard/model_metadata_type.py +414 -396
app.py
CHANGED
@@ -328,7 +328,13 @@ with demo:
|
|
328 |
)
|
329 |
filter_columns = gr.Radio(
|
330 |
label="β Filter model types",
|
331 |
-
choices = [
|
|
|
|
|
|
|
|
|
|
|
|
|
332 |
value="all",
|
333 |
elem_id="filter-columns"
|
334 |
)
|
@@ -404,7 +410,12 @@ with demo:
|
|
404 |
False, label="Private", visible=not IS_PUBLIC
|
405 |
)
|
406 |
model_type = gr.Dropdown(
|
407 |
-
choices=[
|
|
|
|
|
|
|
|
|
|
|
408 |
label="Model type",
|
409 |
multiselect=False,
|
410 |
value=None,
|
|
|
328 |
)
|
329 |
filter_columns = gr.Radio(
|
330 |
label="β Filter model types",
|
331 |
+
choices = [
|
332 |
+
"all",
|
333 |
+
ModelType.PT.to_str(),
|
334 |
+
ModelType.FT.to_str(),
|
335 |
+
ModelType.IFT.to_str(),
|
336 |
+
ModelType.RL.to_str(),
|
337 |
+
],
|
338 |
value="all",
|
339 |
elem_id="filter-columns"
|
340 |
)
|
|
|
410 |
False, label="Private", visible=not IS_PUBLIC
|
411 |
)
|
412 |
model_type = gr.Dropdown(
|
413 |
+
choices=[
|
414 |
+
ModelType.PT.to_str(" : "),
|
415 |
+
ModelType.FT.to_str(" : "),
|
416 |
+
ModelType.IFT.to_str(" : "),
|
417 |
+
ModelType.RL.to_str(" : "),
|
418 |
+
],
|
419 |
label="Model type",
|
420 |
multiselect=False,
|
421 |
value=None,
|
src/assets/text_content.py
CHANGED
@@ -1,3 +1,5 @@
|
|
|
|
|
|
1 |
CHANGELOG_TEXT = f"""
|
2 |
## [2023-06-19]
|
3 |
- Added model type column
|
@@ -62,8 +64,6 @@ INTRODUCTION_TEXT = f"""
|
|
62 |
π€ Anyone from the community can submit a model for automated evaluation on the π€ GPU cluster, as long as it is a π€ Transformers model with weights on the Hub. We also support evaluation of models with delta-weights for non-commercial licensed models, such as the original LLaMa release.
|
63 |
|
64 |
Other cool benchmarks for LLMs are developed at HuggingFace, go check them out: ππ€ [human and GPT4 evals](https://huggingface.co/spaces/HuggingFaceH4/human_eval_llm_leaderboard), π₯οΈ [performance benchmarks](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
65 |
-
|
66 |
-
π’: Base pretrained model β πΆ: Instruction finetuned model β π¦: Model finetuned with RL (read more details in "About" tab)
|
67 |
"""
|
68 |
|
69 |
LLM_BENCHMARKS_TEXT = f"""
|
@@ -131,9 +131,10 @@ To get more information about quantization, see:
|
|
131 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
132 |
|
133 |
### Icons
|
134 |
-
|
135 |
-
|
136 |
-
|
|
|
137 |
If there is no icon, we have not uploaded the information on the model yet, feel free to open an issue with the model information!
|
138 |
|
139 |
|
|
|
1 |
+
from ..auto_leaderboard.model_metadata_type import ModelType
|
2 |
+
|
3 |
CHANGELOG_TEXT = f"""
|
4 |
## [2023-06-19]
|
5 |
- Added model type column
|
|
|
64 |
π€ Anyone from the community can submit a model for automated evaluation on the π€ GPU cluster, as long as it is a π€ Transformers model with weights on the Hub. We also support evaluation of models with delta-weights for non-commercial licensed models, such as the original LLaMa release.
|
65 |
|
66 |
Other cool benchmarks for LLMs are developed at HuggingFace, go check them out: ππ€ [human and GPT4 evals](https://huggingface.co/spaces/HuggingFaceH4/human_eval_llm_leaderboard), π₯οΈ [performance benchmarks](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
|
|
|
|
67 |
"""
|
68 |
|
69 |
LLM_BENCHMARKS_TEXT = f"""
|
|
|
131 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
132 |
|
133 |
### Icons
|
134 |
+
{ModelType.PT.to_str(" : ")} model
|
135 |
+
{ModelType.FT.to_str(" : ")} model
|
136 |
+
{ModelType.IFT.to_str(" : ")} model
|
137 |
+
{ModelType.RL.to_str(" : ")} model
|
138 |
If there is no icon, we have not uploaded the information on the model yet, feel free to open an issue with the model information!
|
139 |
|
140 |
|
src/auto_leaderboard/model_metadata_type.py
CHANGED
@@ -16,129 +16,141 @@ model_type_symbols = {
|
|
16 |
"fine-tuned": "πΆ",
|
17 |
"pretrained": "π’",
|
18 |
"RL-tuned": "π¦",
|
|
|
19 |
}
|
20 |
|
21 |
class ModelType(Enum):
|
22 |
PT = ModelInfo(name="pretrained", symbol="π’")
|
23 |
-
|
|
|
24 |
RL = ModelInfo(name="RL-tuned", symbol="π¦")
|
25 |
|
|
|
|
|
|
|
26 |
|
27 |
TYPE_METADATA: Dict[str, ModelType] = {
|
28 |
-
"notstoic/PygmalionCoT-7b": ModelType.
|
29 |
-
"aisquared/dlite-v1-355m": ModelType.
|
30 |
-
"aisquared/dlite-v1-1_5b": ModelType.
|
31 |
-
"aisquared/dlite-v1-774m": ModelType.
|
32 |
-
"aisquared/dlite-v1-124m": ModelType.
|
33 |
-
"aisquared/chopt-2_7b": ModelType.
|
34 |
-
"aisquared/dlite-v2-124m": ModelType.
|
35 |
-
"aisquared/dlite-v2-774m": ModelType.
|
36 |
-
"aisquared/dlite-v2-1_5b": ModelType.
|
37 |
-
"aisquared/chopt-1_3b": ModelType.
|
38 |
-
"aisquared/dlite-v2-355m": ModelType.
|
39 |
-
"augtoma/qCammel-13": ModelType.
|
40 |
-
"Aspik101/Llama-2-7b-hf-instruct-pl-lora_unload": ModelType.
|
41 |
-
"Aspik101/vicuna-7b-v1.3-instruct-pl-lora_unload": ModelType.
|
42 |
-
"TheBloke/alpaca-lora-65B-HF": ModelType.
|
43 |
-
"TheBloke/tulu-7B-fp16": ModelType.
|
44 |
-
"TheBloke/guanaco-7B-HF": ModelType.
|
45 |
-
"TheBloke/koala-7B-HF": ModelType.
|
46 |
-
"TheBloke/wizardLM-7B-HF": ModelType.
|
47 |
-
"TheBloke/airoboros-13B-HF": ModelType.
|
48 |
-
"TheBloke/koala-13B-HF": ModelType.
|
49 |
-
"TheBloke/Wizard-Vicuna-7B-Uncensored-HF": ModelType.
|
50 |
-
"TheBloke/dromedary-65b-lora-HF": ModelType.
|
51 |
-
"TheBloke/wizardLM-13B-1.0-fp16": ModelType.
|
52 |
-
"TheBloke/
|
53 |
-
"TheBloke/
|
54 |
-
"TheBloke/
|
55 |
-
"TheBloke/
|
56 |
-
"TheBloke/
|
57 |
-
"TheBloke/
|
58 |
-
"TheBloke/guanaco-
|
59 |
-
"TheBloke/
|
|
|
|
|
60 |
"TheBloke/Llama-2-13B-fp16": ModelType.PT,
|
61 |
-
"TheBloke/llama-2-70b-Guanaco-QLoRA-fp16": ModelType.
|
62 |
-
"TheBloke/landmark-attention-llama7b-fp16": ModelType.
|
63 |
-
"TheBloke/Planner-7B-fp16": ModelType.
|
64 |
-
"TheBloke/Wizard-Vicuna-13B-Uncensored-HF": ModelType.
|
65 |
-
"TheBloke/gpt4-alpaca-lora-13B-HF": ModelType.
|
66 |
-
"TheBloke/gpt4-x-vicuna-13B-HF": ModelType.
|
67 |
-
"TheBloke/gpt4-alpaca-lora_mlp-65B-HF": ModelType.
|
68 |
-
"TheBloke/tulu-13B-fp16": ModelType.
|
69 |
-
"TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16": ModelType.
|
70 |
-
"TheBloke/Llama-2-70B-fp16": ModelType.
|
71 |
-
"TheBloke/WizardLM-30B-fp16": ModelType.
|
72 |
-
"TheBloke/robin-13B-v2-fp16": ModelType.
|
73 |
-
"TheBloke/robin-33B-v2-fp16": ModelType.
|
74 |
-
"TheBloke/Vicuna-13B-CoT-fp16": ModelType.
|
75 |
-
"TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16": ModelType.
|
76 |
-
"TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16": ModelType.
|
77 |
-
"
|
78 |
-
"
|
79 |
-
"
|
80 |
-
"
|
81 |
-
"
|
82 |
-
"
|
83 |
-
"
|
84 |
-
"
|
85 |
-
"
|
86 |
-
"
|
87 |
-
"
|
88 |
-
"
|
89 |
-
"digitous/
|
90 |
-
"digitous/
|
91 |
-
"digitous/
|
92 |
-
"digitous/
|
93 |
-
"digitous/
|
94 |
-
"digitous/
|
95 |
-
"digitous/
|
96 |
-
"digitous/
|
|
|
|
|
|
|
|
|
97 |
"SaylorTwift/gpt2_test": ModelType.PT,
|
98 |
-
"anton-l/gpt-j-tiny-random": ModelType.
|
99 |
-
"Andron00e/YetAnother_Open-Llama-3B-LoRA-OpenOrca": ModelType.
|
100 |
-
"Lazycuber/pyg-instruct-wizardlm": ModelType.
|
101 |
-
"Lazycuber/Janemalion-6B": ModelType.
|
102 |
-
"IDEA-CCNL/Ziya-LLaMA-13B-Pretrain-v1": ModelType.
|
103 |
-
"IDEA-CCNL/Ziya-LLaMA-13B-v1": ModelType.
|
104 |
-
"dsvv-cair/alpaca-cleaned-llama-30b-bf16": ModelType.
|
105 |
"gpt2-medium": ModelType.PT,
|
106 |
-
"camel-ai/CAMEL-13B-Combined-Data": ModelType.
|
107 |
-
"camel-ai/CAMEL-13B-Role-Playing-Data": ModelType.
|
108 |
-
"
|
109 |
-
"PygmalionAI/
|
110 |
-
"PygmalionAI/
|
111 |
-
"PygmalionAI/pygmalion-
|
112 |
-
"PygmalionAI/pygmalion-
|
113 |
-
"
|
114 |
-
"
|
115 |
-
"lilloukas/
|
116 |
-
"
|
117 |
-
"
|
118 |
-
"
|
119 |
-
"LoupGarou/WizardCoder-Guanaco-15B-V1.
|
120 |
-
"
|
121 |
-
"hakurei/
|
122 |
-
"
|
123 |
-
"shibing624/chinese-
|
124 |
-
"shibing624/chinese-alpaca-plus-
|
125 |
-
"
|
126 |
-
"mosaicml/mpt-
|
127 |
-
"mosaicml/mpt-
|
128 |
-
"mosaicml/mpt-
|
129 |
-
"mosaicml/mpt-
|
|
|
130 |
"mosaicml/mpt-30b": ModelType.PT,
|
131 |
-
"Corianas/111m": ModelType.
|
132 |
-
"Corianas/Quokka_1.3b": ModelType.
|
133 |
-
"Corianas/256_5epoch": ModelType.
|
134 |
-
"Corianas/Quokka_256m": ModelType.
|
135 |
-
"Corianas/Quokka_590m": ModelType.
|
136 |
-
"Corianas/gpt-j-6B-Dolly": ModelType.
|
137 |
-
"Corianas/Quokka_2.7b": ModelType.
|
138 |
-
"cyberagent/open-calm-7b": ModelType.
|
139 |
-
"Aspik101/Nous-Hermes-13b-pl-lora_unload": ModelType.
|
140 |
-
"THUDM/chatglm2-6b": ModelType.
|
141 |
-
"MetaIX/GPT4-X-Alpasta-30b": ModelType.
|
142 |
"NYTK/PULI-GPTrio": ModelType.PT,
|
143 |
"EleutherAI/pythia-1.3b": ModelType.PT,
|
144 |
"EleutherAI/pythia-2.8b-deduped": ModelType.PT,
|
@@ -165,194 +177,195 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
165 |
"roneneldan/TinyStories-1M": ModelType.PT,
|
166 |
"roneneldan/TinyStories-8M": ModelType.PT,
|
167 |
"roneneldan/TinyStories-3M": ModelType.PT,
|
168 |
-
"jerryjalapeno/nart-100k-7b": ModelType.
|
169 |
-
"lmsys/vicuna-13b-v1.3": ModelType.
|
170 |
-
"lmsys/vicuna-7b-v1.3": ModelType.
|
171 |
-
"lmsys/vicuna-13b-v1.1": ModelType.
|
172 |
-
"lmsys/vicuna-13b-delta-v1.1": ModelType.
|
173 |
-
"lmsys/vicuna-7b-delta-v1.1": ModelType.
|
174 |
-
"abhiramtirumala/DialoGPT-sarcastic-medium": ModelType.
|
175 |
-
"haonan-li/bactrian-x-llama-13b-merged": ModelType.
|
176 |
-
"Gryphe/MythoLogic-13b": ModelType.
|
177 |
-
"Gryphe/MythoBoros-13b": ModelType.
|
178 |
-
"pillowtalks-ai/delta13b": ModelType.
|
179 |
-
"wannaphong/openthaigpt-0.1.0-beta-full-model_for_open_llm_leaderboard": ModelType.
|
|
|
180 |
"bigcode/tiny_starcoder_py": ModelType.PT,
|
181 |
-
"bigcode/starcoderplus": ModelType.
|
182 |
"bigcode/gpt_bigcode-santacoder": ModelType.PT,
|
183 |
"bigcode/starcoder": ModelType.PT,
|
184 |
-
"Open-Orca/OpenOrca-Preview1-13B": ModelType.
|
185 |
-
"microsoft/DialoGPT-large": ModelType.
|
186 |
-
"microsoft/DialoGPT-small": ModelType.
|
187 |
-
"microsoft/DialoGPT-medium": ModelType.
|
188 |
-
"microsoft/CodeGPT-small-py": ModelType.
|
189 |
-
"Tincando/fiction_story_generator": ModelType.
|
190 |
-
"Pirr/pythia-13b-deduped-green_devil": ModelType.
|
191 |
-
"Aeala/GPT4-x-AlpacaDente2-30b": ModelType.
|
192 |
-
"Aeala/GPT4-x-AlpacaDente-30b": ModelType.
|
193 |
-
"Aeala/GPT4-x-Alpasta-13b": ModelType.
|
194 |
-
"Aeala/VicUnlocked-alpaca-30b": ModelType.
|
195 |
-
"Tap-M/Luna-AI-Llama2-Uncensored": ModelType.
|
196 |
-
"illuin/test-custom-llama": ModelType.
|
197 |
-
"dvruette/oasst-llama-13b-2-epochs": ModelType.
|
198 |
-
"dvruette/oasst-gpt-neox-20b-1000-steps": ModelType.
|
199 |
"dvruette/llama-13b-pretrained-dropout": ModelType.PT,
|
200 |
"dvruette/llama-13b-pretrained": ModelType.PT,
|
201 |
"dvruette/llama-13b-pretrained-sft-epoch-1": ModelType.PT,
|
202 |
"dvruette/llama-13b-pretrained-sft-do2": ModelType.PT,
|
203 |
-
"dvruette/oasst-gpt-neox-20b-3000-steps": ModelType.
|
204 |
"dvruette/oasst-pythia-12b-pretrained-sft": ModelType.PT,
|
205 |
-
"dvruette/oasst-pythia-6.9b-4000-steps": ModelType.
|
206 |
-
"dvruette/gpt-neox-20b-full-precision": ModelType.
|
207 |
-
"dvruette/oasst-llama-13b-1000-steps": ModelType.
|
208 |
"openlm-research/open_llama_7b_700bt_preview": ModelType.PT,
|
209 |
"openlm-research/open_llama_7b": ModelType.PT,
|
210 |
"openlm-research/open_llama_7b_v2": ModelType.PT,
|
211 |
"openlm-research/open_llama_3b": ModelType.PT,
|
212 |
"openlm-research/open_llama_13b": ModelType.PT,
|
213 |
"openlm-research/open_llama_3b_v2": ModelType.PT,
|
214 |
-
"PocketDoc/Dans-PileOfSets-Mk1-llama-13b-merged": ModelType.
|
215 |
-
"GeorgiaTechResearchInstitute/galpaca-30b": ModelType.
|
216 |
-
"GeorgiaTechResearchInstitute/starcoder-gpteacher-code-instruct": ModelType.
|
217 |
-
"databricks/dolly-v2-7b": ModelType.
|
218 |
-
"databricks/dolly-v2-3b": ModelType.
|
219 |
-
"databricks/dolly-v2-12b": ModelType.
|
220 |
-
"Rachneet/gpt2-xl-alpaca": ModelType.
|
221 |
-
"Locutusque/gpt2-conversational-or-qa": ModelType.
|
222 |
-
"psyche/kogpt": ModelType.
|
223 |
-
"NbAiLab/nb-gpt-j-6B-alpaca": ModelType.
|
224 |
-
"Mikael110/llama-2-7b-guanaco-fp16": ModelType.
|
225 |
-
"Mikael110/llama-2-13b-guanaco-fp16": ModelType.
|
226 |
-
"Fredithefish/CrimsonPajama": ModelType.
|
227 |
-
"Fredithefish/RedPajama-INCITE-Chat-3B-ShareGPT-11K": ModelType.
|
228 |
-
"Fredithefish/ScarletPajama-3B-HF": ModelType.
|
229 |
-
"Fredithefish/RedPajama-INCITE-Chat-3B-Instruction-Tuning-with-GPT-4": ModelType.
|
230 |
-
"acrastt/RedPajama-INCITE-Chat-Instruct-3B-V1": ModelType.
|
231 |
-
"eachadea/vicuna-13b-1.1": ModelType.
|
232 |
-
"eachadea/vicuna-7b-1.1": ModelType.
|
233 |
-
"eachadea/vicuna-13b": ModelType.
|
234 |
-
"openaccess-ai-collective/wizard-mega-13b": ModelType.
|
235 |
-
"openaccess-ai-collective/manticore-13b": ModelType.
|
236 |
-
"openaccess-ai-collective/manticore-30b-chat-pyg-alpha": ModelType.
|
237 |
-
"openaccess-ai-collective/minotaur-13b": ModelType.
|
238 |
-
"openaccess-ai-collective/minotaur-13b-fixed": ModelType.
|
239 |
-
"openaccess-ai-collective/hippogriff-30b-chat": ModelType.
|
240 |
-
"openaccess-ai-collective/manticore-13b-chat-pyg": ModelType.
|
241 |
-
"pythainlp/wangchanglm-7.5B-sft-enth": ModelType.
|
242 |
-
"pythainlp/wangchanglm-7.5B-sft-en-sharded": ModelType.
|
243 |
-
"euclaise/gpt-neox-122m-minipile-digits": ModelType.
|
244 |
-
"stabilityai/StableBeluga1-Delta": ModelType.
|
245 |
-
"stabilityai/stablelm-tuned-alpha-7b": ModelType.
|
246 |
-
"stabilityai/StableBeluga2": ModelType.
|
247 |
-
"stabilityai/StableBeluga-13B": ModelType.
|
248 |
-
"stabilityai/StableBeluga-7B": ModelType.
|
249 |
"stabilityai/stablelm-base-alpha-7b": ModelType.PT,
|
250 |
"stabilityai/stablelm-base-alpha-3b": ModelType.PT,
|
251 |
-
"stabilityai/stablelm-tuned-alpha-3b": ModelType.
|
252 |
-
"alibidaran/medical_transcription_generator": ModelType.
|
253 |
-
"CalderaAI/30B-Lazarus": ModelType.
|
254 |
-
"CalderaAI/13B-BlueMethod": ModelType.
|
255 |
-
"CalderaAI/13B-Ouroboros": ModelType.
|
256 |
-
"KoboldAI/OPT-13B-Erebus": ModelType.
|
257 |
-
"KoboldAI/GPT-J-6B-Janeway": ModelType.
|
258 |
-
"KoboldAI/GPT-J-6B-Shinen": ModelType.
|
259 |
"KoboldAI/fairseq-dense-2.7B": ModelType.PT,
|
260 |
-
"KoboldAI/OPT-6B-nerys-v2": ModelType.
|
261 |
-
"KoboldAI/GPT-NeoX-20B-Skein": ModelType.
|
262 |
-
"KoboldAI/PPO_Pygway-6b-Mix": ModelType.
|
263 |
"KoboldAI/fairseq-dense-6.7B": ModelType.PT,
|
264 |
"KoboldAI/fairseq-dense-125M": ModelType.PT,
|
265 |
-
"KoboldAI/OPT-13B-Nerybus-Mix": ModelType.
|
266 |
-
"KoboldAI/OPT-2.7B-Erebus": ModelType.
|
267 |
-
"KoboldAI/OPT-350M-Nerys-v2": ModelType.
|
268 |
-
"KoboldAI/OPT-2.7B-Nerys-v2": ModelType.
|
269 |
-
"KoboldAI/OPT-2.7B-Nerybus-Mix": ModelType.
|
270 |
-
"KoboldAI/OPT-13B-Nerys-v2": ModelType.
|
271 |
-
"KoboldAI/GPT-NeoX-20B-Erebus": ModelType.
|
272 |
-
"KoboldAI/OPT-6.7B-Erebus": ModelType.
|
273 |
"KoboldAI/fairseq-dense-355M": ModelType.PT,
|
274 |
-
"KoboldAI/OPT-6.7B-Nerybus-Mix": ModelType.
|
275 |
-
"KoboldAI/GPT-J-6B-Adventure": ModelType.
|
276 |
-
"KoboldAI/OPT-350M-Erebus": ModelType.
|
277 |
-
"KoboldAI/GPT-J-6B-Skein": ModelType.
|
278 |
-
"KoboldAI/OPT-30B-Erebus": ModelType.
|
279 |
"klosax/pythia-160m-deduped-step92k-193bt": ModelType.PT,
|
280 |
"klosax/open_llama_3b_350bt_preview": ModelType.PT,
|
281 |
"klosax/openllama-3b-350bt": ModelType.PT,
|
282 |
"klosax/pythia-70m-deduped-step44k-92bt": ModelType.PT,
|
283 |
"klosax/open_llama_13b_600bt_preview": ModelType.PT,
|
284 |
"klosax/open_llama_7b_400bt_preview": ModelType.PT,
|
285 |
-
"kfkas/Llama-2-ko-7b-Chat": ModelType.
|
286 |
-
"WeOpenML/Alpaca-7B-v1": ModelType.
|
287 |
-
"WeOpenML/PandaLM-Alpaca-7B-v1": ModelType.
|
288 |
-
"TFLai/gpt2-turkish-uncased": ModelType.
|
289 |
-
"ehartford/WizardLM-13B-Uncensored": ModelType.
|
290 |
-
"ehartford/dolphin-llama-13b": ModelType.
|
291 |
-
"ehartford/Wizard-Vicuna-30B-Uncensored": ModelType.
|
292 |
-
"ehartford/WizardLM-30B-Uncensored": ModelType.
|
293 |
-
"ehartford/Wizard-Vicuna-13B-Uncensored": ModelType.
|
294 |
-
"ehartford/WizardLM-7B-Uncensored": ModelType.
|
295 |
-
"ehartford/based-30b": ModelType.
|
296 |
-
"ehartford/Wizard-Vicuna-7B-Uncensored": ModelType.
|
297 |
-
"wahaha1987/llama_7b_sharegpt94k_fastchat": ModelType.
|
298 |
-
"wahaha1987/llama_13b_sharegpt94k_fastchat": ModelType.
|
299 |
-
"OpenAssistant/oasst-sft-1-pythia-12b": ModelType.
|
300 |
-
"OpenAssistant/stablelm-7b-sft-v7-epoch-3": ModelType.
|
301 |
-
"OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5": ModelType.
|
302 |
-
"OpenAssistant/pythia-12b-sft-v8-2.5k-steps": ModelType.
|
303 |
-
"OpenAssistant/pythia-12b-sft-v8-7k-steps": ModelType.
|
304 |
-
"OpenAssistant/pythia-12b-pre-v8-12.5k-steps": ModelType.
|
305 |
-
"OpenAssistant/llama2-13b-orca-8k-3319": ModelType.
|
306 |
-
"junelee/wizard-vicuna-13b": ModelType.
|
307 |
"BreadAi/gpt-YA-1-1_160M": ModelType.PT,
|
308 |
"BreadAi/MuseCan": ModelType.PT,
|
309 |
"BreadAi/MusePy-1-2": ModelType.PT,
|
310 |
"BreadAi/DiscordPy": ModelType.PT,
|
311 |
"BreadAi/PM_modelV2": ModelType.PT,
|
312 |
"BreadAi/gpt-Youtube": ModelType.PT,
|
313 |
-
"BreadAi/StoryPy": ModelType.
|
314 |
-
"julianweng/Llama-2-7b-chat-orcah": ModelType.
|
315 |
-
"AGI-inc/lora_moe_7b_baseline": ModelType.
|
316 |
-
"AGI-inc/lora_moe_7b": ModelType.
|
317 |
-
"togethercomputer/GPT-NeoXT-Chat-Base-20B": ModelType.
|
318 |
-
"togethercomputer/RedPajama-INCITE-Chat-7B-v0.1": ModelType.
|
319 |
-
"togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1": ModelType.
|
320 |
"togethercomputer/RedPajama-INCITE-7B-Base": ModelType.PT,
|
321 |
-
"togethercomputer/RedPajama-INCITE-7B-Instruct": ModelType.
|
322 |
"togethercomputer/RedPajama-INCITE-Base-3B-v1": ModelType.PT,
|
323 |
-
"togethercomputer/Pythia-Chat-Base-7B": ModelType.
|
324 |
"togethercomputer/RedPajama-INCITE-Base-7B-v0.1": ModelType.PT,
|
325 |
-
"togethercomputer/GPT-JT-6B-v1": ModelType.
|
326 |
-
"togethercomputer/GPT-JT-6B-v0": ModelType.
|
327 |
-
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": ModelType.
|
328 |
-
"togethercomputer/RedPajama-INCITE-7B-Chat": ModelType.
|
329 |
-
"togethercomputer/RedPajama-INCITE-Instruct-3B-v1": ModelType.
|
330 |
-
"Writer/camel-5b-hf": ModelType.
|
331 |
"Writer/palmyra-base": ModelType.PT,
|
332 |
-
"MBZUAI/LaMini-GPT-1.5B": ModelType.
|
333 |
-
"MBZUAI/lamini-cerebras-111m": ModelType.
|
334 |
-
"MBZUAI/lamini-neo-1.3b": ModelType.
|
335 |
-
"MBZUAI/lamini-cerebras-1.3b": ModelType.
|
336 |
-
"MBZUAI/lamini-cerebras-256m": ModelType.
|
337 |
-
"MBZUAI/LaMini-GPT-124M": ModelType.
|
338 |
-
"MBZUAI/lamini-neo-125m": ModelType.
|
339 |
-
"TehVenom/DiffMerge-DollyGPT-Pygmalion": ModelType.
|
340 |
-
"TehVenom/PPO_Shygmalion-6b": ModelType.
|
341 |
-
"TehVenom/Dolly_Shygmalion-6b-Dev_V8P2": ModelType.
|
342 |
-
"TehVenom/Pygmalion_AlpacaLora-7b": ModelType.
|
343 |
-
"TehVenom/PPO_Pygway-V8p4_Dev-6b": ModelType.
|
344 |
-
"TehVenom/Dolly_Malion-6b": ModelType.
|
345 |
-
"TehVenom/PPO_Shygmalion-V8p4_Dev-6b": ModelType.
|
346 |
-
"TehVenom/ChanMalion": ModelType.
|
347 |
-
"TehVenom/GPT-J-Pyg_PPO-6B": ModelType.
|
348 |
-
"TehVenom/Pygmalion-13b-Merged": ModelType.
|
349 |
-
"TehVenom/Metharme-13b-Merged": ModelType.
|
350 |
-
"TehVenom/Dolly_Shygmalion-6b": ModelType.
|
351 |
-
"TehVenom/GPT-J-Pyg_PPO-6B-Dev-V8p4": ModelType.
|
352 |
-
"georgesung/llama2_7b_chat_uncensored": ModelType.
|
353 |
-
"vicgalle/gpt2-alpaca": ModelType.
|
354 |
-
"vicgalle/alpaca-7b": ModelType.
|
355 |
-
"vicgalle/gpt2-alpaca-gpt4": ModelType.
|
356 |
"facebook/opt-350m": ModelType.PT,
|
357 |
"facebook/opt-125m": ModelType.PT,
|
358 |
"facebook/xglm-4.5B": ModelType.PT,
|
@@ -364,81 +377,81 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
364 |
"facebook/xglm-7.5B": ModelType.PT,
|
365 |
"facebook/xglm-564M": ModelType.PT,
|
366 |
"facebook/opt-30b": ModelType.PT,
|
367 |
-
"golaxy/gogpt-7b": ModelType.
|
368 |
-
"golaxy/gogpt2-7b": ModelType.
|
369 |
-
"golaxy/gogpt-7b-bloom": ModelType.
|
370 |
-
"golaxy/gogpt-3b-bloom": ModelType.
|
371 |
-
"psmathur/orca_mini_v2_7b": ModelType.
|
372 |
-
"psmathur/orca_mini_7b": ModelType.
|
373 |
-
"psmathur/orca_mini_3b": ModelType.
|
374 |
-
"psmathur/orca_mini_v2_13b": ModelType.
|
375 |
"gpt2-xl": ModelType.PT,
|
376 |
-
"lxe/Cerebras-GPT-2.7B-Alpaca-SP": ModelType.
|
377 |
-
"Monero/Manticore-13b-Chat-Pyg-Guanaco": ModelType.
|
378 |
-
"Monero/WizardLM-Uncensored-SuperCOT-StoryTelling-30b": ModelType.
|
379 |
-
"Monero/WizardLM-13b-OpenAssistant-Uncensored": ModelType.
|
380 |
-
"Monero/WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b": ModelType.
|
381 |
-
"jzjiao/opt-1.3b-rlhf": ModelType.
|
382 |
-
"HuggingFaceH4/starchat-beta": ModelType.
|
383 |
-
"KnutJaegersberg/gpt-2-xl-EvolInstruct": ModelType.
|
384 |
-
"KnutJaegersberg/megatron-GPT-2-345m-EvolInstruct": ModelType.
|
385 |
-
"KnutJaegersberg/galactica-orca-wizardlm-1.3b": ModelType.
|
386 |
-
"openchat/openchat_8192": ModelType.
|
387 |
-
"openchat/openchat_v2": ModelType.
|
388 |
-
"openchat/openchat_v2_w": ModelType.
|
389 |
-
"ausboss/llama-13b-supercot": ModelType.
|
390 |
-
"ausboss/llama-30b-supercot": ModelType.
|
391 |
-
"Neko-Institute-of-Science/metharme-7b": ModelType.
|
392 |
-
"Neko-Institute-of-Science/pygmalion-7b": ModelType.
|
393 |
-
"SebastianSchramm/Cerebras-GPT-111M-instruction": ModelType.
|
394 |
-
"victor123/WizardLM-13B-1.0": ModelType.
|
395 |
-
"OpenBuddy/openbuddy-openllama-13b-v7-fp16": ModelType.
|
396 |
-
"OpenBuddy/openbuddy-llama2-13b-v8.1-fp16": ModelType.
|
397 |
-
"OpenBuddyEA/openbuddy-llama-30b-v7.1-bf16": ModelType.
|
398 |
"baichuan-inc/Baichuan-7B": ModelType.PT,
|
399 |
-
"tiiuae/falcon-40b-instruct": ModelType.
|
400 |
"tiiuae/falcon-40b": ModelType.PT,
|
401 |
"tiiuae/falcon-7b": ModelType.PT,
|
402 |
-
"YeungNLP/firefly-llama-13b": ModelType.
|
403 |
-
"YeungNLP/firefly-llama-13b-v1.2": ModelType.
|
404 |
-
"YeungNLP/firefly-llama2-13b": ModelType.
|
405 |
-
"YeungNLP/firefly-ziya-13b": ModelType.
|
406 |
-
"shaohang/Sparse0.5_OPT-1.3": ModelType.
|
407 |
-
"xzuyn/Alpacino-SuperCOT-13B": ModelType.
|
408 |
-
"xzuyn/MedicWizard-7B": ModelType.
|
409 |
-
"xDAN-AI/xDAN_13b_l2_lora": ModelType.
|
410 |
-
"beomi/KoAlpaca-Polyglot-5.8B": ModelType.
|
411 |
-
"beomi/llama-2-ko-7b": ModelType.
|
412 |
"Salesforce/codegen-6B-multi": ModelType.PT,
|
413 |
"Salesforce/codegen-16B-nl": ModelType.PT,
|
414 |
"Salesforce/codegen-6B-nl": ModelType.PT,
|
415 |
-
"ai-forever/rugpt3large_based_on_gpt2": ModelType.
|
416 |
"gpt2-large": ModelType.PT,
|
417 |
-
"frank098/orca_mini_3b_juniper": ModelType.
|
418 |
-
"frank098/WizardLM_13B_juniper": ModelType.
|
419 |
-
"FPHam/Free_Sydney_13b_HF": ModelType.
|
420 |
"huggingface/llama-13b": ModelType.PT,
|
421 |
"huggingface/llama-7b": ModelType.PT,
|
422 |
"huggingface/llama-65b": ModelType.PT,
|
423 |
"huggingface/llama-65b": ModelType.PT,
|
424 |
"huggingface/llama-30b": ModelType.PT,
|
425 |
-
"Henk717/chronoboros-33B": ModelType.
|
426 |
-
"jondurbin/airoboros-13b-gpt4-1.4": ModelType.
|
427 |
-
"jondurbin/airoboros-7b": ModelType.
|
428 |
-
"jondurbin/airoboros-7b-gpt4": ModelType.
|
429 |
-
"jondurbin/airoboros-7b-gpt4-1.1": ModelType.
|
430 |
-
"jondurbin/airoboros-7b-gpt4-1.2": ModelType.
|
431 |
-
"jondurbin/airoboros-7b-gpt4-1.3": ModelType.
|
432 |
-
"jondurbin/airoboros-7b-gpt4-1.4": ModelType.
|
433 |
-
"jondurbin/airoboros-l2-7b-gpt4-1.4.1": ModelType.
|
434 |
-
"jondurbin/airoboros-l2-13b-gpt4-1.4.1": ModelType.
|
435 |
-
"jondurbin/airoboros-l2-70b-gpt4-1.4.1": ModelType.
|
436 |
-
"jondurbin/airoboros-13b": ModelType.
|
437 |
-
"jondurbin/airoboros-33b-gpt4-1.4": ModelType.
|
438 |
-
"jondurbin/airoboros-33b-gpt4-1.2": ModelType.
|
439 |
-
"jondurbin/airoboros-65b-gpt4-1.2": ModelType.
|
440 |
-
"ariellee/SuperPlatty-30B": ModelType.
|
441 |
-
"danielhanchen/open_llama_3b_600bt_preview": ModelType.
|
442 |
"cerebras/Cerebras-GPT-256M": ModelType.PT,
|
443 |
"cerebras/Cerebras-GPT-1.3B": ModelType.PT,
|
444 |
"cerebras/Cerebras-GPT-13B": ModelType.PT,
|
@@ -446,86 +459,92 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
446 |
"cerebras/Cerebras-GPT-111M": ModelType.PT,
|
447 |
"cerebras/Cerebras-GPT-6.7B": ModelType.PT,
|
448 |
"Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-hf": ModelType.RL,
|
449 |
-
"Yhyu13/llama-30B-hf-openassitant": ModelType.
|
450 |
-
"NousResearch/Nous-Hermes-Llama2-13b": ModelType.
|
451 |
-
"NousResearch/Nous-Hermes-llama-2-7b": ModelType.
|
452 |
-
"NousResearch/Redmond-Puffin-13B": ModelType.
|
453 |
-
"NousResearch/Nous-Hermes-13b": ModelType.
|
454 |
-
"project-baize/baize-v2-7b": ModelType.
|
455 |
-
"project-baize/baize-v2-13b": ModelType.
|
456 |
-
"LLMs/WizardLM-13B-V1.0": ModelType.
|
457 |
-
"LLMs/AlpacaGPT4-7B-elina": ModelType.
|
458 |
-
"wenge-research/yayi-7b
|
459 |
-
"wenge-research/yayi-
|
460 |
-
"
|
461 |
-
"
|
462 |
-
"
|
463 |
-
"huggingtweets/
|
464 |
-
"huggingtweets/
|
|
|
465 |
"huggyllama/llama-13b": ModelType.PT,
|
466 |
"huggyllama/llama-65b": ModelType.PT,
|
467 |
"FabbriSimo01/Facebook_opt_1.3b_Quantized": ModelType.PT,
|
468 |
-
"upstage/Llama-2-70b-instruct": ModelType.
|
469 |
-
"upstage/Llama-2-70b-instruct-1024": ModelType.
|
470 |
-
"upstage/llama-65b-instruct": ModelType.
|
471 |
-
"upstage/llama-30b-instruct-2048": ModelType.
|
472 |
-
"upstage/llama-30b-instruct": ModelType.
|
473 |
-
"WizardLM/WizardLM-13B-1.0": ModelType.
|
474 |
-
"WizardLM/WizardLM-13B-V1.1": ModelType.
|
475 |
-
"WizardLM/WizardLM-13B-V1.2": ModelType.
|
476 |
-
"WizardLM/WizardLM-30B-V1.0": ModelType.
|
477 |
-
"WizardLM/WizardCoder-15B-V1.0": ModelType.
|
478 |
"gpt2": ModelType.PT,
|
479 |
-
"keyfan/vicuna-chinese-replication-v1.1": ModelType.
|
480 |
-
"nthngdy/pythia-owt2-70m-100k": ModelType.
|
481 |
-
"nthngdy/pythia-owt2-70m-50k": ModelType.
|
482 |
-
"quantumaikr/KoreanLM-hf": ModelType.
|
483 |
-
"quantumaikr/open_llama_7b_hf": ModelType.
|
484 |
-
"
|
485 |
-
"MayaPH/FinOPT-
|
486 |
-
"MayaPH/
|
487 |
-
"MayaPH/GodziLLa-30B
|
488 |
-
"MayaPH/
|
489 |
-
"
|
490 |
-
"
|
491 |
-
"
|
492 |
-
"CobraMamba/mamba-gpt-3b
|
493 |
-
"
|
494 |
-
"
|
495 |
-
"
|
496 |
-
"
|
497 |
-
"heegyu/
|
|
|
|
|
498 |
"meta-llama/Llama-2-7b-chat-hf": ModelType.RL,
|
499 |
"meta-llama/Llama-2-7b-hf": ModelType.PT,
|
500 |
"meta-llama/Llama-2-13b-chat-hf": ModelType.RL,
|
501 |
"meta-llama/Llama-2-13b-hf": ModelType.PT,
|
502 |
"meta-llama/Llama-2-70b-chat-hf": ModelType.RL,
|
503 |
"meta-llama/Llama-2-70b-hf": ModelType.PT,
|
504 |
-
"xhyi/PT_GPTNEO350_ATG": ModelType.
|
505 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-20b": ModelType.
|
506 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-open-llama-7b-preview-400bt": ModelType.
|
507 |
-
"h2oai/h2ogpt-oig-oasst1-512-6_9b": ModelType.
|
508 |
-
"h2oai/h2ogpt-oasst1-512-12b": ModelType.
|
509 |
-
"h2oai/h2ogpt-oig-oasst1-256-6_9b": ModelType.
|
510 |
-
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt": ModelType.
|
511 |
-
"h2oai/h2ogpt-oasst1-512-20b": ModelType.
|
512 |
-
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2": ModelType.
|
513 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-12b": ModelType.
|
514 |
-
"h2oai/h2ogpt-gm-oasst1-multilang-1024-20b": ModelType.
|
515 |
-
"bofenghuang/vigogne-13b-instruct": ModelType.
|
516 |
-
"bofenghuang/vigogne-13b-chat": ModelType.
|
517 |
-
"bofenghuang/vigogne-2-7b-instruct": ModelType.
|
518 |
-
"bofenghuang/vigogne-7b-instruct": ModelType.
|
519 |
-
"bofenghuang/vigogne-7b-chat": ModelType.
|
520 |
-
"Vmware/open-llama-7b-v2-open-instruct": ModelType.
|
521 |
-
"VMware/open-llama-0.7T-7B-open-instruct-v1.1": ModelType.
|
522 |
-
"ewof/koishi-instruct-3b": ModelType.
|
523 |
-
"gywy/llama2-13b-chinese-v1": ModelType.
|
524 |
-
"GOAT-AI/GOAT-7B-Community": ModelType.
|
525 |
-
"psyche/kollama2-7b": ModelType.
|
526 |
-
"TheTravellingEngineer/llama2-7b-hf-guanaco": ModelType.
|
527 |
-
"beaugogh/pythia-1.4b-deduped-sharegpt": ModelType.
|
528 |
-
"
|
|
|
|
|
|
|
529 |
}
|
530 |
|
531 |
|
@@ -559,4 +578,3 @@ def get_model_type(leaderboard_data: List[dict]):
|
|
559 |
except Exception:
|
560 |
model_data[AutoEvalColumn.model_type.name] = "Unknown"
|
561 |
model_data[AutoEvalColumn.model_type_symbol.name] = "Add type to request files"
|
562 |
-
|
|
|
16 |
"fine-tuned": "πΆ",
|
17 |
"pretrained": "π’",
|
18 |
"RL-tuned": "π¦",
|
19 |
+
"instruction-tuned": "β",
|
20 |
}
|
21 |
|
22 |
class ModelType(Enum):
|
23 |
PT = ModelInfo(name="pretrained", symbol="π’")
|
24 |
+
FT = ModelInfo(name="fine-tuned", symbol="πΆ")
|
25 |
+
IFT = ModelInfo(name="instruction-tuned", symbol="β")
|
26 |
RL = ModelInfo(name="RL-tuned", symbol="π¦")
|
27 |
|
28 |
+
def to_str(self, separator = " "):
|
29 |
+
return f"{self.value.symbol}{separator}{self.value.name}"
|
30 |
+
|
31 |
|
32 |
TYPE_METADATA: Dict[str, ModelType] = {
|
33 |
+
"notstoic/PygmalionCoT-7b": ModelType.FT,
|
34 |
+
"aisquared/dlite-v1-355m": ModelType.FT,
|
35 |
+
"aisquared/dlite-v1-1_5b": ModelType.FT,
|
36 |
+
"aisquared/dlite-v1-774m": ModelType.FT,
|
37 |
+
"aisquared/dlite-v1-124m": ModelType.FT,
|
38 |
+
"aisquared/chopt-2_7b": ModelType.FT,
|
39 |
+
"aisquared/dlite-v2-124m": ModelType.FT,
|
40 |
+
"aisquared/dlite-v2-774m": ModelType.FT,
|
41 |
+
"aisquared/dlite-v2-1_5b": ModelType.FT,
|
42 |
+
"aisquared/chopt-1_3b": ModelType.FT,
|
43 |
+
"aisquared/dlite-v2-355m": ModelType.FT,
|
44 |
+
"augtoma/qCammel-13": ModelType.FT,
|
45 |
+
"Aspik101/Llama-2-7b-hf-instruct-pl-lora_unload": ModelType.FT,
|
46 |
+
"Aspik101/vicuna-7b-v1.3-instruct-pl-lora_unload": ModelType.FT,
|
47 |
+
"TheBloke/alpaca-lora-65B-HF": ModelType.FT,
|
48 |
+
"TheBloke/tulu-7B-fp16": ModelType.FT,
|
49 |
+
"TheBloke/guanaco-7B-HF": ModelType.FT,
|
50 |
+
"TheBloke/koala-7B-HF": ModelType.FT,
|
51 |
+
"TheBloke/wizardLM-7B-HF": ModelType.FT,
|
52 |
+
"TheBloke/airoboros-13B-HF": ModelType.FT,
|
53 |
+
"TheBloke/koala-13B-HF": ModelType.FT,
|
54 |
+
"TheBloke/Wizard-Vicuna-7B-Uncensored-HF": ModelType.FT,
|
55 |
+
"TheBloke/dromedary-65b-lora-HF": ModelType.FT,
|
56 |
+
"TheBloke/wizardLM-13B-1.0-fp16": ModelType.FT,
|
57 |
+
"TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-fp16": ModelType.FT,
|
58 |
+
"TheBloke/Wizard-Vicuna-30B-Uncensored-fp16": ModelType.FT,
|
59 |
+
"TheBloke/wizard-vicuna-13B-HF": ModelType.FT,
|
60 |
+
"TheBloke/UltraLM-13B-fp16": ModelType.FT,
|
61 |
+
"TheBloke/OpenAssistant-FT-7-Llama-30B-HF": ModelType.FT,
|
62 |
+
"TheBloke/vicuna-13B-1.1-HF": ModelType.FT,
|
63 |
+
"TheBloke/guanaco-13B-HF": ModelType.FT,
|
64 |
+
"TheBloke/guanaco-65B-HF": ModelType.FT,
|
65 |
+
"TheBloke/airoboros-7b-gpt4-fp16": ModelType.FT,
|
66 |
+
"TheBloke/llama-30b-supercot-SuperHOT-8K-fp16": ModelType.FT,
|
67 |
"TheBloke/Llama-2-13B-fp16": ModelType.PT,
|
68 |
+
"TheBloke/llama-2-70b-Guanaco-QLoRA-fp16": ModelType.FT,
|
69 |
+
"TheBloke/landmark-attention-llama7b-fp16": ModelType.FT,
|
70 |
+
"TheBloke/Planner-7B-fp16": ModelType.FT,
|
71 |
+
"TheBloke/Wizard-Vicuna-13B-Uncensored-HF": ModelType.FT,
|
72 |
+
"TheBloke/gpt4-alpaca-lora-13B-HF": ModelType.FT,
|
73 |
+
"TheBloke/gpt4-x-vicuna-13B-HF": ModelType.FT,
|
74 |
+
"TheBloke/gpt4-alpaca-lora_mlp-65B-HF": ModelType.FT,
|
75 |
+
"TheBloke/tulu-13B-fp16": ModelType.FT,
|
76 |
+
"TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16": ModelType.FT,
|
77 |
+
"TheBloke/Llama-2-70B-fp16": ModelType.FT,
|
78 |
+
"TheBloke/WizardLM-30B-fp16": ModelType.FT,
|
79 |
+
"TheBloke/robin-13B-v2-fp16": ModelType.FT,
|
80 |
+
"TheBloke/robin-33B-v2-fp16": ModelType.FT,
|
81 |
+
"TheBloke/Vicuna-13B-CoT-fp16": ModelType.FT,
|
82 |
+
"TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16": ModelType.FT,
|
83 |
+
"TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16": ModelType.FT,
|
84 |
+
"TheBloke/Nous-Hermes-13B-SuperHOT-8K-fp16": ModelType.FT,
|
85 |
+
"TheBloke/GPlatty-30B-SuperHOT-8K-fp16": ModelType.FT,
|
86 |
+
"TheBloke/CAMEL-33B-Combined-Data-SuperHOT-8K-fp16": ModelType.FT,
|
87 |
+
"TheBloke/Chinese-Alpaca-33B-SuperHOT-8K-fp16": ModelType.FT,
|
88 |
+
"jphme/orca_mini_v2_ger_7b": ModelType.FT,
|
89 |
+
"Ejafa/vicuna_7B_vanilla_1.1": ModelType.FT,
|
90 |
+
"kevinpro/Vicuna-13B-CoT": ModelType.FT,
|
91 |
+
"AlekseyKorshuk/pygmalion-6b-vicuna-chatml": ModelType.FT,
|
92 |
+
"AlekseyKorshuk/chatml-pyg-v1": ModelType.FT,
|
93 |
+
"concedo/Vicuzard-30B-Uncensored": ModelType.FT,
|
94 |
+
"concedo/OPT-19M-ChatSalad": ModelType.FT,
|
95 |
+
"concedo/Pythia-70M-ChatSalad": ModelType.FT,
|
96 |
+
"digitous/13B-HyperMantis": ModelType.FT,
|
97 |
+
"digitous/Adventien-GPTJ": ModelType.FT,
|
98 |
+
"digitous/Alpacino13b": ModelType.FT,
|
99 |
+
"digitous/GPT-R": ModelType.FT,
|
100 |
+
"digitous/Javelin-R": ModelType.FT,
|
101 |
+
"digitous/Javalion-GPTJ": ModelType.FT,
|
102 |
+
"digitous/Javalion-R": ModelType.FT,
|
103 |
+
"digitous/Skegma-GPTJ": ModelType.FT,
|
104 |
+
"digitous/Alpacino30b": ModelType.FT,
|
105 |
+
"digitous/Janin-GPTJ": ModelType.FT,
|
106 |
+
"digitous/Janin-R": ModelType.FT,
|
107 |
+
"digitous/Javelin-GPTJ": ModelType.FT,
|
108 |
"SaylorTwift/gpt2_test": ModelType.PT,
|
109 |
+
"anton-l/gpt-j-tiny-random": ModelType.FT,
|
110 |
+
"Andron00e/YetAnother_Open-Llama-3B-LoRA-OpenOrca": ModelType.FT,
|
111 |
+
"Lazycuber/pyg-instruct-wizardlm": ModelType.FT,
|
112 |
+
"Lazycuber/Janemalion-6B": ModelType.FT,
|
113 |
+
"IDEA-CCNL/Ziya-LLaMA-13B-Pretrain-v1": ModelType.FT,
|
114 |
+
"IDEA-CCNL/Ziya-LLaMA-13B-v1": ModelType.FT,
|
115 |
+
"dsvv-cair/alpaca-cleaned-llama-30b-bf16": ModelType.FT,
|
116 |
"gpt2-medium": ModelType.PT,
|
117 |
+
"camel-ai/CAMEL-13B-Combined-Data": ModelType.FT,
|
118 |
+
"camel-ai/CAMEL-13B-Role-Playing-Data": ModelType.FT,
|
119 |
+
"camel-ai/CAMEL-33B-Combined-Data": ModelType.FT,
|
120 |
+
"PygmalionAI/pygmalion-6b": ModelType.FT,
|
121 |
+
"PygmalionAI/metharme-1.3b": ModelType.FT,
|
122 |
+
"PygmalionAI/pygmalion-1.3b": ModelType.FT,
|
123 |
+
"PygmalionAI/pygmalion-350m": ModelType.FT,
|
124 |
+
"PygmalionAI/pygmalion-2.7b": ModelType.FT,
|
125 |
+
"medalpaca/medalpaca-7b": ModelType.FT,
|
126 |
+
"lilloukas/Platypus-30B": ModelType.FT,
|
127 |
+
"lilloukas/GPlatty-30B": ModelType.FT,
|
128 |
+
"mncai/chatdoctor": ModelType.FT,
|
129 |
+
"chaoyi-wu/MedLLaMA_13B": ModelType.FT,
|
130 |
+
"LoupGarou/WizardCoder-Guanaco-15B-V1.0": ModelType.FT,
|
131 |
+
"LoupGarou/WizardCoder-Guanaco-15B-V1.1": ModelType.FT,
|
132 |
+
"hakurei/instruct-12b": ModelType.FT,
|
133 |
+
"hakurei/lotus-12B": ModelType.FT,
|
134 |
+
"shibing624/chinese-llama-plus-13b-hf": ModelType.FT,
|
135 |
+
"shibing624/chinese-alpaca-plus-7b-hf": ModelType.FT,
|
136 |
+
"shibing624/chinese-alpaca-plus-13b-hf": ModelType.FT,
|
137 |
+
"mosaicml/mpt-7b-instruct": ModelType.FT,
|
138 |
+
"mosaicml/mpt-30b-chat": ModelType.FT,
|
139 |
+
"mosaicml/mpt-7b-storywriter": ModelType.FT,
|
140 |
+
"mosaicml/mpt-30b-instruct": ModelType.FT,
|
141 |
+
"mosaicml/mpt-7b-chat": ModelType.FT,
|
142 |
"mosaicml/mpt-30b": ModelType.PT,
|
143 |
+
"Corianas/111m": ModelType.FT,
|
144 |
+
"Corianas/Quokka_1.3b": ModelType.FT,
|
145 |
+
"Corianas/256_5epoch": ModelType.FT,
|
146 |
+
"Corianas/Quokka_256m": ModelType.FT,
|
147 |
+
"Corianas/Quokka_590m": ModelType.FT,
|
148 |
+
"Corianas/gpt-j-6B-Dolly": ModelType.FT,
|
149 |
+
"Corianas/Quokka_2.7b": ModelType.FT,
|
150 |
+
"cyberagent/open-calm-7b": ModelType.FT,
|
151 |
+
"Aspik101/Nous-Hermes-13b-pl-lora_unload": ModelType.FT,
|
152 |
+
"THUDM/chatglm2-6b": ModelType.FT,
|
153 |
+
"MetaIX/GPT4-X-Alpasta-30b": ModelType.FT,
|
154 |
"NYTK/PULI-GPTrio": ModelType.PT,
|
155 |
"EleutherAI/pythia-1.3b": ModelType.PT,
|
156 |
"EleutherAI/pythia-2.8b-deduped": ModelType.PT,
|
|
|
177 |
"roneneldan/TinyStories-1M": ModelType.PT,
|
178 |
"roneneldan/TinyStories-8M": ModelType.PT,
|
179 |
"roneneldan/TinyStories-3M": ModelType.PT,
|
180 |
+
"jerryjalapeno/nart-100k-7b": ModelType.FT,
|
181 |
+
"lmsys/vicuna-13b-v1.3": ModelType.FT,
|
182 |
+
"lmsys/vicuna-7b-v1.3": ModelType.FT,
|
183 |
+
"lmsys/vicuna-13b-v1.1": ModelType.FT,
|
184 |
+
"lmsys/vicuna-13b-delta-v1.1": ModelType.FT,
|
185 |
+
"lmsys/vicuna-7b-delta-v1.1": ModelType.FT,
|
186 |
+
"abhiramtirumala/DialoGPT-sarcastic-medium": ModelType.FT,
|
187 |
+
"haonan-li/bactrian-x-llama-13b-merged": ModelType.FT,
|
188 |
+
"Gryphe/MythoLogic-13b": ModelType.FT,
|
189 |
+
"Gryphe/MythoBoros-13b": ModelType.FT,
|
190 |
+
"pillowtalks-ai/delta13b": ModelType.FT,
|
191 |
+
"wannaphong/openthaigpt-0.1.0-beta-full-model_for_open_llm_leaderboard": ModelType.FT,
|
192 |
+
"bigscience/bloom-7b1": ModelType.PT,
|
193 |
"bigcode/tiny_starcoder_py": ModelType.PT,
|
194 |
+
"bigcode/starcoderplus": ModelType.FT,
|
195 |
"bigcode/gpt_bigcode-santacoder": ModelType.PT,
|
196 |
"bigcode/starcoder": ModelType.PT,
|
197 |
+
"Open-Orca/OpenOrca-Preview1-13B": ModelType.FT,
|
198 |
+
"microsoft/DialoGPT-large": ModelType.FT,
|
199 |
+
"microsoft/DialoGPT-small": ModelType.FT,
|
200 |
+
"microsoft/DialoGPT-medium": ModelType.FT,
|
201 |
+
"microsoft/CodeGPT-small-py": ModelType.FT,
|
202 |
+
"Tincando/fiction_story_generator": ModelType.FT,
|
203 |
+
"Pirr/pythia-13b-deduped-green_devil": ModelType.FT,
|
204 |
+
"Aeala/GPT4-x-AlpacaDente2-30b": ModelType.FT,
|
205 |
+
"Aeala/GPT4-x-AlpacaDente-30b": ModelType.FT,
|
206 |
+
"Aeala/GPT4-x-Alpasta-13b": ModelType.FT,
|
207 |
+
"Aeala/VicUnlocked-alpaca-30b": ModelType.FT,
|
208 |
+
"Tap-M/Luna-AI-Llama2-Uncensored": ModelType.FT,
|
209 |
+
"illuin/test-custom-llama": ModelType.FT,
|
210 |
+
"dvruette/oasst-llama-13b-2-epochs": ModelType.FT,
|
211 |
+
"dvruette/oasst-gpt-neox-20b-1000-steps": ModelType.FT,
|
212 |
"dvruette/llama-13b-pretrained-dropout": ModelType.PT,
|
213 |
"dvruette/llama-13b-pretrained": ModelType.PT,
|
214 |
"dvruette/llama-13b-pretrained-sft-epoch-1": ModelType.PT,
|
215 |
"dvruette/llama-13b-pretrained-sft-do2": ModelType.PT,
|
216 |
+
"dvruette/oasst-gpt-neox-20b-3000-steps": ModelType.FT,
|
217 |
"dvruette/oasst-pythia-12b-pretrained-sft": ModelType.PT,
|
218 |
+
"dvruette/oasst-pythia-6.9b-4000-steps": ModelType.FT,
|
219 |
+
"dvruette/gpt-neox-20b-full-precision": ModelType.FT,
|
220 |
+
"dvruette/oasst-llama-13b-1000-steps": ModelType.FT,
|
221 |
"openlm-research/open_llama_7b_700bt_preview": ModelType.PT,
|
222 |
"openlm-research/open_llama_7b": ModelType.PT,
|
223 |
"openlm-research/open_llama_7b_v2": ModelType.PT,
|
224 |
"openlm-research/open_llama_3b": ModelType.PT,
|
225 |
"openlm-research/open_llama_13b": ModelType.PT,
|
226 |
"openlm-research/open_llama_3b_v2": ModelType.PT,
|
227 |
+
"PocketDoc/Dans-PileOfSets-Mk1-llama-13b-merged": ModelType.FT,
|
228 |
+
"GeorgiaTechResearchInstitute/galpaca-30b": ModelType.FT,
|
229 |
+
"GeorgiaTechResearchInstitute/starcoder-gpteacher-code-instruct": ModelType.FT,
|
230 |
+
"databricks/dolly-v2-7b": ModelType.FT,
|
231 |
+
"databricks/dolly-v2-3b": ModelType.FT,
|
232 |
+
"databricks/dolly-v2-12b": ModelType.FT,
|
233 |
+
"Rachneet/gpt2-xl-alpaca": ModelType.FT,
|
234 |
+
"Locutusque/gpt2-conversational-or-qa": ModelType.FT,
|
235 |
+
"psyche/kogpt": ModelType.FT,
|
236 |
+
"NbAiLab/nb-gpt-j-6B-alpaca": ModelType.FT,
|
237 |
+
"Mikael110/llama-2-7b-guanaco-fp16": ModelType.FT,
|
238 |
+
"Mikael110/llama-2-13b-guanaco-fp16": ModelType.FT,
|
239 |
+
"Fredithefish/CrimsonPajama": ModelType.FT,
|
240 |
+
"Fredithefish/RedPajama-INCITE-Chat-3B-ShareGPT-11K": ModelType.FT,
|
241 |
+
"Fredithefish/ScarletPajama-3B-HF": ModelType.FT,
|
242 |
+
"Fredithefish/RedPajama-INCITE-Chat-3B-Instruction-Tuning-with-GPT-4": ModelType.FT,
|
243 |
+
"acrastt/RedPajama-INCITE-Chat-Instruct-3B-V1": ModelType.FT,
|
244 |
+
"eachadea/vicuna-13b-1.1": ModelType.FT,
|
245 |
+
"eachadea/vicuna-7b-1.1": ModelType.FT,
|
246 |
+
"eachadea/vicuna-13b": ModelType.FT,
|
247 |
+
"openaccess-ai-collective/wizard-mega-13b": ModelType.FT,
|
248 |
+
"openaccess-ai-collective/manticore-13b": ModelType.FT,
|
249 |
+
"openaccess-ai-collective/manticore-30b-chat-pyg-alpha": ModelType.FT,
|
250 |
+
"openaccess-ai-collective/minotaur-13b": ModelType.FT,
|
251 |
+
"openaccess-ai-collective/minotaur-13b-fixed": ModelType.FT,
|
252 |
+
"openaccess-ai-collective/hippogriff-30b-chat": ModelType.FT,
|
253 |
+
"openaccess-ai-collective/manticore-13b-chat-pyg": ModelType.FT,
|
254 |
+
"pythainlp/wangchanglm-7.5B-sft-enth": ModelType.FT,
|
255 |
+
"pythainlp/wangchanglm-7.5B-sft-en-sharded": ModelType.FT,
|
256 |
+
"euclaise/gpt-neox-122m-minipile-digits": ModelType.FT,
|
257 |
+
"stabilityai/StableBeluga1-Delta": ModelType.FT,
|
258 |
+
"stabilityai/stablelm-tuned-alpha-7b": ModelType.FT,
|
259 |
+
"stabilityai/StableBeluga2": ModelType.FT,
|
260 |
+
"stabilityai/StableBeluga-13B": ModelType.FT,
|
261 |
+
"stabilityai/StableBeluga-7B": ModelType.FT,
|
262 |
"stabilityai/stablelm-base-alpha-7b": ModelType.PT,
|
263 |
"stabilityai/stablelm-base-alpha-3b": ModelType.PT,
|
264 |
+
"stabilityai/stablelm-tuned-alpha-3b": ModelType.FT,
|
265 |
+
"alibidaran/medical_transcription_generator": ModelType.FT,
|
266 |
+
"CalderaAI/30B-Lazarus": ModelType.FT,
|
267 |
+
"CalderaAI/13B-BlueMethod": ModelType.FT,
|
268 |
+
"CalderaAI/13B-Ouroboros": ModelType.FT,
|
269 |
+
"KoboldAI/OPT-13B-Erebus": ModelType.FT,
|
270 |
+
"KoboldAI/GPT-J-6B-Janeway": ModelType.FT,
|
271 |
+
"KoboldAI/GPT-J-6B-Shinen": ModelType.FT,
|
272 |
"KoboldAI/fairseq-dense-2.7B": ModelType.PT,
|
273 |
+
"KoboldAI/OPT-6B-nerys-v2": ModelType.FT,
|
274 |
+
"KoboldAI/GPT-NeoX-20B-Skein": ModelType.FT,
|
275 |
+
"KoboldAI/PPO_Pygway-6b-Mix": ModelType.FT,
|
276 |
"KoboldAI/fairseq-dense-6.7B": ModelType.PT,
|
277 |
"KoboldAI/fairseq-dense-125M": ModelType.PT,
|
278 |
+
"KoboldAI/OPT-13B-Nerybus-Mix": ModelType.FT,
|
279 |
+
"KoboldAI/OPT-2.7B-Erebus": ModelType.FT,
|
280 |
+
"KoboldAI/OPT-350M-Nerys-v2": ModelType.FT,
|
281 |
+
"KoboldAI/OPT-2.7B-Nerys-v2": ModelType.FT,
|
282 |
+
"KoboldAI/OPT-2.7B-Nerybus-Mix": ModelType.FT,
|
283 |
+
"KoboldAI/OPT-13B-Nerys-v2": ModelType.FT,
|
284 |
+
"KoboldAI/GPT-NeoX-20B-Erebus": ModelType.FT,
|
285 |
+
"KoboldAI/OPT-6.7B-Erebus": ModelType.FT,
|
286 |
"KoboldAI/fairseq-dense-355M": ModelType.PT,
|
287 |
+
"KoboldAI/OPT-6.7B-Nerybus-Mix": ModelType.FT,
|
288 |
+
"KoboldAI/GPT-J-6B-Adventure": ModelType.FT,
|
289 |
+
"KoboldAI/OPT-350M-Erebus": ModelType.FT,
|
290 |
+
"KoboldAI/GPT-J-6B-Skein": ModelType.FT,
|
291 |
+
"KoboldAI/OPT-30B-Erebus": ModelType.FT,
|
292 |
"klosax/pythia-160m-deduped-step92k-193bt": ModelType.PT,
|
293 |
"klosax/open_llama_3b_350bt_preview": ModelType.PT,
|
294 |
"klosax/openllama-3b-350bt": ModelType.PT,
|
295 |
"klosax/pythia-70m-deduped-step44k-92bt": ModelType.PT,
|
296 |
"klosax/open_llama_13b_600bt_preview": ModelType.PT,
|
297 |
"klosax/open_llama_7b_400bt_preview": ModelType.PT,
|
298 |
+
"kfkas/Llama-2-ko-7b-Chat": ModelType.FT,
|
299 |
+
"WeOpenML/Alpaca-7B-v1": ModelType.FT,
|
300 |
+
"WeOpenML/PandaLM-Alpaca-7B-v1": ModelType.FT,
|
301 |
+
"TFLai/gpt2-turkish-uncased": ModelType.FT,
|
302 |
+
"ehartford/WizardLM-13B-Uncensored": ModelType.FT,
|
303 |
+
"ehartford/dolphin-llama-13b": ModelType.FT,
|
304 |
+
"ehartford/Wizard-Vicuna-30B-Uncensored": ModelType.FT,
|
305 |
+
"ehartford/WizardLM-30B-Uncensored": ModelType.FT,
|
306 |
+
"ehartford/Wizard-Vicuna-13B-Uncensored": ModelType.FT,
|
307 |
+
"ehartford/WizardLM-7B-Uncensored": ModelType.FT,
|
308 |
+
"ehartford/based-30b": ModelType.FT,
|
309 |
+
"ehartford/Wizard-Vicuna-7B-Uncensored": ModelType.FT,
|
310 |
+
"wahaha1987/llama_7b_sharegpt94k_fastchat": ModelType.FT,
|
311 |
+
"wahaha1987/llama_13b_sharegpt94k_fastchat": ModelType.FT,
|
312 |
+
"OpenAssistant/oasst-sft-1-pythia-12b": ModelType.FT,
|
313 |
+
"OpenAssistant/stablelm-7b-sft-v7-epoch-3": ModelType.FT,
|
314 |
+
"OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5": ModelType.FT,
|
315 |
+
"OpenAssistant/pythia-12b-sft-v8-2.5k-steps": ModelType.FT,
|
316 |
+
"OpenAssistant/pythia-12b-sft-v8-7k-steps": ModelType.FT,
|
317 |
+
"OpenAssistant/pythia-12b-pre-v8-12.5k-steps": ModelType.FT,
|
318 |
+
"OpenAssistant/llama2-13b-orca-8k-3319": ModelType.FT,
|
319 |
+
"junelee/wizard-vicuna-13b": ModelType.FT,
|
320 |
"BreadAi/gpt-YA-1-1_160M": ModelType.PT,
|
321 |
"BreadAi/MuseCan": ModelType.PT,
|
322 |
"BreadAi/MusePy-1-2": ModelType.PT,
|
323 |
"BreadAi/DiscordPy": ModelType.PT,
|
324 |
"BreadAi/PM_modelV2": ModelType.PT,
|
325 |
"BreadAi/gpt-Youtube": ModelType.PT,
|
326 |
+
"BreadAi/StoryPy": ModelType.FT,
|
327 |
+
"julianweng/Llama-2-7b-chat-orcah": ModelType.FT,
|
328 |
+
"AGI-inc/lora_moe_7b_baseline": ModelType.FT,
|
329 |
+
"AGI-inc/lora_moe_7b": ModelType.FT,
|
330 |
+
"togethercomputer/GPT-NeoXT-Chat-Base-20B": ModelType.FT,
|
331 |
+
"togethercomputer/RedPajama-INCITE-Chat-7B-v0.1": ModelType.FT,
|
332 |
+
"togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1": ModelType.FT,
|
333 |
"togethercomputer/RedPajama-INCITE-7B-Base": ModelType.PT,
|
334 |
+
"togethercomputer/RedPajama-INCITE-7B-Instruct": ModelType.FT,
|
335 |
"togethercomputer/RedPajama-INCITE-Base-3B-v1": ModelType.PT,
|
336 |
+
"togethercomputer/Pythia-Chat-Base-7B": ModelType.FT,
|
337 |
"togethercomputer/RedPajama-INCITE-Base-7B-v0.1": ModelType.PT,
|
338 |
+
"togethercomputer/GPT-JT-6B-v1": ModelType.FT,
|
339 |
+
"togethercomputer/GPT-JT-6B-v0": ModelType.FT,
|
340 |
+
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": ModelType.FT,
|
341 |
+
"togethercomputer/RedPajama-INCITE-7B-Chat": ModelType.FT,
|
342 |
+
"togethercomputer/RedPajama-INCITE-Instruct-3B-v1": ModelType.FT,
|
343 |
+
"Writer/camel-5b-hf": ModelType.FT,
|
344 |
"Writer/palmyra-base": ModelType.PT,
|
345 |
+
"MBZUAI/LaMini-GPT-1.5B": ModelType.FT,
|
346 |
+
"MBZUAI/lamini-cerebras-111m": ModelType.FT,
|
347 |
+
"MBZUAI/lamini-neo-1.3b": ModelType.FT,
|
348 |
+
"MBZUAI/lamini-cerebras-1.3b": ModelType.FT,
|
349 |
+
"MBZUAI/lamini-cerebras-256m": ModelType.FT,
|
350 |
+
"MBZUAI/LaMini-GPT-124M": ModelType.FT,
|
351 |
+
"MBZUAI/lamini-neo-125m": ModelType.FT,
|
352 |
+
"TehVenom/DiffMerge-DollyGPT-Pygmalion": ModelType.FT,
|
353 |
+
"TehVenom/PPO_Shygmalion-6b": ModelType.FT,
|
354 |
+
"TehVenom/Dolly_Shygmalion-6b-Dev_V8P2": ModelType.FT,
|
355 |
+
"TehVenom/Pygmalion_AlpacaLora-7b": ModelType.FT,
|
356 |
+
"TehVenom/PPO_Pygway-V8p4_Dev-6b": ModelType.FT,
|
357 |
+
"TehVenom/Dolly_Malion-6b": ModelType.FT,
|
358 |
+
"TehVenom/PPO_Shygmalion-V8p4_Dev-6b": ModelType.FT,
|
359 |
+
"TehVenom/ChanMalion": ModelType.FT,
|
360 |
+
"TehVenom/GPT-J-Pyg_PPO-6B": ModelType.FT,
|
361 |
+
"TehVenom/Pygmalion-13b-Merged": ModelType.FT,
|
362 |
+
"TehVenom/Metharme-13b-Merged": ModelType.FT,
|
363 |
+
"TehVenom/Dolly_Shygmalion-6b": ModelType.FT,
|
364 |
+
"TehVenom/GPT-J-Pyg_PPO-6B-Dev-V8p4": ModelType.FT,
|
365 |
+
"georgesung/llama2_7b_chat_uncensored": ModelType.FT,
|
366 |
+
"vicgalle/gpt2-alpaca": ModelType.FT,
|
367 |
+
"vicgalle/alpaca-7b": ModelType.FT,
|
368 |
+
"vicgalle/gpt2-alpaca-gpt4": ModelType.FT,
|
369 |
"facebook/opt-350m": ModelType.PT,
|
370 |
"facebook/opt-125m": ModelType.PT,
|
371 |
"facebook/xglm-4.5B": ModelType.PT,
|
|
|
377 |
"facebook/xglm-7.5B": ModelType.PT,
|
378 |
"facebook/xglm-564M": ModelType.PT,
|
379 |
"facebook/opt-30b": ModelType.PT,
|
380 |
+
"golaxy/gogpt-7b": ModelType.FT,
|
381 |
+
"golaxy/gogpt2-7b": ModelType.FT,
|
382 |
+
"golaxy/gogpt-7b-bloom": ModelType.FT,
|
383 |
+
"golaxy/gogpt-3b-bloom": ModelType.FT,
|
384 |
+
"psmathur/orca_mini_v2_7b": ModelType.FT,
|
385 |
+
"psmathur/orca_mini_7b": ModelType.FT,
|
386 |
+
"psmathur/orca_mini_3b": ModelType.FT,
|
387 |
+
"psmathur/orca_mini_v2_13b": ModelType.FT,
|
388 |
"gpt2-xl": ModelType.PT,
|
389 |
+
"lxe/Cerebras-GPT-2.7B-Alpaca-SP": ModelType.FT,
|
390 |
+
"Monero/Manticore-13b-Chat-Pyg-Guanaco": ModelType.FT,
|
391 |
+
"Monero/WizardLM-Uncensored-SuperCOT-StoryTelling-30b": ModelType.FT,
|
392 |
+
"Monero/WizardLM-13b-OpenAssistant-Uncensored": ModelType.FT,
|
393 |
+
"Monero/WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b": ModelType.FT,
|
394 |
+
"jzjiao/opt-1.3b-rlhf": ModelType.FT,
|
395 |
+
"HuggingFaceH4/starchat-beta": ModelType.FT,
|
396 |
+
"KnutJaegersberg/gpt-2-xl-EvolInstruct": ModelType.FT,
|
397 |
+
"KnutJaegersberg/megatron-GPT-2-345m-EvolInstruct": ModelType.FT,
|
398 |
+
"KnutJaegersberg/galactica-orca-wizardlm-1.3b": ModelType.FT,
|
399 |
+
"openchat/openchat_8192": ModelType.FT,
|
400 |
+
"openchat/openchat_v2": ModelType.FT,
|
401 |
+
"openchat/openchat_v2_w": ModelType.FT,
|
402 |
+
"ausboss/llama-13b-supercot": ModelType.FT,
|
403 |
+
"ausboss/llama-30b-supercot": ModelType.FT,
|
404 |
+
"Neko-Institute-of-Science/metharme-7b": ModelType.FT,
|
405 |
+
"Neko-Institute-of-Science/pygmalion-7b": ModelType.FT,
|
406 |
+
"SebastianSchramm/Cerebras-GPT-111M-instruction": ModelType.FT,
|
407 |
+
"victor123/WizardLM-13B-1.0": ModelType.FT,
|
408 |
+
"OpenBuddy/openbuddy-openllama-13b-v7-fp16": ModelType.FT,
|
409 |
+
"OpenBuddy/openbuddy-llama2-13b-v8.1-fp16": ModelType.FT,
|
410 |
+
"OpenBuddyEA/openbuddy-llama-30b-v7.1-bf16": ModelType.FT,
|
411 |
"baichuan-inc/Baichuan-7B": ModelType.PT,
|
412 |
+
"tiiuae/falcon-40b-instruct": ModelType.FT,
|
413 |
"tiiuae/falcon-40b": ModelType.PT,
|
414 |
"tiiuae/falcon-7b": ModelType.PT,
|
415 |
+
"YeungNLP/firefly-llama-13b": ModelType.FT,
|
416 |
+
"YeungNLP/firefly-llama-13b-v1.2": ModelType.FT,
|
417 |
+
"YeungNLP/firefly-llama2-13b": ModelType.FT,
|
418 |
+
"YeungNLP/firefly-ziya-13b": ModelType.FT,
|
419 |
+
"shaohang/Sparse0.5_OPT-1.3": ModelType.FT,
|
420 |
+
"xzuyn/Alpacino-SuperCOT-13B": ModelType.FT,
|
421 |
+
"xzuyn/MedicWizard-7B": ModelType.FT,
|
422 |
+
"xDAN-AI/xDAN_13b_l2_lora": ModelType.FT,
|
423 |
+
"beomi/KoAlpaca-Polyglot-5.8B": ModelType.FT,
|
424 |
+
"beomi/llama-2-ko-7b": ModelType.FT,
|
425 |
"Salesforce/codegen-6B-multi": ModelType.PT,
|
426 |
"Salesforce/codegen-16B-nl": ModelType.PT,
|
427 |
"Salesforce/codegen-6B-nl": ModelType.PT,
|
428 |
+
"ai-forever/rugpt3large_based_on_gpt2": ModelType.FT,
|
429 |
"gpt2-large": ModelType.PT,
|
430 |
+
"frank098/orca_mini_3b_juniper": ModelType.FT,
|
431 |
+
"frank098/WizardLM_13B_juniper": ModelType.FT,
|
432 |
+
"FPHam/Free_Sydney_13b_HF": ModelType.FT,
|
433 |
"huggingface/llama-13b": ModelType.PT,
|
434 |
"huggingface/llama-7b": ModelType.PT,
|
435 |
"huggingface/llama-65b": ModelType.PT,
|
436 |
"huggingface/llama-65b": ModelType.PT,
|
437 |
"huggingface/llama-30b": ModelType.PT,
|
438 |
+
"Henk717/chronoboros-33B": ModelType.FT,
|
439 |
+
"jondurbin/airoboros-13b-gpt4-1.4": ModelType.FT,
|
440 |
+
"jondurbin/airoboros-7b": ModelType.FT,
|
441 |
+
"jondurbin/airoboros-7b-gpt4": ModelType.FT,
|
442 |
+
"jondurbin/airoboros-7b-gpt4-1.1": ModelType.FT,
|
443 |
+
"jondurbin/airoboros-7b-gpt4-1.2": ModelType.FT,
|
444 |
+
"jondurbin/airoboros-7b-gpt4-1.3": ModelType.FT,
|
445 |
+
"jondurbin/airoboros-7b-gpt4-1.4": ModelType.FT,
|
446 |
+
"jondurbin/airoboros-l2-7b-gpt4-1.4.1": ModelType.FT,
|
447 |
+
"jondurbin/airoboros-l2-13b-gpt4-1.4.1": ModelType.FT,
|
448 |
+
"jondurbin/airoboros-l2-70b-gpt4-1.4.1": ModelType.FT,
|
449 |
+
"jondurbin/airoboros-13b": ModelType.FT,
|
450 |
+
"jondurbin/airoboros-33b-gpt4-1.4": ModelType.FT,
|
451 |
+
"jondurbin/airoboros-33b-gpt4-1.2": ModelType.FT,
|
452 |
+
"jondurbin/airoboros-65b-gpt4-1.2": ModelType.FT,
|
453 |
+
"ariellee/SuperPlatty-30B": ModelType.FT,
|
454 |
+
"danielhanchen/open_llama_3b_600bt_preview": ModelType.FT,
|
455 |
"cerebras/Cerebras-GPT-256M": ModelType.PT,
|
456 |
"cerebras/Cerebras-GPT-1.3B": ModelType.PT,
|
457 |
"cerebras/Cerebras-GPT-13B": ModelType.PT,
|
|
|
459 |
"cerebras/Cerebras-GPT-111M": ModelType.PT,
|
460 |
"cerebras/Cerebras-GPT-6.7B": ModelType.PT,
|
461 |
"Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-hf": ModelType.RL,
|
462 |
+
"Yhyu13/llama-30B-hf-openassitant": ModelType.FT,
|
463 |
+
"NousResearch/Nous-Hermes-Llama2-13b": ModelType.FT,
|
464 |
+
"NousResearch/Nous-Hermes-llama-2-7b": ModelType.FT,
|
465 |
+
"NousResearch/Redmond-Puffin-13B": ModelType.FT,
|
466 |
+
"NousResearch/Nous-Hermes-13b": ModelType.FT,
|
467 |
+
"project-baize/baize-v2-7b": ModelType.FT,
|
468 |
+
"project-baize/baize-v2-13b": ModelType.FT,
|
469 |
+
"LLMs/WizardLM-13B-V1.0": ModelType.FT,
|
470 |
+
"LLMs/AlpacaGPT4-7B-elina": ModelType.FT,
|
471 |
+
"wenge-research/yayi-7b": ModelType.FT,
|
472 |
+
"wenge-research/yayi-7b-llama2": ModelType.FT,
|
473 |
+
"wenge-research/yayi-13b-llama2": ModelType.FT,
|
474 |
+
"yhyhy3/open_llama_7b_v2_med_instruct": ModelType.FT,
|
475 |
+
"llama-anon/instruct-13b": ModelType.FT,
|
476 |
+
"huggingtweets/jerma985": ModelType.FT,
|
477 |
+
"huggingtweets/gladosystem": ModelType.FT,
|
478 |
+
"huggingtweets/bladeecity-jerma985": ModelType.FT,
|
479 |
"huggyllama/llama-13b": ModelType.PT,
|
480 |
"huggyllama/llama-65b": ModelType.PT,
|
481 |
"FabbriSimo01/Facebook_opt_1.3b_Quantized": ModelType.PT,
|
482 |
+
"upstage/Llama-2-70b-instruct": ModelType.FT,
|
483 |
+
"upstage/Llama-2-70b-instruct-1024": ModelType.FT,
|
484 |
+
"upstage/llama-65b-instruct": ModelType.FT,
|
485 |
+
"upstage/llama-30b-instruct-2048": ModelType.FT,
|
486 |
+
"upstage/llama-30b-instruct": ModelType.FT,
|
487 |
+
"WizardLM/WizardLM-13B-1.0": ModelType.FT,
|
488 |
+
"WizardLM/WizardLM-13B-V1.1": ModelType.FT,
|
489 |
+
"WizardLM/WizardLM-13B-V1.2": ModelType.FT,
|
490 |
+
"WizardLM/WizardLM-30B-V1.0": ModelType.FT,
|
491 |
+
"WizardLM/WizardCoder-15B-V1.0": ModelType.FT,
|
492 |
"gpt2": ModelType.PT,
|
493 |
+
"keyfan/vicuna-chinese-replication-v1.1": ModelType.FT,
|
494 |
+
"nthngdy/pythia-owt2-70m-100k": ModelType.FT,
|
495 |
+
"nthngdy/pythia-owt2-70m-50k": ModelType.FT,
|
496 |
+
"quantumaikr/KoreanLM-hf": ModelType.FT,
|
497 |
+
"quantumaikr/open_llama_7b_hf": ModelType.FT,
|
498 |
+
"quantumaikr/QuantumLM-70B-hf": ModelType.FT,
|
499 |
+
"MayaPH/FinOPT-Lincoln": ModelType.FT,
|
500 |
+
"MayaPH/FinOPT-Franklin": ModelType.FT,
|
501 |
+
"MayaPH/GodziLLa-30B": ModelType.FT,
|
502 |
+
"MayaPH/GodziLLa-30B-plus": ModelType.FT,
|
503 |
+
"MayaPH/FinOPT-Washington": ModelType.FT,
|
504 |
+
"ogimgio/gpt-neo-125m-neurallinguisticpioneers": ModelType.FT,
|
505 |
+
"layoric/llama-2-13b-code-alpaca": ModelType.FT,
|
506 |
+
"CobraMamba/mamba-gpt-3b": ModelType.FT,
|
507 |
+
"CobraMamba/mamba-gpt-3b-v2": ModelType.FT,
|
508 |
+
"CobraMamba/mamba-gpt-3b-v3": ModelType.FT,
|
509 |
+
"timdettmers/guanaco-33b-merged": ModelType.FT,
|
510 |
+
"elinas/chronos-33b": ModelType.FT,
|
511 |
+
"heegyu/RedTulu-Uncensored-3B-0719": ModelType.FT,
|
512 |
+
"heegyu/WizardVicuna-Uncensored-3B-0719": ModelType.FT,
|
513 |
+
"heegyu/WizardVicuna-3B-0719": ModelType.FT,
|
514 |
"meta-llama/Llama-2-7b-chat-hf": ModelType.RL,
|
515 |
"meta-llama/Llama-2-7b-hf": ModelType.PT,
|
516 |
"meta-llama/Llama-2-13b-chat-hf": ModelType.RL,
|
517 |
"meta-llama/Llama-2-13b-hf": ModelType.PT,
|
518 |
"meta-llama/Llama-2-70b-chat-hf": ModelType.RL,
|
519 |
"meta-llama/Llama-2-70b-hf": ModelType.PT,
|
520 |
+
"xhyi/PT_GPTNEO350_ATG": ModelType.FT,
|
521 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-20b": ModelType.FT,
|
522 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-open-llama-7b-preview-400bt": ModelType.FT,
|
523 |
+
"h2oai/h2ogpt-oig-oasst1-512-6_9b": ModelType.FT,
|
524 |
+
"h2oai/h2ogpt-oasst1-512-12b": ModelType.FT,
|
525 |
+
"h2oai/h2ogpt-oig-oasst1-256-6_9b": ModelType.FT,
|
526 |
+
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt": ModelType.FT,
|
527 |
+
"h2oai/h2ogpt-oasst1-512-20b": ModelType.FT,
|
528 |
+
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2": ModelType.FT,
|
529 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-12b": ModelType.FT,
|
530 |
+
"h2oai/h2ogpt-gm-oasst1-multilang-1024-20b": ModelType.FT,
|
531 |
+
"bofenghuang/vigogne-13b-instruct": ModelType.FT,
|
532 |
+
"bofenghuang/vigogne-13b-chat": ModelType.FT,
|
533 |
+
"bofenghuang/vigogne-2-7b-instruct": ModelType.FT,
|
534 |
+
"bofenghuang/vigogne-7b-instruct": ModelType.FT,
|
535 |
+
"bofenghuang/vigogne-7b-chat": ModelType.FT,
|
536 |
+
"Vmware/open-llama-7b-v2-open-instruct": ModelType.FT,
|
537 |
+
"VMware/open-llama-0.7T-7B-open-instruct-v1.1": ModelType.FT,
|
538 |
+
"ewof/koishi-instruct-3b": ModelType.FT,
|
539 |
+
"gywy/llama2-13b-chinese-v1": ModelType.FT,
|
540 |
+
"GOAT-AI/GOAT-7B-Community": ModelType.FT,
|
541 |
+
"psyche/kollama2-7b": ModelType.FT,
|
542 |
+
"TheTravellingEngineer/llama2-7b-hf-guanaco": ModelType.FT,
|
543 |
+
"beaugogh/pythia-1.4b-deduped-sharegpt": ModelType.FT,
|
544 |
+
"augtoma/qCammel-70-x": ModelType.FT,
|
545 |
+
"Lajonbot/Llama-2-7b-chat-hf-instruct-pl-lora_unload": ModelType.FT,
|
546 |
+
"anhnv125/pygmalion-6b-roleplay": ModelType.FT,
|
547 |
+
"64bits/LexPodLM-13B": ModelType.FT
|
548 |
}
|
549 |
|
550 |
|
|
|
578 |
except Exception:
|
579 |
model_data[AutoEvalColumn.model_type.name] = "Unknown"
|
580 |
model_data[AutoEvalColumn.model_type_symbol.name] = "Add type to request files"
|
|