diff --git a/0-hero/Matter-0.2-7B-DPO_eval_request_False_bfloat16_Original.json b/0-hero/Matter-0.2-7B-DPO_eval_request_False_bfloat16_Original.json index ba57c8bfd59c35ace23c51a5500f2809abd499c2..eb68668692b625dd7187777405a91208c1ac2f78 100644 --- a/0-hero/Matter-0.2-7B-DPO_eval_request_False_bfloat16_Original.json +++ b/0-hero/Matter-0.2-7B-DPO_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "0-hero/Matter-0.2-7B-DPO", "base_model": "", "revision": "26a66f0d862e2024ce4ad0a09c37052ac36e8af6", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:58:40Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "0-hero/Matter-0.2-7B-DPO", + "base_model": "", + "revision": "26a66f0d862e2024ce4ad0a09c37052ac36e8af6", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:58:40Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/01-ai/Yi-1.5-34B-Chat-16K_eval_request_False_bfloat16_Original.json b/01-ai/Yi-1.5-34B-Chat-16K_eval_request_False_bfloat16_Original.json index 5279aea08680c0012b40bdc8c5eeb269d5a6f7a7..21116a6b0b89250cb723c40d192b8b56f236b713 100644 --- a/01-ai/Yi-1.5-34B-Chat-16K_eval_request_False_bfloat16_Original.json +++ b/01-ai/Yi-1.5-34B-Chat-16K_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-12T12:06:53Z", "model_type": "chat", "job_id": "6658010", "job_start_time": "2024-06-22T18:39:45.456325", "use_chat_template": true -} \ No newline at end of file +} diff --git a/152334H/miqu-1-70b-sf_eval_request_False_float16_Original.json b/152334H/miqu-1-70b-sf_eval_request_False_float16_Original.json index 2411aa9e080d4a19bf688cf33fcbeac5de346d35..27eac7432557456ee96edba14fd436a89e0c62ec 100644 --- a/152334H/miqu-1-70b-sf_eval_request_False_float16_Original.json +++ b/152334H/miqu-1-70b-sf_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 68.977, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T15:32:34Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7208639", "job_start_time": "2024-06-26T19:24:39.425080", "use_chat_template": false -} \ No newline at end of file +} diff --git a/AI-Sweden-Models/gpt-sw3-40b_eval_request_False_float16_Original.json b/AI-Sweden-Models/gpt-sw3-40b_eval_request_False_float16_Original.json index be0fb8c0167b2f4df159d7e93a420a7b055d764e..da0ee4bdd88f1de07cac6eacb216e3ee698ccd01 100644 --- a/AI-Sweden-Models/gpt-sw3-40b_eval_request_False_float16_Original.json +++ b/AI-Sweden-Models/gpt-sw3-40b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "AI-Sweden-Models/gpt-sw3-40b", "base_model": "", "revision": "1af27994df1287a7fac1b10d60e40ca43a22a385", "precision": "float16", "params": 39.927, "architectures": "GPT2LMHeadModel", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:13:04Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "AI-Sweden-Models/gpt-sw3-40b", + "base_model": "", + "revision": "1af27994df1287a7fac1b10d60e40ca43a22a385", + "precision": "float16", + "params": 39.927, + "architectures": "GPT2LMHeadModel", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:13:04Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/AiMavenAi/Athena-70B-L3_eval_request_False_float16_Original.json b/AiMavenAi/Athena-70B-L3_eval_request_False_float16_Original.json index 86877bd2c3603ef1e627102e67c730390b919f90..a8e79827ec6949fe25874ae1e5520bdf270a8096 100644 --- a/AiMavenAi/Athena-70B-L3_eval_request_False_float16_Original.json +++ b/AiMavenAi/Athena-70B-L3_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "AiMavenAi/Athena-70B-L3", "base_model": "AiMavenAi/Athena-70B-L3", "revision": "ed35274b035d52760c4bf00ecd28d04045a25f97", "precision": "float16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:55:48Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "AiMavenAi/Athena-70B-L3", + "base_model": "AiMavenAi/Athena-70B-L3", + "revision": "ed35274b035d52760c4bf00ecd28d04045a25f97", + "precision": "float16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:55:48Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu_eval_request_False_bfloat16_Original.json b/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu_eval_request_False_bfloat16_Original.json index 5c6b697ea2823d8d8dc60514f45608da9728c35f..bb56e83ca4ea38fc10edd22f47607c11430744a6 100644 --- a/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu_eval_request_False_bfloat16_Original.json +++ b/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 0.218, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:34:07Z", - "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", + "model_type": "🟩 : 🟩 continuously pretrained", "job_id": "7205446", "job_start_time": "2024-06-26T14:59:08.621087", "use_chat_template": false -} \ No newline at end of file +} diff --git a/BEE-spoke-data/smol_llama-220M-GQA_eval_request_False_bfloat16_Original.json b/BEE-spoke-data/smol_llama-220M-GQA_eval_request_False_bfloat16_Original.json index d53ec49e5df04d4e4c9a420988f6308a44c9af65..ddc7794194da9ed63fed400ed2c48e457e70b02d 100644 --- a/BEE-spoke-data/smol_llama-220M-GQA_eval_request_False_bfloat16_Original.json +++ b/BEE-spoke-data/smol_llama-220M-GQA_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 0.218, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-26T14:33:46Z", - "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", + "model_type": "🟢 : 🟢 pretrained", "job_id": "7205417", "job_start_time": "2024-06-26T14:48:57.752173", "use_chat_template": false -} \ No newline at end of file +} diff --git a/BEE-spoke-data/smol_llama-220M-openhermes_eval_request_False_bfloat16_Original.json b/BEE-spoke-data/smol_llama-220M-openhermes_eval_request_False_bfloat16_Original.json index 9c3b6574c550f4ffc5c9963d46f1a9758f9cb4d0..c8444368a4b1e97184fe80141f2e6c03db5d807f 100644 --- a/BEE-spoke-data/smol_llama-220M-openhermes_eval_request_False_bfloat16_Original.json +++ b/BEE-spoke-data/smol_llama-220M-openhermes_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 0.218, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:33:23Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7205418", "job_start_time": "2024-06-26T14:49:07.666028", "use_chat_template": false -} \ No newline at end of file +} diff --git a/BEE-spoke-data/zephyr-220m-dpo-full_eval_request_False_bfloat16_Original.json b/BEE-spoke-data/zephyr-220m-dpo-full_eval_request_False_bfloat16_Original.json index 88a486a354b5feef709950316b400b194903cbcf..9c16826a9a9ac082d697b5c14f957a8eab36a25a 100644 --- a/BEE-spoke-data/zephyr-220m-dpo-full_eval_request_False_bfloat16_Original.json +++ b/BEE-spoke-data/zephyr-220m-dpo-full_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 0.218, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:34:45Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7205414", "job_start_time": "2024-06-26T14:46:28.859979", "use_chat_template": true -} \ No newline at end of file +} diff --git a/Ba2han/Llama-Phi-3_DoRA_eval_request_False_bfloat16_Original.json b/Ba2han/Llama-Phi-3_DoRA_eval_request_False_bfloat16_Original.json index 7f16504c23fffd414384716a64998560330b03a5..ba96fa54b29ab35df75a231eb27f2c8e759f5a85 100644 --- a/Ba2han/Llama-Phi-3_DoRA_eval_request_False_bfloat16_Original.json +++ b/Ba2han/Llama-Phi-3_DoRA_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Ba2han/Llama-Phi-3_DoRA", "base_model": "", "revision": "36f99064a7be8ba475c2ee5c5424e95c263ccb87", "precision": "bfloat16", "params": 3.821, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:34:56Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Ba2han/Llama-Phi-3_DoRA", + "base_model": "", + "revision": "36f99064a7be8ba475c2ee5c5424e95c263ccb87", + "precision": "bfloat16", + "params": 3.821, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:34:56Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/BarraHome/Mistroll-7B-v2.2_eval_request_False_bfloat16_Original.json b/BarraHome/Mistroll-7B-v2.2_eval_request_False_bfloat16_Original.json index 48afb2cfaf3c4a6936ca8f172826b03eb7dab333..2e632e8abde3a26729bb2a3095ab0fc2113bf0f0 100644 --- a/BarraHome/Mistroll-7B-v2.2_eval_request_False_bfloat16_Original.json +++ b/BarraHome/Mistroll-7B-v2.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "BarraHome/Mistroll-7B-v2.2", "base_model": "yam-peleg/Experiment26-7B", "revision": "755df0d9ed26d10744ec1f9dbad8cab88882ce73", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:40:40Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "BarraHome/Mistroll-7B-v2.2", + "base_model": "yam-peleg/Experiment26-7B", + "revision": "755df0d9ed26d10744ec1f9dbad8cab88882ce73", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:40:40Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/CausalLM/34b-beta_eval_request_False_bfloat16_Original.json b/CausalLM/34b-beta_eval_request_False_bfloat16_Original.json index 936b4a2c2c6eb5a763436044ad34ec8a95586dd6..2a3e8c30b3aea3c2597514a301417417212167fd 100644 --- a/CausalLM/34b-beta_eval_request_False_bfloat16_Original.json +++ b/CausalLM/34b-beta_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "CausalLM/34b-beta", "base_model": "", "revision": "0429951eb30ccdfff3515e711aaa7649a8a7364c", "precision": "bfloat16", "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:02:34Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "CausalLM/34b-beta", + "base_model": "", + "revision": "0429951eb30ccdfff3515e711aaa7649a8a7364c", + "precision": "bfloat16", + "params": 34.389, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:02:34Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/CausalLM/34b-beta_eval_request_False_float16_Original.json b/CausalLM/34b-beta_eval_request_False_float16_Original.json index 56df86fd8dd739e18e1500c48444fa9cb41a8029..09451b97d6f04cd740fb33be5fba535471b21656 100644 --- a/CausalLM/34b-beta_eval_request_False_float16_Original.json +++ b/CausalLM/34b-beta_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "CausalLM/34b-beta", "base_model": "", "revision": "0429951eb30ccdfff3515e711aaa7649a8a7364c", "precision": "float16", "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:02:48Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "CausalLM/34b-beta", + "base_model": "", + "revision": "0429951eb30ccdfff3515e711aaa7649a8a7364c", + "precision": "float16", + "params": 34.389, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:02:48Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/CausalLM/35b-beta-long_eval_request_False_bfloat16_Original.json b/CausalLM/35b-beta-long_eval_request_False_bfloat16_Original.json index 48550a60e10631266116e6e00fd695e03b2402fa..00fcb981dfcf6b84778e95b651cda39d04436f10 100644 --- a/CausalLM/35b-beta-long_eval_request_False_bfloat16_Original.json +++ b/CausalLM/35b-beta-long_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "CausalLM/35b-beta-long", "base_model": "", "revision": "36fa24a5e2288e2d81092523f14adb7d2b027a3b", "precision": "bfloat16", "params": 34.981, "architectures": "CohereForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:29:57Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "CausalLM/35b-beta-long", + "base_model": "", + "revision": "36fa24a5e2288e2d81092523f14adb7d2b027a3b", + "precision": "bfloat16", + "params": 34.981, + "architectures": "CohereForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:29:57Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/ClaudioItaly/TopEvolutionWiz_eval_request_False_bfloat16_Original.json b/ClaudioItaly/TopEvolutionWiz_eval_request_False_bfloat16_Original.json index 088d7bbca283b876e0f3e2152e3f4e9d1fbb580d..552122465ac9d382efeed4e408742a65eca09f6c 100644 --- a/ClaudioItaly/TopEvolutionWiz_eval_request_False_bfloat16_Original.json +++ b/ClaudioItaly/TopEvolutionWiz_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "ClaudioItaly/TopEvolutionWiz", "base_model": "", "revision": "a207b871d09091847f806a51c2ec879ce91c6040", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:32:08Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "ClaudioItaly/TopEvolutionWiz", + "base_model": "", + "revision": "a207b871d09091847f806a51c2ec879ce91c6040", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:32:08Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/CortexLM/btlm-7b-base-v0.1_eval_request_False_bfloat16_Original.json b/CortexLM/btlm-7b-base-v0.1_eval_request_False_bfloat16_Original.json index f17d2bb5522fe613c2bdf1103d8627db3d835592..fbbddcf8110ca285ce333f9a5a55f859fc20223f 100644 --- a/CortexLM/btlm-7b-base-v0.1_eval_request_False_bfloat16_Original.json +++ b/CortexLM/btlm-7b-base-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "CortexLM/btlm-7b-base-v0.1", "base_model": "", "revision": "876db86b226fe6eec6f1b02e6c082a6dfe0317e0", "precision": "bfloat16", "params": 6.893, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:28:41Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "CortexLM/btlm-7b-base-v0.1", + "base_model": "", + "revision": "876db86b226fe6eec6f1b02e6c082a6dfe0317e0", + "precision": "bfloat16", + "params": 6.893, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:28:41Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/CortexLM/btlm-7b-base-v0.2_eval_request_False_bfloat16_Original.json b/CortexLM/btlm-7b-base-v0.2_eval_request_False_bfloat16_Original.json index 9902653046f053e7ae12f3420195db47f7c49110..1014e0fc10246efbdd196ad71a24e68c7f6f2ca2 100644 --- a/CortexLM/btlm-7b-base-v0.2_eval_request_False_bfloat16_Original.json +++ b/CortexLM/btlm-7b-base-v0.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "CortexLM/btlm-7b-base-v0.2", "base_model": "", "revision": "eda8b4298365a26c8981316e09427c237b11217f", "precision": "bfloat16", "params": 6.885, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:28:25Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "CortexLM/btlm-7b-base-v0.2", + "base_model": "", + "revision": "eda8b4298365a26c8981316e09427c237b11217f", + "precision": "bfloat16", + "params": 6.885, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:28:25Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/DevQuasar/coma-7B-v0.1_eval_request_False_float16_Original.json b/DevQuasar/coma-7B-v0.1_eval_request_False_float16_Original.json index c8fff18300d7b0b95919e839db2d57721414cfe5..ac740b88d4d3852ca8e75a799b4f4fdb4d7bb42e 100644 --- a/DevQuasar/coma-7B-v0.1_eval_request_False_float16_Original.json +++ b/DevQuasar/coma-7B-v0.1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "DevQuasar/coma-7B-v0.1", "base_model": "meta-llama/Llama-2-7b-chat-hf", "revision": "8358359ac3152fba1d284b2dcd00a4efc205cc63", "precision": "float16", "params": 6.738, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:45:12Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "DevQuasar/coma-7B-v0.1", + "base_model": "meta-llama/Llama-2-7b-chat-hf", + "revision": "8358359ac3152fba1d284b2dcd00a4efc205cc63", + "precision": "float16", + "params": 6.738, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:45:12Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1_eval_request_False_float16_Original.json b/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1_eval_request_False_float16_Original.json index 7f8e4de5476ad666a59616ec57842357c65a0a21..733b0c5e2b24f8b085bbc23d13bedca3032d2f1f 100644 --- a/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1_eval_request_False_float16_Original.json +++ b/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1", "base_model": "", "revision": "47239ced888d8be13ff0423bb5148693de2add5a", "precision": "float16", "params": 8.031, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:14:23Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.1", + "base_model": "", + "revision": "47239ced888d8be13ff0423bb5148693de2add5a", + "precision": "float16", + "params": 8.031, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:14:23Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3_eval_request_False_bfloat16_Original.json b/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3_eval_request_False_bfloat16_Original.json index 4c5a0b0a05f80547fc0b19ca3e9539441382e2da..1f35a8f5244179309b77848e90b6b1bd1519c110 100644 --- a/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3_eval_request_False_bfloat16_Original.json +++ b/Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3", "base_model": "", "revision": "cf29b8b484a909132e3a1f85ce891d28347c0d13", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:57:58Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3", + "base_model": "", + "revision": "cf29b8b484a909132e3a1f85ce891d28347c0d13", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:57:58Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/FallenMerick/Chewy-Lemon-Cookie-11B_eval_request_False_bfloat16_Original.json b/FallenMerick/Chewy-Lemon-Cookie-11B_eval_request_False_bfloat16_Original.json index d1c7f8c967ce9ecbc5baa57c521dde72a200f822..ef39b407381feb4f65ad7277ba9cd7542b46a525 100644 --- a/FallenMerick/Chewy-Lemon-Cookie-11B_eval_request_False_bfloat16_Original.json +++ b/FallenMerick/Chewy-Lemon-Cookie-11B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "FallenMerick/Chewy-Lemon-Cookie-11B", "base_model": "", "revision": "0f5d0d6d218b3ef034f58eba32d6fe7ac4c237ae", "precision": "bfloat16", "params": 10.732, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T23:15:20Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "FallenMerick/Chewy-Lemon-Cookie-11B", + "base_model": "", + "revision": "0f5d0d6d218b3ef034f58eba32d6fe7ac4c237ae", + "precision": "bfloat16", + "params": 10.732, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T23:15:20Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/HiroseKoichi/Llama-Salad-4x8B-V3_eval_request_False_bfloat16_Original.json b/HiroseKoichi/Llama-Salad-4x8B-V3_eval_request_False_bfloat16_Original.json index 48722080a5416b370c28d77e3b671f1633e8ca6a..ca1fe8593553e6b096364c97292e70a0fa91b6cb 100644 --- a/HiroseKoichi/Llama-Salad-4x8B-V3_eval_request_False_bfloat16_Original.json +++ b/HiroseKoichi/Llama-Salad-4x8B-V3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "HiroseKoichi/Llama-Salad-4x8B-V3", "base_model": "", "revision": "a343915429779efbd1478f01ba1f7fd9d8d226c0", "precision": "bfloat16", "params": 24.942, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:36:42Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "HiroseKoichi/Llama-Salad-4x8B-V3", + "base_model": "", + "revision": "a343915429779efbd1478f01ba1f7fd9d8d226c0", + "precision": "bfloat16", + "params": 24.942, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:36:42Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Josephgflowers/Cinder-Phi-2-V1-F16-gguf_eval_request_False_float16_Original.json b/Josephgflowers/Cinder-Phi-2-V1-F16-gguf_eval_request_False_float16_Original.json index f7f362c85858d7dd9194926057a3a5f322a0026d..d469e27fdc1b42bee70ac1f1a42c74016d93f352 100644 --- a/Josephgflowers/Cinder-Phi-2-V1-F16-gguf_eval_request_False_float16_Original.json +++ b/Josephgflowers/Cinder-Phi-2-V1-F16-gguf_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "Josephgflowers/Cinder-Phi-2-V1-F16-gguf", "base_model": "", "revision": "85629ec9b18efee31d07630664e7a3815121badf", "precision": "float16", "params": 2.78, "architectures": "PhiForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:45:09Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Josephgflowers/Cinder-Phi-2-V1-F16-gguf", + "base_model": "", + "revision": "85629ec9b18efee31d07630664e7a3815121badf", + "precision": "float16", + "params": 2.78, + "architectures": "PhiForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:45:09Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Josephgflowers/TinyLlama-Cinder-Agent-v1_eval_request_False_float16_Original.json b/Josephgflowers/TinyLlama-Cinder-Agent-v1_eval_request_False_float16_Original.json index 599a685e51b7e35d2a89e134db4cc8cb51540de8..9bbf61f37279578e823c717382795623ecd4e14b 100644 --- a/Josephgflowers/TinyLlama-Cinder-Agent-v1_eval_request_False_float16_Original.json +++ b/Josephgflowers/TinyLlama-Cinder-Agent-v1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "Josephgflowers/TinyLlama-Cinder-Agent-v1", "base_model": "", "revision": "a9cd8b48bfe30f29bb1f819213da9a4c41eee67f", "precision": "float16", "params": 1.1, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:40:30Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Josephgflowers/TinyLlama-Cinder-Agent-v1", + "base_model": "", + "revision": "a9cd8b48bfe30f29bb1f819213da9a4c41eee67f", + "precision": "float16", + "params": 1.1, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:40:30Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/LLM360/K2-Chat_eval_request_False_float32_Original.json b/LLM360/K2-Chat_eval_request_False_float32_Original.json index 397b4a4c22f098663f4ec051e43af5ca8f30926f..c2ab00eeb9623d2173136586e5191bc8c9a2b314 100644 --- a/LLM360/K2-Chat_eval_request_False_float32_Original.json +++ b/LLM360/K2-Chat_eval_request_False_float32_Original.json @@ -6,10 +6,10 @@ "params": 65.286, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-12T12:07:30Z", "model_type": "chat", "job_id": "5757327", "job_start_time": "2024-06-15T12:56:23.549769", "use_chat_template": true -} \ No newline at end of file +} diff --git a/LLM360/K2_eval_request_False_float16_Original.json b/LLM360/K2_eval_request_False_float16_Original.json index f101a81fe694281e6451f820fee6d05c98b9b10a..09c7c2933acad9a03dc2be5d3eaa698236e15e61 100644 --- a/LLM360/K2_eval_request_False_float16_Original.json +++ b/LLM360/K2_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "LLM360/K2", "base_model": "", "revision": "49d159b6f2b64d562e745f0ff06e65b9a4c28ead", "precision": "float16", "params": 65.286, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:18:19Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "LLM360/K2", + "base_model": "", + "revision": "49d159b6f2b64d562e745f0ff06e65b9a4c28ead", + "precision": "float16", + "params": 65.286, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:18:19Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/Locutusque/Llama-3-NeuralHercules-5.0-8B_eval_request_False_bfloat16_Original.json b/Locutusque/Llama-3-NeuralHercules-5.0-8B_eval_request_False_bfloat16_Original.json index 49953bc3044f5dc619583819bd0b74ffb89cd208..0cf297fa4f46023dcca9ef463836bd017f59f1a2 100644 --- a/Locutusque/Llama-3-NeuralHercules-5.0-8B_eval_request_False_bfloat16_Original.json +++ b/Locutusque/Llama-3-NeuralHercules-5.0-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Locutusque/Llama-3-NeuralHercules-5.0-8B", "base_model": "", "revision": "2bbb675e592a1772f2389fe2d58a5b610d479d94", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:43:56Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Locutusque/Llama-3-NeuralHercules-5.0-8B", + "base_model": "", + "revision": "2bbb675e592a1772f2389fe2d58a5b610d479d94", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:43:56Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Locutusque/Llama-3-Yggdrasil-2.0-8B_eval_request_False_bfloat16_Original.json b/Locutusque/Llama-3-Yggdrasil-2.0-8B_eval_request_False_bfloat16_Original.json index 26618979ed877404e28e5c95a904d38f42c382fb..3e2cdef395e16bf90775a89aace8101226025079 100644 --- a/Locutusque/Llama-3-Yggdrasil-2.0-8B_eval_request_False_bfloat16_Original.json +++ b/Locutusque/Llama-3-Yggdrasil-2.0-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Locutusque/Llama-3-Yggdrasil-2.0-8B", "base_model": "", "revision": "ec2329946ccc81a7c1ae36210728f717bc4f01d8", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:47:48Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Locutusque/Llama-3-Yggdrasil-2.0-8B", + "base_model": "", + "revision": "ec2329946ccc81a7c1ae36210728f717bc4f01d8", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:47:48Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Goku-8x22B-v0.1_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Goku-8x22B-v0.1_eval_request_False_bfloat16_Original.json index eab18c32d7cfba9fb15c1d00a4285c3c45f86587..5f6e4070402ac4129266b2c160358289290f0d72 100644 --- a/MaziyarPanahi/Goku-8x22B-v0.1_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Goku-8x22B-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Goku-8x22B-v0.1", "base_model": "", "revision": "f75ebf5ceb71046184aaf969cb85daaac320e4a7", "precision": "bfloat16", "params": 140.621, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:28:47Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Goku-8x22B-v0.1", + "base_model": "", + "revision": "f75ebf5ceb71046184aaf969cb85daaac320e4a7", + "precision": "bfloat16", + "params": 140.621, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:28:47Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Goku-8x22B-v0.2_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Goku-8x22B-v0.2_eval_request_False_bfloat16_Original.json index c008b8d357e6b8cdf5217e7306b5585cf711b295..e0ea69dc484af2b1b1a9ead06f37aa1d4a63eb13 100644 --- a/MaziyarPanahi/Goku-8x22B-v0.2_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Goku-8x22B-v0.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Goku-8x22B-v0.2", "base_model": "", "revision": "215a4d7c13ac4b69540300d37a231f61968acaa6", "precision": "bfloat16", "params": 140.621, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:29:06Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Goku-8x22B-v0.2", + "base_model": "", + "revision": "215a4d7c13ac4b69540300d37a231f61968acaa6", + "precision": "bfloat16", + "params": 140.621, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:29:06Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1_eval_request_False_bfloat16_Original.json index e986bf3509482161dcbfab3f3b257d847794971c..5e5efb4b0189ac5ac1dba086b1a2b33f76f180a3 100644 --- a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1", "base_model": "", "revision": "a584e3f14a5fdddced8def6e4bbe93da83e4b971", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:22:17Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.1", + "base_model": "", + "revision": "a584e3f14a5fdddced8def6e4bbe93da83e4b971", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:22:17Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2_eval_request_False_bfloat16_Original.json index de39b5e5e2ee9f07ecf629d797f489cf078498ee..c401881b682aa17ffcfb7322d0970f1e5f78be9b 100644 --- a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2", "base_model": "", "revision": "95366b974baedee4d95c1e841bc3d15e94753804", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:21:18Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.2", + "base_model": "", + "revision": "95366b974baedee4d95c1e841bc3d15e94753804", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:21:18Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3_eval_request_False_bfloat16_Original.json index a88dd7741f403ca3a51810a52cb5fc494bb33540..ab7c1ebb2dbcdd43eea4ab550700f8b74c953216 100644 --- a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3", "base_model": "", "revision": "a1b2e5df828ab055d02cb2cbfb357557ce37881d", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:21:50Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.3", + "base_model": "", + "revision": "a1b2e5df828ab055d02cb2cbfb357557ce37881d", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:21:50Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4_eval_request_False_bfloat16_Original.json index 9f158b0626f54dfeb880eab3b4d9ab591075907b..f4038a7b1d0d95e2c410efad83e7d689df939ab0 100644 --- a/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4", "base_model": "", "revision": "cb03e4d810b82d86e7cb01ab146bade09a5d06d1", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:48Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-70B-Instruct-DPO-v0.4", + "base_model": "", + "revision": "cb03e4d810b82d86e7cb01ab146bade09a5d06d1", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:48Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-70B-Instruct-v0.1_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-70B-Instruct-v0.1_eval_request_False_bfloat16_Original.json index 5dbf6c513d8fd7dfc67b18e010a87efaa55f4eba..d3a14f73594405c0359d7f8764984f6e185287b7 100644 --- a/MaziyarPanahi/Llama-3-70B-Instruct-v0.1_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-70B-Instruct-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-70B-Instruct-v0.1", "base_model": "", "revision": "6db1cb4256525fc5429734ddc0eb941d08d0be30", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:08Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-70B-Instruct-v0.1", + "base_model": "", + "revision": "6db1cb4256525fc5429734ddc0eb941d08d0be30", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:08Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-8B-Instruct-v0.10_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-8B-Instruct-v0.10_eval_request_False_bfloat16_Original.json index 8d38d3e055cdf900ec364fa313596bca025eb9bf..38b78693e3abc5b84fb0338e0968bc0a241acdee 100644 --- a/MaziyarPanahi/Llama-3-8B-Instruct-v0.10_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-8B-Instruct-v0.10_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.10", "base_model": "", "revision": "4411eb9f6f5e4c462a6bdbc64c26dcc123100b66", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:41:58Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.10", + "base_model": "", + "revision": "4411eb9f6f5e4c462a6bdbc64c26dcc123100b66", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:41:58Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-8B-Instruct-v0.8_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-8B-Instruct-v0.8_eval_request_False_bfloat16_Original.json index 004b3a205a01652e0afd7e8ba488a3dd0b2e3d60..8752ad069c3df90457ef8f0caaeb7a3907c13f89 100644 --- a/MaziyarPanahi/Llama-3-8B-Instruct-v0.8_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-8B-Instruct-v0.8_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.8", "base_model": "", "revision": "94d222b8447b600b9836da4036df9490b59fe966", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:41:21Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.8", + "base_model": "", + "revision": "94d222b8447b600b9836da4036df9490b59fe966", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:41:21Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Llama-3-8B-Instruct-v0.9_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Llama-3-8B-Instruct-v0.9_eval_request_False_bfloat16_Original.json index 1f75ffa20d1393b04dd86ff393116555f0679c7a..716816e421373acd7d2cd3ffd9b2a820a52ef833 100644 --- a/MaziyarPanahi/Llama-3-8B-Instruct-v0.9_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Llama-3-8B-Instruct-v0.9_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.9", "base_model": "", "revision": "ddf91fdc0a3ab5e5d76864f1c4cf44e5adacd565", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:41:38Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Llama-3-8B-Instruct-v0.9", + "base_model": "", + "revision": "ddf91fdc0a3ab5e5d76864f1c4cf44e5adacd565", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:41:38Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1_eval_request_False_bfloat16_Original.json index 3d3c97f3fd660e1f96bb82ff5fa3991d10d1108f..ca10f30992c46711c4401603c51f6624ba8247bd 100644 --- a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1", "base_model": "", "revision": "6764c79badacba5fa3584d2d2593d762caa1d17d", "precision": "bfloat16", "params": 3.821, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:30:49Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.1", + "base_model": "", + "revision": "6764c79badacba5fa3584d2d2593d762caa1d17d", + "precision": "bfloat16", + "params": 3.821, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:30:49Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2_eval_request_False_bfloat16_Original.json index f5522ea19d1b9b1fbae8ad82b92a931c22610998..7e26a3b139d9fe6e395ecd57b802601edd49a72f 100644 --- a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2", "base_model": "", "revision": "c0a366a4c01d7e724ceba7e2f2c19251983423fe", "precision": "bfloat16", "params": 3.821, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:31:11Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.2", + "base_model": "", + "revision": "c0a366a4c01d7e724ceba7e2f2c19251983423fe", + "precision": "bfloat16", + "params": 3.821, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:31:11Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3_eval_request_False_bfloat16_Original.json index 6c9a7c7095c918f1a4b956f5979113756f055602..311f99764f846a79b791a7f299db526544ccbfa6 100644 --- a/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3", "base_model": "", "revision": "e1f70c3724c728aadd1c7c1bb279487494f7059e", "precision": "bfloat16", "params": 3.821, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:32:08Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Phi-3-mini-4k-instruct-v0.3", + "base_model": "", + "revision": "e1f70c3724c728aadd1c7c1bb279487494f7059e", + "precision": "bfloat16", + "params": 3.821, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:32:08Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/MaziyarPanahi/Qwen2-72B-Instruct-v0.1_eval_request_False_bfloat16_Original.json b/MaziyarPanahi/Qwen2-72B-Instruct-v0.1_eval_request_False_bfloat16_Original.json index 6e3b1e7c5d5390080cb218168a07570a8e3803f2..093f47e853165611718dd837bb47d5a401c546cb 100644 --- a/MaziyarPanahi/Qwen2-72B-Instruct-v0.1_eval_request_False_bfloat16_Original.json +++ b/MaziyarPanahi/Qwen2-72B-Instruct-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "MaziyarPanahi/Qwen2-72B-Instruct-v0.1", "base_model": "", "revision": "0369c39770f45f2464587918f2dbdb8449ea3a0d", "precision": "bfloat16", "params": 72.699, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:23:39Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "MaziyarPanahi/Qwen2-72B-Instruct-v0.1", + "base_model": "", + "revision": "0369c39770f45f2464587918f2dbdb8449ea3a0d", + "precision": "bfloat16", + "params": 72.699, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:23:39Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Naveenpoliasetty/llama3-8B-V2_eval_request_False_float16_Original.json b/Naveenpoliasetty/llama3-8B-V2_eval_request_False_float16_Original.json index ad35fde6e27ee0f6b7ff64733708a09af1e033ca..5c21a7c8870f5849d5548189d30069eba0bb90b0 100644 --- a/Naveenpoliasetty/llama3-8B-V2_eval_request_False_float16_Original.json +++ b/Naveenpoliasetty/llama3-8B-V2_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "Naveenpoliasetty/llama3-8B-V2", "base_model": "", "revision": "e0458381d02bc411b9e576796d185f23dcc11f71", "precision": "float16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:54:47Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "Naveenpoliasetty/llama3-8B-V2", + "base_model": "", + "revision": "e0458381d02bc411b9e576796d185f23dcc11f71", + "precision": "float16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:54:47Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/NeverSleep/CausalLM-RP-34B_eval_request_False_float16_Original.json b/NeverSleep/CausalLM-RP-34B_eval_request_False_float16_Original.json index 8f3f89613cce63825a2b36a02cd7ede34853a9e1..3214c09875cc5adb34ff36aa15c6a046fabfcde7 100644 --- a/NeverSleep/CausalLM-RP-34B_eval_request_False_float16_Original.json +++ b/NeverSleep/CausalLM-RP-34B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "NeverSleep/CausalLM-RP-34B", "base_model": "", "revision": "e2a033646231bd947a3948d3aac198d34d04ea38", "precision": "float16", "params": 34.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:57:33Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "NeverSleep/CausalLM-RP-34B", + "base_model": "", + "revision": "e2a033646231bd947a3948d3aac198d34d04ea38", + "precision": "float16", + "params": 34, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:57:33Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/Nitral-AI/Hathor_Stable-v0.2-L3-8B_eval_request_False_bfloat16_Original.json b/Nitral-AI/Hathor_Stable-v0.2-L3-8B_eval_request_False_bfloat16_Original.json index 0c1a88440028eeee501c72f7825f8c3a1aedfbe7..8f93b30637d1e4f56fd7adc3b0af8fb58cf5a27c 100644 --- a/Nitral-AI/Hathor_Stable-v0.2-L3-8B_eval_request_False_bfloat16_Original.json +++ b/Nitral-AI/Hathor_Stable-v0.2-L3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Nitral-AI/Hathor_Stable-v0.2-L3-8B", "base_model": "", "revision": "6109d7624cadc4ddba5f23ed0abe99f6b29c9139", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:59:23Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Nitral-AI/Hathor_Stable-v0.2-L3-8B", + "base_model": "", + "revision": "6109d7624cadc4ddba5f23ed0abe99f6b29c9139", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:59:23Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Nitral-AI/Poppy_Porpoise-0.72-L3-8B_eval_request_False_bfloat16_Original.json b/Nitral-AI/Poppy_Porpoise-0.72-L3-8B_eval_request_False_bfloat16_Original.json index a54edce349aebe0010a3d98f1ae99d1baea2dd93..4bad44b9c5e2d13f9f2c4fd746d71776b3c7bceb 100644 --- a/Nitral-AI/Poppy_Porpoise-0.72-L3-8B_eval_request_False_bfloat16_Original.json +++ b/Nitral-AI/Poppy_Porpoise-0.72-L3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Nitral-AI/Poppy_Porpoise-0.72-L3-8B", "base_model": "", "revision": "022d166a2aa5d323cc6d683728e867694cefa513", "precision": "bfloat16", "params": 8.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:05:27Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Nitral-AI/Poppy_Porpoise-0.72-L3-8B", + "base_model": "", + "revision": "022d166a2aa5d323cc6d683728e867694cefa513", + "precision": "bfloat16", + "params": 8, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:05:27Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/NousResearch/Hermes-2-Theta-Llama-3-70B_eval_request_False_bfloat16_Original.json b/NousResearch/Hermes-2-Theta-Llama-3-70B_eval_request_False_bfloat16_Original.json index 2f6191b155461fdc452aaaff683b24d5cb944c4f..70a41204e99731aa19a391b1b19fca40432ed7b1 100644 --- a/NousResearch/Hermes-2-Theta-Llama-3-70B_eval_request_False_bfloat16_Original.json +++ b/NousResearch/Hermes-2-Theta-Llama-3-70B_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T15:14:27Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7206065", "job_start_time": "2024-06-26T16:41:39.162018", "use_chat_template": true -} \ No newline at end of file +} diff --git a/NousResearch/Yarn-Llama-2-13b-128k_eval_request_False_bfloat16_Original.json b/NousResearch/Yarn-Llama-2-13b-128k_eval_request_False_bfloat16_Original.json index 6f9c349bbf6c715dd1a01ad7eaab18bdb948ed6e..cb5d4aaedd5d6cad42e7c1a0685e931e9bd4eace 100644 --- a/NousResearch/Yarn-Llama-2-13b-128k_eval_request_False_bfloat16_Original.json +++ b/NousResearch/Yarn-Llama-2-13b-128k_eval_request_False_bfloat16_Original.json @@ -3,13 +3,13 @@ "base_model": null, "revision": "4e3e87a067f64f8814c83dd5e3bad92dcf8a2391", "precision": "bfloat16", - "params": 13.0, + "params": 13, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-13T18:09:47Z", "model_type": "pretrained", "job_id": "6660890", "job_start_time": "2024-06-22T18:36:33.519447", "use_chat_template": false -} \ No newline at end of file +} diff --git a/NucleusAI/nucleus-22B-token-500B_eval_request_False_bfloat16_Original.json b/NucleusAI/nucleus-22B-token-500B_eval_request_False_bfloat16_Original.json index 0c3b640073cc0585b36a2ae1b1745e975cab2253..5e02885cb71f004cb9572d3a4b5025d00178ba4d 100644 --- a/NucleusAI/nucleus-22B-token-500B_eval_request_False_bfloat16_Original.json +++ b/NucleusAI/nucleus-22B-token-500B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "NucleusAI/nucleus-22B-token-500B", "base_model": "", "revision": "49bb1a47c0d32b4bfa6630a4eff04a857adcd4ca", "precision": "bfloat16", "params": 21.828, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:29:04Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "NucleusAI/nucleus-22B-token-500B", + "base_model": "", + "revision": "49bb1a47c0d32b4bfa6630a4eff04a857adcd4ca", + "precision": "bfloat16", + "params": 21.828, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:29:04Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/OpenBuddy/openbuddy-deepseek-67b-v18.1-4k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-deepseek-67b-v18.1-4k_eval_request_False_bfloat16_Original.json index 727e67590548cc739047eb24fba55da3737c7b1a..46faee8fc5c3b27258fe2c0d39afa6f9e0ed3ac0 100644 --- a/OpenBuddy/openbuddy-deepseek-67b-v18.1-4k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-deepseek-67b-v18.1-4k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-deepseek-67b-v18.1-4k", "base_model": "", "revision": "897fd7543d9b2abe04b9a5b92db79c090bc169ce", "precision": "bfloat16", "params": 67.425, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:30:09Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-deepseek-67b-v18.1-4k", + "base_model": "", + "revision": "897fd7543d9b2abe04b9a5b92db79c090bc169ce", + "precision": "bfloat16", + "params": 67.425, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:30:09Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-llama3-70b-v21.2-32k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-llama3-70b-v21.2-32k_eval_request_False_bfloat16_Original.json index f380d43b4d846c4f41fb5831213478e475e89866..8a0c4e1e76d195a0d9cc5363598518a645b9e138 100644 --- a/OpenBuddy/openbuddy-llama3-70b-v21.2-32k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-llama3-70b-v21.2-32k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-llama3-70b-v21.2-32k", "base_model": "", "revision": "e79a2f16c052fc76eeafb5b51d16261b2b981d0f", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:28:17Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-llama3-70b-v21.2-32k", + "base_model": "", + "revision": "e79a2f16c052fc76eeafb5b51d16261b2b981d0f", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:28:17Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-llama3-8b-v21.1-8k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-llama3-8b-v21.1-8k_eval_request_False_bfloat16_Original.json index bb4d0e11fc4558ad17da98672a2dfe0f4f5c3cdd..c8b459a375680ebecf3f3db8786964ab24544957 100644 --- a/OpenBuddy/openbuddy-llama3-8b-v21.1-8k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-llama3-8b-v21.1-8k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-llama3-8b-v21.1-8k", "base_model": "", "revision": "658508bce03ccd61cea9657e0357bd4cd10503ba", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:33:43Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-llama3-8b-v21.1-8k", + "base_model": "", + "revision": "658508bce03ccd61cea9657e0357bd4cd10503ba", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:33:43Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-llama3-8b-v21.2-32k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-llama3-8b-v21.2-32k_eval_request_False_bfloat16_Original.json index 766488f0a3a8bb879c8bd1293bbc677a70e3361f..15d4f86ead6dc7dc2eda5b3d6823bab98620b56d 100644 --- a/OpenBuddy/openbuddy-llama3-8b-v21.2-32k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-llama3-8b-v21.2-32k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-llama3-8b-v21.2-32k", "base_model": "", "revision": "f3ea2dec2533a3dd97df32db2376b17875cafda2", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:26:51Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-llama3-8b-v21.2-32k", + "base_model": "", + "revision": "f3ea2dec2533a3dd97df32db2376b17875cafda2", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:26:51Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k_eval_request_False_bfloat16_Original.json index 821dbdc7edde71adccd717b619b37cdc1c8030f5..860a06420586e1deee8bc25cf1d65dc5ff61105c 100644 --- a/OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k", "base_model": "", "revision": "98596b6731058cc9cca85f3b8ac9077342cb60ae", "precision": "bfloat16", "params": 46.741, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:31:31Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-mixtral-7bx8-v18.1-32k", + "base_model": "", + "revision": "98596b6731058cc9cca85f3b8ac9077342cb60ae", + "precision": "bfloat16", + "params": 46.741, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:31:31Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16_eval_request_False_float16_Original.json b/OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16_eval_request_False_float16_Original.json index 5b0aea2d332e88c6868f493bdbe7f54a38eab968..e4348876f8feccc93ad1e1ea8a38640dc0705911 100644 --- a/OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16_eval_request_False_float16_Original.json +++ b/OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16", "base_model": "", "revision": "abcc0e98eb7f82268a379c5ccf1fd043a07d3460", "precision": "float16", "params": 34.393, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:27:23Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-yi1.5-34b-v21.6-32k-fp16", + "base_model": "", + "revision": "abcc0e98eb7f82268a379c5ccf1fd043a07d3460", + "precision": "float16", + "params": 34.393, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:27:23Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-zero-3b-v21.2-32k_eval_request_False_bfloat16_Original.json b/OpenBuddy/openbuddy-zero-3b-v21.2-32k_eval_request_False_bfloat16_Original.json index 763eac7b9e871c948a864a1d63f70062199a2e5d..83c4d875bb98fdfd081a1d8ec6a7de6ce5193fec 100644 --- a/OpenBuddy/openbuddy-zero-3b-v21.2-32k_eval_request_False_bfloat16_Original.json +++ b/OpenBuddy/openbuddy-zero-3b-v21.2-32k_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-zero-3b-v21.2-32k", "base_model": "", "revision": "74e1d168c5e917219d668d1483f6355dd0464a31", "precision": "bfloat16", "params": 4.769, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:30:53Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-zero-3b-v21.2-32k", + "base_model": "", + "revision": "74e1d168c5e917219d668d1483f6355dd0464a31", + "precision": "bfloat16", + "params": 4.769, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:30:53Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/OpenBuddy/openbuddy-zero-56b-v21.2-32k_eval_request_False_float16_Original.json b/OpenBuddy/openbuddy-zero-56b-v21.2-32k_eval_request_False_float16_Original.json index 39e3aa557c40dd58f5c9e9895f04188ff3d6eebd..02f00c2e01a0b27e2d6597ba13319c4c37508528 100644 --- a/OpenBuddy/openbuddy-zero-56b-v21.2-32k_eval_request_False_float16_Original.json +++ b/OpenBuddy/openbuddy-zero-56b-v21.2-32k_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "OpenBuddy/openbuddy-zero-56b-v21.2-32k", "base_model": "", "revision": "c7a1a4a6e798f75d1d3219ab9ff9f2692e29f7d5", "precision": "float16", "params": 56.707, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:43:45Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "OpenBuddy/openbuddy-zero-56b-v21.2-32k", + "base_model": "", + "revision": "c7a1a4a6e798f75d1d3219ab9ff9f2692e29f7d5", + "precision": "float16", + "params": 56.707, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:43:45Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Qwen/Qwen1.5-32B-Chat_eval_request_False_bfloat16_Original.json b/Qwen/Qwen1.5-32B-Chat_eval_request_False_bfloat16_Original.json index 8aa4e14617d50b714db34964046ef0b12be57dc1..5e11dce96c118297ab38f210a4cb4897e2db44c6 100644 --- a/Qwen/Qwen1.5-32B-Chat_eval_request_False_bfloat16_Original.json +++ b/Qwen/Qwen1.5-32B-Chat_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 32.512, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-12T09:20:33Z", "model_type": "chat", "job_id": "6693759", "job_start_time": "2024-06-23T04:22:38.528472", "use_chat_template": true -} \ No newline at end of file +} diff --git a/Qwen/Qwen1.5-32B_eval_request_False_bfloat16_Original.json b/Qwen/Qwen1.5-32B_eval_request_False_bfloat16_Original.json index 6b1edd3f8d7eb1a5a4c92d8a34e101a643a124d6..0fa6ec86deea240fe77456ec4643fe19f57ef36d 100644 --- a/Qwen/Qwen1.5-32B_eval_request_False_bfloat16_Original.json +++ b/Qwen/Qwen1.5-32B_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 32.512, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-13T17:56:14Z", "model_type": "pretrained", "job_id": "6676113", "job_start_time": "2024-06-22T21:35:25.308158", "use_chat_template": false -} \ No newline at end of file +} diff --git a/Qwen/Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json b/Qwen/Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json index 21a69c2baf5d94c7f5cc5a1004fc9ffc079efcd1..b36ff49ddc0eee95550dc1c965a2d5f2cb825140 100644 --- a/Qwen/Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json +++ b/Qwen/Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Qwen/Qwen2-72B-Instruct", "base_model": "", "revision": "1af63c698f59c4235668ec9c1395468cb7cd7e79", "precision": "bfloat16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:26:20Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "Qwen/Qwen2-72B-Instruct", + "base_model": "", + "revision": "1af63c698f59c4235668ec9c1395468cb7cd7e79", + "precision": "bfloat16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:26:20Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/Qwen/Qwen2-72B_eval_request_False_bfloat16_Original.json b/Qwen/Qwen2-72B_eval_request_False_bfloat16_Original.json index 84717413b74516260d0c9c8c4e6f628e4c0d3ddc..7bd02a6f8d889369646471de5401b43378a606c4 100644 --- a/Qwen/Qwen2-72B_eval_request_False_bfloat16_Original.json +++ b/Qwen/Qwen2-72B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Qwen/Qwen2-72B", "base_model": "", "revision": "87993795c78576318087f70b43fbf530eb7789e7", "precision": "bfloat16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "FINISHED", "submitted_time": "2024-06-26T14:53:03Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "Qwen/Qwen2-72B", + "base_model": "", + "revision": "87993795c78576318087f70b43fbf530eb7789e7", + "precision": "bfloat16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:53:03Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/Qwen/Qwen2-72B_eval_request_False_float16_Original.json b/Qwen/Qwen2-72B_eval_request_False_float16_Original.json index 7d464f2898334049d44879ef2996ecd76c00952e..fb62324a4683aaa72eda759f5d1afdde4d2691d0 100644 --- a/Qwen/Qwen2-72B_eval_request_False_float16_Original.json +++ b/Qwen/Qwen2-72B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "Qwen/Qwen2-72B", "base_model": "", "revision": "87993795c78576318087f70b43fbf530eb7789e7", "precision": "float16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T13:58:36Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "Qwen/Qwen2-72B", + "base_model": "", + "revision": "87993795c78576318087f70b43fbf530eb7789e7", + "precision": "float16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T13:58:36Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/RLHFlow/LLaMA3-iterative-DPO-final_eval_request_False_bfloat16_Original.json b/RLHFlow/LLaMA3-iterative-DPO-final_eval_request_False_bfloat16_Original.json index 61bbda73d449ec0afa622767da918c4704123976..18ff8669cc52603e96d654a4dfecc2fd07fc4413 100644 --- a/RLHFlow/LLaMA3-iterative-DPO-final_eval_request_False_bfloat16_Original.json +++ b/RLHFlow/LLaMA3-iterative-DPO-final_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T15:00:01Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7215719", "job_start_time": "2024-06-26T23:30:42.096525", "use_chat_template": true -} \ No newline at end of file +} diff --git a/Replete-AI/Replete-Coder-Llama3-8B_eval_request_False_bfloat16_Original.json b/Replete-AI/Replete-Coder-Llama3-8B_eval_request_False_bfloat16_Original.json index 38c98c3daf046acc6dd998a09da750dc2cf14c51..c4632f1aec83f4aaf3b41c4fa2266446c9c75019 100644 --- a/Replete-AI/Replete-Coder-Llama3-8B_eval_request_False_bfloat16_Original.json +++ b/Replete-AI/Replete-Coder-Llama3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Replete-AI/Replete-Coder-Llama3-8B", "base_model": "", "revision": "2aca75c53e7eb2f523889ab1a279e349b8f1b0e8", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:33:22Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Replete-AI/Replete-Coder-Llama3-8B", + "base_model": "", + "revision": "2aca75c53e7eb2f523889ab1a279e349b8f1b0e8", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:33:22Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Replete-AI/Replete-Coder-Qwen2-1.5b_eval_request_False_bfloat16_Original.json b/Replete-AI/Replete-Coder-Qwen2-1.5b_eval_request_False_bfloat16_Original.json index 2d086b72be7f68e96a18edd7c1359a76d0c8c64a..20b3346f0fe872815ec81a65d35793b26888613b 100644 --- a/Replete-AI/Replete-Coder-Qwen2-1.5b_eval_request_False_bfloat16_Original.json +++ b/Replete-AI/Replete-Coder-Qwen2-1.5b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Replete-AI/Replete-Coder-Qwen2-1.5b", "base_model": "", "revision": "86fcccbf921b7eb8a4d348e4a3cde0beb63d6626", "precision": "bfloat16", "params": 1.544, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:33:56Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Replete-AI/Replete-Coder-Qwen2-1.5b", + "base_model": "", + "revision": "86fcccbf921b7eb8a4d348e4a3cde0beb63d6626", + "precision": "bfloat16", + "params": 1.544, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:33:56Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Sao10K/L3-70B-Euryale-v2.1_eval_request_False_float16_Original.json b/Sao10K/L3-70B-Euryale-v2.1_eval_request_False_float16_Original.json index ed8bd70fb0a63796448453409e750607ddb1b833..b49b930865cfe1ca9aecb71b7d3958dfbdd6e8e8 100644 --- a/Sao10K/L3-70B-Euryale-v2.1_eval_request_False_float16_Original.json +++ b/Sao10K/L3-70B-Euryale-v2.1_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T15:05:07Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7205746", "job_start_time": "2024-06-26T15:16:24.819249", "use_chat_template": true -} \ No newline at end of file +} diff --git a/Sao10K/L3-8B-Stheno-v3.3-32K_eval_request_False_bfloat16_Original.json b/Sao10K/L3-8B-Stheno-v3.3-32K_eval_request_False_bfloat16_Original.json index a4fcffd2e02b131ac78c671a607f51759e45dc99..c0daa5839a46d6608e3d77d4df1ba9633ae44cbb 100644 --- a/Sao10K/L3-8B-Stheno-v3.3-32K_eval_request_False_bfloat16_Original.json +++ b/Sao10K/L3-8B-Stheno-v3.3-32K_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:58:11Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7205447", "job_start_time": "2024-06-26T15:00:23.923841", "use_chat_template": true -} \ No newline at end of file +} diff --git a/Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json b/Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json index 04d83f9cba304857c085cad96899d75199ff16b2..615e3082d9ed7042c47b28e23a38a993327505bc 100644 --- a/Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json +++ b/Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo", "base_model": "", "revision": "c4e37d0923641867e4822c4d4181e6d4397041db", "precision": "bfloat16", "params": 3.723, "architectures": "Phi3Model", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:28:58Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Syed-Hasan-8503/Phi-3-mini-128K-instruct-cpo-simpo", + "base_model": "", + "revision": "c4e37d0923641867e4822c4d4181e6d4397041db", + "precision": "bfloat16", + "params": 3.723, + "architectures": "Phi3Model", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:28:58Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json b/Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json index 2dd9ed3a2bb630ce054bc5b8a92cfb730bf8eae4..4190b1bc6130c82b4207f96c9fb4c46500f4996a 100644 --- a/Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json +++ b/Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo", "base_model": "", "revision": "4c0f5ef387f5e81ffba80e00550140c2febeecd9", "precision": "bfloat16", "params": 3.821, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:22:07Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Syed-Hasan-8503/Phi-3-mini-4K-instruct-cpo-simpo", + "base_model": "", + "revision": "4c0f5ef387f5e81ffba80e00550140c2febeecd9", + "precision": "bfloat16", + "params": 3.821, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:22:07Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/TIGER-Lab/MAmmoTH2-8x7B-Plus_eval_request_False_bfloat16_Original.json b/TIGER-Lab/MAmmoTH2-8x7B-Plus_eval_request_False_bfloat16_Original.json index d20de857e39bd11d73efbcadf4e4a43019382f2d..d1930e3ffac5f22b7345a7d0baff07bdc67f4952 100644 --- a/TIGER-Lab/MAmmoTH2-8x7B-Plus_eval_request_False_bfloat16_Original.json +++ b/TIGER-Lab/MAmmoTH2-8x7B-Plus_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "TIGER-Lab/MAmmoTH2-8x7B-Plus", "base_model": "", "revision": "2874911ada99fd46688ad40516e61bc30a702e20", "precision": "bfloat16", "params": 46.703, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:11:21Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "TIGER-Lab/MAmmoTH2-8x7B-Plus", + "base_model": "", + "revision": "2874911ada99fd46688ad40516e61bc30a702e20", + "precision": "bfloat16", + "params": 46.703, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:11:21Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_bfloat16_Original.json b/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_bfloat16_Original.json index 7562c3acfefa643398ebff19afeeb58d46ebb567..49bd63c9ef106d0368a73fa9ee6cadd90cf85a77 100644 --- a/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_bfloat16_Original.json +++ b/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T18:31:20Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7214813", "job_start_time": "2024-06-26T20:10:41.147848", "use_chat_template": true -} \ No newline at end of file +} diff --git a/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_float16_Original.json b/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_float16_Original.json index abd0d8c2d7795cc2711181ef848e5f62c673de38..f1437c7085c39335704745b8e69df34ebc09b513 100644 --- a/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_float16_Original.json +++ b/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T16:09:20Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7214854", "job_start_time": "2024-06-26T20:11:42.951002", "use_chat_template": true -} \ No newline at end of file +} diff --git a/VAGOsolutions/Llama-3-SauerkrautLM-70b-Instruct_eval_request_False_bfloat16_Original.json b/VAGOsolutions/Llama-3-SauerkrautLM-70b-Instruct_eval_request_False_bfloat16_Original.json index 17756f5e1b8967a39603018848e47d6f328873fa..6c1c8f9e5b102edd524ae31a99ffd7e0c0f0c410 100644 --- a/VAGOsolutions/Llama-3-SauerkrautLM-70b-Instruct_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/Llama-3-SauerkrautLM-70b-Instruct_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:08:39Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7205835", "job_start_time": "2024-06-26T16:22:44.162400", "use_chat_template": false -} \ No newline at end of file +} diff --git a/VAGOsolutions/SauerkrautLM-1.5b_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-1.5b_eval_request_False_bfloat16_Original.json index db5315e458886a9c1fb7caaf8e9cd09a4a2a710f..4329b613607ec49c9521b2c881a7a9fb1af76982 100644 --- a/VAGOsolutions/SauerkrautLM-1.5b_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-1.5b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-1.5b", "base_model": "", "revision": "8f5170f03e6b0355dd920adc3a7e65d0417ee14e", "precision": "bfloat16", "params": 1.544, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:11:13Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-1.5b", + "base_model": "", + "revision": "8f5170f03e6b0355dd920adc3a7e65d0417ee14e", + "precision": "bfloat16", + "params": 1.544, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:11:13Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-7b-HerO_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-7b-HerO_eval_request_False_bfloat16_Original.json index 3d6610b6341ebbb755b4f997e9bff31005a94dcd..0cb9e533dc65e376a2954abebc6add873d4c6036 100644 --- a/VAGOsolutions/SauerkrautLM-7b-HerO_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-7b-HerO_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-7b-HerO", "base_model": "", "revision": "3a14b437e2f375b74de3b6923e171662133347bb", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:11:55Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-7b-HerO", + "base_model": "", + "revision": "3a14b437e2f375b74de3b6923e171662133347bb", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:11:55Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-7b-LaserChat_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-7b-LaserChat_eval_request_False_bfloat16_Original.json index 054b8e93367b2f4938a02a4b7db900c581b0c2f0..8e442ce9b4b4d7416b582acb0668ed2e9980c50b 100644 --- a/VAGOsolutions/SauerkrautLM-7b-LaserChat_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-7b-LaserChat_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-7b-LaserChat", "base_model": "", "revision": "cb759636a3d5b0768df2f43a3d3da9b17e10e7b9", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:09:32Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-7b-LaserChat", + "base_model": "", + "revision": "cb759636a3d5b0768df2f43a3d3da9b17e10e7b9", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:09:32Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-Gemma-2b_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-Gemma-2b_eval_request_False_bfloat16_Original.json index 492f036a9cba3a3fca46268a0ee98ffbe33a7b95..fa39460e7019247811917294965990cfd48ced4d 100644 --- a/VAGOsolutions/SauerkrautLM-Gemma-2b_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-Gemma-2b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-Gemma-2b", "base_model": "", "revision": "f9d5575c23da96f33ce77dea3b0776746b9469bc", "precision": "bfloat16", "params": 2.506, "architectures": "GemmaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:11:34Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-Gemma-2b", + "base_model": "", + "revision": "f9d5575c23da96f33ce77dea3b0776746b9469bc", + "precision": "bfloat16", + "params": 2.506, + "architectures": "GemmaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:11:34Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-Gemma-7b_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-Gemma-7b_eval_request_False_bfloat16_Original.json index 262c38e52669c178e5dd6bd20475126bef0a81ba..418cede04c06f3692a4762e225c77382de6a6b03 100644 --- a/VAGOsolutions/SauerkrautLM-Gemma-7b_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-Gemma-7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-Gemma-7b", "base_model": "", "revision": "4296bdabf82e900235b094e5348be03ebb0ec891", "precision": "bfloat16", "params": 8.538, "architectures": "GemmaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:10:54Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-Gemma-7b", + "base_model": "", + "revision": "4296bdabf82e900235b094e5348be03ebb0ec891", + "precision": "bfloat16", + "params": 8.538, + "architectures": "GemmaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:10:54Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct_eval_request_False_bfloat16_Original.json index f984a8a7987b322442aff2c067a35a441a4635af..620c0a425ee88bc3969060b4590de856eb7185db 100644 --- a/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct", "base_model": "", "revision": "30ed549de7d84f68b4c6cb619f73275c99af23cc", "precision": "bfloat16", "params": 46.703, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:10:24Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct", + "base_model": "", + "revision": "30ed549de7d84f68b4c6cb619f73275c99af23cc", + "precision": "bfloat16", + "params": 46.703, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:10:24Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-Qwen-32b_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-Qwen-32b_eval_request_False_bfloat16_Original.json index 2df788fb35c832e2a85359a52e113d4e39c65322..9b0bfbc69847350c40b7b83562cf13311b10fb27 100644 --- a/VAGOsolutions/SauerkrautLM-Qwen-32b_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-Qwen-32b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-Qwen-32b", "base_model": "", "revision": "f5306490ea218e74a88e7431fe2f4c2d7f16b721", "precision": "bfloat16", "params": 32.512, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:09:09Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-Qwen-32b", + "base_model": "", + "revision": "f5306490ea218e74a88e7431fe2f4c2d7f16b721", + "precision": "bfloat16", + "params": 32.512, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:09:09Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/VAGOsolutions/SauerkrautLM-SOLAR-Instruct_eval_request_False_bfloat16_Original.json b/VAGOsolutions/SauerkrautLM-SOLAR-Instruct_eval_request_False_bfloat16_Original.json index 0367672d7c6426bcde4d6d4cc6b1fd4c709f5d98..4eac3cd23d3a4915f6cae11130608eaddf19cd0c 100644 --- a/VAGOsolutions/SauerkrautLM-SOLAR-Instruct_eval_request_False_bfloat16_Original.json +++ b/VAGOsolutions/SauerkrautLM-SOLAR-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "VAGOsolutions/SauerkrautLM-SOLAR-Instruct", "base_model": "", "revision": "2665d7600ccd253728453433d2434844e6f702bd", "precision": "bfloat16", "params": 10.732, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:10:00Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "VAGOsolutions/SauerkrautLM-SOLAR-Instruct", + "base_model": "", + "revision": "2665d7600ccd253728453433d2434844e6f702bd", + "precision": "bfloat16", + "params": 10.732, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:10:00Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/ValiantLabs/Llama3-70B-Fireplace_eval_request_False_float16_Original.json b/ValiantLabs/Llama3-70B-Fireplace_eval_request_False_float16_Original.json index d5f35b6644f488454fffcb9a8592e28d6311be50..d0da5972e2b6cf603a74ade38a0db48537ca144f 100644 --- a/ValiantLabs/Llama3-70B-Fireplace_eval_request_False_float16_Original.json +++ b/ValiantLabs/Llama3-70B-Fireplace_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "ValiantLabs/Llama3-70B-Fireplace", "base_model": "", "revision": "220079e4115733991eb19c30d5480db9696a665e", "precision": "float16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:01:44Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "ValiantLabs/Llama3-70B-Fireplace", + "base_model": "", + "revision": "220079e4115733991eb19c30d5480db9696a665e", + "precision": "float16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:01:44Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/ValiantLabs/Llama3-70B-ShiningValiant2_eval_request_False_float16_Original.json b/ValiantLabs/Llama3-70B-ShiningValiant2_eval_request_False_float16_Original.json index 26c3487c80e7674cba6dac0256c975d6741b1a5f..e978bdfb4ff6f2dd21faf846dfc920e8d354e539 100644 --- a/ValiantLabs/Llama3-70B-ShiningValiant2_eval_request_False_float16_Original.json +++ b/ValiantLabs/Llama3-70B-ShiningValiant2_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "ValiantLabs/Llama3-70B-ShiningValiant2", "base_model": "", "revision": "bd6cce8da08ccefe9ec58cae3df4bf75c97d8950", "precision": "float16", "params": 70.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:01:26Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "ValiantLabs/Llama3-70B-ShiningValiant2", + "base_model": "", + "revision": "bd6cce8da08ccefe9ec58cae3df4bf75c97d8950", + "precision": "float16", + "params": 70, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:01:26Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Bagel-Hermes-2x34B_eval_request_False_bfloat16_Original.json b/Weyaxi/Bagel-Hermes-2x34B_eval_request_False_bfloat16_Original.json index 4515fa4729f0afc8c5ce687280ad791698f0eb06..e6b6e7f8b7e0dfdee3d60350bd894ae05b6ceecb 100644 --- a/Weyaxi/Bagel-Hermes-2x34B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Bagel-Hermes-2x34B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Bagel-Hermes-2x34B", "base_model": "", "revision": "dc989cab883f222a7d8219436ca2376806735452", "precision": "bfloat16", "params": 60.814, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:21:15Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Bagel-Hermes-2x34B", + "base_model": "", + "revision": "dc989cab883f222a7d8219436ca2376806735452", + "precision": "bfloat16", + "params": 60.814, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:21:15Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Einstein-v4-7B_eval_request_False_bfloat16_Original.json b/Weyaxi/Einstein-v4-7B_eval_request_False_bfloat16_Original.json index 47aad2d979c5c4c1046d4ed54744435ee15448a9..55dd2367dc7ade07366e12104410646076054596 100644 --- a/Weyaxi/Einstein-v4-7B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Einstein-v4-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Einstein-v4-7B", "base_model": "", "revision": "7eecd9833b8a012e23ac1df789884888b047baa0", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:19:31Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Einstein-v4-7B", + "base_model": "", + "revision": "7eecd9833b8a012e23ac1df789884888b047baa0", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:19:31Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Einstein-v4-Qwen-1.5-32B_eval_request_False_bfloat16_Original.json b/Weyaxi/Einstein-v4-Qwen-1.5-32B_eval_request_False_bfloat16_Original.json index b4026ea4edde14e1b5a1c1c99eee6b65da53b888..b2e71e80e5e381dc7b7ebb6222a4d50a7ee31365 100644 --- a/Weyaxi/Einstein-v4-Qwen-1.5-32B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Einstein-v4-Qwen-1.5-32B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Einstein-v4-Qwen-1.5-32B", "base_model": "", "revision": "c68a8b73748dd1301a9601abb26b3ad966049023", "precision": "bfloat16", "params": 32.512, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:56Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Einstein-v4-Qwen-1.5-32B", + "base_model": "", + "revision": "c68a8b73748dd1301a9601abb26b3ad966049023", + "precision": "bfloat16", + "params": 32.512, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:56Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Einstein-v6.1-Llama3-8B_eval_request_False_bfloat16_Original.json b/Weyaxi/Einstein-v6.1-Llama3-8B_eval_request_False_bfloat16_Original.json index 336e5051efeeed01abd2fc1db481950ede504b49..55398ee71b47275a3073b841160ac05be0fd1c6c 100644 --- a/Weyaxi/Einstein-v6.1-Llama3-8B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Einstein-v6.1-Llama3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Einstein-v6.1-Llama3-8B", "base_model": "", "revision": "5cab6d54666b6024d0f745d61abf1842edb934e0", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:18:51Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Einstein-v6.1-Llama3-8B", + "base_model": "", + "revision": "5cab6d54666b6024d0f745d61abf1842edb934e0", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:18:51Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B_eval_request_False_bfloat16_Original.json b/Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B_eval_request_False_bfloat16_Original.json index 68b54e8492c0a6b488914ae4bb30b3a4482db197..ba1903084235375f48b05eb2fddb2752d2546444 100644 --- a/Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B", "base_model": "", "revision": "b7507e94146c0832c26609e9ab8115934d3e25b3", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:08Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Einstein-v6.1-developed-by-Weyaxi-Llama3-8B", + "base_model": "", + "revision": "b7507e94146c0832c26609e9ab8115934d3e25b3", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:08Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/Einstein-v7-Qwen2-7B_eval_request_False_bfloat16_Original.json b/Weyaxi/Einstein-v7-Qwen2-7B_eval_request_False_bfloat16_Original.json index 10a109fd9c743778bb1807c3162501d48b0d36e6..07e198ee6dabd188e648e9fb16a233e5e6e69562 100644 --- a/Weyaxi/Einstein-v7-Qwen2-7B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/Einstein-v7-Qwen2-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/Einstein-v7-Qwen2-7B", "base_model": "", "revision": "d5a2f245bf98a40d196821bc378e10f35b4da81a", "precision": "bfloat16", "params": 7.616, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:57:30Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/Einstein-v7-Qwen2-7B", + "base_model": "", + "revision": "d5a2f245bf98a40d196821bc378e10f35b4da81a", + "precision": "bfloat16", + "params": 7.616, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:57:30Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B_eval_request_False_bfloat16_Original.json b/Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B_eval_request_False_bfloat16_Original.json index 5d8cad174f1a4c4f2efa7c0c838e488665f2605c..0357c5c147277292cbaeb4805b1f5d9d846aed00 100644 --- a/Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B", "base_model": "", "revision": "42cd6a0c65ece5b3bae4465fab96b436d970ae09", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:38Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B", + "base_model": "", + "revision": "42cd6a0c65ece5b3bae4465fab96b436d970ae09", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:38Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp_eval_request_False_bfloat16_Original.json b/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp_eval_request_False_bfloat16_Original.json index ec82423c9cb5d40bc597a597e3652f2174693b15..3c99ca20032607d52b7585b839e0a0a46d3b88af 100644 --- a/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp", "base_model": "", "revision": "e21f57085e0593ce2bafe0b1f7784d473fa79c77", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:20:28Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp", + "base_model": "", + "revision": "e21f57085e0593ce2bafe0b1f7784d473fa79c77", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:20:28Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct_eval_request_False_bfloat16_Original.json b/Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct_eval_request_False_bfloat16_Original.json index 31a0971201ab23c5363b61aa3a7a826f939c9091..977a8d8dd088a33a794ed8366cc9dc0f4bc0d9eb 100644 --- a/Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct_eval_request_False_bfloat16_Original.json +++ b/Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct", "base_model": "", "revision": "9678b9ca952abe0083dbfc772a56b849866bfa1a", "precision": "bfloat16", "params": 10.732, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:21:24Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Weyaxi/SauerkrautLM-UNA-SOLAR-Instruct", + "base_model": "", + "revision": "9678b9ca952abe0083dbfc772a56b849866bfa1a", + "precision": "bfloat16", + "params": 10.732, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:21:24Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/WizardLMTeam/WizardLM-13B-V1.0_eval_request_False_bfloat16_Original.json b/WizardLMTeam/WizardLM-13B-V1.0_eval_request_False_bfloat16_Original.json index 3e430af50bff49896fb200b5f28394f5d27fffcf..79dec23d19123fd2a6f863abed99a94380bce583 100644 --- a/WizardLMTeam/WizardLM-13B-V1.0_eval_request_False_bfloat16_Original.json +++ b/WizardLMTeam/WizardLM-13B-V1.0_eval_request_False_bfloat16_Original.json @@ -3,13 +3,13 @@ "base_model": null, "revision": "964a93aa2e78da377115bb856075a69ebe8aefa4", "precision": "bfloat16", - "params": 13.0, + "params": 13, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-13T18:10:32Z", "model_type": "finetuned", "job_id": "6693785", "job_start_time": "2024-06-23T04:40:08.857677", "use_chat_template": false -} \ No newline at end of file +} diff --git a/Yuma42/KangalKhan-RawRuby-7B_eval_request_False_bfloat16_Original.json b/Yuma42/KangalKhan-RawRuby-7B_eval_request_False_bfloat16_Original.json index a4e33a3f69e0b5e6650d1c7724a2d79bb7082f08..aa77b9618384b8390c68c0acecee89ad98290851 100644 --- a/Yuma42/KangalKhan-RawRuby-7B_eval_request_False_bfloat16_Original.json +++ b/Yuma42/KangalKhan-RawRuby-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "Yuma42/KangalKhan-RawRuby-7B", "base_model": "", "revision": "54f56d4c6889eaf43fdd5f7d6dcef3c2ebe51929", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:26:31Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "Yuma42/KangalKhan-RawRuby-7B", + "base_model": "", + "revision": "54f56d4c6889eaf43fdd5f7d6dcef3c2ebe51929", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:26:31Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/aaditya/Llama3-OpenBioLLM-70B_eval_request_False_float16_Original.json b/aaditya/Llama3-OpenBioLLM-70B_eval_request_False_float16_Original.json index 8a356e0d68114cd725988552cd069256a18a77b2..16615f72b560d1ca117b72e44d8f4b3b655407bb 100644 --- a/aaditya/Llama3-OpenBioLLM-70B_eval_request_False_float16_Original.json +++ b/aaditya/Llama3-OpenBioLLM-70B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "aaditya/Llama3-OpenBioLLM-70B", "base_model": "", "revision": "5f79deaf38bc5f662943d304d59cb30357e8e5bd", "precision": "float16", "params": 70.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:56:10Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "aaditya/Llama3-OpenBioLLM-70B", + "base_model": "", + "revision": "5f79deaf38bc5f662943d304d59cb30357e8e5bd", + "precision": "float16", + "params": 70, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:56:10Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/abacusai/Smaug-Llama-3-70B-Instruct_eval_request_False_bfloat16_Original.json b/abacusai/Smaug-Llama-3-70B-Instruct_eval_request_False_bfloat16_Original.json index f683a3e5e2e456890cd423f04ce20f4bcbea950d..d0bfadade86e64aee0c110c40c03f5cad9b44432 100644 --- a/abacusai/Smaug-Llama-3-70B-Instruct_eval_request_False_bfloat16_Original.json +++ b/abacusai/Smaug-Llama-3-70B-Instruct_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T16:29:36Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7215718", "job_start_time": "2024-06-26T23:20:53.509263", "use_chat_template": true -} \ No newline at end of file +} diff --git a/abacusai/Smaug-Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json b/abacusai/Smaug-Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json index bcbd6da45b0774d29018b677971f37cd3623231a..42bbb8da5f21e86560bd13182b31dce36f478123 100644 --- a/abacusai/Smaug-Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json +++ b/abacusai/Smaug-Qwen2-72B-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "abacusai/Smaug-Qwen2-72B-Instruct", "base_model": "", "revision": "c8a5ba00b275b58df6951210361fbe708981c907", "precision": "bfloat16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T18:50:17Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "abacusai/Smaug-Qwen2-72B-Instruct", + "base_model": "", + "revision": "c8a5ba00b275b58df6951210361fbe708981c907", + "precision": "bfloat16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T18:50:17Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/abhishek/autotrain-llama3-orpo-v2_eval_request_False_bfloat16_Original.json b/abhishek/autotrain-llama3-orpo-v2_eval_request_False_bfloat16_Original.json index 04cc5212010a27c995e468d7b7202b0de102f294..e83374f3987ccb5da7fae7616cd655224c32cea1 100644 --- a/abhishek/autotrain-llama3-orpo-v2_eval_request_False_bfloat16_Original.json +++ b/abhishek/autotrain-llama3-orpo-v2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "abhishek/autotrain-llama3-orpo-v2", "base_model": "", "revision": "1655d0683696a5de2eb9a59c339ee469297beb9c", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:27:52Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "abhishek/autotrain-llama3-orpo-v2", + "base_model": "", + "revision": "1655d0683696a5de2eb9a59c339ee469297beb9c", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:27:52Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/adamo1139/Yi-34B-200K-AEZAKMI-v2_eval_request_False_float16_Original.json b/adamo1139/Yi-34B-200K-AEZAKMI-v2_eval_request_False_float16_Original.json index 6beb9ecf65756b5354481df5be0e073f9c90b3ec..ba60ee590854f0e3b2343629a04b10093ed790fd 100644 --- a/adamo1139/Yi-34B-200K-AEZAKMI-v2_eval_request_False_float16_Original.json +++ b/adamo1139/Yi-34B-200K-AEZAKMI-v2_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "adamo1139/Yi-34B-200K-AEZAKMI-v2", "base_model": "", "revision": "189b42b0dae6352fbe7165255aae851961c8e678", "precision": "float16", "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:38:26Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "adamo1139/Yi-34B-200K-AEZAKMI-v2", + "base_model": "", + "revision": "189b42b0dae6352fbe7165255aae851961c8e678", + "precision": "float16", + "params": 34.389, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:38:26Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/alchemonaut/QuartetAnemoi-70B-t0.0001_eval_request_False_float16_Original.json b/alchemonaut/QuartetAnemoi-70B-t0.0001_eval_request_False_float16_Original.json index 136ad8bc98da16a3c08a76b512acfd8c933fe8ce..a3dcfcc6b98534b7708e0cf9c5c919db1a7ba615 100644 --- a/alchemonaut/QuartetAnemoi-70B-t0.0001_eval_request_False_float16_Original.json +++ b/alchemonaut/QuartetAnemoi-70B-t0.0001_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 68.977, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T14:39:27Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7215722", "job_start_time": "2024-06-26T23:49:07.579905", "use_chat_template": false -} \ No newline at end of file +} diff --git a/allenai/OLMo-7B-hf_eval_request_False_bfloat16_Original.json b/allenai/OLMo-7B-hf_eval_request_False_bfloat16_Original.json index 025736de192a99326cb71415ad175f9efd7e9b0e..233efeee05a8754947ac0e9ea4f7388a686e768b 100644 --- a/allenai/OLMo-7B-hf_eval_request_False_bfloat16_Original.json +++ b/allenai/OLMo-7B-hf_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 6.888, "architectures": "OlmoForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-12T13:30:42Z", "model_type": "pretrained", "job_id": "6661018", "job_start_time": "2024-06-22T18:44:57.200718", "use_chat_template": false -} \ No newline at end of file +} diff --git a/allknowingroger/LimyQstar-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/LimyQstar-7B-slerp_eval_request_False_bfloat16_Original.json index e3ef768f14314887a68baad47b304cffc435f6a9..fb90a2c93b1101c01d9a6cd8c5ed800aae2666a9 100644 --- a/allknowingroger/LimyQstar-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/LimyQstar-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/LimyQstar-7B-slerp", "base_model": "", "revision": "6dc557c7bfd6a6f9bc8190bc8a31c3b732deca40", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:33:57Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/LimyQstar-7B-slerp", + "base_model": "", + "revision": "6dc557c7bfd6a6f9bc8190bc8a31c3b732deca40", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:33:57Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Meme-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Meme-7B-slerp_eval_request_False_bfloat16_Original.json index 3d3a55a170ce581920284d5519ec50413164030c..3829536bca92313ae0e31df2eedcf57f49710fb1 100644 --- a/allknowingroger/Meme-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Meme-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Meme-7B-slerp", "base_model": "", "revision": "7836c0f4fce70286382e61003e9a05d7559365d9", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:19:52Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Meme-7B-slerp", + "base_model": "", + "revision": "7836c0f4fce70286382e61003e9a05d7559365d9", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:19:52Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Mistral3mash1-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Mistral3mash1-7B-slerp_eval_request_False_bfloat16_Original.json index 033391938136e25afd369cf4d2c56cbe1552f1c5..ed2bd01e687c37d93a96733c8be57ceb8e0e4ab2 100644 --- a/allknowingroger/Mistral3mash1-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Mistral3mash1-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Mistral3mash1-7B-slerp", "base_model": "", "revision": "c71c23baba04807c077b260ac2785ecdee5be480", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:19:09Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Mistral3mash1-7B-slerp", + "base_model": "", + "revision": "c71c23baba04807c077b260ac2785ecdee5be480", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:19:09Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MixTAO-19B-pass_eval_request_False_bfloat16_Original.json b/allknowingroger/MixTAO-19B-pass_eval_request_False_bfloat16_Original.json index 003382d556f152a02c511dc8fbcefc8ff91ea6e1..6e460d6f782c9bb138e2b8e6bec4e084666fd50b 100644 --- a/allknowingroger/MixTAO-19B-pass_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MixTAO-19B-pass_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MixTAO-19B-pass", "base_model": "", "revision": "a41369cfcfbada9d5387051ba616bf1432b31d31", "precision": "bfloat16", "params": 19.188, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:11:46Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MixTAO-19B-pass", + "base_model": "", + "revision": "a41369cfcfbada9d5387051ba616bf1432b31d31", + "precision": "bfloat16", + "params": 19.188, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:11:46Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MixTaoTruthful-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MixTaoTruthful-13B-slerp_eval_request_False_bfloat16_Original.json index 11186a1f3faf259c42015ffdd0b06d3a27008a4d..23762f3a5179429c7dfb04db510de8d81c803820 100644 --- a/allknowingroger/MixTaoTruthful-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MixTaoTruthful-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MixTaoTruthful-13B-slerp", "base_model": "", "revision": "3324d37e138c6bf0d6891e54b6dd839c8d2f35ec", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:17:33Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MixTaoTruthful-13B-slerp", + "base_model": "", + "revision": "3324d37e138c6bf0d6891e54b6dd839c8d2f35ec", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:17:33Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiCalm-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiCalm-7B-slerp_eval_request_False_bfloat16_Original.json index 9bf2431a6c6cdd3499e2654f49105abc5fcf8534..9442bb9698188aa177d852847824da4b91c7af71 100644 --- a/allknowingroger/MultiCalm-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiCalm-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiCalm-7B-slerp", "base_model": "", "revision": "1c23540e907fab4dfe0ef66edd0003e764bfe568", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:26:58Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiCalm-7B-slerp", + "base_model": "", + "revision": "1c23540e907fab4dfe0ef66edd0003e764bfe568", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:26:58Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash-12B-slerp_eval_request_False_bfloat16_Original.json index f14e5210ff39ace41327974367ca3d5b162b41f4..602caa4819ae2a27dd4fcb76522a754af8661d77 100644 --- a/allknowingroger/MultiMash-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash-12B-slerp", "base_model": "", "revision": "91a6d0fe6b9271000ca713ee9ab414c782ba4c50", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:26:21Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash-12B-slerp", + "base_model": "", + "revision": "91a6d0fe6b9271000ca713ee9ab414c782ba4c50", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:26:21Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash10-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash10-13B-slerp_eval_request_False_bfloat16_Original.json index ef201d11bae915631c5d7282182dd7f705c6fda9..8fe5e760f944a470d0b7e685d8daec8e9f0f6d5d 100644 --- a/allknowingroger/MultiMash10-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash10-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash10-13B-slerp", "base_model": "", "revision": "6def2fd1a11d4c380a19b7a3bdf263a6b80cd8f3", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:15:05Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash10-13B-slerp", + "base_model": "", + "revision": "6def2fd1a11d4c380a19b7a3bdf263a6b80cd8f3", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:15:05Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash11-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash11-13B-slerp_eval_request_False_bfloat16_Original.json index f22576eaa5bf2b21703f517d0f81ea992ce6a5e0..968b0110cb2790e17c85deb4ac349e879eaaf2e9 100644 --- a/allknowingroger/MultiMash11-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash11-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash11-13B-slerp", "base_model": "", "revision": "1134a0adabef4a26e1d49c302baff74c4a7e9f46", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:14:26Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash11-13B-slerp", + "base_model": "", + "revision": "1134a0adabef4a26e1d49c302baff74c4a7e9f46", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:14:26Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash12-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash12-13B-slerp_eval_request_False_bfloat16_Original.json index 6c6d835487aff74b661d323268cd10fec6d64bf6..124576caa24ea97f42652803006151e9df380f14 100644 --- a/allknowingroger/MultiMash12-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash12-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash12-13B-slerp", "base_model": "", "revision": "8b140b6c524496fd21adae1b4ad67df254f9faad", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:13:37Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash12-13B-slerp", + "base_model": "", + "revision": "8b140b6c524496fd21adae1b4ad67df254f9faad", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:13:37Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash2-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash2-12B-slerp_eval_request_False_bfloat16_Original.json index 49df5dd2a4e329063a138981dfb062ba2b841189..12d848418110b9b8f728e42ccf1ce87ad430753b 100644 --- a/allknowingroger/MultiMash2-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash2-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash2-12B-slerp", "base_model": "", "revision": "e44e9563368699f753a4474b068c059d233ddee3", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:25:30Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash2-12B-slerp", + "base_model": "", + "revision": "e44e9563368699f753a4474b068c059d233ddee3", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:25:30Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash5-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash5-12B-slerp_eval_request_False_bfloat16_Original.json index 125bbdb09a4333a5d3e6dc128bed41a578cd49b4..9e2562986b23097b42b46a1dc55adaa44d4d6aa3 100644 --- a/allknowingroger/MultiMash5-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash5-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash5-12B-slerp", "base_model": "", "revision": "15ef0301c7ce939208d55ad13fa840662f92bce6", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:23:02Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash5-12B-slerp", + "base_model": "", + "revision": "15ef0301c7ce939208d55ad13fa840662f92bce6", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:23:02Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash6-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash6-12B-slerp_eval_request_False_bfloat16_Original.json index 54ba6d2e0074684f8e8d6f37d71374694581f2f2..36046fcd160ee4c43786822d5b0019ee80fe1061 100644 --- a/allknowingroger/MultiMash6-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash6-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash6-12B-slerp", "base_model": "", "revision": "a04856a12b85e986e1b540cf0c7510e9ce2df09b", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:22:15Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash6-12B-slerp", + "base_model": "", + "revision": "a04856a12b85e986e1b540cf0c7510e9ce2df09b", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:22:15Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash7-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash7-12B-slerp_eval_request_False_bfloat16_Original.json index 39ce922a1a2ce07401aa38da33766d4150bda158..35029a6e24bf722aeaf1560509e552318619220c 100644 --- a/allknowingroger/MultiMash7-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash7-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash7-12B-slerp", "base_model": "", "revision": "5f91dd41fb4b58e76c52b03ed15477a046b079df", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:20:49Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash7-12B-slerp", + "base_model": "", + "revision": "5f91dd41fb4b58e76c52b03ed15477a046b079df", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:20:49Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash8-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash8-13B-slerp_eval_request_False_bfloat16_Original.json index 2607d328016d371674f60489fb98cdc9afbfd559..a0f20669de1fb989e5a98e986011ee4a717d1e4f 100644 --- a/allknowingroger/MultiMash8-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash8-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash8-13B-slerp", "base_model": "", "revision": "5590ccd99f74301951f450f9d0271a99e97728c8", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:16:35Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash8-13B-slerp", + "base_model": "", + "revision": "5590ccd99f74301951f450f9d0271a99e97728c8", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:16:35Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMash9-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMash9-13B-slerp_eval_request_False_bfloat16_Original.json index 9240da296f93508b668c0b3d6545f952893ba33d..1ead740c822a55d9a638c4fd5a486acedc2e9e82 100644 --- a/allknowingroger/MultiMash9-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMash9-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMash9-13B-slerp", "base_model": "", "revision": "56dac45f387669baa04a8997ebb9ea63c65fbbd1", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:15:47Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMash9-13B-slerp", + "base_model": "", + "revision": "56dac45f387669baa04a8997ebb9ea63c65fbbd1", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:15:47Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiMerge-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiMerge-7B-slerp_eval_request_False_bfloat16_Original.json index bbb25b6ef27b4f61e91b93a8ccf8c466d0520c18..07e7ad80a8c1b0decef37cc1553efa7eae2e0fa9 100644 --- a/allknowingroger/MultiMerge-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiMerge-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiMerge-7B-slerp", "base_model": "", "revision": "a026bbea09f0b1880deed62b9081e3708be0dec2", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:32:08Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiMerge-7B-slerp", + "base_model": "", + "revision": "a026bbea09f0b1880deed62b9081e3708be0dec2", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:32:08Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Multimash3-12B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Multimash3-12B-slerp_eval_request_False_bfloat16_Original.json index c7d27bcfdd322805a40317490b90014eb6702007..1cfdf1f73e3f94175623a458e462e97d26685ec2 100644 --- a/allknowingroger/Multimash3-12B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Multimash3-12B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Multimash3-12B-slerp", "base_model": "", "revision": "0b90bf0b5230d02b4ba63879fc3bf0b85d46c3ce", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:24:40Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Multimash3-12B-slerp", + "base_model": "", + "revision": "0b90bf0b5230d02b4ba63879fc3bf0b85d46c3ce", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:24:40Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Multimerge-19B-pass_eval_request_False_bfloat16_Original.json b/allknowingroger/Multimerge-19B-pass_eval_request_False_bfloat16_Original.json index a90a4f4a4883d3c7955233a8d9232e19f90e18ca..bb661c7083a84f5632d0903b0321c0b1c81eefda 100644 --- a/allknowingroger/Multimerge-19B-pass_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Multimerge-19B-pass_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Multimerge-19B-pass", "base_model": "", "revision": "e75918ed5601f400f62601cf6c0887aa936e8a52", "precision": "bfloat16", "params": 19.188, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:10:27Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Multimerge-19B-pass", + "base_model": "", + "revision": "e75918ed5601f400f62601cf6c0887aa936e8a52", + "precision": "bfloat16", + "params": 19.188, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:10:27Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/MultiverseEx26-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/MultiverseEx26-7B-slerp_eval_request_False_bfloat16_Original.json index e59fc5d0c4cad373b1ed38eeccd364d93939d124..7d4b1dc039199dd26518ad01e9df04ae76ee8211 100644 --- a/allknowingroger/MultiverseEx26-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/MultiverseEx26-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/MultiverseEx26-7B-slerp", "base_model": "", "revision": "43f18d84e025693f00e9be335bf12fce96089b2f", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:36:57Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/MultiverseEx26-7B-slerp", + "base_model": "", + "revision": "43f18d84e025693f00e9be335bf12fce96089b2f", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:36:57Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/NeuralWestSeverus-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/NeuralWestSeverus-7B-slerp_eval_request_False_bfloat16_Original.json index a7f24c7c763d76858ca3226ee6a9b37e7d73c245..9d48f57fb94e1f2faba45767935c7f5cd754b06c 100644 --- a/allknowingroger/NeuralWestSeverus-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/NeuralWestSeverus-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/NeuralWestSeverus-7B-slerp", "base_model": "", "revision": "5ee5d6a11ffc4f9733e78994169a2e1614d5e16e", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:30:06Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/NeuralWestSeverus-7B-slerp", + "base_model": "", + "revision": "5ee5d6a11ffc4f9733e78994169a2e1614d5e16e", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:30:06Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Neuralcoven-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Neuralcoven-7B-slerp_eval_request_False_bfloat16_Original.json index f9f507349d4bc3da8381dc2d2f6db7d085692e69..c928e08867d0df7fc13d370f668d14dd81d07527 100644 --- a/allknowingroger/Neuralcoven-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Neuralcoven-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Neuralcoven-7B-slerp", "base_model": "", "revision": "129b40a7fd816f679ef5d4ab29fc77345f33a7b1", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:28:14Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Neuralcoven-7B-slerp", + "base_model": "", + "revision": "129b40a7fd816f679ef5d4ab29fc77345f33a7b1", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:28:14Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Neuralgamma-2.5-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Neuralgamma-2.5-7B-slerp_eval_request_False_bfloat16_Original.json index 75d48ffb20b9c0891d7c0737b093bb065763de3b..a0aa9055003377074f0adf5820f526f5d086e6fc 100644 --- a/allknowingroger/Neuralgamma-2.5-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Neuralgamma-2.5-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Neuralgamma-2.5-7B-slerp", "base_model": "", "revision": "13f38b8e3ca17cee9a7546b103ffa0854a5a29b9", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:33:17Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Neuralgamma-2.5-7B-slerp", + "base_model": "", + "revision": "13f38b8e3ca17cee9a7546b103ffa0854a5a29b9", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:33:17Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Neuralmultiverse-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Neuralmultiverse-7B-slerp_eval_request_False_bfloat16_Original.json index 4d98d7ad0fb66ea18a5c9b94ca96f827e601c65a..2eec80bfa0ab5b41d35a94ea4b7d0a0bd2acd1a7 100644 --- a/allknowingroger/Neuralmultiverse-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Neuralmultiverse-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Neuralmultiverse-7B-slerp", "base_model": "", "revision": "a65fe05e26e10a488b08264ac8ed73a49c3f263a", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:27:40Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Neuralmultiverse-7B-slerp", + "base_model": "", + "revision": "a65fe05e26e10a488b08264ac8ed73a49c3f263a", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:27:40Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Ph3unsloth-3B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Ph3unsloth-3B-slerp_eval_request_False_bfloat16_Original.json index 8ec0dfac1c9d0d87ebbc1ec4278056afb76a78ad..6637c9678953abae2e3f74cc53fb7b935451b128 100644 --- a/allknowingroger/Ph3unsloth-3B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Ph3unsloth-3B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Ph3unsloth-3B-slerp", "base_model": "", "revision": "465444b3cdd43876717f7386ea2f3357c5fe8e53", "precision": "bfloat16", "params": 3.821, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:06:18Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Ph3unsloth-3B-slerp", + "base_model": "", + "revision": "465444b3cdd43876717f7386ea2f3357c5fe8e53", + "precision": "bfloat16", + "params": 3.821, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:06:18Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/ROGERphi-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/ROGERphi-7B-slerp_eval_request_False_bfloat16_Original.json index 08c1520f9140a364bbf7edf98043bb4156ff5432..c19684ddb819755442b036c6183c2700d60cef39 100644 --- a/allknowingroger/ROGERphi-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/ROGERphi-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/ROGERphi-7B-slerp", "base_model": "", "revision": "a92f90ae5e4286daa2399df4951a3347aaf414e1", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:32:40Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/ROGERphi-7B-slerp", + "base_model": "", + "revision": "a92f90ae5e4286daa2399df4951a3347aaf414e1", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:32:40Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/RogerMerge-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/RogerMerge-7B-slerp_eval_request_False_bfloat16_Original.json index 082248f096c72b6d9fd0abc0b71aa183b02c8e30..78c17f1734f569c8437a75e9ab6231b1c52ac1ec 100644 --- a/allknowingroger/RogerMerge-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/RogerMerge-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/RogerMerge-7B-slerp", "base_model": "", "revision": "397f5c0b52a536c130982ca2a7c3056358bbdf92", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:09:44Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/RogerMerge-7B-slerp", + "base_model": "", + "revision": "397f5c0b52a536c130982ca2a7c3056358bbdf92", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:09:44Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Strangecoven-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Strangecoven-7B-slerp_eval_request_False_bfloat16_Original.json index 1c4b7fa76a6f4023d5518e888c86c401d7f6f06e..d2b47e5f35672b5f9f92a55b1d5f7245e1471d14 100644 --- a/allknowingroger/Strangecoven-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Strangecoven-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Strangecoven-7B-slerp", "base_model": "", "revision": "8bc9d8f972d15fdd3e02c602ef4f549493bf2208", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:28:52Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Strangecoven-7B-slerp", + "base_model": "", + "revision": "8bc9d8f972d15fdd3e02c602ef4f549493bf2208", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:28:52Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/WestlakeMaziyar-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/WestlakeMaziyar-7B-slerp_eval_request_False_bfloat16_Original.json index 4ed86f353f44c6fee41959df7e78c0a9f03e642d..c6361e4658eaec142a50550f9827714e79eb1bf4 100644 --- a/allknowingroger/WestlakeMaziyar-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/WestlakeMaziyar-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/WestlakeMaziyar-7B-slerp", "base_model": "", "revision": "751534a844b0d439fe62f98bf8882fe9ab9872e0", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:29:29Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/WestlakeMaziyar-7B-slerp", + "base_model": "", + "revision": "751534a844b0d439fe62f98bf8882fe9ab9872e0", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:29:29Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/YamMaths-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/YamMaths-7B-slerp_eval_request_False_bfloat16_Original.json index bd7ce0516c83a36c6c36decabcd485dbd6367c07..23ed44a707ade2123238f5fb66b3655a67307f80 100644 --- a/allknowingroger/YamMaths-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/YamMaths-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/YamMaths-7B-slerp", "base_model": "", "revision": "bd4ac9d63ca88c80d34fa60ef5cbb56d60a39077", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:12:41Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/YamMaths-7B-slerp", + "base_model": "", + "revision": "bd4ac9d63ca88c80d34fa60ef5cbb56d60a39077", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:12:41Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/Yunconglong-13B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/Yunconglong-13B-slerp_eval_request_False_bfloat16_Original.json index 3c3d7ae46f3671f092f0aa621da97c4fed0b1272..ab3e62a28d34177ffc8d54f89bf8dc68f9c3b452 100644 --- a/allknowingroger/Yunconglong-13B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/Yunconglong-13B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/Yunconglong-13B-slerp", "base_model": "", "revision": "dead687b7342d875bd8ac73bfcd34b88a2e5564c", "precision": "bfloat16", "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:18:16Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/Yunconglong-13B-slerp", + "base_model": "", + "revision": "dead687b7342d875bd8ac73bfcd34b88a2e5564c", + "precision": "bfloat16", + "params": 12.879, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:18:16Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/allknowingroger/limyClown-7B-slerp_eval_request_False_bfloat16_Original.json b/allknowingroger/limyClown-7B-slerp_eval_request_False_bfloat16_Original.json index 21720b2f09c5e4ce5049a2fdb8a6790345c6f3e6..334bfce01daf3341faecb6ab63f4e8aacd671097 100644 --- a/allknowingroger/limyClown-7B-slerp_eval_request_False_bfloat16_Original.json +++ b/allknowingroger/limyClown-7B-slerp_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "allknowingroger/limyClown-7B-slerp", "base_model": "", "revision": "732a1ed0c2c7007297ad9d9797793073825f65ca", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:34:46Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "allknowingroger/limyClown-7B-slerp", + "base_model": "", + "revision": "732a1ed0c2c7007297ad9d9797793073825f65ca", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:34:46Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/aloobun/Meta-Llama-3-7B-28Layers_eval_request_False_bfloat16_Original.json b/aloobun/Meta-Llama-3-7B-28Layers_eval_request_False_bfloat16_Original.json index 46ab383f68a54d21aa60f2c72901d64d7c77b35c..3ff1ad20433ccaee05940b7bb64583bdaccac3fa 100644 --- a/aloobun/Meta-Llama-3-7B-28Layers_eval_request_False_bfloat16_Original.json +++ b/aloobun/Meta-Llama-3-7B-28Layers_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "aloobun/Meta-Llama-3-7B-28Layers", "base_model": "", "revision": "9822e6b8d4de0c0f2964d299f6fcef72385a0341", "precision": "bfloat16", "params": 7.158, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:04:44Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "aloobun/Meta-Llama-3-7B-28Layers", + "base_model": "", + "revision": "9822e6b8d4de0c0f2964d299f6fcef72385a0341", + "precision": "bfloat16", + "params": 7.158, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:04:44Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/alpindale/WizardLM-2-8x22B_eval_request_False_float16_Original.json b/alpindale/WizardLM-2-8x22B_eval_request_False_float16_Original.json index 01eec533b5bec72ac4fe35d01e8f9dfe8ca57cef..33eeef076714c43f8e51180977bfe918ff2c3b94 100644 --- a/alpindale/WizardLM-2-8x22B_eval_request_False_float16_Original.json +++ b/alpindale/WizardLM-2-8x22B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "alpindale/WizardLM-2-8x22B", "base_model": "petra", "revision": "087834da175523cffd66a7e19583725e798c1b4f", "precision": "float16", "params": 140.621, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:15:18Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "alpindale/WizardLM-2-8x22B", + "base_model": "petra", + "revision": "087834da175523cffd66a7e19583725e798c1b4f", + "precision": "float16", + "params": 140.621, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:15:18Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/altomek/CodeRosa-70B-AB1_eval_request_False_float16_Original.json b/altomek/CodeRosa-70B-AB1_eval_request_False_float16_Original.json index a10f4598010d1d43aa87488e432778d6891e2cd2..5567c0ad7b8c882afd47a7db12cfa08a9dd7c8a0 100644 --- a/altomek/CodeRosa-70B-AB1_eval_request_False_float16_Original.json +++ b/altomek/CodeRosa-70B-AB1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "altomek/CodeRosa-70B-AB1", "base_model": "codellama/CodeLlama-70b-Python-hf", "revision": "92ace14feef21ba45347b1d63d58dd7c8c045ad9", "precision": "float16", "params": 68.977, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:13:54Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "altomek/CodeRosa-70B-AB1", + "base_model": "codellama/CodeLlama-70b-Python-hf", + "revision": "92ace14feef21ba45347b1d63d58dd7c8c045ad9", + "precision": "float16", + "params": 68.977, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:13:54Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/altomek/YiSM-34B-0rn_eval_request_False_float16_Original.json b/altomek/YiSM-34B-0rn_eval_request_False_float16_Original.json index 8b22bd08e944188d94944b927bf6a37979731916..96651523ec5498884293b732b1a506601808887b 100644 --- a/altomek/YiSM-34B-0rn_eval_request_False_float16_Original.json +++ b/altomek/YiSM-34B-0rn_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "altomek/YiSM-34B-0rn", "base_model": "01-ai/Yi-1.5-34B-Chat", "revision": "7a481c67cbdd5c846d6aaab5ef9f1eebfad812c2", "precision": "float16", "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:15:17Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "altomek/YiSM-34B-0rn", + "base_model": "01-ai/Yi-1.5-34B-Chat", + "revision": "7a481c67cbdd5c846d6aaab5ef9f1eebfad812c2", + "precision": "float16", + "params": 34.389, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:15:17Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/arcee-ai/Arcee-Spark_eval_request_False_bfloat16_Original.json b/arcee-ai/Arcee-Spark_eval_request_False_bfloat16_Original.json index 80cd7b9f85a9a595519db138936901bfc00d9ee6..c1ea630b0cba43f23a0fdf0a79fee3ca7d8b9134 100644 --- a/arcee-ai/Arcee-Spark_eval_request_False_bfloat16_Original.json +++ b/arcee-ai/Arcee-Spark_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "arcee-ai/Arcee-Spark", "base_model": "", "revision": "3fe368ea5fd32bc4a8d1bcf42510416f7fa28668", "precision": "bfloat16", "params": 7.616, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:57:03Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "arcee-ai/Arcee-Spark", + "base_model": "", + "revision": "3fe368ea5fd32bc4a8d1bcf42510416f7fa28668", + "precision": "bfloat16", + "params": 7.616, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:57:03Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/arcee-ai/Arcee-Spark_eval_request_False_float16_Original.json b/arcee-ai/Arcee-Spark_eval_request_False_float16_Original.json index d10a4f2375841a6c4a9cbb5f135525ec6b2657f2..53bc2ebe8d9517416df9bcd6d9eec20ff1624d01 100644 --- a/arcee-ai/Arcee-Spark_eval_request_False_float16_Original.json +++ b/arcee-ai/Arcee-Spark_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "arcee-ai/Arcee-Spark", "base_model": "", "revision": "3fe368ea5fd32bc4a8d1bcf42510416f7fa28668", "precision": "float16", "params": 7.616, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:06:23Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "arcee-ai/Arcee-Spark", + "base_model": "", + "revision": "3fe368ea5fd32bc4a8d1bcf42510416f7fa28668", + "precision": "float16", + "params": 7.616, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:06:23Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/ayan-sh003/hermes-2-theta-neural-llama3_eval_request_False_bfloat16_Original.json b/ayan-sh003/hermes-2-theta-neural-llama3_eval_request_False_bfloat16_Original.json index 16d21a3b4538c2e7158f4440fdda2fcae6cde770..d073dd47fbc165cb9e265e31e59ddc279fa2f43a 100644 --- a/ayan-sh003/hermes-2-theta-neural-llama3_eval_request_False_bfloat16_Original.json +++ b/ayan-sh003/hermes-2-theta-neural-llama3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "ayan-sh003/hermes-2-theta-neural-llama3", "base_model": "", "revision": "40cd411cc2713f6be69307390173e9dc8fe00b28", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:55:20Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "ayan-sh003/hermes-2-theta-neural-llama3", + "base_model": "", + "revision": "40cd411cc2713f6be69307390173e9dc8fe00b28", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:55:20Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/bertin-project/Gromenauer-7B-Instruct_eval_request_False_bfloat16_Original.json b/bertin-project/Gromenauer-7B-Instruct_eval_request_False_bfloat16_Original.json index 07afb8f41c952199485998a5d4e7796080bea0ba..0b6720456c5ec5d0a99904fea587940c14e726b5 100644 --- a/bertin-project/Gromenauer-7B-Instruct_eval_request_False_bfloat16_Original.json +++ b/bertin-project/Gromenauer-7B-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "bertin-project/Gromenauer-7B-Instruct", "base_model": "", "revision": "4cee56ca72ee98ec67f4532010851c14d6b0d4e9", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T23:21:20Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "bertin-project/Gromenauer-7B-Instruct", + "base_model": "", + "revision": "4cee56ca72ee98ec67f4532010851c14d6b0d4e9", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T23:21:20Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/bertin-project/Gromenauer-7B_eval_request_False_bfloat16_Original.json b/bertin-project/Gromenauer-7B_eval_request_False_bfloat16_Original.json index 0728a8416a4214918e6cabcb130388ca21928a6c..56746aff788c9135ea1b01adda5b6cdefa6abf0e 100644 --- a/bertin-project/Gromenauer-7B_eval_request_False_bfloat16_Original.json +++ b/bertin-project/Gromenauer-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "bertin-project/Gromenauer-7B", "base_model": "", "revision": "aaff2b37b64b0cdf4ed5694ea5ee483b898a6c77", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T23:22:01Z", "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "bertin-project/Gromenauer-7B", + "base_model": "", + "revision": "aaff2b37b64b0cdf4ed5694ea5ee483b898a6c77", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T23:22:01Z", + "model_type": "🟩 : 🟩 continuously pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/bhenrym14/platypus-yi-34b_eval_request_False_float16_Original.json b/bhenrym14/platypus-yi-34b_eval_request_False_float16_Original.json index c5ab5bd6a798a29dc83cabe0152ec67a76638cc3..3e1cc7a785527b3aded2a2d342cde394a439658d 100644 --- a/bhenrym14/platypus-yi-34b_eval_request_False_float16_Original.json +++ b/bhenrym14/platypus-yi-34b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "bhenrym14/platypus-yi-34b", "base_model": "", "revision": "66abec7cba89b35c7b6cab2140c3532049de0157", "precision": "float16", "params": 34.389, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:55:25Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "bhenrym14/platypus-yi-34b", + "base_model": "", + "revision": "66abec7cba89b35c7b6cab2140c3532049de0157", + "precision": "float16", + "params": 34.389, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:55:25Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/bosonai/Higgs-Llama-3-70B_eval_request_False_bfloat16_Original.json b/bosonai/Higgs-Llama-3-70B_eval_request_False_bfloat16_Original.json index 2434c5d73a198cb675a2f6222a813190f866bad8..ceec18f8ae1256e793451543a625443f232bac80 100644 --- a/bosonai/Higgs-Llama-3-70B_eval_request_False_bfloat16_Original.json +++ b/bosonai/Higgs-Llama-3-70B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "bosonai/Higgs-Llama-3-70B", "base_model": "", "revision": "b2c7540768046dfdae7a0cb846a7da6c41d826b1", "precision": "bfloat16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:15:10Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "bosonai/Higgs-Llama-3-70B", + "base_model": "", + "revision": "b2c7540768046dfdae7a0cb846a7da6c41d826b1", + "precision": "bfloat16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:15:10Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/bosonai/Higgs-Llama-3-70B_eval_request_False_float16_Original.json b/bosonai/Higgs-Llama-3-70B_eval_request_False_float16_Original.json index 65fd9b5c8c0c4f06f8ec94bdc342f90d128253be..90f68925586a279f72fdfafdd2f26594ebd4fce5 100644 --- a/bosonai/Higgs-Llama-3-70B_eval_request_False_float16_Original.json +++ b/bosonai/Higgs-Llama-3-70B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "bosonai/Higgs-Llama-3-70B", "base_model": "", "revision": "b2c7540768046dfdae7a0cb846a7da6c41d826b1", "precision": "float16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:14:48Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "bosonai/Higgs-Llama-3-70B", + "base_model": "", + "revision": "b2c7540768046dfdae7a0cb846a7da6c41d826b1", + "precision": "float16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:14:48Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/bunnycore/RemixLlama3-8B_eval_request_False_bfloat16_Original.json b/bunnycore/RemixLlama3-8B_eval_request_False_bfloat16_Original.json index f19d98f31379f1908eebf6f76d5be498f1b7e656..dca365d4e32f056d2fd014f2533f997884e8bb24 100644 --- a/bunnycore/RemixLlama3-8B_eval_request_False_bfloat16_Original.json +++ b/bunnycore/RemixLlama3-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "bunnycore/RemixLlama3-8B", "base_model": "bunnycore/RemixLlama3-8B", "revision": "77562598abaf43c39eb10845596eaa66037aee0c", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:27:48Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "bunnycore/RemixLlama3-8B", + "base_model": "bunnycore/RemixLlama3-8B", + "revision": "77562598abaf43c39eb10845596eaa66037aee0c", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:27:48Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/cgato/TheSalt-L3-8b-v0.3.2_eval_request_False_bfloat16_Original.json b/cgato/TheSalt-L3-8b-v0.3.2_eval_request_False_bfloat16_Original.json index 12937606ccb3be4fcb86f9303f529738d923c761..ecb4eb7306b955cc13cc21f4a0542e481a57d529 100644 --- a/cgato/TheSalt-L3-8b-v0.3.2_eval_request_False_bfloat16_Original.json +++ b/cgato/TheSalt-L3-8b-v0.3.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "cgato/TheSalt-L3-8b-v0.3.2", "base_model": "", "revision": "5cf08e2bf9590ebcd14ba021e113def28c65afa2", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:58:35Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "cgato/TheSalt-L3-8b-v0.3.2", + "base_model": "", + "revision": "5cf08e2bf9590ebcd14ba021e113def28c65afa2", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:58:35Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/chargoddard/prometheus-2-llama-3-8b_eval_request_False_bfloat16_Original.json b/chargoddard/prometheus-2-llama-3-8b_eval_request_False_bfloat16_Original.json index 64e854bfd5cc081ea29385a55a52a757455890fc..f667f50071ef8308562bc3b3b3c99b2f48e6c5fe 100644 --- a/chargoddard/prometheus-2-llama-3-8b_eval_request_False_bfloat16_Original.json +++ b/chargoddard/prometheus-2-llama-3-8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "chargoddard/prometheus-2-llama-3-8b", "base_model": "", "revision": "90a728ac98e5b4169f88ae4945e357cf45477568", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:15:25Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "chargoddard/prometheus-2-llama-3-8b", + "base_model": "", + "revision": "90a728ac98e5b4169f88ae4945e357cf45477568", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:15:25Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO_eval_request_False_bfloat16_Original.json b/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO_eval_request_False_bfloat16_Original.json index c7c1d71065bb786b1e05c5636908cdc931af424b..35dc35fb94877623423b40e527d6711451ee519b 100644 --- a/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO_eval_request_False_bfloat16_Original.json +++ b/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO", "base_model": "", "revision": "3fcaa9fe99691659eb197487e9a343f601bf63f2", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:08:15Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO", + "base_model": "", + "revision": "3fcaa9fe99691659eb197487e9a343f601bf63f2", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:08:15Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/cloudyu/Phoenix_DPO_60B_eval_request_False_float16_Original.json b/cloudyu/Phoenix_DPO_60B_eval_request_False_float16_Original.json index 4c0216192aebb4eef77aaee47da077b43d67579d..4c37184af9e324948a52c717e45191bfa066fab1 100644 --- a/cloudyu/Phoenix_DPO_60B_eval_request_False_float16_Original.json +++ b/cloudyu/Phoenix_DPO_60B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "cloudyu/Phoenix_DPO_60B", "base_model": "", "revision": "af72ef1ac31af544a122b6037523cc6131eee9b5", "precision": "float16", "params": 60.814, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:56:58Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "cloudyu/Phoenix_DPO_60B", + "base_model": "", + "revision": "af72ef1ac31af544a122b6037523cc6131eee9b5", + "precision": "float16", + "params": 60.814, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:56:58Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated_eval_request_False_float16_Original.json b/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated_eval_request_False_float16_Original.json index ccd793fa2e88036625f86c2a1cc2be1da78ac20e..b5a7fd0a1e5ea5f0b4713e765e0b58c5d17137aa 100644 --- a/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated_eval_request_False_float16_Original.json +++ b/cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated", "base_model": "", "revision": "d50be5f22ca9745a2a3175996611d6a840318b7f", "precision": "float16", "params": 13.96, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:02:23Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "cognitivecomputations/dolphin-2.9.2-Phi-3-Medium-abliterated", + "base_model": "", + "revision": "d50be5f22ca9745a2a3175996611d6a840318b7f", + "precision": "float16", + "params": 13.96, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:02:23Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/cognitivecomputations/dolphin-2.9.2-qwen2-72b_eval_request_False_float16_Original.json b/cognitivecomputations/dolphin-2.9.2-qwen2-72b_eval_request_False_float16_Original.json index 077d96d61d22a3d74491d0105b09fc5671453f73..5d1c54c083f00fc6b52e3c209b35d21316bcddec 100644 --- a/cognitivecomputations/dolphin-2.9.2-qwen2-72b_eval_request_False_float16_Original.json +++ b/cognitivecomputations/dolphin-2.9.2-qwen2-72b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "cognitivecomputations/dolphin-2.9.2-qwen2-72b", "base_model": "Qwen/Qwen2-72B", "revision": "e79582577c2bf2af304221af0e8308b7e7d46ca1", "precision": "float16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:18:47Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "cognitivecomputations/dolphin-2.9.2-qwen2-72b", + "base_model": "Qwen/Qwen2-72B", + "revision": "e79582577c2bf2af304221af0e8308b7e7d46ca1", + "precision": "float16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:18:47Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k_eval_request_False_float16_Original.json b/cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k_eval_request_False_float16_Original.json index f8f961cb0ecf298154d9f7bcdc2406eade26b3ed..10278be1e0010d73c492cd21c39e1e3ffe36da11 100644 --- a/cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k_eval_request_False_float16_Original.json +++ b/cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k", "base_model": "01-ai/Yi-1.5-34B-32K", "revision": "ff4eee6438194a670a95dff3118b5231eb568610", "precision": "float16", "params": 34.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:20:53Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "cognitivecomputations/dolphin-2.9.3-Yi-1.5-34B-32k", + "base_model": "01-ai/Yi-1.5-34B-32K", + "revision": "ff4eee6438194a670a95dff3118b5231eb568610", + "precision": "float16", + "params": 34, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:20:53Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/deepseek-ai/deepseek-llm-67b-chat_eval_request_False_bfloat16_Original.json b/deepseek-ai/deepseek-llm-67b-chat_eval_request_False_bfloat16_Original.json index fe48f267d90c56b6ddf55d8a0eb970ca6e407ffa..9f4e1d37c014ac273e74ab4d976d36897e56aa83 100644 --- a/deepseek-ai/deepseek-llm-67b-chat_eval_request_False_bfloat16_Original.json +++ b/deepseek-ai/deepseek-llm-67b-chat_eval_request_False_bfloat16_Original.json @@ -3,13 +3,13 @@ "base_model": null, "revision": "79648bef7658bb824e4630740f6e1484c1b0620b", "precision": "bfloat16", - "params": 67.0, + "params": 67, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-12T12:10:26Z", "model_type": "chat", "job_id": "6693769", "job_start_time": "2024-06-23T04:30:09.740911", "use_chat_template": true -} \ No newline at end of file +} diff --git a/gpt2_eval_request_False_float16_Original.json b/gpt2_eval_request_False_float16_Original.json index 361d37a4c624abd573a3b9c5bc04304bdceedbb2..040f530fb263774be06fc1b1accf7d088300b016 100644 --- a/gpt2_eval_request_False_float16_Original.json +++ b/gpt2_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 0.137, "architectures": "GPT2LMHeadModel", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T15:25:23Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7206095", "job_start_time": "2024-06-26T16:51:58.720075", "use_chat_template": false -} \ No newline at end of file +} diff --git a/grimjim/llama-3-Nephilim-v1-8B_eval_request_False_bfloat16_Original.json b/grimjim/llama-3-Nephilim-v1-8B_eval_request_False_bfloat16_Original.json index e6617139dfb53b7b76dcc22a42108d6ac8b98193..df37c0f23eba7b7451fb06119b06d91734b77d63 100644 --- a/grimjim/llama-3-Nephilim-v1-8B_eval_request_False_bfloat16_Original.json +++ b/grimjim/llama-3-Nephilim-v1-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "grimjim/llama-3-Nephilim-v1-8B", "base_model": "", "revision": "642799c8c768c53e831a03a1224db875116be866", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:54:14Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "grimjim/llama-3-Nephilim-v1-8B", + "base_model": "", + "revision": "642799c8c768c53e831a03a1224db875116be866", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:54:14Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/h4rz3rk4s3/TinyNewsLlama-1.1B_eval_request_False_bfloat16_Original.json b/h4rz3rk4s3/TinyNewsLlama-1.1B_eval_request_False_bfloat16_Original.json index 65ef5bb03ffa6a4a024d173f032cdd9906a99f96..09b1c6c632ab4b4fd11e65f57929eb4933cfcd00 100644 --- a/h4rz3rk4s3/TinyNewsLlama-1.1B_eval_request_False_bfloat16_Original.json +++ b/h4rz3rk4s3/TinyNewsLlama-1.1B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "h4rz3rk4s3/TinyNewsLlama-1.1B", "base_model": "", "revision": "a4e7c60302a70746c6bfc4a79d85f040c27c675d", "precision": "bfloat16", "params": 1.1, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:10:02Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "h4rz3rk4s3/TinyNewsLlama-1.1B", + "base_model": "", + "revision": "a4e7c60302a70746c6bfc4a79d85f040c27c675d", + "precision": "bfloat16", + "params": 1.1, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:10:02Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/h4rz3rk4s3/TinyParlaMintLlama-1.1B_eval_request_False_bfloat16_Original.json b/h4rz3rk4s3/TinyParlaMintLlama-1.1B_eval_request_False_bfloat16_Original.json index f49299333aac2b4b024fe8e436d2d905730af6db..bf18faf1d87ea144b66269980edbccdbae1377fb 100644 --- a/h4rz3rk4s3/TinyParlaMintLlama-1.1B_eval_request_False_bfloat16_Original.json +++ b/h4rz3rk4s3/TinyParlaMintLlama-1.1B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "h4rz3rk4s3/TinyParlaMintLlama-1.1B", "base_model": "", "revision": "0c9aa196c68732bf1b563dcfb4d9c6f835087e9e", "precision": "bfloat16", "params": 1.1, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:11:31Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "h4rz3rk4s3/TinyParlaMintLlama-1.1B", + "base_model": "", + "revision": "0c9aa196c68732bf1b563dcfb4d9c6f835087e9e", + "precision": "bfloat16", + "params": 1.1, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:11:31Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/h4rz3rk4s3/TinyPoliticaLlama-1.1B_eval_request_False_bfloat16_Original.json b/h4rz3rk4s3/TinyPoliticaLlama-1.1B_eval_request_False_bfloat16_Original.json index 7a7e5d598eae16ee793f8077a9d34ec10689695b..1d8c665c99b4a1a90256041ce16bb509c7a22a30 100644 --- a/h4rz3rk4s3/TinyPoliticaLlama-1.1B_eval_request_False_bfloat16_Original.json +++ b/h4rz3rk4s3/TinyPoliticaLlama-1.1B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "h4rz3rk4s3/TinyPoliticaLlama-1.1B", "base_model": "", "revision": "8838d8f094dee1078572cf127f835cdb32117d6f", "precision": "bfloat16", "params": 1.1, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:11:14Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "h4rz3rk4s3/TinyPoliticaLlama-1.1B", + "base_model": "", + "revision": "8838d8f094dee1078572cf127f835cdb32117d6f", + "precision": "bfloat16", + "params": 1.1, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:11:14Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/huggyllama/llama-65b_eval_request_False_float16_Original.json b/huggyllama/llama-65b_eval_request_False_float16_Original.json index d306e02217b23269165f8832e266b969a8141148..127e267fa1bd22095a20c9a7a1dbf10a2deeb10c 100644 --- a/huggyllama/llama-65b_eval_request_False_float16_Original.json +++ b/huggyllama/llama-65b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "huggyllama/llama-65b", "base_model": "", "revision": "49707c5313d34d1c5a846e29cf2a2a650c22c8ee", "precision": "float16", "params": 65.286, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:18:54Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "huggyllama/llama-65b", + "base_model": "", + "revision": "49707c5313d34d1c5a846e29cf2a2a650c22c8ee", + "precision": "float16", + "params": 65.286, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:18:54Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/iRyanBell/ARC1-II_eval_request_False_bfloat16_Original.json b/iRyanBell/ARC1-II_eval_request_False_bfloat16_Original.json index 68376d15f433370201c3b2eec2d4518d395ef314..245fbedf9d3ed7572301cc8f16febeabbe62f546 100644 --- a/iRyanBell/ARC1-II_eval_request_False_bfloat16_Original.json +++ b/iRyanBell/ARC1-II_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "iRyanBell/ARC1-II", "base_model": "", "revision": "c81076b9bdaac0722b33e411a49b07a296e8fae8", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:35:31Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "iRyanBell/ARC1-II", + "base_model": "", + "revision": "c81076b9bdaac0722b33e411a49b07a296e8fae8", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:35:31Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/iRyanBell/ARC1_eval_request_False_bfloat16_Original.json b/iRyanBell/ARC1_eval_request_False_bfloat16_Original.json index 189012fa776a1f60ce9a24a640b741732f1de85e..8ad8c15db4a496220ff6ed74691e008b87fc1f44 100644 --- a/iRyanBell/ARC1_eval_request_False_bfloat16_Original.json +++ b/iRyanBell/ARC1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "iRyanBell/ARC1", "base_model": "", "revision": "28176c0fb77fa43e1410766faf35d2a2681566e9", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:34:58Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "iRyanBell/ARC1", + "base_model": "", + "revision": "28176c0fb77fa43e1410766faf35d2a2681566e9", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:34:58Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/icefog72/IceCocoaRP-7b_eval_request_False_float16_Original.json b/icefog72/IceCocoaRP-7b_eval_request_False_float16_Original.json index 96a72bdbaa2e64bee1d220c613517b2296741e45..09d942510bfd6fc698b7c8de6ffc2dcf1aa6eb10 100644 --- a/icefog72/IceCocoaRP-7b_eval_request_False_float16_Original.json +++ b/icefog72/IceCocoaRP-7b_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:07:59Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205376", "job_start_time": "2024-06-26T14:23:59.225587", "use_chat_template": false -} \ No newline at end of file +} diff --git a/icefog72/IceCoffeeRP-7b_eval_request_False_float16_Original.json b/icefog72/IceCoffeeRP-7b_eval_request_False_float16_Original.json index f689278492f674a7ec5db2932affd93702d7dd6f..814b657b2b1eb0dc4a3aca1a5faa38a555ee76c5 100644 --- a/icefog72/IceCoffeeRP-7b_eval_request_False_float16_Original.json +++ b/icefog72/IceCoffeeRP-7b_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:20:21Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205375", "job_start_time": "2024-06-26T14:23:44.942991", "use_chat_template": false -} \ No newline at end of file +} diff --git a/icefog72/IceSakeV4RP-7b_eval_request_False_bfloat16_Original.json b/icefog72/IceSakeV4RP-7b_eval_request_False_bfloat16_Original.json index 2806e4eed95c7a4d0b6681831151c05311feae4b..1f0aa3fc7bff1add2fdb73feb51208774387f0e3 100644 --- a/icefog72/IceSakeV4RP-7b_eval_request_False_bfloat16_Original.json +++ b/icefog72/IceSakeV4RP-7b_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:04:47Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205373", "job_start_time": "2024-06-26T14:19:56.043855", "use_chat_template": false -} \ No newline at end of file +} diff --git a/icefog72/IceSakeV6RP-7b_eval_request_False_float16_Original.json b/icefog72/IceSakeV6RP-7b_eval_request_False_float16_Original.json index 9b572aeea084bd665ad262d0ca12b00e8b22c506..9d99c5691ac3c69bb195b32b128e65819d7a9f9f 100644 --- a/icefog72/IceSakeV6RP-7b_eval_request_False_float16_Original.json +++ b/icefog72/IceSakeV6RP-7b_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:05:48Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205372", "job_start_time": "2024-06-26T14:19:23.813475", "use_chat_template": false -} \ No newline at end of file +} diff --git a/icefog72/IceSakeV8RP-7b_eval_request_False_float16_Original.json b/icefog72/IceSakeV8RP-7b_eval_request_False_float16_Original.json index 155bd38de2a115f7bfb4cf9447a16cdc0e1bacce..412a5474d6234d188b6475236f26bcc785bcef56 100644 --- a/icefog72/IceSakeV8RP-7b_eval_request_False_float16_Original.json +++ b/icefog72/IceSakeV8RP-7b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "icefog72/IceSakeV8RP-7b", "base_model": "", "revision": "0f8f73fe356583e561479c689aa6597435327f4e", "precision": "float16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T17:26:45Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "icefog72/IceSakeV8RP-7b", + "base_model": "", + "revision": "0f8f73fe356583e561479c689aa6597435327f4e", + "precision": "float16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T17:26:45Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/jetmoe/jetmoe-8b_eval_request_False_bfloat16_Original.json b/jetmoe/jetmoe-8b_eval_request_False_bfloat16_Original.json index c01a6748952fdec1212c6274a9f549b75d06d1c7..5a5dc0a9889d8b4009ac1db5ace71c3fa19d1b4b 100644 --- a/jetmoe/jetmoe-8b_eval_request_False_bfloat16_Original.json +++ b/jetmoe/jetmoe-8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "jetmoe/jetmoe-8b", "base_model": "", "revision": "d8fd02ccf7911aa8148a63c7984ffd2e465b0352", "precision": "bfloat16", "params": 8.522, "architectures": "JetMoEForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T21:16:08Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "jetmoe/jetmoe-8b", + "base_model": "", + "revision": "d8fd02ccf7911aa8148a63c7984ffd2e465b0352", + "precision": "bfloat16", + "params": 8.522, + "architectures": "JetMoEForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T21:16:08Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/jieliu/Storm-7B_eval_request_False_bfloat16_Original.json b/jieliu/Storm-7B_eval_request_False_bfloat16_Original.json index 225ef6982fd42d38846030559aed8439e1342c41..4d5f8aec8a076078a941d97f24dc19b1b02bea30 100644 --- a/jieliu/Storm-7B_eval_request_False_bfloat16_Original.json +++ b/jieliu/Storm-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "jieliu/Storm-7B", "base_model": "", "revision": "71edab8ee6c2578e428b0359158fb0d43133e989", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:10:59Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "jieliu/Storm-7B", + "base_model": "", + "revision": "71edab8ee6c2578e428b0359158fb0d43133e989", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:10:59Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json index 3bd527a3e845a2c81199f50945a09c4ddc954a8b..4f31f517bf51ea664083ebc65da7240fb09ae52e 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01", "base_model": "", "revision": "f4ebbf27d586e94c63f0a7293f565cbd947b824f", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:48:40Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.01", + "base_model": "", + "revision": "f4ebbf27d586e94c63f0a7293f565cbd947b824f", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:48:40Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json index 6361fad95c4fe04b1680358fff0d4f7fca227fd5..c9b946ac828a39b91ad7e254ebf067449776c986 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1", "base_model": "", "revision": "66c7330e9d04b13a68ea7dcf25bc0a71d144221a", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:47:35Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.1-gamma-0.1", + "base_model": "", + "revision": "66c7330e9d04b13a68ea7dcf25bc0a71d144221a", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:47:35Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json index 5629c49502f7aa033a4092cc74cf8462397310b1..fb40400a4e6712768015a793642c07dfc0be085e 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01", "base_model": "", "revision": "4a432be239528ffc654955338982f1f32eb12901", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:45:40Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.01", + "base_model": "", + "revision": "4a432be239528ffc654955338982f1f32eb12901", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:45:40Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json index b3a3afa668dd01b70d7a648edd0760d3cc45aa8e..ff3574f8d891514226032c5213064977ed78ed9a 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1", "base_model": "", "revision": "d6f8ed8dc4b7f74b4312bc0d24aaac275c61958d", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:50:48Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1", + "base_model": "", + "revision": "d6f8ed8dc4b7f74b4312bc0d24aaac275c61958d", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:50:48Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json index a810deedd49d19e2a4215fb2e47c171982c9f7d1..55b5fa78f93414ab6c4ac704ea40fd7a0584d97a 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01", "base_model": "", "revision": "6ab1392c825907b08eff8fbed4c97a3e6e0d6dd9", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:44:48Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.01", + "base_model": "", + "revision": "6ab1392c825907b08eff8fbed4c97a3e6e0d6dd9", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:44:48Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json index 6b315e863304e51f07ab20464165f52d340cecbf..593f4c4b3b67028240a45d15407807c814e03955 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1", "base_model": "", "revision": "a481edaceeaab34f4dc0e90c4d8ec0f72658bbdd", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:49:31Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.5-gamma-0.1", + "base_model": "", + "revision": "a481edaceeaab34f4dc0e90c4d8ec0f72658bbdd", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:49:31Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json index 6147d1fc3d2a5c20b87f6e3bc4efcdb14dd7a546..8506b1c45b761bde19edae58594d4040b4ddf2cc 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01", "base_model": "", "revision": "61f4b44fb917cdb46f0ade9f8fc2a382e0cf67af", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:50:09Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.01", + "base_model": "", + "revision": "61f4b44fb917cdb46f0ade9f8fc2a382e0cf67af", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:50:09Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json index a3a35294593c0f471e53d3a7c8a24d6feb794187..ccff1a0df5b66399256d1b881fd0239969276c8b 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1", "base_model": "", "revision": "139a9bccd0ffb284e670a181a5986a01b1420c6c", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:48:27Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.7-gamma-0.1", + "base_model": "", + "revision": "139a9bccd0ffb284e670a181a5986a01b1420c6c", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:48:27Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json index f43685680e9d510a1a242cd705ace2896572b965..2a676749bdd07e03a62dbffcf3246db18fc62a01 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01", "base_model": "", "revision": "c88c6b65f751156e7bc04c738947387eb55747e9", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:47:23Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.01", + "base_model": "", + "revision": "c88c6b65f751156e7bc04c738947387eb55747e9", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:47:23Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json index c324a8e356d9a38502ad54eb8b566bd7f1573cde..513b46a881524885ca139bacc5bc99e0f0b3880f 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1", "base_model": "", "revision": "818f7e586444b551200862fb234c39bd48d69ae8", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:46:32Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.9-gamma-0.1", + "base_model": "", + "revision": "818f7e586444b551200862fb234c39bd48d69ae8", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:46:32Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json index e819e99df3f43ea19eeb01db9abae3a7046213a5..5065170984fb44f1215703dac98ccfd9f9e4c075 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01", "base_model": "", "revision": "861347cd643d396877d8e560367cf0717c671228", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:45:14Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.01", + "base_model": "", + "revision": "861347cd643d396877d8e560367cf0717c671228", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:45:14Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json index 13084044379cce3173481b67d4bcb13d92e303d3..473e7404ccf933d703cf5e3d25efcf68938e5ff8 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1", "base_model": "", "revision": "2647bc863e6ee686e7174366107eecbd4b37f62e", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:47:10Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.1-gamma-0.1", + "base_model": "", + "revision": "2647bc863e6ee686e7174366107eecbd4b37f62e", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:47:10Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json index 3e0d6c8ac4c2db6ff7bd4a527d995a39d2923a54..6a20709e71e7e35c2ab3af2fbc46fefc4c011c4e 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01", "base_model": "", "revision": "fa77530fe3723d7b15b06b88c3ca6110a8421742", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:50:35Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.01", + "base_model": "", + "revision": "fa77530fe3723d7b15b06b88c3ca6110a8421742", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:50:35Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json index acd136e0ce89b13a5f6c3a3708d176b2498881be..025eab87aabe6127dd2776418092a6773ab7dad1 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1", "base_model": "", "revision": "6fe73aa7f9c5b59297739166e9557089d39e5fc7", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:51:13Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.3-gamma-0.1", + "base_model": "", + "revision": "6fe73aa7f9c5b59297739166e9557089d39e5fc7", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:51:13Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json index 39c6f0ba109a3547ae10e2e1e3e70dd03adc4921..596803be16870d5c6d54e15fa17f4a0c3ed6b6eb 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01", "base_model": "", "revision": "a31f86b538ba8b2983620cc27a741bc9a81a7e2f", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:48:52Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.01", + "base_model": "", + "revision": "a31f86b538ba8b2983620cc27a741bc9a81a7e2f", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:48:52Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json index 77647760fef426ef2dfdaca1e493a04eb713fb1f..d7c3bcdc51c3b5b87bae45d582a32c79df9f4d6f 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1", "base_model": "", "revision": "f9d5bab1c1d0d6890e89b513225d13f68a1c6d75", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:49:44Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.5-gamma-0.1", + "base_model": "", + "revision": "f9d5bab1c1d0d6890e89b513225d13f68a1c6d75", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:49:44Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json index 207254beb48da21a062bd21db6ede448199dd8c8..2ae1c411aed98269eca081a79f0f1b3ed6bebd3b 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01", "base_model": "", "revision": "d30c75506feaec957dc73bc5c040159c310ecf4c", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:45:53Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.01", + "base_model": "", + "revision": "d30c75506feaec957dc73bc5c040159c310ecf4c", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:45:53Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json index 0508333b8097c54b507abe53abeec56924ccc30c..4cab681dc46adb72860a1b477f01ded3a6c64c20 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1", "base_model": "", "revision": "cd52bafe64e82d466d0bc590da5399f2299d24e1", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:48:01Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.7-gamma-0.1", + "base_model": "", + "revision": "cd52bafe64e82d466d0bc590da5399f2299d24e1", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:48:01Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json index 9ae640024eb8ca47b838b0bc8258caa16b231f65..6b9bfc025c99202780f5ccbcadabc528053a9506 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01", "base_model": "", "revision": "4c30fdbe0708afefe50788ea640c3dfab294c77f", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:44:35Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.01", + "base_model": "", + "revision": "4c30fdbe0708afefe50788ea640c3dfab294c77f", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:44:35Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json index f08bf7a5df8f5cdbcd7f61d1aeec26e2f2695069..88f1be7ce5281218b22c1e0a11dcbf58eaedf09f 100644 --- a/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1", "base_model": "", "revision": "378a7cad3e34a1a8b11e77edd95b02ff0d228da2", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:46:57Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_breadcrumbs_ties-density-0.9-gamma-0.1", + "base_model": "", + "revision": "378a7cad3e34a1a8b11e77edd95b02ff0d228da2", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:46:57Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_linear_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_linear_eval_request_False_bfloat16_Original.json index 2b5856ddb29f871a4685edb06a36e40edd8cee93..ec40c1e1f266d286857df22ce09697c32369eabc 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_linear_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_linear_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_linear", "base_model": "", "revision": "abb81fd8fdc2ad32f65befcb7ae369c9837cd563", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:49:18Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_linear", + "base_model": "", + "revision": "abb81fd8fdc2ad32f65befcb7ae369c9837cd563", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:49:18Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1_eval_request_False_bfloat16_Original.json index 0ec332d9514d0d2f048fd882bbe36076d1d7f64a..c8c95d3b911127e087b94718358b49ee2e852061 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1", "base_model": "", "revision": "e7a3a3b955d945f53da8301b958f0b90a28a62d3", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:45:27Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.1", + "base_model": "", + "revision": "e7a3a3b955d945f53da8301b958f0b90a28a62d3", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:45:27Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3_eval_request_False_bfloat16_Original.json index 95f55765b494a0921a2f22d2d14400f5ea17bc93..b9958abec75deed131047bb9d8f53e4a69270b0c 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3", "base_model": "", "revision": "6f966d14d7236f3da6d1ea9ce3bd9b20808e02a9", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:47:49Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.3", + "base_model": "", + "revision": "6f966d14d7236f3da6d1ea9ce3bd9b20808e02a9", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:47:49Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5_eval_request_False_bfloat16_Original.json index 5206920c5a0b04571b33ecadca132dc827661a9a..2a3cc139c586129ac1ddacf3675f7e36bc0a91ce 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5", "base_model": "", "revision": "dad090558e450aec69fd5233a38cdac38792c770", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:49:05Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.5", + "base_model": "", + "revision": "dad090558e450aec69fd5233a38cdac38792c770", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:49:05Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7_eval_request_False_bfloat16_Original.json index 9e99935c6686025d6b876c861a1b88235f9cd6fb..75a639cb905e2556275f0e1bbb59825f11920b82 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7", "base_model": "", "revision": "b14b5cd07feb749e42b0567b1e387b390bed033e", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:44:23Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.7", + "base_model": "", + "revision": "b14b5cd07feb749e42b0567b1e387b390bed033e", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:44:23Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9_eval_request_False_bfloat16_Original.json index 0c12211f82bb593200607d957f3e61d4b669ac3d..bd6d1f73aa39cc6e8a0945791db806aba6d748a0 100644 --- a/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9", "base_model": "", "revision": "3b7a98f31ccc055cca0f821392abdc1cafbb7e25", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:49:56Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_dare_ties-density-0.9", + "base_model": "", + "revision": "3b7a98f31ccc055cca0f821392abdc1cafbb7e25", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:49:56Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_linear_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_linear_eval_request_False_bfloat16_Original.json index 421befa34076ea9463cfaed308f7ed6b3b979259..80a2f594e8e3eb4460f4b9fc1ec87a2f8bcae8d1 100644 --- a/johnsutor/Llama-3-8B-Instruct_linear_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_linear_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_linear", "base_model": "", "revision": "7449157fbc2e8b02e5b6e8ad56b4b2bd7ea82e9d", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:51:00Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_linear", + "base_model": "", + "revision": "7449157fbc2e8b02e5b6e8ad56b4b2bd7ea82e9d", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:51:00Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_ties-density-0.1_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_ties-density-0.1_eval_request_False_bfloat16_Original.json index ef12af6b05f683f80ecfe37fae4fff782b897059..855591e11b0725e0dfe7b84b703145a32f62f8ed 100644 --- a/johnsutor/Llama-3-8B-Instruct_ties-density-0.1_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_ties-density-0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.1", "base_model": "", "revision": "84793f89ebe3be5b5bd9a797d4bbdf374c07419d", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:50:22Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.1", + "base_model": "", + "revision": "84793f89ebe3be5b5bd9a797d4bbdf374c07419d", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:50:22Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_ties-density-0.3_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_ties-density-0.3_eval_request_False_bfloat16_Original.json index f24af976a45c3b39b14ef9569d09b7400cc425ba..ab55a0ffc1bb651b7047a9ba4f9252ded4d18fbc 100644 --- a/johnsutor/Llama-3-8B-Instruct_ties-density-0.3_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_ties-density-0.3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.3", "base_model": "", "revision": "8d051f3eec3fc93a4521073c2d290c4ff9144fc1", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:45:01Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.3", + "base_model": "", + "revision": "8d051f3eec3fc93a4521073c2d290c4ff9144fc1", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:45:01Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_ties-density-0.5_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_ties-density-0.5_eval_request_False_bfloat16_Original.json index 8b6f82e8b2381682a31aad98b369ca6ca2b639c6..51706a8ef0fe5bf5fcb63d65b99c20a1e15c9e73 100644 --- a/johnsutor/Llama-3-8B-Instruct_ties-density-0.5_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_ties-density-0.5_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.5", "base_model": "", "revision": "c857e33c30016960f114e3a049f5dae41d68bfe7", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:46:06Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.5", + "base_model": "", + "revision": "c857e33c30016960f114e3a049f5dae41d68bfe7", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:46:06Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_ties-density-0.7_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_ties-density-0.7_eval_request_False_bfloat16_Original.json index 2989fc2da1d74ef0ea076144be9ad31a262b177b..3a8435e64fcf3d64af74f858029fd62bf5c64f20 100644 --- a/johnsutor/Llama-3-8B-Instruct_ties-density-0.7_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_ties-density-0.7_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.7", "base_model": "", "revision": "8d7d8bbb1e8cba5e51337f97bc3d6d8ae40544d5", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:48:14Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.7", + "base_model": "", + "revision": "8d7d8bbb1e8cba5e51337f97bc3d6d8ae40544d5", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:48:14Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/johnsutor/Llama-3-8B-Instruct_ties-density-0.9_eval_request_False_bfloat16_Original.json b/johnsutor/Llama-3-8B-Instruct_ties-density-0.9_eval_request_False_bfloat16_Original.json index 79f8c275535b89f94ab1d608a3822c609443c408..668345dd7da356de34789c86b1e043f51583c42b 100644 --- a/johnsutor/Llama-3-8B-Instruct_ties-density-0.9_eval_request_False_bfloat16_Original.json +++ b/johnsutor/Llama-3-8B-Instruct_ties-density-0.9_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.9", "base_model": "", "revision": "57c280ce43fe81a23c966b48de6db7f4a85383a3", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:46:19Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "johnsutor/Llama-3-8B-Instruct_ties-density-0.9", + "base_model": "", + "revision": "57c280ce43fe81a23c966b48de6db7f4a85383a3", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:46:19Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/jpacifico/Chocolatine-8B-Instruct-DPO-v1.0_eval_request_False_float16_Original.json b/jpacifico/Chocolatine-8B-Instruct-DPO-v1.0_eval_request_False_float16_Original.json index 97c74e7a1fedb97aeda7a6ff09c7913e719eee5d..09e6da3c5e00100f3620d81659fd09c3e066587d 100644 --- a/jpacifico/Chocolatine-8B-Instruct-DPO-v1.0_eval_request_False_float16_Original.json +++ b/jpacifico/Chocolatine-8B-Instruct-DPO-v1.0_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "jpacifico/Chocolatine-8B-Instruct-DPO-v1.0", "base_model": "", "revision": "20bc1e67d5385ed119578b4038c07970570fb274", "precision": "float16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:19:28Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "jpacifico/Chocolatine-8B-Instruct-DPO-v1.0", + "base_model": "", + "revision": "20bc1e67d5385ed119578b4038c07970570fb274", + "precision": "float16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:19:28Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/jrahn/llama-3-8b-claudstruct-v3_eval_request_False_4bit_Adapter.json b/jrahn/llama-3-8b-claudstruct-v3_eval_request_False_4bit_Adapter.json index 716c7b4e793b6f1b3ba09d0655512596e37e8834..804f5c3521059bca498a3b302e607fb142e8fb47 100644 --- a/jrahn/llama-3-8b-claudstruct-v3_eval_request_False_4bit_Adapter.json +++ b/jrahn/llama-3-8b-claudstruct-v3_eval_request_False_4bit_Adapter.json @@ -3,13 +3,13 @@ "base_model": "meta-llama/Meta-Llama-3-8B-Instruct", "revision": "2edff46653527ecc201b63ad22043cdf7fc7d5ea", "precision": "4bit", - "params": 8.0, + "params": 8, "architectures": "?", "weight_type": "Adapter", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T15:02:13Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7205462", "job_start_time": "2024-06-26T15:04:48.490247", "use_chat_template": true -} \ No newline at end of file +} diff --git a/jrahn/llama-3-8b-infinistruct-v1_eval_request_False_4bit_Adapter.json b/jrahn/llama-3-8b-infinistruct-v1_eval_request_False_4bit_Adapter.json index 3578a03fcd5af6340e7d1fb5b7ffc33d23491502..d4b3880bb056c40dc3d4c035723d8ec7c40a562b 100644 --- a/jrahn/llama-3-8b-infinistruct-v1_eval_request_False_4bit_Adapter.json +++ b/jrahn/llama-3-8b-infinistruct-v1_eval_request_False_4bit_Adapter.json @@ -3,13 +3,13 @@ "base_model": "meta-llama/Meta-Llama-3-8B-Instruct", "revision": "e6543ca2e253c72ae18946baac665daa3b123034", "precision": "4bit", - "params": 8.0, + "params": 8, "architectures": "?", "weight_type": "Adapter", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:49:23Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7205445", "job_start_time": "2024-06-26T14:58:55.768587", "use_chat_template": true -} \ No newline at end of file +} diff --git a/kenhktsui/nano-phi-115M-v0.1_eval_request_False_float16_Original.json b/kenhktsui/nano-phi-115M-v0.1_eval_request_False_float16_Original.json index 8878802f7ec09fe6bc8148eec8ba46b0f65415cc..daf3e96394222703026bbadffeee82d210d25c7d 100644 --- a/kenhktsui/nano-phi-115M-v0.1_eval_request_False_float16_Original.json +++ b/kenhktsui/nano-phi-115M-v0.1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "kenhktsui/nano-phi-115M-v0.1", "base_model": "", "revision": "554e4f046a7991fb9e158a900793b945fea1c6a6", "precision": "float16", "params": 0.115, "architectures": "PhiForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:21:36Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "kenhktsui/nano-phi-115M-v0.1", + "base_model": "", + "revision": "554e4f046a7991fb9e158a900793b945fea1c6a6", + "precision": "float16", + "params": 0.115, + "architectures": "PhiForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:21:36Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/kenhktsui/nano-phi-192M-v0.1_eval_request_False_float16_Original.json b/kenhktsui/nano-phi-192M-v0.1_eval_request_False_float16_Original.json index fecdfadd682bdf4d6ab2c0a34363eeef439575ef..e82f80cb3fe5e5ef8b74fd1b833c653bc8f852f5 100644 --- a/kenhktsui/nano-phi-192M-v0.1_eval_request_False_float16_Original.json +++ b/kenhktsui/nano-phi-192M-v0.1_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "kenhktsui/nano-phi-192M-v0.1", "base_model": "", "revision": "e5c1dbe43d90c34fa07c1c19251a557a90a55ea8", "precision": "float16", "params": 0.192, "architectures": "PhiForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:30:13Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "kenhktsui/nano-phi-192M-v0.1", + "base_model": "", + "revision": "e5c1dbe43d90c34fa07c1c19251a557a90a55ea8", + "precision": "float16", + "params": 0.192, + "architectures": "PhiForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:30:13Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/kevin009/llamaRAGdrama_eval_request_False_bfloat16_Original.json b/kevin009/llamaRAGdrama_eval_request_False_bfloat16_Original.json index 1b0608c33b7c1e3d08af311be972f10ac0f8f8b5..058e478a76ef17aa54e55acc5ca0fd911979d4d3 100644 --- a/kevin009/llamaRAGdrama_eval_request_False_bfloat16_Original.json +++ b/kevin009/llamaRAGdrama_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "kevin009/llamaRAGdrama", "base_model": "", "revision": "8c103ca8fa6dd9a8d3dab81b319408095e9a1ad8", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:15:57Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "kevin009/llamaRAGdrama", + "base_model": "", + "revision": "8c103ca8fa6dd9a8d3dab81b319408095e9a1ad8", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:15:57Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/lemon07r/Llama-3-MahouDevil-8B_eval_request_False_bfloat16_Original.json b/lemon07r/Llama-3-MahouDevil-8B_eval_request_False_bfloat16_Original.json index 669c1c4f2bfc3460c4982ca60716f15bd17230ff..9214f3cbd085c1a48e9a7dbfa325de9eab57b3af 100644 --- a/lemon07r/Llama-3-MahouDevil-8B_eval_request_False_bfloat16_Original.json +++ b/lemon07r/Llama-3-MahouDevil-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "lemon07r/Llama-3-MahouDevil-8B", "base_model": "", "revision": "c0fd11bc0511c42488ed03495ceafdaea3ff4114", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:22:31Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "lemon07r/Llama-3-MahouDevil-8B", + "base_model": "", + "revision": "c0fd11bc0511c42488ed03495ceafdaea3ff4114", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:22:31Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/lemon07r/Llama-3-RedMagic2-8B_eval_request_False_bfloat16_Original.json b/lemon07r/Llama-3-RedMagic2-8B_eval_request_False_bfloat16_Original.json index 1f73f5def8e8b9d9fbc4511c40d5c6ef99081827..c03ab43a9c624885bb33fe90527be5758e3a74e0 100644 --- a/lemon07r/Llama-3-RedMagic2-8B_eval_request_False_bfloat16_Original.json +++ b/lemon07r/Llama-3-RedMagic2-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "lemon07r/Llama-3-RedMagic2-8B", "base_model": "", "revision": "82269f41b3645882c1d503bd66f78b6128173948", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:23:00Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "lemon07r/Llama-3-RedMagic2-8B", + "base_model": "", + "revision": "82269f41b3645882c1d503bd66f78b6128173948", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:23:00Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/lemon07r/Llama-3-RedMagic4-8B_eval_request_False_bfloat16_Original.json b/lemon07r/Llama-3-RedMagic4-8B_eval_request_False_bfloat16_Original.json index cc6e223cb828cce8548b394aab23941738495cd3..94b26074b6793862a611568748ae0e558d79df74 100644 --- a/lemon07r/Llama-3-RedMagic4-8B_eval_request_False_bfloat16_Original.json +++ b/lemon07r/Llama-3-RedMagic4-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "lemon07r/Llama-3-RedMagic4-8B", "base_model": "", "revision": "65ee08a0434f1903a8971640fc3cca6c8ae8590e", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:21:25Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "lemon07r/Llama-3-RedMagic4-8B", + "base_model": "", + "revision": "65ee08a0434f1903a8971640fc3cca6c8ae8590e", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:21:25Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/lemon07r/Lllama-3-RedElixir-8B_eval_request_False_bfloat16_Original.json b/lemon07r/Lllama-3-RedElixir-8B_eval_request_False_bfloat16_Original.json index 6adbd2e2a4492a3dc73f530fa725e3c8807a65c8..f2285f7c606622567620739f30e54fb8fe3eadb0 100644 --- a/lemon07r/Lllama-3-RedElixir-8B_eval_request_False_bfloat16_Original.json +++ b/lemon07r/Lllama-3-RedElixir-8B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "lemon07r/Lllama-3-RedElixir-8B", "base_model": "", "revision": "4ecf1a154f1312be06c18f37230969b6cbdfe925", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:21:44Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "lemon07r/Lllama-3-RedElixir-8B", + "base_model": "", + "revision": "4ecf1a154f1312be06c18f37230969b6cbdfe925", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:21:44Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/lemon07r/llama-3-NeuralMahou-8b_eval_request_False_bfloat16_Original.json b/lemon07r/llama-3-NeuralMahou-8b_eval_request_False_bfloat16_Original.json index 6a762e368340f422d27fde65835e975cb4d43ba7..296cdb7e245402e651c7f194aad0c6a0b72e4a77 100644 --- a/lemon07r/llama-3-NeuralMahou-8b_eval_request_False_bfloat16_Original.json +++ b/lemon07r/llama-3-NeuralMahou-8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "lemon07r/llama-3-NeuralMahou-8b", "base_model": "", "revision": "59a0937df85f9d6d65d15dbb4a7c06b6ad8a0305", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:22:49Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "lemon07r/llama-3-NeuralMahou-8b", + "base_model": "", + "revision": "59a0937df85f9d6d65d15dbb4a7c06b6ad8a0305", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:22:49Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/llm-mle/arb_mle_pretrained_8b_eval_request_False_bfloat16_Original.json b/llm-mle/arb_mle_pretrained_8b_eval_request_False_bfloat16_Original.json index e0252f70bdf8bc8580c1fc6c58787bc8651aff6a..5459cd031af28a7c99547bc0eb69e28a5f9f0c79 100644 --- a/llm-mle/arb_mle_pretrained_8b_eval_request_False_bfloat16_Original.json +++ b/llm-mle/arb_mle_pretrained_8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "llm-mle/arb_mle_pretrained_8b", "base_model": "meta-llama/Meta-Llama-3-8B", "revision": "3a47b6e7df4433065a7c9ee02229458c68ba122e", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T13:57:55Z", "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "llm-mle/arb_mle_pretrained_8b", + "base_model": "meta-llama/Meta-Llama-3-8B", + "revision": "3a47b6e7df4433065a7c9ee02229458c68ba122e", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T13:57:55Z", + "model_type": "🟩 : 🟩 continuously pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/maldv/badger-lambda-llama-3-8b_eval_request_False_bfloat16_Original.json b/maldv/badger-lambda-llama-3-8b_eval_request_False_bfloat16_Original.json index 29a3ea65f707b856441c7e4a1071bc3a201411f2..d2e274551440dd10efe1be799bcfa25f25f9acb9 100644 --- a/maldv/badger-lambda-llama-3-8b_eval_request_False_bfloat16_Original.json +++ b/maldv/badger-lambda-llama-3-8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "maldv/badger-lambda-llama-3-8b", "base_model": "", "revision": "8ef157d0d3c12212ca5e70d354869aed90e03f22", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:39:30Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "maldv/badger-lambda-llama-3-8b", + "base_model": "", + "revision": "8ef157d0d3c12212ca5e70d354869aed90e03f22", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:39:30Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/maldv/badger-writer-llama-3-8b_eval_request_False_bfloat16_Original.json b/maldv/badger-writer-llama-3-8b_eval_request_False_bfloat16_Original.json index 6e9c68e3e4e87d552bad4ea577e05e349b11317e..051ad96db342b434040c8d2c80d4244154ac239d 100644 --- a/maldv/badger-writer-llama-3-8b_eval_request_False_bfloat16_Original.json +++ b/maldv/badger-writer-llama-3-8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "maldv/badger-writer-llama-3-8b", "base_model": "", "revision": "1d8134d01af87e994571ae16ccd7b31cce42418f", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:40:29Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "maldv/badger-writer-llama-3-8b", + "base_model": "", + "revision": "1d8134d01af87e994571ae16ccd7b31cce42418f", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:40:29Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/microsoft/Phi-3-medium-128k-instruct_eval_request_False_float16_Original.json b/microsoft/Phi-3-medium-128k-instruct_eval_request_False_float16_Original.json index 41149149f4aae11e8c2bda94c0204c761a6a955d..f171eb26ae2466760e7205883948b704c4bac8e2 100644 --- a/microsoft/Phi-3-medium-128k-instruct_eval_request_False_float16_Original.json +++ b/microsoft/Phi-3-medium-128k-instruct_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "microsoft/Phi-3-medium-128k-instruct", "base_model": "", "revision": "cae1d42b5577398fd1be9f0746052562ae552886", "precision": "float16", "params": 13.96, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:55:42Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "microsoft/Phi-3-medium-128k-instruct", + "base_model": "", + "revision": "cae1d42b5577398fd1be9f0746052562ae552886", + "precision": "float16", + "params": 13.96, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:55:42Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/migtissera/Llama-3-70B-Synthia-v3.5_eval_request_False_float16_Original.json b/migtissera/Llama-3-70B-Synthia-v3.5_eval_request_False_float16_Original.json index 577a8a45bd550e6279667f190b8f9ae8d9f0add9..7335ab556f541b1b0bde9c481dae2b778b1479fb 100644 --- a/migtissera/Llama-3-70B-Synthia-v3.5_eval_request_False_float16_Original.json +++ b/migtissera/Llama-3-70B-Synthia-v3.5_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "migtissera/Llama-3-70B-Synthia-v3.5", "base_model": "", "revision": "8744db0bccfc18f1847633da9d29fc89b35b4190", "precision": "float16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:35:39Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "migtissera/Llama-3-70B-Synthia-v3.5", + "base_model": "", + "revision": "8744db0bccfc18f1847633da9d29fc89b35b4190", + "precision": "float16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:35:39Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/migtissera/Tess-2.0-Llama-3-70B-v0.2_eval_request_False_float16_Original.json b/migtissera/Tess-2.0-Llama-3-70B-v0.2_eval_request_False_float16_Original.json index 7b51b58f80b469255935b2a0a1e37bb21e2311c9..8cb38d9141ae031cb857d28aed26aeb311bc5e7a 100644 --- a/migtissera/Tess-2.0-Llama-3-70B-v0.2_eval_request_False_float16_Original.json +++ b/migtissera/Tess-2.0-Llama-3-70B-v0.2_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "migtissera/Tess-2.0-Llama-3-70B-v0.2", "base_model": "", "revision": "8a51cb60508869528cf1570bd63c02b3ad3b3f2e", "precision": "float16", "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:35:57Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "migtissera/Tess-2.0-Llama-3-70B-v0.2", + "base_model": "", + "revision": "8a51cb60508869528cf1570bd63c02b3ad3b3f2e", + "precision": "float16", + "params": 70.554, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:35:57Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/migtissera/Tess-v2.5-Phi-3-medium-128k-14B_eval_request_False_float16_Original.json b/migtissera/Tess-v2.5-Phi-3-medium-128k-14B_eval_request_False_float16_Original.json index c7e57d9e49a1ac59f316688c117512f1c1ec715e..6aa025120fc099fcd36297246bd1929705fd0674 100644 --- a/migtissera/Tess-v2.5-Phi-3-medium-128k-14B_eval_request_False_float16_Original.json +++ b/migtissera/Tess-v2.5-Phi-3-medium-128k-14B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "migtissera/Tess-v2.5-Phi-3-medium-128k-14B", "base_model": "", "revision": "3a4dbce32e765f659d418c57f0040d290b8b480d", "precision": "float16", "params": 13.96, "architectures": "Phi3ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:35:09Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "migtissera/Tess-v2.5-Phi-3-medium-128k-14B", + "base_model": "", + "revision": "3a4dbce32e765f659d418c57f0040d290b8b480d", + "precision": "float16", + "params": 13.96, + "architectures": "Phi3ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:35:09Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/migtissera/Tess-v2.5.2-Qwen2-72B_eval_request_False_float16_Original.json b/migtissera/Tess-v2.5.2-Qwen2-72B_eval_request_False_float16_Original.json index 8f54baf8ecfc12bcdc580d9bf7e33236872ce64f..d768d4e3923e65c83815d37ea373790134cdd5dd 100644 --- a/migtissera/Tess-v2.5.2-Qwen2-72B_eval_request_False_float16_Original.json +++ b/migtissera/Tess-v2.5.2-Qwen2-72B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "migtissera/Tess-v2.5.2-Qwen2-72B", "base_model": "", "revision": "0435e634ad9bc8b1172395a535b78e6f25f3594f", "precision": "float16", "params": 72.0, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:34:57Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "migtissera/Tess-v2.5.2-Qwen2-72B", + "base_model": "", + "revision": "0435e634ad9bc8b1172395a535b78e6f25f3594f", + "precision": "float16", + "params": 72, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:34:57Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/mistral-community/mixtral-8x22B-v0.3_eval_request_False_bfloat16_Original.json b/mistral-community/mixtral-8x22B-v0.3_eval_request_False_bfloat16_Original.json index c2a5c999be022afad8fd99130e4ab531c65e7a9b..243d2c482486bd8352e26adedde74ab01cd0a480 100644 --- a/mistral-community/mixtral-8x22B-v0.3_eval_request_False_bfloat16_Original.json +++ b/mistral-community/mixtral-8x22B-v0.3_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 140.63, "architectures": "MixtralForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-13T18:10:12Z", "model_type": "pretrained", "job_id": "5781904", "job_start_time": "2024-06-17T15:34:34.188120", "use_chat_template": false -} \ No newline at end of file +} diff --git a/mistralai/Mixtral-8x22B-Instruct-v0.1_eval_request_False_bfloat16_Original.json b/mistralai/Mixtral-8x22B-Instruct-v0.1_eval_request_False_bfloat16_Original.json index b6829427168315b864261b55f74158d0dd66b8a7..b8435142bf4cabf4d39a24eb4ff10a6592997186 100644 --- a/mistralai/Mixtral-8x22B-Instruct-v0.1_eval_request_False_bfloat16_Original.json +++ b/mistralai/Mixtral-8x22B-Instruct-v0.1_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 140.621, "architectures": "MixtralForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-12T13:38:49Z", "model_type": "pretrained", "job_id": "7191306", "job_start_time": "2024-06-22T18:26:25.180632", "use_chat_template": true -} \ No newline at end of file +} diff --git a/mistralai/Mixtral-8x22B-v0.1_eval_request_False_bfloat16_Original.json b/mistralai/Mixtral-8x22B-v0.1_eval_request_False_bfloat16_Original.json index 10af6c07dec2cdf30cb16bdc2d2bee27ebbda19b..8b654db8304070a449440546e3c453c57d7ec9b5 100644 --- a/mistralai/Mixtral-8x22B-v0.1_eval_request_False_bfloat16_Original.json +++ b/mistralai/Mixtral-8x22B-v0.1_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 140.621, "architectures": "MixtralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-12T13:38:49Z", "model_type": "pretrained", "job_id": "6660774", "job_start_time": "2024-06-22T18:26:25.180632", "use_chat_template": false -} \ No newline at end of file +} diff --git a/nlpguy/StarFusion-alpha1_eval_request_False_bfloat16_Original.json b/nlpguy/StarFusion-alpha1_eval_request_False_bfloat16_Original.json index 2b4ba946a4c041495a908664302431fc006d6152..e8ed248aea82998d5539233bcd0e47e5000b5f24 100644 --- a/nlpguy/StarFusion-alpha1_eval_request_False_bfloat16_Original.json +++ b/nlpguy/StarFusion-alpha1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "nlpguy/StarFusion-alpha1", "base_model": "", "revision": "dccad965a710d7bee001b6387c8307e7c320291e", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:32:05Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "nlpguy/StarFusion-alpha1", + "base_model": "", + "revision": "dccad965a710d7bee001b6387c8307e7c320291e", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:32:05Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/numfa/numfa_v2-3b_eval_request_False_float16_Original.json b/numfa/numfa_v2-3b_eval_request_False_float16_Original.json index cbf7776eafb0a270aefbd68b0222cec2796ed9c6..5c1232cbc354363fb7996514104de10ed4b9a987 100644 --- a/numfa/numfa_v2-3b_eval_request_False_float16_Original.json +++ b/numfa/numfa_v2-3b_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "numfa/numfa_v2-3b", "base_model": "", "revision": "93861f828e68299185b6bacf34655af9e2eae974", "precision": "float16", "params": 3.633, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:35:54Z", "model_type": "\ud83d\udfe2 : \ud83d\udfe2 pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "numfa/numfa_v2-3b", + "base_model": "", + "revision": "93861f828e68299185b6bacf34655af9e2eae974", + "precision": "float16", + "params": 3.633, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:35:54Z", + "model_type": "🟢 : 🟢 pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/openchat/openchat-3.5-0106_eval_request_False_bfloat16_Original.json b/openchat/openchat-3.5-0106_eval_request_False_bfloat16_Original.json index 3e7185c95f6ee496aac71c2f05bb72291cb457c9..b16b864d5fcc9409a30f1dec00e73b7b91dc106f 100644 --- a/openchat/openchat-3.5-0106_eval_request_False_bfloat16_Original.json +++ b/openchat/openchat-3.5-0106_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T14:58:53Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7213573", "job_start_time": "2024-06-26T19:49:29.870973", "use_chat_template": true -} \ No newline at end of file +} diff --git a/openchat/openchat-3.6-8b-20240522_eval_request_False_bfloat16_Original.json b/openchat/openchat-3.6-8b-20240522_eval_request_False_bfloat16_Original.json index e9c2729fbac1bee0211f14afa112b58b6ed92c84..5366caef525665ab8a982a54d73f793375fa03a9 100644 --- a/openchat/openchat-3.6-8b-20240522_eval_request_False_bfloat16_Original.json +++ b/openchat/openchat-3.6-8b-20240522_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T14:59:40Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7208726", "job_start_time": "2024-06-26T19:26:27.987933", "use_chat_template": true -} \ No newline at end of file +} diff --git a/paloalma/ECE-TW3-JRGL-V1_eval_request_False_float16_Original.json b/paloalma/ECE-TW3-JRGL-V1_eval_request_False_float16_Original.json index fe30ff18e92b88b5bc424430d7ef9c88481e3990..c068db2ee11ecf5034f9bb6a6d37aae9832f5374 100644 --- a/paloalma/ECE-TW3-JRGL-V1_eval_request_False_float16_Original.json +++ b/paloalma/ECE-TW3-JRGL-V1_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 68.977, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:09:43Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205390", "job_start_time": "2024-06-26T14:26:31.225324", "use_chat_template": true -} \ No newline at end of file +} diff --git a/paloalma/ECE-TW3-JRGL-V3_eval_request_False_bfloat16_Original.json b/paloalma/ECE-TW3-JRGL-V3_eval_request_False_bfloat16_Original.json index 71eae0127b68ca5f0ab36b3496a216269fb2bddc..9aae682b113eb7785c4b8112726f5be6d6975a75 100644 --- a/paloalma/ECE-TW3-JRGL-V3_eval_request_False_bfloat16_Original.json +++ b/paloalma/ECE-TW3-JRGL-V3_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 72.286, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:10:12Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205747", "job_start_time": "2024-06-26T15:17:17.733778", "use_chat_template": true -} \ No newline at end of file +} diff --git a/paloalma/Le_Triomphant-ECE-TW3_eval_request_False_bfloat16_Original.json b/paloalma/Le_Triomphant-ECE-TW3_eval_request_False_bfloat16_Original.json index bec9b4617cb11e716af38eda23f6b21134a34756..5095cc2dd3d63964d24577f0e029798001bc040c 100644 --- a/paloalma/Le_Triomphant-ECE-TW3_eval_request_False_bfloat16_Original.json +++ b/paloalma/Le_Triomphant-ECE-TW3_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "paloalma/Le_Triomphant-ECE-TW3", "base_model": "", "revision": "f72399253bb3e65c0f55e50461488c098f658a49", "precision": "bfloat16", "params": 72.289, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:04:48Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "paloalma/Le_Triomphant-ECE-TW3", + "base_model": "", + "revision": "f72399253bb3e65c0f55e50461488c098f658a49", + "precision": "bfloat16", + "params": 72.289, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:04:48Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/paloalma/TW3-JRGL-v2_eval_request_False_bfloat16_Original.json b/paloalma/TW3-JRGL-v2_eval_request_False_bfloat16_Original.json index b31a27349db2996d72c78220cea61bcb078c3ea3..26d1a779ed768140205632151ae56c60dd306568 100644 --- a/paloalma/TW3-JRGL-v2_eval_request_False_bfloat16_Original.json +++ b/paloalma/TW3-JRGL-v2_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 72.289, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:09:10Z", - "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", + "model_type": "🤝 : 🤝 base merges and moerges", "job_id": "7205448", "job_start_time": "2024-06-26T15:00:33.644318", "use_chat_template": true -} \ No newline at end of file +} diff --git a/pankajmathur/Al_Dente_v1_8b_eval_request_False_bfloat16_Original.json b/pankajmathur/Al_Dente_v1_8b_eval_request_False_bfloat16_Original.json index 0422ee80a6bdf4317e75a40886065bb6f4e50c5c..d3e9249bea12b0e34b0d4e77d8c8f5a17dc98c69 100644 --- a/pankajmathur/Al_Dente_v1_8b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/Al_Dente_v1_8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/Al_Dente_v1_8b", "base_model": "", "revision": "149d70e04085ecd90510a60f916efc55da1294e7", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:30:39Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/Al_Dente_v1_8b", + "base_model": "", + "revision": "149d70e04085ecd90510a60f916efc55da1294e7", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:30:39Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/Lima_Unchained_70b_eval_request_False_bfloat16_Original.json b/pankajmathur/Lima_Unchained_70b_eval_request_False_bfloat16_Original.json index a546adfe7fe3ae23f49cf609e171b00778b42405..e3a21609acca9fd8598a9a6ae4bf7b6d84ccdaf5 100644 --- a/pankajmathur/Lima_Unchained_70b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/Lima_Unchained_70b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/Lima_Unchained_70b", "base_model": "", "revision": "6b018d5a3bd497011af245efcdd339c6d21717fb", "precision": "bfloat16", "params": 70.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:48Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/Lima_Unchained_70b", + "base_model": "", + "revision": "6b018d5a3bd497011af245efcdd339c6d21717fb", + "precision": "bfloat16", + "params": 70, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:48Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/Mistral-7B-model_45k6e2e4_eval_request_False_bfloat16_Original.json b/pankajmathur/Mistral-7B-model_45k6e2e4_eval_request_False_bfloat16_Original.json index 8c7d40879fd884912502b05cdbff10ff551e9eb2..9eadf13309dfa219d1324f0a3fa976644b25a027 100644 --- a/pankajmathur/Mistral-7B-model_45k6e2e4_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/Mistral-7B-model_45k6e2e4_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/Mistral-7B-model_45k6e2e4", "base_model": "", "revision": "45cc48ccb82991e09f54b22c08cc9c0814d01557", "precision": "bfloat16", "params": 7.0, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:40:52Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/Mistral-7B-model_45k6e2e4", + "base_model": "", + "revision": "45cc48ccb82991e09f54b22c08cc9c0814d01557", + "precision": "bfloat16", + "params": 7, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:40:52Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_007_13b_v2_eval_request_False_bfloat16_Original.json b/pankajmathur/model_007_13b_v2_eval_request_False_bfloat16_Original.json index 9370d42262038d8778e58b85458963cb626ffa60..d126f2edf7b95c3a70a60b2bb117314fa92d36b3 100644 --- a/pankajmathur/model_007_13b_v2_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_007_13b_v2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_007_13b_v2", "base_model": "", "revision": "2c6ddf25cdb134f22e2543121b5a36b41342a9e2", "precision": "bfloat16", "params": 13.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:15Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_007_13b_v2", + "base_model": "", + "revision": "2c6ddf25cdb134f22e2543121b5a36b41342a9e2", + "precision": "bfloat16", + "params": 13, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:15Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_007_eval_request_False_bfloat16_Original.json b/pankajmathur/model_007_eval_request_False_bfloat16_Original.json index 4c046edc8bc8dc30239b0f859be8a95423a34ebb..ab00b7ce76005367ceb923a52f49af8d6968d937 100644 --- a/pankajmathur/model_007_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_007_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_007", "base_model": "", "revision": "c4261ec1043c2c27d844bdfdac3c6d24baf4e85c", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:40:15Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_007", + "base_model": "", + "revision": "c4261ec1043c2c27d844bdfdac3c6d24baf4e85c", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:40:15Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_007_preview_eval_request_False_bfloat16_Original.json b/pankajmathur/model_007_preview_eval_request_False_bfloat16_Original.json index 6f82c46c9ca97c919b5fe8b95314a7126fbeea99..d8d17e7f15a6b462c46d98f03298bbf5514a393b 100644 --- a/pankajmathur/model_007_preview_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_007_preview_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_007_preview", "base_model": "", "revision": "b138ff68415f86ee8a781e747b325718d7827d6b", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:38:16Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_007_preview", + "base_model": "", + "revision": "b138ff68415f86ee8a781e747b325718d7827d6b", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:38:16Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_009_eval_request_False_bfloat16_Original.json b/pankajmathur/model_009_eval_request_False_bfloat16_Original.json index 9310569fe0700f556229cff02755d27726591c03..a97b67b2134969073ef96abac6718989b3a0fa12 100644 --- a/pankajmathur/model_009_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_009_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_009", "base_model": "", "revision": "6f3324d49125daffe68ae30ec56f2d9fdae233f7", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:38:38Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_009", + "base_model": "", + "revision": "6f3324d49125daffe68ae30ec56f2d9fdae233f7", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:38:38Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_101_eval_request_False_bfloat16_Original.json b/pankajmathur/model_101_eval_request_False_bfloat16_Original.json index ecea3b52de75488272c0ebd7bff2b93ef1301733..91066a4a3e4ab2e2f12341ddf45833ed8e9f3883 100644 --- a/pankajmathur/model_101_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_101_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_101", "base_model": "", "revision": "3be9263c488692cd037bdddafcaf484d6acc54af", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:01Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_101", + "base_model": "", + "revision": "3be9263c488692cd037bdddafcaf484d6acc54af", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:01Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_420_eval_request_False_bfloat16_Original.json b/pankajmathur/model_420_eval_request_False_bfloat16_Original.json index 9315fabf35a9cc9ad651f71140a7076ffca2850a..7702ad2387e607e6d2093550a1fe2c1070a147e7 100644 --- a/pankajmathur/model_420_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_420_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_420", "base_model": "", "revision": "bbd7997e1fd3955c5b67ade070a4cf80988624d2", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:24Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_420", + "base_model": "", + "revision": "bbd7997e1fd3955c5b67ade070a4cf80988624d2", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:24Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_420_preview_eval_request_False_bfloat16_Original.json b/pankajmathur/model_420_preview_eval_request_False_bfloat16_Original.json index 8d30ed3d0bd9172cbf9a28dd4f9b4672eb90e217..5a9e9e400be172ec9383d589da2a2c70157bbdd0 100644 --- a/pankajmathur/model_420_preview_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_420_preview_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_420_preview", "base_model": "", "revision": "fc4e7e9167dc130c649734772e94dd93e838c8c1", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:37:38Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_420_preview", + "base_model": "", + "revision": "fc4e7e9167dc130c649734772e94dd93e838c8c1", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:37:38Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/model_51_eval_request_False_bfloat16_Original.json b/pankajmathur/model_51_eval_request_False_bfloat16_Original.json index 71530ccef75da868d7cb27048380ff7f1b969481..7b24656e16568e3b0c35f2f3747ce44135a20978 100644 --- a/pankajmathur/model_51_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/model_51_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/model_51", "base_model": "", "revision": "c285b3dd15327ffe294c4d7e4fbf80eafacd9808", "precision": "bfloat16", "params": 0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:38:26Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/model_51", + "base_model": "", + "revision": "c285b3dd15327ffe294c4d7e4fbf80eafacd9808", + "precision": "bfloat16", + "params": 0, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:38:26Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_13b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_13b_eval_request_False_bfloat16_Original.json index 3f646140419fb4ead91648a99871f8bc2b79c68f..55a88a2dcf819c8073107cb9072e65a5d8f040b4 100644 --- a/pankajmathur/orca_mini_13b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_13b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_13b", "base_model": "", "revision": "8de6376c340be029ab46c8f5ca09eb40ab99357e", "precision": "bfloat16", "params": 13.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:38Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_13b", + "base_model": "", + "revision": "8de6376c340be029ab46c8f5ca09eb40ab99357e", + "precision": "bfloat16", + "params": 13, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:38Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_3b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_3b_eval_request_False_bfloat16_Original.json index 7ca2c8ae7e8a59972a88de922c26c566bdc93847..58f97d751c0640b264bff5c20608485889b4792c 100644 --- a/pankajmathur/orca_mini_3b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_3b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_3b", "base_model": "", "revision": "31e1a7bc3f7ea2f247b432d60036d975b8d590e9", "precision": "bfloat16", "params": 3.426, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:36:37Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_3b", + "base_model": "", + "revision": "31e1a7bc3f7ea2f247b432d60036d975b8d590e9", + "precision": "bfloat16", + "params": 3.426, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:36:37Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_7b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_7b_eval_request_False_bfloat16_Original.json index 131f5c85b2aceee83836d943c5e22a9421c0b67b..8824776950de18ab14a99334654e9f6b9e94c722 100644 --- a/pankajmathur/orca_mini_7b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_7b", "base_model": "", "revision": "fec86e316b7b98d7be6cf74e98fb927092077abb", "precision": "bfloat16", "params": 7.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:38:04Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_7b", + "base_model": "", + "revision": "fec86e316b7b98d7be6cf74e98fb927092077abb", + "precision": "bfloat16", + "params": 7, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:38:04Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v2_13b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v2_13b_eval_request_False_bfloat16_Original.json index b6e403bfba01906f6be21463a15022d892836ecd..a372867c4c69f50f2861e98932263170b9fe57e5 100644 --- a/pankajmathur/orca_mini_v2_13b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v2_13b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v2_13b", "base_model": "", "revision": "60d75cc8478ed4b0be8d6ece6d73bac0909eec8e", "precision": "bfloat16", "params": 13.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:38:51Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v2_13b", + "base_model": "", + "revision": "60d75cc8478ed4b0be8d6ece6d73bac0909eec8e", + "precision": "bfloat16", + "params": 13, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:38:51Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v2_7b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v2_7b_eval_request_False_bfloat16_Original.json index 922ed7bfb12290794a2e0d44daa425a7c55c6482..cb3b68e8adc4fcca8fd1c763a1c9699c4331da60 100644 --- a/pankajmathur/orca_mini_v2_7b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v2_7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v2_7b", "base_model": "", "revision": "66d3f32a4a6bca0a2a261f1bdb54d2582028f75f", "precision": "bfloat16", "params": 7.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:36:54Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v2_7b", + "base_model": "", + "revision": "66d3f32a4a6bca0a2a261f1bdb54d2582028f75f", + "precision": "bfloat16", + "params": 7, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:36:54Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v3_13b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v3_13b_eval_request_False_bfloat16_Original.json index 2ec871befa3577e0926a7df50dc85f3d7541771a..9e7cd58de261f630744fbc6f0437e2cb8c6b8e8e 100644 --- a/pankajmathur/orca_mini_v3_13b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v3_13b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v3_13b", "base_model": "", "revision": "7d6e567d24ce2f228beaf54e89c17b0e750bfe99", "precision": "bfloat16", "params": 13.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:37:18Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v3_13b", + "base_model": "", + "revision": "7d6e567d24ce2f228beaf54e89c17b0e750bfe99", + "precision": "bfloat16", + "params": 13, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:37:18Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v3_70b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v3_70b_eval_request_False_bfloat16_Original.json index 5ebf058251b4410fcd43727b6e12c32c0fd50d47..efa29ea5f6fb131772fda71503d626d2ec819ff1 100644 --- a/pankajmathur/orca_mini_v3_70b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v3_70b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v3_70b", "base_model": "", "revision": "e8e856dfb5c737d1906b50f9e65fd3a4f8d77422", "precision": "bfloat16", "params": 70.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:37:48Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v3_70b", + "base_model": "", + "revision": "e8e856dfb5c737d1906b50f9e65fd3a4f8d77422", + "precision": "bfloat16", + "params": 70, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:37:48Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v3_7b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v3_7b_eval_request_False_bfloat16_Original.json index 342f6733b937a47fa79b5c9985807531a2125ad7..c6ea8415a8eb546a392642ae796ca014efffb5e0 100644 --- a/pankajmathur/orca_mini_v3_7b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v3_7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v3_7b", "base_model": "", "revision": "6252eb7ca29da8d951ae7d2bca948bf84e04a2b9", "precision": "bfloat16", "params": 7.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:39:59Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v3_7b", + "base_model": "", + "revision": "6252eb7ca29da8d951ae7d2bca948bf84e04a2b9", + "precision": "bfloat16", + "params": 7, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:39:59Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pankajmathur/orca_mini_v4_8b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v4_8b_eval_request_False_bfloat16_Original.json index d621722b2d25f25e27776abd0bf7043c55a4d8cc..e97eac0a836ec5ceacbc196f133cdfb13f780fd5 100644 --- a/pankajmathur/orca_mini_v4_8b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v4_8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v4_8b", "base_model": "", "revision": "33ff13689d4402e4993fe1fd2f4fa68e8501b8dd", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:31:39Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v4_8b", + "base_model": "", + "revision": "33ff13689d4402e4993fe1fd2f4fa68e8501b8dd", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:31:39Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v5_8b_dpo_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v5_8b_dpo_eval_request_False_bfloat16_Original.json index 45b4838b4c5566497f3fc3f51dfa720c0a30a8bb..8b1def48f74e8c975f3b2fb0929347ade6172fad 100644 --- a/pankajmathur/orca_mini_v5_8b_dpo_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v5_8b_dpo_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v5_8b_dpo", "base_model": "", "revision": "fdc0d0aaa85a58f1abaf2c24ce0ddca10c08f0f1", "precision": "bfloat16", "params": 8.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:31:26Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v5_8b_dpo", + "base_model": "", + "revision": "fdc0d0aaa85a58f1abaf2c24ce0ddca10c08f0f1", + "precision": "bfloat16", + "params": 8, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:31:26Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v5_8b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v5_8b_eval_request_False_bfloat16_Original.json index 8a3c58a7aab1670a189c0dbd25d6b5f4047d07ec..a091dd89b920369b7450ef418b9274ae61b3a1a1 100644 --- a/pankajmathur/orca_mini_v5_8b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v5_8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v5_8b", "base_model": "", "revision": "f57c84d4cc0b3b74549458c0d38e868bd7fffad1", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:31:58Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v5_8b", + "base_model": "", + "revision": "f57c84d4cc0b3b74549458c0d38e868bd7fffad1", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:31:58Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v5_8b_orpo_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v5_8b_orpo_eval_request_False_bfloat16_Original.json index 46cc6aeae81a069f3704e82d80998e1b1c0f39ec..d9e3b47404fd7fb63836f9693132e4512153911e 100644 --- a/pankajmathur/orca_mini_v5_8b_orpo_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v5_8b_orpo_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v5_8b_orpo", "base_model": "", "revision": "4cdc018043ef439f15bd8a09c4f09c6bc528dfc7", "precision": "bfloat16", "params": 8.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:31:00Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v5_8b_orpo", + "base_model": "", + "revision": "4cdc018043ef439f15bd8a09c4f09c6bc528dfc7", + "precision": "bfloat16", + "params": 8, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:31:00Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v6_8b_dpo_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v6_8b_dpo_eval_request_False_bfloat16_Original.json index 2a9c640c89f965bd438b7b0323a09bf3cb2d4fbe..3df91f0fd8288b1dd2995437d6144da5df05b4d5 100644 --- a/pankajmathur/orca_mini_v6_8b_dpo_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v6_8b_dpo_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v6_8b_dpo", "base_model": "", "revision": "ebb11b63839d38e8c03c7ecac012e047fcb2346e", "precision": "bfloat16", "params": 8.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:32:10Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v6_8b_dpo", + "base_model": "", + "revision": "ebb11b63839d38e8c03c7ecac012e047fcb2346e", + "precision": "bfloat16", + "params": 8, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:32:10Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v6_8b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v6_8b_eval_request_False_bfloat16_Original.json index 8da76a6d75f5a43c3fa48c3cce5b567ac70cb89e..cfc195b3dd1c88a4df45faa8d2b715dee3e13e27 100644 --- a/pankajmathur/orca_mini_v6_8b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v6_8b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v6_8b", "base_model": "", "revision": "e95dc8e4c6b6ca5957b657cc2d905683142eaf3e", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:30:47Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v6_8b", + "base_model": "", + "revision": "e95dc8e4c6b6ca5957b657cc2d905683142eaf3e", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:30:47Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v7_72b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v7_72b_eval_request_False_bfloat16_Original.json index 03be098bb15b5fd602a717891701594ea1ecd599..94ccf984d4908833687f08fa8022599b2e6b2a5e 100644 --- a/pankajmathur/orca_mini_v7_72b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v7_72b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v7_72b", "base_model": "", "revision": "447f11912cfa496e32e188a55214043a05760d3a", "precision": "bfloat16", "params": 72.0, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:29:21Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v7_72b", + "base_model": "", + "revision": "447f11912cfa496e32e188a55214043a05760d3a", + "precision": "bfloat16", + "params": 72, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:29:21Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pankajmathur/orca_mini_v7_7b_eval_request_False_bfloat16_Original.json b/pankajmathur/orca_mini_v7_7b_eval_request_False_bfloat16_Original.json index d9ec2414f9d479c6d8be8429a51039a6a1088eae..c91e166c4834c9df5b77f56263005142e7a3e0b8 100644 --- a/pankajmathur/orca_mini_v7_7b_eval_request_False_bfloat16_Original.json +++ b/pankajmathur/orca_mini_v7_7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pankajmathur/orca_mini_v7_7b", "base_model": "", "revision": "f5e84ff6ea25fb4585908ea45d1520bac416d803", "precision": "bfloat16", "params": 7.616, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T22:30:27Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "pankajmathur/orca_mini_v7_7b", + "base_model": "", + "revision": "f5e84ff6ea25fb4585908ea45d1520bac416d803", + "precision": "bfloat16", + "params": 7.616, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T22:30:27Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/paulml/ECE-ILAB-Q1_eval_request_False_bfloat16_Original.json b/paulml/ECE-ILAB-Q1_eval_request_False_bfloat16_Original.json index d46fda67ef2e59c604a369784b2321e5a7e0cc4e..73585030fa74b4f73fdf2d51cfbe6330f00db13c 100644 --- a/paulml/ECE-ILAB-Q1_eval_request_False_bfloat16_Original.json +++ b/paulml/ECE-ILAB-Q1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "paulml/ECE-ILAB-Q1", "base_model": "", "revision": "393bea0ee85e4c752acd5fd77ce07f577fc13bd9", "precision": "bfloat16", "params": 72.706, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:05:59Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "paulml/ECE-ILAB-Q1", + "base_model": "", + "revision": "393bea0ee85e4c752acd5fd77ce07f577fc13bd9", + "precision": "bfloat16", + "params": 72.706, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:05:59Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/paulml/Qwen2-125B-Instruct_eval_request_False_float16_Original.json b/paulml/Qwen2-125B-Instruct_eval_request_False_float16_Original.json index 95db507585e80fb3f964b8bf0bf1711efee3a6c1..9ab3f5b9d1b23eb95b37590ddc9e6dc407980674 100644 --- a/paulml/Qwen2-125B-Instruct_eval_request_False_float16_Original.json +++ b/paulml/Qwen2-125B-Instruct_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "paulml/Qwen2-125B-Instruct", "base_model": "", "revision": "724cc525e9745ce6f05952c16104cf04bdca39fb", "precision": "float16", "params": 125.367, "architectures": "Qwen2ForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:07:21Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "paulml/Qwen2-125B-Instruct", + "base_model": "", + "revision": "724cc525e9745ce6f05952c16104cf04bdca39fb", + "precision": "float16", + "params": 125.367, + "architectures": "Qwen2ForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:07:21Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/pszemraj/Llama-3-6.3b-v0.1_eval_request_False_bfloat16_Original.json b/pszemraj/Llama-3-6.3b-v0.1_eval_request_False_bfloat16_Original.json index 479209553def9dd8e458a299fb0d1ba61b109db9..1cf05c88126bea9157035c9245fc4d817feffd2b 100644 --- a/pszemraj/Llama-3-6.3b-v0.1_eval_request_False_bfloat16_Original.json +++ b/pszemraj/Llama-3-6.3b-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pszemraj/Llama-3-6.3b-v0.1", "base_model": "", "revision": "7000b39346162f95f19aa4ca3975242db61902d7", "precision": "bfloat16", "params": 6.3, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:41:05Z", "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pszemraj/Llama-3-6.3b-v0.1", + "base_model": "", + "revision": "7000b39346162f95f19aa4ca3975242db61902d7", + "precision": "bfloat16", + "params": 6.3, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:41:05Z", + "model_type": "🟩 : 🟩 continuously pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pszemraj/Mistral-v0.3-6B_eval_request_False_bfloat16_Original.json b/pszemraj/Mistral-v0.3-6B_eval_request_False_bfloat16_Original.json index a26befb1c3c6c972c1869c0a58d0e84d01e506ab..4e2352bdf26b1fe2daff94db3a2512267d1fda68 100644 --- a/pszemraj/Mistral-v0.3-6B_eval_request_False_bfloat16_Original.json +++ b/pszemraj/Mistral-v0.3-6B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pszemraj/Mistral-v0.3-6B", "base_model": "", "revision": "ae11a699012b83996361f04808f4d45debf3b01c", "precision": "bfloat16", "params": 5.939, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:43:17Z", "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pszemraj/Mistral-v0.3-6B", + "base_model": "", + "revision": "ae11a699012b83996361f04808f4d45debf3b01c", + "precision": "bfloat16", + "params": 5.939, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:43:17Z", + "model_type": "🟩 : 🟩 continuously pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/pszemraj/stablelm-4e1t-2b-v0.1_eval_request_False_bfloat16_Original.json b/pszemraj/stablelm-4e1t-2b-v0.1_eval_request_False_bfloat16_Original.json index 524d2c08928a76a382bd73152d1d2dc1e311c8c3..4001aaeaa25dff75764281a3ec407350380e3d5d 100644 --- a/pszemraj/stablelm-4e1t-2b-v0.1_eval_request_False_bfloat16_Original.json +++ b/pszemraj/stablelm-4e1t-2b-v0.1_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "pszemraj/stablelm-4e1t-2b-v0.1", "base_model": "", "revision": "e502a49fbf6df3866f4e1bd54d3f0fe1716aae43", "precision": "bfloat16", "params": 2.002, "architectures": "StableLmForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:43:29Z", "model_type": "\ud83d\udfe9 : \ud83d\udfe9 continuously pretrained", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "pszemraj/stablelm-4e1t-2b-v0.1", + "base_model": "", + "revision": "e502a49fbf6df3866f4e1bd54d3f0fe1716aae43", + "precision": "bfloat16", + "params": 2.002, + "architectures": "StableLmForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:43:29Z", + "model_type": "🟩 : 🟩 continuously pretrained", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/rwitz/go-bruins-v2_eval_request_False_float16_Original.json b/rwitz/go-bruins-v2_eval_request_False_float16_Original.json index 9ee8ccc6f33aa82eac1791a373276d4780483e38..956e5f65897b66d324c93d0939baafe2bc2cd929 100644 --- a/rwitz/go-bruins-v2_eval_request_False_float16_Original.json +++ b/rwitz/go-bruins-v2_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T15:06:03Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7205488", "job_start_time": "2024-06-26T15:08:57.890128", "use_chat_template": true -} \ No newline at end of file +} diff --git a/shyamieee/Padma-v7.0_eval_request_False_bfloat16_Original.json b/shyamieee/Padma-v7.0_eval_request_False_bfloat16_Original.json index edd8edfd95abee7faa3a2735364e1e800520486f..f57274d8fbaeb6e5cb3b9042d70893c64e1ac673 100644 --- a/shyamieee/Padma-v7.0_eval_request_False_bfloat16_Original.json +++ b/shyamieee/Padma-v7.0_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "shyamieee/Padma-v7.0", "base_model": "", "revision": "caf70bd6e2f819cc6a18dda8516f2cbdc101fdde", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T20:20:45Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "shyamieee/Padma-v7.0", + "base_model": "", + "revision": "caf70bd6e2f819cc6a18dda8516f2cbdc101fdde", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T20:20:45Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/softwareweaver/Twilight-Miqu-146B_eval_request_False_float16_Original.json b/softwareweaver/Twilight-Miqu-146B_eval_request_False_float16_Original.json index ad41d7143ef8b9a5ab0a3f275a4a4dce1d5e924a..92c70a338aec2d1c63963e8dd543061006883212 100644 --- a/softwareweaver/Twilight-Miqu-146B_eval_request_False_float16_Original.json +++ b/softwareweaver/Twilight-Miqu-146B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "softwareweaver/Twilight-Miqu-146B", "base_model": "", "revision": "4483a5d228a9f9471cfa2014e1bf68d452f15cba", "precision": "float16", "params": 145.986, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:11:33Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "softwareweaver/Twilight-Miqu-146B", + "base_model": "", + "revision": "4483a5d228a9f9471cfa2014e1bf68d452f15cba", + "precision": "float16", + "params": 145.986, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:11:33Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/stabilityai/StableBeluga2_eval_request_False_bfloat16_Original.json b/stabilityai/StableBeluga2_eval_request_False_bfloat16_Original.json index b5d05e9d4dc877fb36c37014a27ff0f779ff5b22..8fdb51e5214631b4a7bfe19868ebc097386c508e 100644 --- a/stabilityai/StableBeluga2_eval_request_False_bfloat16_Original.json +++ b/stabilityai/StableBeluga2_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 68.977, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-13T18:10:18Z", "model_type": "finetuned", "job_id": "6660953", "job_start_time": "2024-06-22T18:41:32.210506", "use_chat_template": false -} \ No newline at end of file +} diff --git a/tenyx/Llama3-TenyxChat-70B_eval_request_False_bfloat16_Original.json b/tenyx/Llama3-TenyxChat-70B_eval_request_False_bfloat16_Original.json index 999de1442f826b06845d443fdeb702faa139af8d..cddd5164d0ef99cd64694da395f4deba1c5f9f1c 100644 --- a/tenyx/Llama3-TenyxChat-70B_eval_request_False_bfloat16_Original.json +++ b/tenyx/Llama3-TenyxChat-70B_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 70.554, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T17:13:02Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7208628", "job_start_time": "2024-06-26T19:24:21.156071", "use_chat_template": true -} \ No newline at end of file +} diff --git a/tiiuae/falcon-180B-chat_eval_request_False_float16_Original.json b/tiiuae/falcon-180B-chat_eval_request_False_float16_Original.json index 508dc1c815315d3c3487cd89ea1be12af53df210..a2a4906db832486a18adca773c01a7767fdf1fee 100644 --- a/tiiuae/falcon-180B-chat_eval_request_False_float16_Original.json +++ b/tiiuae/falcon-180B-chat_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "tiiuae/falcon-180B-chat", "base_model": "", "revision": "6882299beda36df40e683c4478edfcc94f2cde5f", "precision": "float16", "params": 179.523, "architectures": "FalconForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T19:59:30Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "tiiuae/falcon-180B-chat", + "base_model": "", + "revision": "6882299beda36df40e683c4478edfcc94f2cde5f", + "precision": "float16", + "params": 179.523, + "architectures": "FalconForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T19:59:30Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/togethercomputer/GPT-JT-6B-v1_eval_request_False_float16_Original.json b/togethercomputer/GPT-JT-6B-v1_eval_request_False_float16_Original.json index f14682ca61984582327bd76311485fcb470f3789..3211f80f5c2eaa65641880010decc560c558587c 100644 --- a/togethercomputer/GPT-JT-6B-v1_eval_request_False_float16_Original.json +++ b/togethercomputer/GPT-JT-6B-v1_eval_request_False_float16_Original.json @@ -3,13 +3,13 @@ "base_model": null, "revision": "f34aa35f906895602c1f86f5685e598afdea8051", "precision": "float16", - "params": 6.0, + "params": 6, "architectures": "GPTJForCausalLM", "weight_type": "Original", - "status": "FINISHED", + "status": "PENDING", "submitted_time": "2024-06-12T12:07:34Z", "model_type": "finetuned", "job_id": "6660824", "job_start_time": "2024-06-22T18:30:58.591597", "use_chat_template": false -} \ No newline at end of file +} diff --git a/uukuguy/speechless-code-mistral-7b-v1.0_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-code-mistral-7b-v1.0_eval_request_False_bfloat16_Original.json index 262a5f698d78ff616ccc3920db683583ee7a597f..7ebf768e0317f4c991f1fe36f51fc674015c73bb 100644 --- a/uukuguy/speechless-code-mistral-7b-v1.0_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-code-mistral-7b-v1.0_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-code-mistral-7b-v1.0", "base_model": "", "revision": "1862e0a712efc6002112e9c1235a197d58419b37", "precision": "bfloat16", "params": 7.0, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:15:32Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-code-mistral-7b-v1.0", + "base_model": "", + "revision": "1862e0a712efc6002112e9c1235a197d58419b37", + "precision": "bfloat16", + "params": 7, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:15:32Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-codellama-34b-v2.0_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-codellama-34b-v2.0_eval_request_False_bfloat16_Original.json index f8293c86ac1effc6e5cfd7b1c725f02e781cd3ea..7a78d7ab6edc4dc9beab738de69b05276c513211 100644 --- a/uukuguy/speechless-codellama-34b-v2.0_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-codellama-34b-v2.0_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-codellama-34b-v2.0", "base_model": "", "revision": "419bc42a254102d6a5486a1a854068e912c4047c", "precision": "bfloat16", "params": 34.0, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:16:03Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-codellama-34b-v2.0", + "base_model": "", + "revision": "419bc42a254102d6a5486a1a854068e912c4047c", + "precision": "bfloat16", + "params": 34, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:16:03Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-coder-ds-6.7b_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-coder-ds-6.7b_eval_request_False_bfloat16_Original.json index 9d57ddfa55b1c2bfb05d0684a51a95fb4b905b99..ddcef1492c6bb72c15cba2609e2752b533df96ff 100644 --- a/uukuguy/speechless-coder-ds-6.7b_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-coder-ds-6.7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-coder-ds-6.7b", "base_model": "", "revision": "c813a5268c6dfe267a720ad3b51773f1ab0feb59", "precision": "bfloat16", "params": 6.7, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:21:31Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-coder-ds-6.7b", + "base_model": "", + "revision": "c813a5268c6dfe267a720ad3b51773f1ab0feb59", + "precision": "bfloat16", + "params": 6.7, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:21:31Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-instruct-mistral-7b-v0.2_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-instruct-mistral-7b-v0.2_eval_request_False_bfloat16_Original.json index 446445009dfa58512ba04d0fd1bd18a292044a22..0ff44e051e1816fe59358972d7e3559d9cc7842a 100644 --- a/uukuguy/speechless-instruct-mistral-7b-v0.2_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-instruct-mistral-7b-v0.2_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-instruct-mistral-7b-v0.2", "base_model": "", "revision": "87a4d214f7d028d61c3dc013a7410b3c34a24072", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:14:34Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-instruct-mistral-7b-v0.2", + "base_model": "", + "revision": "87a4d214f7d028d61c3dc013a7410b3c34a24072", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:14:34Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-llama2-13b_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-llama2-13b_eval_request_False_bfloat16_Original.json index 90f1df97649ef2ca7dbbbbfb3451db6b9405c2d4..c9e2918dd27e0880cd915357133fc9759481d760 100644 --- a/uukuguy/speechless-llama2-13b_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-llama2-13b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-llama2-13b", "base_model": "", "revision": "3587df1ae1e5cc877ba7f35ba745f15468416eb3", "precision": "bfloat16", "params": 13.016, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:16:46Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-llama2-13b", + "base_model": "", + "revision": "3587df1ae1e5cc877ba7f35ba745f15468416eb3", + "precision": "bfloat16", + "params": 13.016, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:16:46Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b_eval_request_False_bfloat16_Original.json index 90ae50aeea707e28d1be4f90a16cb58292d48ca7..498c5b45743a37440fca06051d6eadd73d05c776 100644 --- a/uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b", "base_model": "", "revision": "954cc87b0ed5fa280126de546daf648861031512", "precision": "bfloat16", "params": 13.016, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:19:50Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-llama2-hermes-orca-platypus-wizardlm-13b", + "base_model": "", + "revision": "954cc87b0ed5fa280126de546daf648861031512", + "precision": "bfloat16", + "params": 13.016, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:19:50Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b_eval_request_False_bfloat16_Original.json index 9265696ab9f2f216fc5584e4ba8108ada76ba9d5..e02535da4984647d659618bd65b65d2919474911 100644 --- a/uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b", "base_model": "", "revision": "b1de043468a15198b55a6509293a4ee585139043", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:22:16Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-mistral-dolphin-orca-platypus-samantha-7b", + "base_model": "", + "revision": "b1de043468a15198b55a6509293a4ee585139043", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:22:16Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/uukuguy/speechless-zephyr-code-functionary-7b_eval_request_False_bfloat16_Original.json b/uukuguy/speechless-zephyr-code-functionary-7b_eval_request_False_bfloat16_Original.json index f1d741c69cb7c0feb9f35543c1e27f08e790ada4..1280dc8a8547ce10ea763992bf27a0346d4a14b0 100644 --- a/uukuguy/speechless-zephyr-code-functionary-7b_eval_request_False_bfloat16_Original.json +++ b/uukuguy/speechless-zephyr-code-functionary-7b_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "uukuguy/speechless-zephyr-code-functionary-7b", "base_model": "", "revision": "d66fc775ece679966e352195c42444e9c70af7fa", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T16:18:18Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": false} \ No newline at end of file +{ + "model": "uukuguy/speechless-zephyr-code-functionary-7b", + "base_model": "", + "revision": "d66fc775ece679966e352195c42444e9c70af7fa", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T16:18:18Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": false +} diff --git a/vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original.json b/vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original.json index 09382754ac4f8dcbcd5483391dcfc1a35af7232c..bd6dc33ab5d63968caf4ff7fd264ef1ded2fd656 100644 --- a/vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original.json +++ b/vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/CarbonBeagle-11B", "base_model": "", "revision": "3fe9bf5327606d013b182fed17a472f5f043759b", "precision": "float16", "params": 10.732, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:42:47Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/CarbonBeagle-11B", + "base_model": "", + "revision": "3fe9bf5327606d013b182fed17a472f5f043759b", + "precision": "float16", + "params": 10.732, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:42:47Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/Configurable-Yi-1.5-9B-Chat_eval_request_False_float16_Original.json b/vicgalle/Configurable-Yi-1.5-9B-Chat_eval_request_False_float16_Original.json index 34197a6c56a62e1dc9c2182f9f6a7f90b23eee88..70ca86f3d493e400c9dc6bb72f94e2c86f522345 100644 --- a/vicgalle/Configurable-Yi-1.5-9B-Chat_eval_request_False_float16_Original.json +++ b/vicgalle/Configurable-Yi-1.5-9B-Chat_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/Configurable-Yi-1.5-9B-Chat", "base_model": "", "revision": "992cb2232caae78eff6a836b2e0642f7cbf6018e", "precision": "float16", "params": 8.829, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:41:52Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/Configurable-Yi-1.5-9B-Chat", + "base_model": "", + "revision": "992cb2232caae78eff6a836b2e0642f7cbf6018e", + "precision": "float16", + "params": 8.829, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:41:52Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/ConfigurableBeagle-11B_eval_request_False_float16_Original.json b/vicgalle/ConfigurableBeagle-11B_eval_request_False_float16_Original.json index e33ee4429e0239c98be4731a6faa7d82aceef468..f2abd84369f829fd7f7b4917b48e2317d71c4810 100644 --- a/vicgalle/ConfigurableBeagle-11B_eval_request_False_float16_Original.json +++ b/vicgalle/ConfigurableBeagle-11B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/ConfigurableBeagle-11B", "base_model": "", "revision": "bbc16dbf94b8e8a99bb3e2ada6755faf9c2990dd", "precision": "float16", "params": 10.732, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:29:49Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/ConfigurableBeagle-11B", + "base_model": "", + "revision": "bbc16dbf94b8e8a99bb3e2ada6755faf9c2990dd", + "precision": "float16", + "params": 10.732, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:29:49Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/ConfigurableHermes-7B_eval_request_False_float16_Original.json b/vicgalle/ConfigurableHermes-7B_eval_request_False_float16_Original.json index 390b85dbef3c3d0ce14f2e4b40ce6f606bea7221..c65a58e9d3b30afde1cff827faa621abbe716550 100644 --- a/vicgalle/ConfigurableHermes-7B_eval_request_False_float16_Original.json +++ b/vicgalle/ConfigurableHermes-7B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/ConfigurableHermes-7B", "base_model": "", "revision": "1333a88eaf6591836b2d9825d1eaec7260f336c9", "precision": "float16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:39:56Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/ConfigurableHermes-7B", + "base_model": "", + "revision": "1333a88eaf6591836b2d9825d1eaec7260f336c9", + "precision": "float16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:39:56Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/ConfigurableSOLAR-10.7B_eval_request_False_float16_Original.json b/vicgalle/ConfigurableSOLAR-10.7B_eval_request_False_float16_Original.json index b1939e21237f817b99664f3613c10c135d033434..93a33c6cef69fe5227201b20b9d07101a5afda19 100644 --- a/vicgalle/ConfigurableSOLAR-10.7B_eval_request_False_float16_Original.json +++ b/vicgalle/ConfigurableSOLAR-10.7B_eval_request_False_float16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/ConfigurableSOLAR-10.7B", "base_model": "", "revision": "9d9baad88ea9dbaa61881f15e4f0d16e931033b4", "precision": "float16", "params": 10.732, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:39:18Z", "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/ConfigurableSOLAR-10.7B", + "base_model": "", + "revision": "9d9baad88ea9dbaa61881f15e4f0d16e931033b4", + "precision": "float16", + "params": 10.732, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:39:18Z", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/Merge-Mistral-Prometheus-7B_eval_request_False_bfloat16_Original.json b/vicgalle/Merge-Mistral-Prometheus-7B_eval_request_False_bfloat16_Original.json index e545d7745bd716cb54ee0217b11e43077cd14b25..f407a2ef9fe6c9ff27ee8ebd969a1f60f0bb8fbd 100644 --- a/vicgalle/Merge-Mistral-Prometheus-7B_eval_request_False_bfloat16_Original.json +++ b/vicgalle/Merge-Mistral-Prometheus-7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/Merge-Mistral-Prometheus-7B", "base_model": "", "revision": "a7083581b508ce83c74f9267f07024bd462e7161", "precision": "bfloat16", "params": 7.242, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:34:10Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/Merge-Mistral-Prometheus-7B", + "base_model": "", + "revision": "a7083581b508ce83c74f9267f07024bd462e7161", + "precision": "bfloat16", + "params": 7.242, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:34:10Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/Merge-Mixtral-Prometheus-8x7B_eval_request_False_bfloat16_Original.json b/vicgalle/Merge-Mixtral-Prometheus-8x7B_eval_request_False_bfloat16_Original.json index dc7be2a7502b85524f27a7c43b374d0a3ceeb5cc..dda2d540666d4ffb85eee8608800a1b970867dee 100644 --- a/vicgalle/Merge-Mixtral-Prometheus-8x7B_eval_request_False_bfloat16_Original.json +++ b/vicgalle/Merge-Mixtral-Prometheus-8x7B_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "vicgalle/Merge-Mixtral-Prometheus-8x7B", "base_model": "", "revision": "ba53ee5b52a81e56b01e919c069a0d045cfd4e83", "precision": "bfloat16", "params": 46.703, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:33:36Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "vicgalle/Merge-Mixtral-Prometheus-8x7B", + "base_model": "", + "revision": "ba53ee5b52a81e56b01e919c069a0d045cfd4e83", + "precision": "bfloat16", + "params": 46.703, + "architectures": "MixtralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:33:36Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/vicgalle/Roleplay-Llama-3-8B_eval_request_False_float16_Original.json b/vicgalle/Roleplay-Llama-3-8B_eval_request_False_float16_Original.json index 0e18522ad7df4fac5b0ada58f204ea34eaf24ebe..ba86fc53776301d9f3bcb7d37414740651bc763e 100644 --- a/vicgalle/Roleplay-Llama-3-8B_eval_request_False_float16_Original.json +++ b/vicgalle/Roleplay-Llama-3-8B_eval_request_False_float16_Original.json @@ -6,10 +6,10 @@ "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", - "status": "RUNNING", + "status": "PENDING", "submitted_time": "2024-06-26T15:35:53Z", - "model_type": "\ud83d\udcac : \ud83d\udcac chat models (RLHF, DPO, IFT, ...)", + "model_type": "💬 : 💬 chat models (RLHF, DPO, IFT, ...)", "job_id": "7213574", "job_start_time": "2024-06-26T19:49:59.601419", "use_chat_template": true -} \ No newline at end of file +} diff --git a/win10/Breeze-13B-32k-Instruct-v1_0_eval_request_False_bfloat16_Original.json b/win10/Breeze-13B-32k-Instruct-v1_0_eval_request_False_bfloat16_Original.json index 17a315f4c451402e5afe44f038ee2b0c6c8b2f1a..bb34caf101f618e3a41cf949d590de889d56da36 100644 --- a/win10/Breeze-13B-32k-Instruct-v1_0_eval_request_False_bfloat16_Original.json +++ b/win10/Breeze-13B-32k-Instruct-v1_0_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "win10/Breeze-13B-32k-Instruct-v1_0", "base_model": "", "revision": "220c957cf5d9c534a4ef75c11a18221c461de40a", "precision": "bfloat16", "params": 12.726, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:21:14Z", "model_type": "\ud83e\udd1d : \ud83e\udd1d base merges and moerges", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "win10/Breeze-13B-32k-Instruct-v1_0", + "base_model": "", + "revision": "220c957cf5d9c534a4ef75c11a18221c461de40a", + "precision": "bfloat16", + "params": 12.726, + "architectures": "MistralForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:21:14Z", + "model_type": "🤝 : 🤝 base merges and moerges", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/win10/llama3-13.45b-Instruct_eval_request_False_bfloat16_Original.json b/win10/llama3-13.45b-Instruct_eval_request_False_bfloat16_Original.json index fc29adce9f90cbd067643881756aedb2eaea00dc..3ff7d5730d63d2dbed91fa28d3d2574ce56a41d7 100644 --- a/win10/llama3-13.45b-Instruct_eval_request_False_bfloat16_Original.json +++ b/win10/llama3-13.45b-Instruct_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "win10/llama3-13.45b-Instruct", "base_model": "", "revision": "94cc0f415e355c6d3d47168a6ff5239ca586904a", "precision": "bfloat16", "params": 13.265, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T14:09:24Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "win10/llama3-13.45b-Instruct", + "base_model": "", + "revision": "94cc0f415e355c6d3d47168a6ff5239ca586904a", + "precision": "bfloat16", + "params": 13.265, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T14:09:24Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/winglian/Llama-3-8b-64k-PoSE_eval_request_False_bfloat16_Original.json b/winglian/Llama-3-8b-64k-PoSE_eval_request_False_bfloat16_Original.json index 0259cbdfdd94b028882b84d136021f14b221937b..5ec9d30ba3b286ff9dd2acacf72d246e8c61ba82 100644 --- a/winglian/Llama-3-8b-64k-PoSE_eval_request_False_bfloat16_Original.json +++ b/winglian/Llama-3-8b-64k-PoSE_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "winglian/Llama-3-8b-64k-PoSE", "base_model": "", "revision": "5481d9b74a3ec5a95789673e194c8ff86e2bc2bc", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:03:45Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "winglian/Llama-3-8b-64k-PoSE", + "base_model": "", + "revision": "5481d9b74a3ec5a95789673e194c8ff86e2bc2bc", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:03:45Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/winglian/llama-3-8b-256k-PoSE_eval_request_False_bfloat16_Original.json b/winglian/llama-3-8b-256k-PoSE_eval_request_False_bfloat16_Original.json index f4997cec3e23a5bd434d5b3a1476274cd7d3a43d..d02cbf979e5fb1e7f952ab073a5dfe532f00237a 100644 --- a/winglian/llama-3-8b-256k-PoSE_eval_request_False_bfloat16_Original.json +++ b/winglian/llama-3-8b-256k-PoSE_eval_request_False_bfloat16_Original.json @@ -1 +1,15 @@ -{"model": "winglian/llama-3-8b-256k-PoSE", "base_model": "", "revision": "93e7b0b6433c96583ffcef3bc47203e6fdcbbe8b", "precision": "bfloat16", "params": 8.03, "architectures": "LlamaForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-26T15:04:29Z", "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", "job_id": -1, "job_start_time": null, "use_chat_template": true} \ No newline at end of file +{ + "model": "winglian/llama-3-8b-256k-PoSE", + "base_model": "", + "revision": "93e7b0b6433c96583ffcef3bc47203e6fdcbbe8b", + "precision": "bfloat16", + "params": 8.03, + "architectures": "LlamaForCausalLM", + "weight_type": "Original", + "status": "PENDING", + "submitted_time": "2024-06-26T15:04:29Z", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", + "job_id": -1, + "job_start_time": null, + "use_chat_template": true +} diff --git a/zhengr/MixTAO-7Bx2-MoE-v8.1_eval_request_False_bfloat16_Original.json b/zhengr/MixTAO-7Bx2-MoE-v8.1_eval_request_False_bfloat16_Original.json index 5b988c4d77ff8cafe5fbb558f362726c09e938f4..2ae0cd29a5894d7095c59d342adc5b8798bd812e 100644 --- a/zhengr/MixTAO-7Bx2-MoE-v8.1_eval_request_False_bfloat16_Original.json +++ b/zhengr/MixTAO-7Bx2-MoE-v8.1_eval_request_False_bfloat16_Original.json @@ -6,10 +6,10 @@ "params": 12.879, "architectures": "MixtralForCausalLM", "weight_type": "Original", - "status": "FAILED", + "status": "PENDING", "submitted_time": "2024-06-26T14:28:52Z", - "model_type": "\ud83d\udd36 : \ud83d\udd36 fine-tuned on domain-specific datasets", + "model_type": "🔶 : 🔶 fine-tuned on domain-specific datasets", "job_id": "7205400", "job_start_time": "2024-06-26T14:31:44.377511", "use_chat_template": false -} \ No newline at end of file +}