Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -45,11 +45,11 @@ We've kept these datasets within the 200K-300K range for your convenience. We fo
|
|
45 |
|
46 |
The full list of filtered datasets can be found [here](https://github.com/magpie-align/magpie/blob/main/navigation.md).
|
47 |
|
48 |
-
|Model Name | Dataset | Type | Description |
|
49 |
-
|
50 |
-
| [Llama 3.1 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct) | [Magpie-Llama-3.1-Pro-MT-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered) | SFT | (π Flexible License! π) Select 300K high quality multi-turn conversations from Magpie-Llama-3.1-Pro-MT-500K.
|
51 |
-
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Pro-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-300K-Filtered) | SFT | Apply a filter and select 300K high quality conversations from Magpie-Pro-1M.
|
52 |
-
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Pro-MT-300K](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-MT-300K-v0.1) | SFT | Select 300K difficult questions from Magpie-Pro-1M and extend to multi-turn conversations.
|
53 |
-
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Reasoning-150K](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-150K) | SFT | Reasoning booster with 150K math + code + reasoning conversations. Recommend mixing with Magpie-Pro-MT-300K.
|
54 |
-
| [Qwen2 72B Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct) | [Magpie-Qwen2-Pro-200K-Chinese](https://huggingface.co/datasets/Magpie-Align/Magpie-Qwen2-Pro-200K-Chinese) | SFT | Apply a filter and select 200K high quality Chinese conversations from Magpie-Qwen2-Pro-1M.
|
55 |
-
| [Gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it) | [Magpie-Gemma2-Pro-200K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Gemma2-Pro-200K-Filtered) | SFT | (π Flexible License! π) Apply a filter and select 200K conversations from Magpie-Gemma2-Pro-534K.
|
|
|
45 |
|
46 |
The full list of filtered datasets can be found [here](https://github.com/magpie-align/magpie/blob/main/navigation.md).
|
47 |
|
48 |
+
|Model Name | Dataset | Size | Type | Description |
|
49 |
+
|-------------|:-------|:-------|:-------|:-------|
|
50 |
+
| [Llama 3.1 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct) | [Magpie-Llama-3.1-Pro-MT-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered) | 300K | SFT | (π Flexible License! π) Select 300K high quality multi-turn conversations from Magpie-Llama-3.1-Pro-MT-500K.
|
51 |
+
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Pro-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-300K-Filtered) | 300K | SFT | Apply a filter and select 300K high quality conversations from Magpie-Pro-1M.
|
52 |
+
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Pro-MT-300K](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-MT-300K-v0.1) | 300K | SFT | Select 300K difficult questions from Magpie-Pro-1M and extend to multi-turn conversations.
|
53 |
+
| [Llama 3 70B Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) | [Magpie-Reasoning-150K](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-150K) | 150K | SFT | Reasoning booster with 150K math + code + reasoning conversations. Recommend mixing with Magpie-Pro-MT-300K.
|
54 |
+
| [Qwen2 72B Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct) | [Magpie-Qwen2-Pro-200K-Chinese](https://huggingface.co/datasets/Magpie-Align/Magpie-Qwen2-Pro-200K-Chinese) | 200K | SFT | Apply a filter and select 200K high quality Chinese conversations from Magpie-Qwen2-Pro-1M.
|
55 |
+
| [Gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it) | [Magpie-Gemma2-Pro-200K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Gemma2-Pro-200K-Filtered) | 200K | SFT | (π Flexible License! π) Apply a filter and select 200K conversations from Magpie-Gemma2-Pro-534K.
|