lucyknada commited on
Commit
50051b7
·
verified ·
1 Parent(s): bbe6b06

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +167 -0
README.md ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: mrl
4
+ language:
5
+ - en
6
+ tags:
7
+ - chat
8
+ pipeline_tag: text-generation
9
+ library_name: transformers
10
+ ---
11
+
12
+
13
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/-UC6YN1Gt3e1FDh8EqyaB.png)
14
+ ## This repo contains GGUF quants of the model. If you need the original weights, please find them [here](https://huggingface.co/anthracite-org/magnum-v4-12b).
15
+
16
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
17
+
18
+
19
+
20
+ This model is fine-tuned on top of [mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407).
21
+
22
+ ## Prompting
23
+ A typical input would look like this:
24
+
25
+ ```py
26
+ <s>[INST] SYSTEM MESSAGE
27
+ USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
28
+ ```
29
+
30
+ ## SillyTavern templates
31
+
32
+ Below are Instruct and Context templates for use within SillyTavern.
33
+
34
+ <details><summary>context template</summary>
35
+
36
+ ```yaml
37
+ default SillyTavern template works fine
38
+ ```
39
+
40
+ </details><br>
41
+ <details><summary>instruct template</summary>
42
+
43
+ ```yaml
44
+ default SillyTavern template works fine
45
+ ```
46
+
47
+ </details><br>
48
+
49
+ ## Axolotl config
50
+
51
+ <details><summary>See axolotl config</summary>
52
+
53
+ ```yaml
54
+ base_model: mistralai/Mistral-Nemo-Instruct-2407
55
+ model_type: AutoModelForCausalLM
56
+ tokenizer_type: AutoTokenizer
57
+
58
+ hub_model_id: anthracite-org/magnum-v4-12b-r2
59
+ hub_strategy: "all_checkpoints"
60
+ push_dataset_to_hub:
61
+ hf_use_auth_token: true
62
+
63
+ plugins:
64
+ - axolotl.integrations.liger.LigerPlugin
65
+ liger_rope: true
66
+ liger_rms_norm: true
67
+ liger_swiglu: true
68
+ liger_fused_linear_cross_entropy: true
69
+
70
+ load_in_8bit: false
71
+ load_in_4bit: false
72
+ strict: false
73
+
74
+ datasets:
75
+ - path: anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system
76
+ type: custommistralv3tekken
77
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
78
+ type: custommistralv3tekken
79
+ - path: anthracite-org/kalo-opus-instruct-3k-filtered-no-system
80
+ type: custommistralv3tekken
81
+ - path: anthracite-org/nopm_claude_writing_fixed
82
+ type: custommistralv3tekken
83
+ - path: anthracite-org/kalo_opus_misc_240827_no_system
84
+ type: custommistralv3tekken
85
+ - path: anthracite-org/kalo_misc_part2_no_system
86
+ type: custommistralv3tekken
87
+ #chat_template: chatml
88
+ shuffle_merged_datasets: true
89
+ #default_system_message: "You are an assistant that responds to the user."
90
+ dataset_prepared_path: /workspace/data/magnum-12b-data
91
+ val_set_size: 0.0
92
+ output_dir: /workspace/data/12b-fft-out
93
+
94
+ sequence_len: 32768
95
+ sample_packing: true
96
+ pad_to_sequence_len: true
97
+
98
+ adapter:
99
+ lora_model_dir:
100
+ lora_r:
101
+ lora_alpha:
102
+ lora_dropout:
103
+ lora_target_linear:
104
+ lora_fan_in_fan_out:
105
+
106
+ wandb_project: 12b-magnum-fft
107
+ wandb_entity:
108
+ wandb_watch:
109
+ wandb_name: v4-r2-attempt-01
110
+ wandb_log_model:
111
+
112
+ gradient_accumulation_steps: 2
113
+ micro_batch_size: 1
114
+ num_epochs: 2
115
+ optimizer: adamw_bnb_8bit
116
+ lr_scheduler: cosine
117
+ learning_rate: 0.00001
118
+
119
+ train_on_inputs: false
120
+ group_by_length: false
121
+ bf16: auto
122
+ fp16:
123
+ tf32: false
124
+
125
+ gradient_checkpointing: true
126
+ early_stopping_patience:
127
+ resume_from_checkpoint:
128
+ local_rank:
129
+ logging_steps: 1
130
+ xformers_attention:
131
+ flash_attention: true
132
+
133
+ warmup_steps: 40
134
+ evals_per_epoch:
135
+ eval_table_size:
136
+ eval_max_new_tokens:
137
+ saves_per_epoch: 2
138
+ debug:
139
+ deepspeed: deepspeed_configs/zero2.json
140
+ weight_decay: 0.1
141
+ fsdp:
142
+ fsdp_config:
143
+ special_tokens:
144
+ pad_token: <pad>
145
+ ```
146
+ </details><br>
147
+
148
+ ## Credits
149
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
150
+
151
+ We would also like to thank all members of Anthracite who made this finetune possible.
152
+
153
+ ## Datasets
154
+ - [anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system)
155
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
156
+ - [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
157
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
158
+ - [anthracite-org/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827_no_system)
159
+ - [anthracite-org/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2_no_system)
160
+
161
+ ## Training
162
+ The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
163
+
164
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
165
+
166
+ ## Safety
167
+ ...