alpindale commited on
Commit
2a7341d
1 Parent(s): 81b8bc7

Create axolotl_config/config.yml

Browse files
Files changed (1) hide show
  1. axolotl_config/config.yml +70 -0
axolotl_config/config.yml ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: Qwen/Qwen2-72B-Instruct
2
+ model_type: AutoModelForCausalLM
3
+ tokenizer_type: AutoTokenizer
4
+
5
+ trust_remote_code: true
6
+
7
+ load_in_8bit: false
8
+ load_in_4bit: false
9
+ strict: false
10
+
11
+ datasets:
12
+ - path: data/data.json
13
+ type: sharegpt
14
+ conversation: chatml
15
+ chat_template: chatml
16
+ dataset_prepared_path:
17
+ val_set_size: 0.0
18
+ output_dir: ./qwen2-72b-instruct
19
+
20
+ sequence_len: 8192
21
+ sample_packing: true
22
+ eval_sample_packing: false
23
+ pad_to_sequence_len:
24
+
25
+ adapter:
26
+ lora_model_dir:
27
+ lora_r:
28
+ lora_alpha:
29
+ lora_dropout:
30
+ lora_target_linear:
31
+ lora_fan_in_fan_out:
32
+
33
+ wandb_project: qwen2-72b-opus-v0.1
34
+ wandb_entity:
35
+ wandb_watch:
36
+ wandb_name: attempt-3
37
+ wandb_log_model:
38
+
39
+ gradient_accumulation_steps: 2
40
+ micro_batch_size: 1
41
+ num_epochs: 4
42
+ optimizer: paged_adamw_8bit
43
+ lr_scheduler: cosine
44
+ learning_rate: 0.000001
45
+
46
+ train_on_inputs: false
47
+ group_by_length: false
48
+ bf16: auto
49
+ fp16:
50
+ tf32: false
51
+
52
+ gradient_checkpointing: unsloth
53
+ early_stopping_patience:
54
+ resume_from_checkpoint:
55
+ local_rank:
56
+ logging_steps: 1
57
+ xformers_attention:
58
+ flash_attention: true
59
+
60
+ warmup_steps: 10
61
+ evals_per_epoch:
62
+ eval_table_size:
63
+ eval_max_new_tokens:
64
+ saves_per_epoch: 4
65
+ debug:
66
+ deepspeed: deepspeed_configs/zero3_bf16.json
67
+ weight_decay: 0.0
68
+ fsdp:
69
+ fsdp_config:
70
+ special_tokens: