plaguss HF staff commited on
Commit
72dff2a
1 Parent(s): 2b28e0c

Upload train-config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. train-config.yaml +49 -0
train-config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_args:
2
+ path: distilabel-internal-testing/ultrafeedback-critique-sft-v0.1
3
+
4
+ format_args:
5
+ prompt_format: chatml
6
+
7
+ model_args:
8
+ pretrained_model_name_or_path: teknium/OpenHermes-2.5-Mistral-7B
9
+ torch_dtype: bfloat16
10
+
11
+ wandb_args:
12
+ entity: argilla-io
13
+ project: criticon
14
+ name: criticon-sft-full-v0.1
15
+
16
+ training_args:
17
+ # `trl.SFTTrainer`
18
+ max_seq_length: 2048
19
+ # `transformers.Trainer`
20
+ bf16: true
21
+ do_eval: true
22
+ do_train: true
23
+ evaluation_strategy: steps
24
+ eval_steps: 500
25
+ gradient_accumulation_steps: 2
26
+ gradient_checkpointing: true
27
+ hub_model_id: distilabel-internal-testing/criticon-sft-v0.1
28
+ hub_model_revision: v1.0
29
+ hub_strategy: every_save
30
+ hub_private_repo: true
31
+ push_to_hub: true
32
+ learning_rate: 5.0e-06
33
+ logging_steps: 10
34
+ lr_scheduler_type: cosine
35
+ num_train_epochs: 2
36
+ optim: paged_adamw_32bit
37
+ output_dir: data/criticon
38
+ per_device_train_batch_size: 8
39
+ per_device_eval_batch_size: 8
40
+ save_strategy: epoch
41
+ save_total_limit: null
42
+ # load_best_model_at_end: true
43
+ seed: 42
44
+ warmup_ratio: 0.1
45
+ report_to:
46
+ - wandb
47
+
48
+ use_accelerate: true
49
+ use_unsloth: false