winglian commited on
Commit
8303cac
·
1 Parent(s): de0f0c2

Update configs/omega.yml

Browse files
Files changed (1) hide show
  1. configs/omega.yml +83 -0
configs/omega.yml CHANGED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: winglian/phi-1_5-2x-fib
2
+ base_model_config: winglian/phi-1_5-2x-fib
3
+ model_type: AutoModelForCausalLM
4
+ tokenizer_type: AutoTokenizer
5
+ is_llama_derived_model: false
6
+ trust_remote_code: true
7
+
8
+ load_in_8bit: false
9
+ load_in_4bit: false
10
+ strict: false
11
+
12
+ datasets:
13
+ - path: nampdn-ai/tiny-lessons
14
+ type: completion
15
+ field: textbook
16
+ - path: nampdn-ai/tiny-textbooks
17
+ type: completion
18
+ field: textbook
19
+
20
+ dataset_prepared_path: last_run_prepared
21
+ val_set_size: 0.01
22
+ output_dir: ./phi-2x-out
23
+ hub_model_id: winglian/phi-2x-pt
24
+
25
+ sequence_len: 2048
26
+ sample_packing: false
27
+ pad_to_sequence_len:
28
+ resize_token_embeddings_to_32x: true
29
+
30
+ adapter:
31
+ lora_model_dir:
32
+ lora_r: 32
33
+ lora_alpha: 16
34
+ lora_dropout: 0.05
35
+ lora_target_linear: true
36
+ lora_fan_in_fan_out:
37
+
38
+ wandb_project: phi-2x-pt-tiny
39
+ wandb_entity:
40
+ wandb_watch:
41
+ wandb_run_id:
42
+ wandb_log_model:
43
+
44
+ gradient_accumulation_steps: 2
45
+ micro_batch_size: 1
46
+ num_epochs: 1
47
+ optimizer: adamw_torch
48
+ adam_beta2: 0.95
49
+ adam_epsilon: 0.00001
50
+ max_grad_norm: 1.0
51
+ lr_scheduler: cosine
52
+ learning_rate: 0.00003
53
+
54
+ train_on_inputs: false
55
+ group_by_length: false
56
+ bf16: true
57
+ fp16: false
58
+ tf32: true
59
+
60
+ gradient_checkpointing:
61
+ early_stopping_patience:
62
+ resume_from_checkpoint:
63
+ local_rank:
64
+ logging_steps: 1
65
+ xformers_attention:
66
+ flash_attention:
67
+
68
+ warmup_steps: 1000
69
+ eval_steps: 0.05
70
+ save_steps: 0.10
71
+ do_bench_eval: true
72
+ bench_dataset: "pharaouk/dharma-1/dharma_1_full.json"
73
+ save_total_limit: 20
74
+ debug:
75
+ deepspeed:
76
+ weight_decay: 0.1
77
+ fsdp:
78
+ fsdp_config:
79
+ special_tokens:
80
+ bos_token: "<|endoftext|>"
81
+ eos_token: "<|endoftext|>"
82
+ unk_token: "<|endoftext|>"
83
+ pad_token: "<|endoftext|>"