Glavin001 commited on
Commit
3f84d0d
·
1 Parent(s): 71febab

Create coqar-questions-7b-lora.yml

Browse files
Files changed (1) hide show
  1. coqar-questions-7b-lora.yml +72 -0
coqar-questions-7b-lora.yml ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # accelerate launch scripts/finetune.py config/coqar-questions-7b-lora.yml
2
+ # accelerate launch scripts/finetune.py config/coqar-questions-7b-lora.yml --inference --lora_model_dir="./outputs/lora-out"
3
+ # CUDA_VISIBLE_DEVICES="" accelerate launch scripts/finetune.py config/coqar-questions-7b-lora.yml --merge_lora --lora_model_dir="./outputs/lora-out" --load_in_8bit=False --load_in_4bit=False
4
+
5
+ base_model: meta-llama/Llama-2-7b-hf
6
+ base_model_config: meta-llama/Llama-2-7b-hf
7
+ model_type: LlamaForCausalLM
8
+ tokenizer_type: LlamaTokenizer
9
+ is_llama_derived_model: true
10
+
11
+ load_in_8bit: true
12
+ load_in_4bit: false
13
+ strict: false
14
+
15
+ datasets:
16
+ - path: Glavin001/generate-questions-v0.1
17
+ type: alpaca
18
+ dataset_prepared_path: last_run_prepared
19
+ val_set_size: 0.01
20
+ # output_dir: ./lora-out
21
+ output_dir: ./outputs/lora-out
22
+
23
+ sequence_len: 4096
24
+ sample_packing: true
25
+
26
+ adapter: lora
27
+ lora_model_dir:
28
+ lora_r: 32
29
+ lora_alpha: 16
30
+ lora_dropout: 0.05
31
+ lora_target_linear: true
32
+ lora_fan_in_fan_out:
33
+
34
+ wandb_project: coqar-questions
35
+ wandb_entity:
36
+ wandb_watch:
37
+ wandb_run_id:
38
+ wandb_log_model:
39
+
40
+ gradient_accumulation_steps: 4
41
+ micro_batch_size: 2
42
+ num_epochs: 3
43
+ optimizer: adamw_bnb_8bit
44
+ lr_scheduler: cosine
45
+ learning_rate: 0.0002
46
+
47
+ train_on_inputs: false
48
+ group_by_length: false
49
+ bf16: true
50
+ fp16: false
51
+ tf32: false
52
+
53
+ gradient_checkpointing: true
54
+ early_stopping_patience:
55
+ resume_from_checkpoint:
56
+ local_rank:
57
+ logging_steps: 1
58
+ xformers_attention:
59
+ flash_attention: true
60
+
61
+ warmup_steps: 10
62
+ eval_steps: 20
63
+ save_steps:
64
+ debug:
65
+ deepspeed:
66
+ weight_decay: 0.0
67
+ fsdp:
68
+ fsdp_config:
69
+ special_tokens:
70
+ bos_token: "<s>"
71
+ eos_token: "</s>"
72
+ unk_token: "<unk>"