inflaton commited on
Commit
6b4da82
·
1 Parent(s): 5fc4c1a

internlm p2 r2

Browse files
llama-factory/config/internlm2_5_7b_lora_sft_bf16_p2_full_r2.yaml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ### model
2
+ model_name_or_path: internlm/internlm2_5-7b-chat-1m
3
+
4
+ ### method
5
+ stage: sft
6
+ do_train: true
7
+ finetuning_type: lora
8
+ lora_target: all
9
+ # quantization_bit: 4 # use 4-bit QLoRA
10
+ loraplus_lr_ratio: 16.0 # use LoRA+ with lambda=16.0
11
+ # use_unsloth: true # use UnslothAI's LoRA optimization for 2x faster training
12
+ upcast_layernorm: true
13
+
14
+ ### dataset
15
+ dataset: alpaca_mgtv_p2
16
+ template: intern2
17
+ cutoff_len: 8192
18
+ max_samples: 25000
19
+ overwrite_cache: true
20
+ preprocessing_num_workers: 16
21
+
22
+ ### output
23
+ output_dir: saves/internlm2_5_7b/lora/sft_bf16_p2_full_r2
24
+ logging_steps: 10
25
+ save_steps: 175
26
+ plot_loss: true
27
+ # overwrite_output_dir: true
28
+
29
+ ### train
30
+ per_device_train_batch_size: 16
31
+ gradient_accumulation_steps: 8
32
+ learning_rate: 1.0e-4
33
+ num_train_epochs: 4.0
34
+ lr_scheduler_type: cosine
35
+ warmup_ratio: 0.1
36
+ bf16: true
37
+ ddp_timeout: 180000000
38
+
39
+ ### eval
40
+ val_size: 0.1
41
+ per_device_eval_batch_size: 1
42
+ eval_strategy: epoch
43
+ eval_steps: 1
44
+
45
+ report_to: wandb
46
+ run_name: internlm2_5_7b_p2_l40_r2 # optional
results/mgtv-results_internlm_best.csv CHANGED
The diff for this file is too large to render. See raw diff
 
scripts/eval-mgtv-best.sh CHANGED
@@ -14,9 +14,9 @@ lscpu
14
  grep MemTotal /proc/meminfo
15
 
16
  #pip install -r requirements.txt
17
- #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
 
18
 
19
- pip install transformers==4.41.2
20
  export USING_LLAMA_FACTORY=true
21
 
22
  export MODEL_NAME=internlm/internlm2_5-7b-chat-1m
 
14
  grep MemTotal /proc/meminfo
15
 
16
  #pip install -r requirements.txt
17
+ #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes] && cd $BASEDIR
18
+ #pip install transformers==4.41.2
19
 
 
20
  export USING_LLAMA_FACTORY=true
21
 
22
  export MODEL_NAME=internlm/internlm2_5-7b-chat-1m
scripts/tune-mgtv-internlm.sh CHANGED
@@ -23,7 +23,10 @@ export CONFIG_FILE=config/internlm2_5_7b_lora_sft_bf16_p1_full.yaml
23
  #echo "Tuning with $CONFIG_FILE"
24
  #$BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
25
 
26
- export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_p2_full.csv
27
- export CONFIG_FILE=config/internlm2_5_7b_lora_sft_bf16_p2_full.yaml
 
 
 
28
  echo "Tuning with $CONFIG_FILE"
29
  $BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
 
23
  #echo "Tuning with $CONFIG_FILE"
24
  #$BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
25
 
26
+ #export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_p2_full.csv
27
+ #export CONFIG_FILE=config/internlm2_5_7b_lora_sft_bf16_p2_full.yaml
28
+ export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_p2_full_r2.csv
29
+ export CONFIG_FILE=config/internlm2_5_7b_lora_sft_bf16_p2_full_r2.yaml
30
+
31
  echo "Tuning with $CONFIG_FILE"
32
  $BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
scripts/tune-mgtv.sh CHANGED
@@ -1 +1 @@
1
- tune-mgtv-glm-4-9b.sh
 
1
+ tune-mgtv-internlm.sh