inflaton commited on
Commit
ad9400c
·
1 Parent(s): c81ec3a

llama3 8b base model results

Browse files
llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml CHANGED
@@ -1,5 +1,6 @@
1
  ### model
2
- model_name_or_path: FlagAlpha/Llama3-Chinese-8B-Instruct
 
3
 
4
  ### method
5
  stage: sft
 
1
  ### model
2
+ model_name_or_path: shenzhi-wang/Llama3-8B-Chinese-Chat
3
+ #model_name_or_path: FlagAlpha/Llama3-Chinese-8B-Instruct
4
 
5
  ### method
6
  stage: sft
llama-factory/config/llama3-8b_lora_sft_bf16-p2.yaml CHANGED
@@ -1,5 +1,6 @@
1
  ### model
2
- model_name_or_path: FlagAlpha/Llama3-Chinese-8B-Instruct
 
3
 
4
  ### method
5
  stage: sft
 
1
  ### model
2
+ model_name_or_path: shenzhi-wang/Llama3-8B-Chinese-Chat
3
+ #model_name_or_path: FlagAlpha/Llama3-Chinese-8B-Instruct
4
 
5
  ### method
6
  stage: sft
scripts/{eval-mgtv-llama3.sh → eval-mgtv-llama3_8b.sh} RENAMED
File without changes
scripts/eval-mgtv.sh CHANGED
@@ -1 +1 @@
1
- eval-mgtv-llama3.sh
 
1
+ eval-mgtv-llama3_8b.sh
scripts/tune-mgtv-llama3_8b.sh CHANGED
@@ -20,24 +20,20 @@ grep MemTotal /proc/meminfo
20
 
21
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
22
 
23
- export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
 
 
24
  export MODEL_PREFIX=llama3-8b_lora_sft_bf16
25
 
26
  export CONFIG_FILE=config/$MODEL_PREFIX-p1.yaml
27
- echo "Tuning with $CONFIG_FILE"
28
- #$BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
29
-
30
- export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
31
- export ADAPTER_PATH_BASE=llama-factory/saves/qwen2-72b/lora/sft_4bit_p1_full
32
- echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
33
- #python llm_toolkit/eval_logical_reasoning_all_epochs.py
34
 
35
 
36
  export CONFIG_FILE=config/$MODEL_PREFIX-p2.yaml
37
- echo "Tuning with $CONFIG_FILE"
38
  $BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
39
 
40
- export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
41
- export ADAPTER_PATH_BASE=llama-factory/saves/qwen2-72b/lora/sft_4bit_p2_full
42
- echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
43
- #python llm_toolkit/eval_logical_reasoning_all_epochs.py
 
20
 
21
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
22
 
23
+ #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
24
+ export MODEL_NAME=shenzhi-wang/Llama3-8B-Chinese-Chat
25
+
26
  export MODEL_PREFIX=llama3-8b_lora_sft_bf16
27
 
28
  export CONFIG_FILE=config/$MODEL_PREFIX-p1.yaml
29
+ echo "Tuning $MODEL_NAME with $CONFIG_FILE"
30
+ $BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
 
 
 
 
 
31
 
32
 
33
  export CONFIG_FILE=config/$MODEL_PREFIX-p2.yaml
34
+ echo "Tuning $MODEL_NAME with $CONFIG_FILE"
35
  $BASEDIR/scripts/tune-lf.sh $CONFIG_FILE
36
 
37
+
38
+ $BASEDIR/scripts/eval-mgtv-llama3_8b.sh
39
+