inflaton commited on
Commit
fc9601b
1 Parent(s): e1e71f5

updated scripts

Browse files
llm_toolkit/eval.py ADDED
@@ -0,0 +1 @@
 
 
1
+ eval_shots.py
scripts/eval-4gpu.sh CHANGED
@@ -17,8 +17,9 @@ grep MemTotal /proc/meminfo
17
 
18
  export BATCH_SIZE=1
19
  # export START_REPETITION_PENALTY=1.06
 
20
 
21
- ./scripts/eval-model.sh Qwen/Qwen2-72B-Instruct
22
 
23
  ./scripts/eval-model.sh shenzhi-wang/Llama3.1-70B-Chinese-Chat
24
 
 
17
 
18
  export BATCH_SIZE=1
19
  # export START_REPETITION_PENALTY=1.06
20
+ export START_NUM_SHOTS=50
21
 
22
+ #./scripts/eval-model.sh Qwen/Qwen2-72B-Instruct
23
 
24
  ./scripts/eval-model.sh shenzhi-wang/Llama3.1-70B-Chinese-Chat
25
 
scripts/eval-mac.sh CHANGED
@@ -11,16 +11,16 @@ cat /etc/os-release
11
  lscpu
12
  grep MemTotal /proc/meminfo
13
 
14
- # pip install torch torchvision torchaudio
15
- # pip install -r requirements.txt
16
 
17
- # export START_NUM_SHOTS=50
18
 
19
- ./scripts/eval-model.sh internlm/internlm2_5-7b-chat
20
 
21
- ./scripts/eval-model.sh Qwen/Qwen2-7B-Instruct
22
 
23
- ./scripts/eval-model.sh shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat
24
 
25
  ./scripts/eval-model.sh shenzhi-wang/Llama3.1-8B-Chinese-Chat
26
 
 
11
  lscpu
12
  grep MemTotal /proc/meminfo
13
 
14
+ pip install torch torchvision torchaudio
15
+ pip install -r requirements.txt
16
 
17
+ export START_NUM_SHOTS=50
18
 
19
+ #./scripts/eval-model.sh internlm/internlm2_5-7b-chat
20
 
21
+ #./scripts/eval-model.sh Qwen/Qwen2-7B-Instruct
22
 
23
+ #./scripts/eval-model.sh shenzhi-wang/Mistral-7B-v0.3-Chinese-Chat
24
 
25
  ./scripts/eval-model.sh shenzhi-wang/Llama3.1-8B-Chinese-Chat
26
 
scripts/tune-lf-4gpu.sh CHANGED
@@ -9,8 +9,10 @@ export ORG_NAME=$1
9
  export MODEL_NAME=$2
10
  export CHAT_TEMPLATE=$3
11
  export DATA_PATH=../datasets/mac/mac.tsv
12
- #export YAML=config/mac_template_4gpu.yaml
13
- export YAML=config/mac_template_qwen2_72b.yaml
 
 
14
 
15
  python ../llm_toolkit/setup_lf.py
16
  llamafactory-cli train config/models/$MODEL_NAME.yaml
 
9
  export MODEL_NAME=$2
10
  export CHAT_TEMPLATE=$3
11
  export DATA_PATH=../datasets/mac/mac.tsv
12
+ export YAML=config/mac_template_4gpu.yaml
13
+ #export YAML=config/mac_template_qwen2_72b.yaml
14
+
15
+ export PYTORCH_CUDA_ALLOC_CONF="expandable_segments:True"
16
 
17
  python ../llm_toolkit/setup_lf.py
18
  llamafactory-cli train config/models/$MODEL_NAME.yaml
scripts/tune-lf.sh CHANGED
@@ -11,5 +11,7 @@ export CHAT_TEMPLATE=$3
11
  export DATA_PATH=../datasets/mac/mac.tsv
12
  export YAML=config/mac_template.yaml
13
 
 
 
14
  python ../llm_toolkit/setup_lf.py
15
  llamafactory-cli train config/models/$MODEL_NAME.yaml
 
11
  export DATA_PATH=../datasets/mac/mac.tsv
12
  export YAML=config/mac_template.yaml
13
 
14
+ export PYTORCH_CUDA_ALLOC_CONF="expandable_segments:True"
15
+
16
  python ../llm_toolkit/setup_lf.py
17
  llamafactory-cli train config/models/$MODEL_NAME.yaml
scripts/tune-mac-4gpu.sh CHANGED
@@ -12,8 +12,8 @@ lscpu
12
  grep MemTotal /proc/meminfo
13
 
14
  #pip install -r requirements.txt
15
- #cd ../LLaMA-Factory && pip install -e .[torch,metrics,vllm] && cd -
16
 
17
- PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True ./scripts/tune-lf-4gpu.sh Qwen Qwen2-72B-Instruct qwen
18
 
19
  #./scripts/tune-lf-4gpu.sh shenzhi-wang Llama3.1-70B-Chinese-Chat llama3
 
12
  grep MemTotal /proc/meminfo
13
 
14
  #pip install -r requirements.txt
15
+ cd ../LLaMA-Factory && pip install -e .[torch,metrics,vllm] && cd -
16
 
17
+ ./scripts/tune-lf-4gpu.sh Qwen Qwen2-72B-Instruct qwen
18
 
19
  #./scripts/tune-lf-4gpu.sh shenzhi-wang Llama3.1-70B-Chinese-Chat llama3