llm-qa-bench / tune_rp.sh
dh-mc's picture
Update tune_rp.sh
886e3cc
#!/bin/sh
BASEDIR=$(dirname "$0")
cd $BASEDIR
echo Current Directory:
pwd
nvidia-smi
uname -a
cat /etc/os-release
lscpu
grep MemTotal /proc/meminfo
pip install -r requirements.txt
# Evaluating Models (from Smallest to Largest) for the MS MARCO Dataset
./eval_modules/eval-hf_v2.sh google gemma-1.1-2b-it true
./scripts/eval-hf_v2.sh microsoft Phi-3-mini-128k-instruct true
./eval_modules/eval-hf_v2.sh google gemma-1.1-7b-it true
./eval_modules/eval-hf_v2.sh mistralai Mistral-7B-Instruct-v0.2 true
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-7b-chat-hf true
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-13b-chat-hf true
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-70b-chat-hf true
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-8B-Instruct true
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-70B-Instruct true
# Evaluating Models (from Smallest to Largest) for the WebQSP Dataset
./eval_modules/eval-hf_v2.sh google gemma-1.1-2b-it false
./scripts/eval-hf_v2.sh microsoft Phi-3-mini-128k-instruct false
./eval_modules/eval-hf_v2.sh google gemma-1.1-7b-it false
./eval_modules/eval-hf_v2.sh mistralai Mistral-7B-Instruct-v0.2 false
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-7b-chat-hf false
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-13b-chat-hf false
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-70b-chat-hf false
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-8B-Instruct false
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-70B-Instruct false