dh-mc commited on
Commit
dc7abea
·
1 Parent(s): ba41152

nv4090 results

Browse files
competition/10e_InternLM_NV4090_eval.ipynb ADDED
The diff for this file is too large to render. See raw diff
 
llm_toolkit/eval_logical_reasoning.py CHANGED
@@ -94,7 +94,7 @@ if adapter_name_or_path is not None:
94
  model_name += "/" + adapter_name_or_path.split("/")[-1]
95
 
96
  save_results(
97
- f"{model_name}_{dtype}{'_4bit' if load_in_4bit else ''}",
98
  results_path,
99
  datasets["test"],
100
  predictions,
 
94
  model_name += "/" + adapter_name_or_path.split("/")[-1]
95
 
96
  save_results(
97
+ f"{model_name}_{dtype}{'_4bit' if load_in_4bit else ''}{'_lf' if using_llama_factory else ''}",
98
  results_path,
99
  datasets["test"],
100
  predictions,
llm_toolkit/llm_utils.py CHANGED
@@ -88,7 +88,9 @@ def check_gpu():
88
  # If we have a GPU available, we'll set our device to GPU. We'll use this device variable later in our code.
89
  if is_cuda:
90
  device = torch.device("cuda")
91
- print("GPU is available")
 
 
92
  elif torch.backends.mps.is_available():
93
  device = torch.device("mps")
94
  print("MPS is available")
 
88
  # If we have a GPU available, we'll set our device to GPU. We'll use this device variable later in our code.
89
  if is_cuda:
90
  device = torch.device("cuda")
91
+ print("CUDA is available, we have found ", torch.cuda.device_count(), " GPU(s)")
92
+ print(torch.cuda.get_device_name(0))
93
+ print("CUDA version: " + torch.version.cuda)
94
  elif torch.backends.mps.is_available():
95
  device = torch.device("mps")
96
  print("MPS is available")
results/mgtv-results_internlm_nv4090.csv ADDED
The diff for this file is too large to render. See raw diff
 
scripts/eval-mgtv-nv4090.sh ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/sh
2
+
3
+ BASEDIR=$(dirname "$0")
4
+ cd $BASEDIR/..
5
+ echo Current Directory:
6
+ pwd
7
+
8
+ BASEDIR=`pwd`
9
+
10
+ nvidia-smi
11
+ uname -a
12
+ cat /etc/os-release
13
+ lscpu
14
+ grep MemTotal /proc/meminfo
15
+
16
+ export LOAD_IN_4BIT=false
17
+
18
+ export MODEL_NAME=internlm/internlm2_5-7b-chat-1m
19
+ export ADAPTER_NAME_OR_PATH=inflaton-ai/InternLM_2_5-7b_LoRA-Adapter
20
+ export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
21
+ export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_internlm_best.csv
22
+
23
+ export USE_FLOAT32_FOR_INFERENCE=false
24
+
25
+ export USE_BF16_FOR_INFERENCE=false
26
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
27
+ python llm_toolkit/eval_logical_reasoning.py
28
+
29
+ export USE_BF16_FOR_INFERENCE=true
30
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
31
+ python llm_toolkit/eval_logical_reasoning.py
32
+
33
+ export LOAD_IN_4BIT=true
34
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
35
+ python llm_toolkit/eval_logical_reasoning.py