dh-mc commited on
Commit
4d1916b
·
1 Parent(s): be560ea

updated best results

Browse files
llm_toolkit/llm_utils.py CHANGED
@@ -183,11 +183,8 @@ def eval_model(
183
  decoded_output = tokenizer.batch_decode(
184
  outputs, skip_special_tokens=True
185
  ) # Skip special tokens for clean output
186
- debug = i == 0
187
- decoded_output = [
188
- extract_answer(output, debug=debug and j == 0)
189
- for j, output in enumerate(decoded_output)
190
- ]
191
  predictions.extend(decoded_output)
192
 
193
  return predictions
 
183
  decoded_output = tokenizer.batch_decode(
184
  outputs, skip_special_tokens=True
185
  ) # Skip special tokens for clean output
186
+ if i == 0:
187
+ print("Batch output:", decoded_output)
 
 
 
188
  predictions.extend(decoded_output)
189
 
190
  return predictions
results/mgtv-results_internlm_best.csv CHANGED
The diff for this file is too large to render. See raw diff