lm1-misc-oscar / 3b977b77b /evaluation /lm1-3b9-77b-oscar-results_lm-eval_global_step73814_2022-12-03-12-11-41.csv
Muennighoff's picture
Add
8ec84fa
raw
history blame contribute delete
530 Bytes
task,metric,value,err,version
arc_challenge,acc,0.23890784982935154,0.012461071376316612,0
arc_challenge,acc_norm,0.27986348122866894,0.013119040897725923,0
arc_easy,acc,0.561026936026936,0.010183076012972067,0
arc_easy,acc_norm,0.5071548821548821,0.01025873302244637,0
boolq,acc,0.5431192660550459,0.008712475433089477,1
hellaswag,acc,0.39543915554670384,0.00487945547466381,0
hellaswag,acc_norm,0.5036845249950209,0.004989645929811447,0
sciq,acc,0.815,0.012285191326386693,0
sciq,acc_norm,0.751,0.013681600278702293,0